vllm/tests/v1/tpu
Chengji Yao a77aea59fd
[TPU] support attention head dim smaller than 128 (#19620)
Signed-off-by: Chengji Yao <chengjiyao@google.com>
Co-authored-by: mgoin <mgoin64@gmail.com>
2025-06-16 06:40:53 +00:00
..
worker [Core] Use tuple for kv cache group block ids (#19175) 2025-06-10 07:01:17 +02:00
__init__.py [V1] TPU - Add tensor parallel support via Ray (#13618) 2025-03-08 08:19:38 -05:00
test_basic.py [TPU] support attention head dim smaller than 128 (#19620) 2025-06-16 06:40:53 +00:00
test_mha_attn.py [Misc] Add SPDX-FileCopyrightText (#19100) 2025-06-03 11:20:17 -07:00
test_multimodal.py [Misc] Add SPDX-FileCopyrightText (#19100) 2025-06-03 11:20:17 -07:00
test_pallas.py [Misc] Add SPDX-FileCopyrightText (#19100) 2025-06-03 11:20:17 -07:00
test_perf.py [Misc] Add SPDX-FileCopyrightText (#19100) 2025-06-03 11:20:17 -07:00
test_sampler.py [Misc] Add SPDX-FileCopyrightText (#19100) 2025-06-03 11:20:17 -07:00
test_spmd_model_weight_loading.py [TPU] Skip hanging tests (#19115) 2025-06-04 01:43:00 -07:00
test_topk_topp_sampler.py [Misc] Add SPDX-FileCopyrightText (#19100) 2025-06-03 11:20:17 -07:00
test_tpu_qkv_linear.py [Hardware][TPU] Initial support of model parallelism with single worker using SPMD (#18011) 2025-06-03 00:06:20 +00:00