From 41e379c10344a99e6893b0659042eddb05b45e28 Mon Sep 17 00:00:00 2001 From: WoosukKwon Date: Fri, 24 Jan 2025 13:50:38 -0800 Subject: [PATCH] Minor Signed-off-by: WoosukKwon --- _posts/2025-01-24-v1.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/_posts/2025-01-24-v1.md b/_posts/2025-01-24-v1.md index 6de2346..47f3ca7 100644 --- a/_posts/2025-01-24-v1.md +++ b/_posts/2025-01-24-v1.md @@ -109,7 +109,7 @@ V1 supports decoder-only Transformers like Llama, mixture-of-experts (MoE) model V1 currently lacks support for log probs, prompt log probs sampling parameters, pipeline parallelism, structured decoding, speculative decoding, prometheus metrics, and LoRA. We are actively working to close this feature gap and add new optimizations. Please stay tuned! **Hardware Support:** -V1 currently supports only Ampere or later NVIDIA GPUs. We are working on support for other hardware backends. +V1 currently supports only Ampere or later NVIDIA GPUs. We are working on support for other hardware backends such as TPU. Finally, please note that you can continue using V0 and maintain backward compatibility by not setting `VLLM_USE_V1=1`.