From e51ece8b3169ef9ac06b81f73871a61d3202f8a9 Mon Sep 17 00:00:00 2001 From: Woosuk Kwon Date: Tue, 14 Nov 2023 12:45:03 -0800 Subject: [PATCH] Update _posts/2023-11-14-notes-vllm-vs-deepspeed.md Co-authored-by: Zhuohan Li --- _posts/2023-11-14-notes-vllm-vs-deepspeed.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/_posts/2023-11-14-notes-vllm-vs-deepspeed.md b/_posts/2023-11-14-notes-vllm-vs-deepspeed.md index eeacbb8..9ea9f60 100644 --- a/_posts/2023-11-14-notes-vllm-vs-deepspeed.md +++ b/_posts/2023-11-14-notes-vllm-vs-deepspeed.md @@ -9,7 +9,7 @@ author: "vLLM Team" - vLLM matches DeepSpeed's speed in common scenarios and surpasses it when handling longer outputs. - DeepSpeed only outperforms vLLM in scenarios with long prompts and short outputs, due to its Dynamic SplitFuse optimization. This optimization is on vLLM’s roadmap. -- vLLM’s mission is to build the fastest and easiest-to-use open-source LLM inference and serving engine. It is Apache 2.0 licensed and driven by a community focus, offering extensive model and optimization support. +- vLLM’s mission is to build the fastest and easiest-to-use open-source LLM inference and serving engine. It is Apache 2.0 and community-owned, offering extensive model and optimization support. ---