This website requires JavaScript.
Explore
Help
Register
Sign In
vLLM
Follow
Repositories
12
Projects
Packages
Search
Sort
Newest
Oldest
Alphabetically
Reverse alphabetically
Recently updated
Least recently updated
Most stars
Fewest stars
Most forks
Fewest forks
vllm-gaudi
Python
0
0
Updated
2025-07-04 16:08:08 +08:00
vllm
Python
0
0
A high-throughput and memory-efficient inference and serving engine for LLMs
llm
mlops
pytorch
cuda
inference
llama
llm-serving
llmops
model-serving
qwen
rocm
tpu
trainium
transformer
amd
xpu
deepseek
gpt
hpu
inferentia
Updated
2025-07-04 16:00:34 +08:00
dashboard
Python
0
0
vLLM performance dashboard
Updated
2025-07-04 15:22:09 +08:00