Go to file
Chendi.Xue 83f99c54ce
[CI] update CI script (#8)
* update CI script

Signed-off-by: Chendi Xue <chendi.xue@intel.com>

* clean up

Signed-off-by: Chendi Xue <chendi.xue@intel.com>

---------

Signed-off-by: Chendi Xue <chendi.xue@intel.com>
2025-07-03 21:11:22 -05:00
.github/workflows [CI] update CI script (#8) 2025-07-03 21:11:22 -05:00
docs Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
examples Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
tests Add hourly CI (#7) 2025-07-03 17:30:08 -05:00
tools Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
vllm_hpu Fix failing due to (#16728) (#5) 2025-07-03 14:50:23 -05:00
.gitignore Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
.pre-commit-config.yaml Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
README.md Update README.md 2025-06-25 18:33:36 +02:00
format.sh Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
mkdocs.yaml Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
pyproject.toml Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
requirements-docs.txt Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
requirements-lint.txt Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
requirements.txt Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
setup.py Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00

README.md

[!IMPORTANT]
This is an early developer preview of the vLLM Gaudi Plugin and is not yet intended for general use. For a more stable experience, consider using the HabanaAI/vllm-fork or the in-tree Gaudi implementation available in vllm-project/vllm.

Welcome to vLLM x Intel Gaudi

vLLM x Intel-Gaudi

vLLM Gaudi plugin (vllm-gaudi) integrates Intel Gaudi accelerators with vLLM to optimize large language model inference.

This plugin follows the [RFC]: Hardware pluggable and [RFC]: Enhancing vLLM Plugin Architecture principles, providing a modular interface for Intel Gaudi hardware.

Learn more:

📚 Intel Gaudi Documentation
🚀 vLLM Plugin System Overview