Go to file
Agata Dobrzyniewicz 86a8acef15
Port: #282 from extension (#52)
https://github.com/HabanaAI/vllm-hpu-extension/pull/282

Signed-off-by: Agata Dobrzyniewicz <adobrzyniewicz@habana.ai>
2025-07-30 11:06:20 +02:00
.github [CI] Report 10 longest unit tests above 1s (#47) 2025-07-28 14:06:34 +02:00
.jenkins [CI] Add models-medium stage (#21) 2025-07-11 16:13:19 +02:00
docs Refactor vllm_hpu -> vllm_gaudi (#2) 2025-07-11 13:11:56 +02:00
examples Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
tests Add HPU model runner & HPU input batch unit tests (#44) 2025-07-28 12:13:27 +02:00
tools Refactor vllm_hpu -> vllm_gaudi (#2) 2025-07-11 13:11:56 +02:00
vllm_gaudi Port: #282 from extension (#52) 2025-07-30 11:06:20 +02:00
.gitignore Refactor vllm_hpu -> vllm_gaudi (#2) 2025-07-11 13:11:56 +02:00
.pre-commit-config.yaml Add pre-commit GHA (#1) 2025-07-11 11:55:04 +02:00
README.md Add Getting Started section to README (#36) 2025-07-17 15:27:30 +02:00
format.sh Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
mkdocs.yaml Refactor vllm_hpu -> vllm_gaudi (#2) 2025-07-11 13:11:56 +02:00
pyproject.toml Refactor vllm_hpu -> vllm_gaudi (#2) 2025-07-11 13:11:56 +02:00
requirements-docs.txt Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
requirements-lint.txt Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
requirements.txt Initial vllm-gaudi commit 2025-06-25 19:19:03 +03:00
setup.py Refactor vllm_hpu -> vllm_gaudi (#2) 2025-07-11 13:11:56 +02:00

README.md

[!IMPORTANT]
This is an early developer preview of the vLLM Gaudi Plugin and is not yet intended for general use. For a more stable experience, consider using the HabanaAI/vllm-fork or the in-tree Gaudi implementation available in vllm-project/vllm.

Welcome to vLLM x Intel Gaudi

vLLM x Intel-Gaudi

vLLM Gaudi plugin (vllm-gaudi) integrates Intel Gaudi accelerators with vLLM to optimize large language model inference.

This plugin follows the [RFC]: Hardware pluggable and [RFC]: Enhancing vLLM Plugin Architecture principles, providing a modular interface for Intel Gaudi hardware.

Learn more:

📚 Intel Gaudi Documentation
🚀 vLLM Plugin System Overview

Getting Started

  1. Install vLLM with pip or from source:

    pip install vllm
    

    or

    # Build vLLM from source for empty platform, reusing existing torch installation
    git clone https://github.com/vllm-project/vllm
    cd vllm
    pip install -r <(sed '/^[torch]/d' requirements/build.txt)
    VLLM_TARGET_DEVICE=empty pip install --no-build-isolation -e .
    cd ..
    
  2. Install vLLM-Gaudi from source:

    git clone https://github.com/vllm-project/vllm-gaudi
    cd vllm-gaudi
    pip install -e .
    

Full installation from source (vLLM and vLLM-Gaudi):

# Build vLLM from source for empty platform, reusing existing torch installation
git clone https://github.com/vllm-project/vllm
cd vllm
pip install -r <(sed '/^[torch]/d' requirements/build.txt)
VLLM_TARGET_DEVICE=empty pip install --no-build-isolation -e .
cd ..

# Build vLLM-Gaudi from source
git clone https://github.com/vllm-project/vllm-gaudi
cd vllm-gaudi
pip install -e .