vllm/docs
Chanh Nguyen 7ea2adb802
[Core] Support full cuda graph in v1 (#16072)
Signed-off-by: Chanh Nguyen <cnguyen@linkedin.com>
Co-authored-by: Chanh Nguyen <cnguyen@linkedin.com>
2025-05-07 22:30:15 -07:00
..
source [Core] Support full cuda graph in v1 (#16072) 2025-05-07 22:30:15 -07:00
Makefile Add full API docs and improve the UX of navigating them (#17485) 2025-05-03 19:42:43 -07:00
README.md [doc] update doc (#15740) 2025-03-29 04:27:22 +00:00
make.bat Add initial sphinx docs (#120) 2023-05-22 17:02:44 -07:00
seed_parameter_behavior.md Fix seed parameter behavior in vLLM (#13007) 2025-02-10 23:26:50 +08:00

README.md

vLLM documents

Build the docs

  • Make sure in docs directory
cd docs
  • Install the dependencies:
pip install -r ../requirements/docs.txt
  • Clean the previous build (optional but recommended):
make clean
  • Generate the HTML documentation:
make html

Open the docs with your browser

  • Serve the documentation locally:
python -m http.server -d build/html/

This will start a local server at http://localhost:8000. You can now open your browser and view the documentation.

If port 8000 is already in use, you can specify a different port, for example:

python -m http.server 3000 -d build/html/