mirror of https://github.com/vllm-project/vllm.git
1.1 KiB
1.1 KiB
(serve-args)=
Server Arguments
The vllm serve
command is used to launch the OpenAI-compatible server.
CLI Arguments
The following are all arguments available from the vllm serve
command:
.. argparse::
:module: vllm.entrypoints.openai.cli_args
:func: create_parser_for_docs
:prog: vllm serve
:nodefaultconst:
:markdownhelp:
Configuration file
You can load CLI arguments via a YAML config file. The argument names must be the long form of those outlined above.
For example:
# config.yaml
model: meta-llama/Llama-3.1-8B-Instruct
host: "127.0.0.1"
port: 6379
uvicorn-log-level: "info"
To use the above config file:
vllm serve --config config.yaml
:::{note}
In case an argument is supplied simultaneously using command line and the config file, the value from the command line will take precedence.
The order of priorities is command line > config file values > defaults
.
e.g. vllm serve SOME_MODEL --config config.yaml
, SOME_MODEL takes precedence over model
in config file.
:::