vllm/docs/source/serving/serve_args.md
Trevor Royer 55f1a468d9
Move cli args docs to its own page (#18228) (#18264)
Signed-off-by: Trevor Royer <troyer@redhat.com>
2025-05-16 19:43:45 -07:00

1.1 KiB

(serve-args)=

Server Arguments

The vllm serve command is used to launch the OpenAI-compatible server.

CLI Arguments

The following are all arguments available from the vllm serve command:

.. argparse::
    :module: vllm.entrypoints.openai.cli_args
    :func: create_parser_for_docs
    :prog: vllm serve
    :nodefaultconst:
    :markdownhelp:

Configuration file

You can load CLI arguments via a YAML config file. The argument names must be the long form of those outlined above.

For example:

# config.yaml

model: meta-llama/Llama-3.1-8B-Instruct
host: "127.0.0.1"
port: 6379
uvicorn-log-level: "info"

To use the above config file:

vllm serve --config config.yaml

:::{note} In case an argument is supplied simultaneously using command line and the config file, the value from the command line will take precedence. The order of priorities is command line > config file values > defaults. e.g. vllm serve SOME_MODEL --config config.yaml, SOME_MODEL takes precedence over model in config file. :::