mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-09 05:24:54 +08:00
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk> Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com> Co-authored-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
10 lines
445 B
Markdown
10 lines
445 B
Markdown
# Using vLLM
|
|
|
|
First, vLLM must be [installed](../getting_started/installation/README.md) for your chosen device in either a Python or Docker environment.
|
|
|
|
Then, vLLM supports the following usage patterns:
|
|
|
|
- [Inference and Serving](../serving/offline_inference.md): Run a single instance of a model.
|
|
- [Deployment](../deployment/docker.md): Scale up model instances for production.
|
|
- [Training](../training/rlhf.md): Train or fine-tune a model.
|