This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2025-12-25 10:56:31 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
docs
/
source
/
serving
History
Cyrus Leung
06386a64dd
[Frontend] Chat-based Embeddings API (
#9759
)
2024-11-01 08:13:35 +00:00
..
compatibility_matrix.rst
[Bugfix][Frontend] Reject guided decoding in multistep mode (
#9892
)
2024-11-01 01:09:46 +00:00
deploying_with_bentoml.rst
…
deploying_with_cerebrium.rst
…
deploying_with_docker.rst
…
deploying_with_dstack.rst
…
deploying_with_k8s.rst
…
deploying_with_kserve.rst
…
deploying_with_lws.rst
…
deploying_with_nginx.rst
[Hardware][Intel CPU][DOC] Update docs for CPU backend (
#6212
)
2024-10-22 10:38:04 -07:00
deploying_with_triton.rst
…
distributed_serving.rst
[doc] update pp support (
#9853
)
2024-10-30 13:36:51 -07:00
env_vars.rst
…
faq.rst
…
integrations.rst
…
metrics.rst
…
openai_compatible_server.md
[Frontend] Chat-based Embeddings API (
#9759
)
2024-11-01 08:13:35 +00:00
run_on_sky.rst
[Misc][OpenAI] deprecate max_tokens in favor of new max_completion_tokens field for chat completion endpoint (
#9837
)
2024-10-30 18:15:56 -07:00
serving_with_langchain.rst
…
serving_with_llamaindex.rst
…
tensorizer.rst
[Doc]: Update tensorizer docs to include vllm[tensorizer] (
#7889
)
2024-10-22 15:43:25 -07:00
usage_stats.md
…