This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-01-01 06:08:43 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
entrypoints
History
Jiaxin Shan
db3bf7c991
[Core] Support load and unload LoRA in api server (
#6566
)
...
Co-authored-by: Jee Jee Li <pandaleefree@gmail.com>
2024-09-05 18:10:33 -07:00
..
openai
[Core] Support load and unload LoRA in api server (
#6566
)
2024-09-05 18:10:33 -07:00
__init__.py
Change the name to vLLM (
#150
)
2023-06-17 03:07:40 -07:00
api_server.py
[mypy] Enable following imports for entrypoints (
#7248
)
2024-08-20 23:28:21 -07:00
chat_utils.py
[MODEL] Qwen Multimodal Support (Qwen-VL / Qwen-VL-Chat) (
#8029
)
2024-09-05 12:48:10 +00:00
launcher.py
[Bugfix][Frontend] Fix Issues Under High Load With
zeromq
Frontend (
#7394
)
2024-08-21 13:34:14 -04:00
llm.py
[Frontend] Multimodal support in offline chat (
#8098
)
2024-09-04 05:22:17 +00:00
logger.py
[Frontend] Refactor prompt processing (
#4028
)
2024-07-22 10:13:53 -07:00