This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-04-08 06:37:02 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
tests
/
models
/
multimodal
History
Isotr0py
55f1fc1b1b
[v1] Add PrefixLM support to TritonAttention backend (
#30386
)
...
(cherry picked from commit 74a1ac38b00a8cf502db085d1bbd77712cf47e41)
2025-12-17 19:57:52 -08:00
..
generation
[v1] Add PrefixLM support to TritonAttention backend (
#30386
)
2025-12-17 19:57:52 -08:00
pooling
Support tokenization_kwargs override (
#29794
)
2025-12-06 09:12:53 +00:00
processing
[Core][MM] Optimize encoder cache manager by operating with embeddings only (
#30475
)
2025-12-16 17:15:49 -08:00
__init__.py
…
test_mapping.py
Revert "[Renderer] Separate out
RendererConfig
from
ModelConfig
(
#30145
)" (
#30199
)
2025-12-07 00:00:22 -08:00