This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2025-12-10 05:04:58 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
tests
/
v1
/
spec_decode
History
Lucia Fang
001e50c92c
[Model] MTP fallback to eager for DeepSeek v32 (
#25982
)
...
Signed-off-by: Lu Fang <fanglu@fb.com>
2025-10-01 01:53:22 +00:00
..
test_eagle.py
[Model] MTP fallback to eager for DeepSeek v32 (
#25982
)
2025-10-01 01:53:22 +00:00
test_max_len.py
[V0 deprecation] Remove _VLLM_V1 suffixes from attention backend names (
#25489
)
2025-09-25 17:37:50 +00:00
test_mtp.py
[Model] MTP fallback to eager for DeepSeek v32 (
#25982
)
2025-10-01 01:53:22 +00:00
test_ngram.py
[Spec Decode] Add Batch Parallel Ngram. Upto 8x lower overhead. (
#24986
)
2025-09-25 15:22:03 -07:00
test_tree_attention.py
[V0 deprecation] Remove _VLLM_V1 suffixes from attention backend names (
#25489
)
2025-09-25 17:37:50 +00:00