This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2025-12-09 23:45:54 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
tests
/
v1
/
e2e
History
Yannick Schnider
f05fea1f5e
[Core] Enable decode of context length equal to max model length (
#26168
)
...
Signed-off-by: Yannick Schnider <yannick.schnider1@ibm.com>
2025-10-04 09:59:26 +00:00
..
__init__.py
[V1] Implement Cascade Attention (
#11635
)
2025-01-01 21:56:46 +09:00
test_cascade_attention.py
[V0 deprecation] Remove _VLLM_V1 suffixes from attention backend names (
#25489
)
2025-09-25 17:37:50 +00:00
test_context_length.py
[Core] Enable decode of context length equal to max model length (
#26168
)
2025-10-04 09:59:26 +00:00
test_correctness_sliding_window.py
[CI] Revert back prepare_prompts and check_answers (
#25087
)
2025-09-17 11:03:16 -07:00
test_kv_sharing_fast_prefill.py
Revert gemma3n fast prefill changes (
#23897
)
2025-08-29 12:16:57 -07:00
test_min_tokens.py
[CI] Add end-to-end V1 min_tokens test coverage (
#22495
)
2025-08-21 22:04:07 -06:00
test_spec_decode.py
[Multimodal][Speculative Decoding]Eagle Eagle3 mm support, enablement on qwen2.5vl (
#22872
)
2025-09-27 03:35:47 +00:00