This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-01-16 09:04:30 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
tests
/
v1
/
attention
History
Chen Zhang
1e50f1be70
[Deepseek v3.2] Support indexer prefill chunking (
#25999
)
...
Signed-off-by: Chen Zhang <zhangch99@outlook.com>
2025-10-02 10:29:12 -07:00
..
test_attention_backends_selection.py
[Attention] Unify mamba and attention backend selection (
#23171
)
2025-08-25 09:09:36 +00:00
test_attention_backends.py
[V0 deprecation] Remove _VLLM_V1 suffixes from attention backend names (
#25489
)
2025-09-25 17:37:50 +00:00
test_attention_splitting.py
[Bug] fix import and unit test (
#25558
)
2025-09-24 10:17:59 +00:00
test_chunked_local_attention.py
fix some typos (
#24071
)
2025-09-02 20:44:50 -07:00
test_mla_backends.py
[New Model] DeepSeek-V3.2 (Rebased to Main) (
#25896
)
2025-09-30 17:14:41 +08:00
test_sparse_mla_backends.py
[Deepseek v3.2] Support indexer prefill chunking (
#25999
)
2025-10-02 10:29:12 -07:00
utils.py
[New Model] DeepSeek-V3.2 (Rebased to Main) (
#25896
)
2025-09-30 17:14:41 +08:00