This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-03-26 05:35:51 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
tests
/
kernels
/
attention
History
elvischenv
58b11b24a6
[Bugfix] Fix workspace buffer None issue for Flashinfer TRTLLM Backend (
#21525
)
...
Signed-off-by: elvischenv <219235043+elvischenv@users.noreply.github.com>
2025-07-29 10:34:00 -04:00
..
conftest.py
…
test_aiter_flash_attn.py
[ROCm][AITER] Enable fp8 kv cache on rocm aiter backend. (
#20295
)
2025-07-25 06:50:21 -07:00
test_attention_selector.py
[V1] Support any head size for FlexAttention backend (
#20467
)
2025-07-06 09:54:36 -07:00
test_attention.py
…
test_cache.py
…
test_cascade_flash_attn.py
…
test_encoder_decoder_attn.py
…
test_flash_attn.py
…
test_flashinfer_trtllm_decode_attention.py
[Bugfix] Fix workspace buffer None issue for Flashinfer TRTLLM Backend (
#21525
)
2025-07-29 10:34:00 -04:00
test_flashinfer.py
[Misc] Add sliding window to flashinfer test (
#21282
)
2025-07-21 08:37:49 -07:00
test_flashmla.py
…
test_lightning_attn.py
…
test_merge_attn_states.py
…
test_mha_attn.py
…
test_mla_decode_cpu.py
…
test_prefix_prefill.py
…
test_rocm_attention_selector.py
[V0 Deprecation] Deprecate BlockSparse Attention & Phi3-Small (
#21217
)
2025-07-19 13:53:17 -07:00
test_triton_decode_attention.py
…
test_triton_unified_attention.py
…