This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2025-12-24 15:25:01 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
attention
/
ops
History
Lucas Wilkinson
2d9ee28cab
[CI/Test Fix] Fix CP tests on Blackwell (
#29338
)
...
Signed-off-by: Lucas Wilkinson <lwilkins@redhat.com>
2025-11-24 20:55:57 -08:00
..
__init__.py
…
chunked_prefill_paged_decode.py
…
common.py
[CI/Test Fix] Fix CP tests on Blackwell (
#29338
)
2025-11-24 20:55:57 -08:00
flashmla.py
…
merge_attn_states.py
…
paged_attn.py
…
pallas_kv_cache_update.py
…
prefix_prefill.py
[Misc] Remove unused attention prefix prefill ops functions (
#26971
)
2025-11-11 18:26:04 +00:00
rocm_aiter_mla_sparse.py
[ROCm] Add AMD GPU support on Deepseek v3.2 and SparseMLA (
#26670
)
2025-11-20 02:54:01 -08:00
rocm_aiter_paged_attn.py
…
triton_decode_attention.py
…
triton_merge_attn_states.py
…
triton_reshape_and_cache_flash.py
[Bugfix][CI/Test][Spec Decode] Fix illegal memory access in offline_inference/spec_decode.py (Issue 27619) (
#28432
)
2025-11-13 22:34:01 -08:00
triton_unified_attention.py
…
vit_attn_wrappers.py
[Core] Deprecate
xformers
(
#29262
)
2025-11-24 04:18:55 +00:00