This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-05-12 00:04:44 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
attention
/
ops
History
Dazhi Jiang
bcb6f5947f
[Perf] Remove sync point in vit torch sdpa attn backend (
#30232
)
...
Signed-off-by: Dazhi Jiang <dazhi_jiang@163.com>
2025-12-08 07:12:42 +00:00
..
__init__.py
…
chunked_prefill_paged_decode.py
…
common.py
bugfix: correct attn output with base 2 or e (
#28840
)
2025-11-29 07:52:12 +08:00
flashmla.py
…
merge_attn_states.py
…
paged_attn.py
[V0 deprecation] Clean up legacy paged attention helper functions (
#28043
)
2025-11-28 16:44:33 +00:00
pallas_kv_cache_update.py
…
prefix_prefill.py
[Misc] Remove unused attention prefix prefill ops functions (
#26971
)
2025-11-11 18:26:04 +00:00
rocm_aiter_mla_sparse.py
[ROCm] Add AMD GPU support on Deepseek v3.2 and SparseMLA (
#26670
)
2025-11-20 02:54:01 -08:00
triton_decode_attention.py
…
triton_merge_attn_states.py
[Performance][MLA][ROCm] Remove redundant D2D copy in deepseek (
#27457
)
2025-11-26 12:45:28 +08:00
triton_reshape_and_cache_flash.py
[Bugfix][CI/Test][Spec Decode] Fix illegal memory access in offline_inference/spec_decode.py (Issue 27619) (
#28432
)
2025-11-13 22:34:01 -08:00
triton_unified_attention.py
…
vit_attn_wrappers.py
[Perf] Remove sync point in vit torch sdpa attn backend (
#30232
)
2025-12-08 07:12:42 +00:00