This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-05-13 16:26:46 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
attention
History
Huamin Li
0cd103e7cb
CP: make correct_attn_out robust to 4‑D views and fix Triton arg binding (
#26509
)
...
Signed-off-by: Huamin Li <3ericli@gmail.com>
2025-10-11 20:50:57 +00:00
..
backends
[Hybrid]: Decouple Kernel Block Size from KV Page Size (
#24486
)
2025-10-08 23:43:39 -07:00
layers
Fix per file ruff ignores related to typing (
#26254
)
2025-10-05 16:37:55 +00:00
ops
CP: make correct_attn_out robust to 4‑D views and fix Triton arg binding (
#26509
)
2025-10-11 20:50:57 +00:00
utils
Convert formatting to use
ruff
instead of
yapf
+
isort
(
#26247
)
2025-10-05 07:06:22 -07:00
__init__.py
Convert formatting to use
ruff
instead of
yapf
+
isort
(
#26247
)
2025-10-05 07:06:22 -07:00
layer.py
Separate MLAAttention class from Attention (
#25103
)
2025-10-08 17:11:11 -07:00
selector.py
[Bugfix] Move current_platform import to avoid python import cache. (
#16601
)
2025-10-09 10:46:19 +00:00