This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2025-12-24 17:16:20 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
attention
History
Chengji Yao
7da296be04
[TPU] kv cache update kernel supports dynamic grid (
#20235
)
...
Signed-off-by: Chengji Yao <chengjiyao@google.com>
2025-07-02 06:33:37 +00:00
..
backends
[Kernel] mark TorchSDPABackend swap_blocks NotImplementedError (
#19749
)
2025-06-20 18:18:11 +00:00
ops
[TPU] kv cache update kernel supports dynamic grid (
#20235
)
2025-07-02 06:33:37 +00:00
utils
Quick Fix by adding conditional import for flash_attn_varlen_func in flash_attn (
#20143
)
2025-06-27 05:48:13 +00:00
__init__.py
[Misc] Add SPDX-FileCopyrightText (
#19100
)
2025-06-03 11:20:17 -07:00
layer.py
[Bugfix][V1][ROCm] Fix AITER Flash Attention Backend (Fix API Break and Local Attention Logic: affecting Llama4) (
#19904
)
2025-06-26 12:42:31 +00:00
selector.py
[Misc] Add SPDX-FileCopyrightText (
#19100
)
2025-06-03 11:20:17 -07:00