This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-04-04 20:07:05 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
model_executor
/
layers
/
mamba
History
tomeras91
1395461f5f
[Hybrid][torch.compile] Refactor mamba2 forward to avoid obscuring linear projections under custom op (
#28587
)
...
Signed-off-by: Tomer Asida <57313761+tomeras91@users.noreply.github.com>
2025-11-18 16:49:36 -08:00
..
ops
[Hybrid] [Kernel] Fix chunk scan kernel when BLOCK_SIZE_DSTATE > 128 (
#28295
)
2025-11-14 22:55:42 +00:00
__init__.py
…
abstract.py
…
linear_attn.py
…
mamba_mixer2.py
[Hybrid][torch.compile] Refactor mamba2 forward to avoid obscuring linear projections under custom op (
#28587
)
2025-11-18 16:49:36 -08:00
mamba_mixer.py
[V1] [Hybrid] Mamba1 Automatic Prefix Caching (
#26377
)
2025-11-02 04:16:23 -08:00
mamba_utils.py
…
short_conv.py
[Chore] Clean up pytorch helper functions in
vllm.utils
(
#26908
)
2025-10-18 09:48:22 -07:00