This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-01-26 18:14:27 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
tests
/
kernels
History
Didier Durand
46876dff32
[Doc]: fixing typos to improve docs (
#24480
)
...
Signed-off-by: Didier Durand <durand.didier@gmail.com>
2025-09-08 23:06:04 -07:00
..
attention
[Flashinfer] Support Flashinfer TRTLLM FP8-qkv BF16/FP16-out Attention Kernel (
#23647
)
2025-09-08 20:53:07 -07:00
core
…
mamba
[Bugfix] Fix mamba2 prefill chunking (
#23279
)
2025-09-08 11:42:41 +00:00
moe
[Bugfix] Fix test_mixtral_moe (
#24371
)
2025-09-06 09:32:03 -07:00
quantization
[Bugfix] Fix unstable silu_mul+nvfp4 quant fusion test (
#24370
)
2025-09-06 20:39:34 +00:00
__init__.py
…
allclose_default.py
…
quant_utils.py
…
test_apply_repetition_penalties.py
…
test_cutlass_mla_decode.py
[Kernel] Support decode context parallelism on Blackwell with CUTLASS MLA (
#24385
)
2025-09-08 09:27:12 +08:00
test_flex_attention.py
Updates to Flex + VLLm integration (
#21416
)
2025-08-25 09:32:42 -04:00
test_fused_quant_activation.py
…
test_onednn.py
[CPU] Refactor CPU unquantized linear (
#24150
)
2025-09-04 14:28:45 +08:00
test_shuffle_rows.py
…
test_triton_flash_attention.py
…
utils.py
[Doc]: fixing typos to improve docs (
#24480
)
2025-09-08 23:06:04 -07:00