This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-05-11 18:05:43 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
model_executor
/
layers
/
fused_moe
History
Gregory Shtrasberg
d4b62d4641
[AMD][Build] Porting dockerfiles from the ROCm/vllm fork (
#11777
)
...
Signed-off-by: Gregory Shtrasberg <Gregory.Shtrasberg@amd.com>
2025-01-21 12:22:23 +08:00
..
configs
[AMD][Build] Porting dockerfiles from the ROCm/vllm fork (
#11777
)
2025-01-21 12:22:23 +08:00
__init__.py
[torch.compile] support moe models (
#9632
)
2024-10-27 21:58:04 -07:00
fused_marlin_moe.py
[optimization] remove python function call for custom op (
#11750
)
2025-01-07 17:04:28 +00:00
fused_moe.py
[Bugfix] Fix deepseekv3 gate bias error (
#12002
)
2025-01-13 13:43:51 -07:00
layer.py
[Hardware][TPU] workaround fix for MoE on TPU (
#11764
)
2025-01-12 10:53:51 -05:00
moe_pallas.py
[Hardware][TPU] Support MoE with Pallas GMM kernel (
#6457
)
2024-07-16 09:56:28 -07:00
moe_torch_iterative.py
[Hardware][TPU] workaround fix for MoE on TPU (
#11764
)
2025-01-12 10:53:51 -05:00