This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-01-22 11:14:28 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
model_executor
History
Jiangyun Zhu
8df98c2161
[perf] Enable concurrent execution of "shared_experts" and "selected_experts" in qwen3-next (
#27578
)
...
Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
2025-10-29 08:12:54 +00:00
..
layers
[Bugfix] Fix non-contiguous tensor error in
rocm_unquantized_gemm_impl
(
#27605
)
2025-10-29 00:00:15 -07:00
model_loader
[Chore]:Extract math and argparse utilities to separate modules (
#27188
)
2025-10-26 04:03:32 -07:00
models
[perf] Enable concurrent execution of "shared_experts" and "selected_experts" in qwen3-next (
#27578
)
2025-10-29 08:12:54 +00:00
warmup
[Bugfix] Fix gpt-oss w4a8 DP/EP on B200 (
#26729
)
2025-10-21 01:51:14 -04:00
__init__.py
…
custom_op.py
…
parameter.py
[Docs] Replace
rst
style double-backtick with
md
single-backtick (
#27091
)
2025-10-17 02:47:34 -07:00
utils.py
[Chore] Clean up pytorch helper functions in
vllm.utils
(
#26908
)
2025-10-18 09:48:22 -07:00