This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-04-05 05:27:04 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
config
History
Li, Jiang
20852c8f4c
[CPU] Refactor CPU WNA16 (
#28826
)
...
Signed-off-by: jiang1.li <jiang1.li@intel.com>
2025-11-19 10:32:00 +08:00
..
__init__.py
…
cache.py
…
compilation.py
[BugFix] Temporary fix for IMA with MTP = 2 and full-cg (
#28315
)
2025-11-17 09:41:22 -05:00
device.py
…
ec_transfer.py
…
kv_events.py
…
kv_transfer.py
…
load.py
…
lora.py
…
model.py
[CPU] Refactor CPU WNA16 (
#28826
)
2025-11-19 10:32:00 +08:00
multimodal.py
…
observability.py
…
parallel.py
[V1] Support MP Executor for multi node distributed inference (
#23691
)
2025-11-16 09:01:21 +00:00
pooler.py
…
scheduler.py
[Misc] Make
SchedulerConfig.max_model_len
init-only (
#28733
)
2025-11-15 01:59:31 -08:00
speculative.py
[Core] Async Scheduling X Spec Decoding Compatibility (
#24799
)
2025-11-17 12:16:20 -08:00
speech_to_text.py
…
structured_outputs.py
…
utils.py
…
vllm.py
[Core] Async Scheduling X Spec Decoding Compatibility (
#24799
)
2025-11-17 12:16:20 -08:00