This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-04-09 12:17:04 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
model_executor
History
Woosuk Kwon
6f038fc4fb
Merge branch 'main' into woosuk/model-runner-v2
2025-09-19 20:30:04 +00:00
..
layers
[Perf] Use FlashInfer RoPE for RotaryEmbedding.forward_cuda when available (
#21126
)
2025-09-19 14:06:49 -06:00
model_loader
Move
ModelConfig
from
config/__init__.py
to
config/model.py
(
#25252
)
2025-09-19 16:22:33 +00:00
models
[Docs] add __init__.py to vllm/model_executor/layers/quantization/compressed_tensors/transform (
#24974
)
2025-09-19 18:32:27 +00:00
warmup
merge
2025-09-18 14:53:32 -07:00
__init__.py
[Misc] Add SPDX-FileCopyrightText (
#19100
)
2025-06-03 11:20:17 -07:00
custom_op.py
[V0 deprecation] Deprecate V0 Neuron backend (
#21159
)
2025-09-06 16:15:18 -07:00
parameter.py
[OOT] Support sync_model_loading for OOT (
#25126
)
2025-09-19 05:41:53 +00:00
sampling_metadata.py
[Doc]: fix typos in Python comments (
#24042
)
2025-09-01 19:07:45 -07:00
utils.py
[OOT] Support sync_model_loading for OOT (
#25126
)
2025-09-19 05:41:53 +00:00