This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-01-05 22:34:10 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
model_executor
History
Matthew Bonanni
2aaa423842
[Attention] Move Backend enum into registry (
#25893
)
...
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
2025-10-02 20:32:24 -07:00
..
layers
Fix MTP with deepep_low_latency (
#25904
)
2025-10-02 21:29:49 +00:00
model_loader
Support RL online quantization with torchao (
#23014
)
2025-10-01 16:39:29 -07:00
models
[Attention] Move Backend enum into registry (
#25893
)
2025-10-02 20:32:24 -07:00
warmup
[V0 deprecation] Remove _VLLM_V1 suffixes from attention backend names (
#25489
)
2025-09-25 17:37:50 +00:00
__init__.py
[V0 Deprecation] Remove V0 sampling metadata (
#25345
)
2025-09-21 10:37:11 -07:00
custom_op.py
[V0 deprecation] Deprecate V0 Neuron backend (
#21159
)
2025-09-06 16:15:18 -07:00
parameter.py
Revert "[Bug] Dynamo Unsupported due to
BasevLLMParameter.torch_function
calling disabled super()" (
#25681
)
2025-09-25 09:45:06 -07:00
utils.py
[OOT] Support sync_model_loading for OOT (
#25126
)
2025-09-19 05:41:53 +00:00