This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-03-19 19:07:19 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
tests
/
models
History
Harry Mellor
4b0da7b60e
Enable hybrid attention models for Transformers backend (
#18494
)
...
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
2025-05-23 10:12:08 +08:00
..
fixtures
…
language
[CI/Build] Update bamba test model location (
#18544
)
2025-05-22 06:01:07 -07:00
multimodal
Re-submit: Fix: Proper RGBA -> RGB conversion for PIL images. (
#18569
)
2025-05-23 01:59:18 +00:00
quantization
[Minor] Rename quantization nvfp4 to modelopt_fp4 (
#18356
)
2025-05-20 09:08:37 -07:00
__init__.py
…
registry.py
[CI/Build] Update bamba test model location (
#18544
)
2025-05-22 06:01:07 -07:00
test_initialization.py
…
test_oot_registration.py
Re-submit: Fix: Proper RGBA -> RGB conversion for PIL images. (
#18569
)
2025-05-23 01:59:18 +00:00
test_registry.py
…
test_transformers.py
Enable hybrid attention models for Transformers backend (
#18494
)
2025-05-23 10:12:08 +08:00
test_utils.py
[Misc] Allow
AutoWeightsLoader
to skip loading weights with specific substr in name (
#18358
)
2025-05-19 20:20:12 -07:00
test_vision.py
…
utils.py
[New Model]: nomic-embed-text-v2-moe (
#17785
)
2025-05-11 00:59:43 -07:00