This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2025-12-28 06:45:16 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
tests
/
lora
History
Austin Veselka
eefeb16464
[Kernel] Full Tensor Parallelism for LoRA Layers (
#3524
)
...
Co-authored-by: Antoni Baum <antoni.baum@protonmail.com>
2024-04-27 00:03:48 -07:00
..
__init__.py
…
conftest.py
[Bugfix] Fix LoRA loading check (
#4138
)
2024-04-19 00:59:54 -07:00
test_baichuan.py
…
test_chatglm3.py
…
test_gemma.py
…
test_layer_variation.py
…
test_layers.py
[Kernel] Full Tensor Parallelism for LoRA Layers (
#3524
)
2024-04-27 00:03:48 -07:00
test_llama.py
…
test_lora_checkpoints.py
[Bugfix] Fix LoRA loading check (
#4138
)
2024-04-19 00:59:54 -07:00
test_lora_manager.py
…
test_lora.py
…
test_mixtral.py
…
test_punica.py
[Kernel] Full Tensor Parallelism for LoRA Layers (
#3524
)
2024-04-27 00:03:48 -07:00
test_quant_model.py
…
test_tokenizer_group.py
…
test_utils.py
…
test_worker.py
[Core] Refactor model loading code (
#4097
)
2024-04-16 11:34:39 -07:00
utils.py
…