mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-10 07:15:01 +08:00
[Bugfix] Fix ReplicatedLinearWithLoRA (#27065)
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
This commit is contained in:
parent
fe3b9372ad
commit
87bc0c492f
@ -56,3 +56,15 @@ class ReplicatedLinearWithLoRA(BaseLinearLayerWithLoRA):
|
|||||||
model_config: PretrainedConfig | None,
|
model_config: PretrainedConfig | None,
|
||||||
) -> bool:
|
) -> bool:
|
||||||
return type(source_layer) is ReplicatedLinear
|
return type(source_layer) is ReplicatedLinear
|
||||||
|
|
||||||
|
def slice_lora_a(
|
||||||
|
self, lora_a: torch.Tensor | list[torch.Tensor | None]
|
||||||
|
) -> torch.Tensor | list[torch.Tensor | None]:
|
||||||
|
"""Slice lora a if splitting for tensor parallelism."""
|
||||||
|
return lora_a
|
||||||
|
|
||||||
|
def slice_lora_b(
|
||||||
|
self, lora_b: torch.Tensor | list[torch.Tensor | None]
|
||||||
|
) -> torch.Tensor | list[torch.Tensor | None]:
|
||||||
|
"""Slice lora b if splitting with tensor parallelism."""
|
||||||
|
return lora_b
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user