mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-15 18:16:46 +08:00
Fix weight loading for some models in Transformers backend (#15544)
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
This commit is contained in:
parent
e64afa455c
commit
0af4d764d6
@ -345,9 +345,11 @@ class TransformersModel(nn.Module):
|
|||||||
params_dict = dict(self.named_parameters())
|
params_dict = dict(self.named_parameters())
|
||||||
loaded_params = set[str]()
|
loaded_params = set[str]()
|
||||||
for name, loaded_weight in weights:
|
for name, loaded_weight in weights:
|
||||||
# Necessary for some models which use remote code
|
# Use "model" instead of base_model_prefix because
|
||||||
if not name.startswith(prefix := self.model.base_model_prefix):
|
# the base model attribute in vLLM is always `model`
|
||||||
name = maybe_prefix(prefix, name)
|
if not name.startswith(prefix := "model."):
|
||||||
|
name = prefix + name
|
||||||
|
|
||||||
if is_pp_missing_parameter(name, self):
|
if is_pp_missing_parameter(name, self):
|
||||||
continue
|
continue
|
||||||
if name in params_dict:
|
if name in params_dict:
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user