Fix weight loading for some models in Transformers backend (#15544)

Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
This commit is contained in:
Harry Mellor 2025-03-26 17:17:53 +00:00 committed by GitHub
parent e64afa455c
commit 0af4d764d6
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -345,9 +345,11 @@ class TransformersModel(nn.Module):
params_dict = dict(self.named_parameters())
loaded_params = set[str]()
for name, loaded_weight in weights:
# Necessary for some models which use remote code
if not name.startswith(prefix := self.model.base_model_prefix):
name = maybe_prefix(prefix, name)
# Use "model" instead of base_model_prefix because
# the base model attribute in vLLM is always `model`
if not name.startswith(prefix := "model."):
name = prefix + name
if is_pp_missing_parameter(name, self):
continue
if name in params_dict: