[Model] Add missing prefix to glm4_1v (#22716)

Signed-off-by: zRzRzRzRzRzRzR <2448370773@qq.com>
This commit is contained in:
Yuxuan Zhang 2025-08-13 16:23:33 +08:00 committed by GitHub
parent d16aa3dae4
commit 9e7e5baaa8
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -453,25 +453,30 @@ class Glm4vPatchMerger(nn.Module):
context_dim: int,
quant_config: Optional[QuantizationConfig] = None,
bias: bool = False,
prefix: str = "",
) -> None:
super().__init__()
self.hidden_size = d_model
self.proj = ColumnParallelLinear(self.hidden_size,
self.hidden_size,
bias=bias,
gather_output=True)
gather_output=True,
quant_config=quant_config,
prefix=f"{prefix}.proj")
self.post_projection_norm = nn.LayerNorm(self.hidden_size)
self.gate_up_proj = MergedColumnParallelLinear(
input_size=self.hidden_size,
output_sizes=[context_dim] * 2,
bias=bias,
quant_config=quant_config,
prefix=f"{prefix}.gate_up_proj",
)
self.down_proj = RowParallelLinear(
context_dim,
self.hidden_size,
bias=bias,
quant_config=quant_config,
prefix=f"{prefix}.down_proj",
)
self.act_fn = SiluAndMul()
self.extra_activation_func = nn.GELU()
@ -661,6 +666,7 @@ class Glm4vVisionTransformer(nn.Module):
context_dim=vision_config.intermediate_size,
quant_config=quant_config,
bias=False,
prefix=f"{prefix}.merger",
)
self.embeddings = Glm4vVisionEmbeddings(vision_config)