[Bugfix] Fix Tensor Parallelism Padding Consistency in Granite Models (#20843)

Signed-off-by: Alex-Brooks <Alex.Brooks@ibm.com>
This commit is contained in:
Alex Brooks 2025-07-12 07:11:30 -06:00 committed by GitHub
parent 2c11a738b3
commit c2a2f19aba
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -273,6 +273,10 @@ class GraniteModel(nn.Module):
self.vocab_size,
config.hidden_size,
org_num_embeddings=config.vocab_size,
padding_size=DEFAULT_VOCAB_PADDING_SIZE
# We need bigger padding if using lora for kernel
# compatibility
if not lora_config else lora_config.lora_vocab_padding_size,
quant_config=quant_config,
)
else: