mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-24 14:06:32 +08:00
Fix nemotron_nas intermediate_size computation (#30795)
Signed-off-by: Grzegorz Karch <gkarch@nvidia.com>
This commit is contained in:
parent
c0a88df7f7
commit
f5db6385a1
@ -169,10 +169,13 @@ class DeciLMDecoderLayer(nn.Module):
|
||||
self.input_layernorm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps)
|
||||
|
||||
if not self._is_no_op_ffn:
|
||||
ffn_mult = block_config.ffn.ffn_mult
|
||||
intermediate_size = _ffn_mult_to_intermediate_size(
|
||||
ffn_mult, config.hidden_size
|
||||
)
|
||||
if hasattr(block_config.ffn, "ffn_mult"):
|
||||
ffn_mult = block_config.ffn.ffn_mult
|
||||
intermediate_size = _ffn_mult_to_intermediate_size(
|
||||
ffn_mult, config.hidden_size
|
||||
)
|
||||
else:
|
||||
intermediate_size = block_config.ffn.intermediate_size
|
||||
|
||||
self.mlp = LlamaMLP(
|
||||
hidden_size=self.hidden_size,
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user