[Minor] Fix small typo in llama.py: QKVParallelLinear -> QuantizationConfig (#4991)

This commit is contained in:
Philipp Moritz 2024-05-22 15:26:56 -07:00 committed by GitHub
parent eb6d3c264d
commit a36de682d4
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -57,7 +57,7 @@ class LlamaMLP(nn.Module):
hidden_size: int,
intermediate_size: int,
hidden_act: str,
quant_config: Optional[QKVParallelLinear] = None,
quant_config: Optional[QuantizationConfig] = None,
bias: bool = False,
) -> None:
super().__init__()