mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-25 22:28:44 +08:00
[Bugfix] fix adding bias twice in ipex GPTQ quantization (#18363)
Signed-off-by: rand-fly <randfly@outlook.com>
This commit is contained in:
parent
d981396778
commit
bca55b556f
@ -181,8 +181,6 @@ class IPEXGPTQLinearMethod(GPTQLinearMethod):
|
||||
bias: Optional[torch.Tensor] = None) -> torch.Tensor:
|
||||
reshaped_x = x.reshape(-1, x.shape[-1])
|
||||
out = layer.ipex_qlinear(reshaped_x)
|
||||
if bias is not None:
|
||||
out.add_(bias)
|
||||
return out.reshape(x.shape[:-1] + (layer.ipex_output_size, ))
|
||||
|
||||
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user