mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-14 05:45:01 +08:00
static_scaled_fp8_quant should not run when scale.numel is not 1 (#20076)
This commit is contained in:
parent
23a04e0895
commit
8b8c209e35
@ -1276,7 +1276,7 @@ def scaled_fp8_quant(
|
|||||||
torch.ops._C.dynamic_scaled_fp8_quant(output, input, scale)
|
torch.ops._C.dynamic_scaled_fp8_quant(output, input, scale)
|
||||||
else:
|
else:
|
||||||
# num_token_padding not implemented for this case
|
# num_token_padding not implemented for this case
|
||||||
assert (scale.numel() == 1 or num_token_padding is None)
|
assert (scale.numel() == 1 and num_token_padding is None)
|
||||||
torch.ops._C.static_scaled_fp8_quant(output, input, scale)
|
torch.ops._C.static_scaled_fp8_quant(output, input, scale)
|
||||||
|
|
||||||
return output, scale
|
return output, scale
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user