mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-24 01:55:02 +08:00
fix(gguf): Disable bfloat16 for GGUF on blackwell device (#30408)
Signed-off-by: Christina <truffle@gmail.com> Signed-off-by: Isotr0py <2037008807@qq.com> Signed-off-by: Christina Norman <christina@example.com> Co-authored-by: Isotr0py <isotr0py@users.noreply.github.com> Co-authored-by: Isotr0py <2037008807@qq.com> Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
This commit is contained in:
parent
3e34adcdfb
commit
dc13c99eed
@ -33,6 +33,7 @@ from vllm.model_executor.layers.vocab_parallel_embedding import (
|
|||||||
)
|
)
|
||||||
from vllm.model_executor.models.utils import WeightsMapper
|
from vllm.model_executor.models.utils import WeightsMapper
|
||||||
from vllm.model_executor.utils import set_weight_attrs
|
from vllm.model_executor.utils import set_weight_attrs
|
||||||
|
from vllm.platforms import current_platform
|
||||||
from vllm.utils.torch_utils import direct_register_custom_op
|
from vllm.utils.torch_utils import direct_register_custom_op
|
||||||
|
|
||||||
logger = init_logger(__name__)
|
logger = init_logger(__name__)
|
||||||
@ -52,6 +53,11 @@ class GGUFConfig(QuantizationConfig):
|
|||||||
return "gguf"
|
return "gguf"
|
||||||
|
|
||||||
def get_supported_act_dtypes(self) -> list[torch.dtype]:
|
def get_supported_act_dtypes(self) -> list[torch.dtype]:
|
||||||
|
# GGUF dequantization kernels use half precision (fp16) internally.
|
||||||
|
# bfloat16 has precision issues on Blackwell devices.
|
||||||
|
if current_platform.has_device_capability(100):
|
||||||
|
logger.warning_once("GGUF has precision issues with bfloat16 on Blackwell.")
|
||||||
|
return [torch.half, torch.float32]
|
||||||
return [torch.half, torch.bfloat16, torch.float32]
|
return [torch.half, torch.bfloat16, torch.float32]
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user