Squelch MLA warning for Compressed-Tensors Models (#12704)

Signed-off-by: Kyle Sayers <kylesayrs@gmail.com>
This commit is contained in:
Kyle Sayers 2025-02-03 16:29:56 -05:00 committed by GitHub
parent c11de33dad
commit 6dd5e52823
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -986,6 +986,9 @@ class ModelConfig:
@property
def use_mla(self) -> bool:
if not self.is_deepseek_mla or envs.VLLM_MLA_DISABLE:
return False
if self.quantization is not None and self.quantization not in [\
"fp8", "compressed-tensors"]:
logger.warning(
@ -1012,8 +1015,7 @@ class ModelConfig:
quant_config)
return False
use_mla = (self.is_deepseek_mla and not envs.VLLM_MLA_DISABLE)
return use_mla
return True
@property
def supported_runner_types(self) -> Set[RunnerType]: