From 948ab03e7ed72d3715f8b7b6398d1d5bb0f9a296 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E2=84=8D=F0=9D=95=A0=F0=9D=95=9D=F0=9D=95=9D=F0=9D=95=A0?= =?UTF-8?q?=F0=9D=95=A8=20=F0=9D=95=84=F0=9D=95=92=F0=9D=95=9F?= Date: Mon, 24 Mar 2025 12:33:12 +0200 Subject: [PATCH] [Bugfix][V1] Avoid importing PreTrainedModel (#15366) Signed-off-by: Hollow Man --- vllm/model_executor/model_loader/utils.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/model_executor/model_loader/utils.py b/vllm/model_executor/model_loader/utils.py index 0b4872827319b..ce90614329725 100644 --- a/vllm/model_executor/model_loader/utils.py +++ b/vllm/model_executor/model_loader/utils.py @@ -32,7 +32,7 @@ def set_default_torch_dtype(dtype: torch.dtype): def is_transformers_impl_compatible( arch: str, - module: Optional[transformers.PreTrainedModel] = None) -> bool: + module: Optional["transformers.PreTrainedModel"] = None) -> bool: mod = module or getattr(transformers, arch, None) if mod is None: return False