From 1d2c53973401b67c07d0fd23c98de16adce19409 Mon Sep 17 00:00:00 2001 From: bk-201 Date: Wed, 10 Dec 2025 16:30:49 +0000 Subject: [PATCH] address ci issue Signed-off-by: bk-201 --- vllm/lora/model_manager.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/vllm/lora/model_manager.py b/vllm/lora/model_manager.py index 7c0903674ce53..a1270f032d2b8 100644 --- a/vllm/lora/model_manager.py +++ b/vllm/lora/model_manager.py @@ -26,6 +26,7 @@ from vllm.lora.utils import ( from_layer_logits_processor, get_supported_lora_modules, is_moe_model, + process_packed_modules_mapping, replace_submodule, ) from vllm.model_executor.layers.fused_moe import FusedMoE @@ -33,7 +34,6 @@ from vllm.model_executor.models import SupportsLoRA, supports_multimodal from vllm.model_executor.models.interfaces import is_pooling_model from vllm.model_executor.models.module_mapping import MultiModelKeys from vllm.model_executor.models.utils import PPMissingLayer -from vllm.model_executor.utils import get_packed_modules_mapping from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.utils.cache import LRUCache from vllm.utils.platform_utils import is_pin_memory_available @@ -102,7 +102,7 @@ class LoRAModelManager: assert self.supported_lora_modules, "No supported LoRA modules found in" f" {self.model.__class__.__name__}." - self.packed_modules_mapping = get_packed_modules_mapping(self.model) + self.packed_modules_mapping = process_packed_modules_mapping(self.model) self._init_multimodal_config(vllm_config) self.is_pooling_model = is_pooling_model(self.model) self.packed_modules: dict[str, list[str]] = {}