From d3c2f3dbe3f99abf3677fad398ac8e40b594cfaf Mon Sep 17 00:00:00 2001 From: bk-201 Date: Wed, 10 Dec 2025 15:25:23 +0000 Subject: [PATCH] address ci issue Signed-off-by: bk-201 --- vllm/lora/model_manager.py | 10 +++++----- vllm/lora/worker_manager.py | 4 ++-- 2 files changed, 7 insertions(+), 7 deletions(-) diff --git a/vllm/lora/model_manager.py b/vllm/lora/model_manager.py index 10164654c816a..7c0903674ce53 100644 --- a/vllm/lora/model_manager.py +++ b/vllm/lora/model_manager.py @@ -65,8 +65,8 @@ class LoRAModelManager: max_num_batched_tokens: int, vocab_size: int, lora_config: LoRAConfig, - vllm_config: VllmConfig, device: torch.device, + vllm_config: VllmConfig | None = None, ): """Create a LoRAModelManager and adapter for a given model. @@ -114,7 +114,7 @@ class LoRAModelManager: self.model.lora_manager = self - def _init_multimodal_config(self, vllm_config: VllmConfig): + def _init_multimodal_config(self, vllm_config: VllmConfig | None = None): # Used to indicate whether the model is a multimodal model self.supports_mm: bool = ( supports_multimodal(self.model) @@ -125,7 +125,7 @@ class LoRAModelManager: self.supports_mm_lora = False - if self.supports_mm: + if self.supports_mm and vllm_config is not None: model_config: ModelConfig = vllm_config.model_config self.mm_mapping: MultiModelKeys = self.model.get_mm_mapping() if self.lora_config.enable_mm_lora: @@ -708,8 +708,8 @@ class LRUCacheLoRAModelManager(LoRAModelManager): max_num_batched_tokens: int, vocab_size: int, lora_config: LoRAConfig, - vllm_config: VllmConfig, device: torch.device, + vllm_config: VllmConfig | None = None, ): super().__init__( model, @@ -717,8 +717,8 @@ class LRUCacheLoRAModelManager(LoRAModelManager): max_num_batched_tokens, vocab_size, lora_config, - vllm_config, device, + vllm_config, ) self._registered_adapters: LoRALRUCache = LoRALRUCache( self.capacity, self.deactivate_adapter diff --git a/vllm/lora/worker_manager.py b/vllm/lora/worker_manager.py index 170c9ec226ab0..967ce458995c6 100644 --- a/vllm/lora/worker_manager.py +++ b/vllm/lora/worker_manager.py @@ -69,7 +69,7 @@ class WorkerLoRAManager: def create_lora_manager( self, model: torch.nn.Module, - vllm_config: VllmConfig, + vllm_config: VllmConfig | None = None, ) -> Any: lora_manager = create_lora_manager( model, @@ -212,7 +212,7 @@ class LRUCacheWorkerLoRAManager(WorkerLoRAManager): def create_lora_manager( self, model: torch.nn.Module, - vllm_config: VllmConfig, + vllm_config: VllmConfig | None = None, ) -> Any: lora_manager = create_lora_manager( model,