From c7b370c603daf287e27d852ac36a7347536b01c6 Mon Sep 17 00:00:00 2001 From: Ning Xie Date: Thu, 19 Jun 2025 13:57:35 +0800 Subject: [PATCH] raise exception for pin_lora (#19809) Signed-off-by: Andy Xie --- vllm/worker/worker_base.py | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/vllm/worker/worker_base.py b/vllm/worker/worker_base.py index 0b37caa71669c..200026dc7282f 100644 --- a/vllm/worker/worker_base.py +++ b/vllm/worker/worker_base.py @@ -202,8 +202,7 @@ class LoRANotSupportedWorkerBase(WorkerBase): raise ValueError(f"{type(self)} does not support LoRA") def pin_lora(self, lora_id: int) -> bool: - return ValueError( - f"{type(self)} does not support LoRA") # type: ignore + raise ValueError(f"{type(self)} does not support LoRA") def list_loras(self) -> Set[int]: raise ValueError(f"{type(self)} does not support LoRA") @@ -398,7 +397,7 @@ class LocalOrDistributedWorkerBase(WorkerBase): model_input, worker_input, kwargs = inputs num_steps = worker_input.num_steps - if (execute_model_req is not None and execute_model_req.spec_step_idx): + if execute_model_req is not None and execute_model_req.spec_step_idx: kwargs["spec_step_idx"] = execute_model_req.spec_step_idx self.execute_worker(worker_input)