From 3cdc57669f231a2c2bdbebb0480164d10cf89edb Mon Sep 17 00:00:00 2001 From: Jee Jee Li Date: Sat, 12 Apr 2025 19:21:37 +0800 Subject: [PATCH] [Misc] Delete redundant code (#16530) Signed-off-by: Jee Jee Li Co-authored-by: Isotr0py --- examples/offline_inference/vision_language.py | 7 ------- examples/offline_inference/vision_language_multi_image.py | 7 ------- 2 files changed, 14 deletions(-) diff --git a/examples/offline_inference/vision_language.py b/examples/offline_inference/vision_language.py index c0799bde61807..f51cef95e8596 100644 --- a/examples/offline_inference/vision_language.py +++ b/examples/offline_inference/vision_language.py @@ -1096,13 +1096,6 @@ def main(args): } llm = LLM(**engine_args) - # To maintain code compatibility in this script, we add LoRA here. - # You can also add LoRA using: - # llm.generate(prompts, lora_request=lora_request,...) - if req_data.lora_requests: - for lora_request in req_data.lora_requests: - llm.llm_engine.add_lora(lora_request=lora_request) - # Don't want to check the flag multiple times, so just hijack `prompts`. prompts = req_data.prompts if args.use_different_prompt_per_request else [ req_data.prompts[0] diff --git a/examples/offline_inference/vision_language_multi_image.py b/examples/offline_inference/vision_language_multi_image.py index 7aff5fd07a355..89818f8b33ee6 100644 --- a/examples/offline_inference/vision_language_multi_image.py +++ b/examples/offline_inference/vision_language_multi_image.py @@ -695,13 +695,6 @@ def run_chat(model: str, question: str, image_urls: list[str], engine_args = asdict(req_data.engine_args) | {"seed": seed} llm = LLM(**engine_args) - # To maintain code compatibility in this script, we add LoRA here. - # You can also add LoRA using: - # llm.generate(prompts, lora_request=lora_request,...) - if req_data.lora_requests: - for lora_request in req_data.lora_requests: - llm.llm_engine.add_lora(lora_request=lora_request) - sampling_params = SamplingParams(temperature=0.0, max_tokens=256, stop_token_ids=req_data.stop_token_ids)