From de906b95f9d0b9669da902785a9012ac96edd578 Mon Sep 17 00:00:00 2001 From: Gregory Shtrasberg <156009573+gshtras@users.noreply.github.com> Date: Tue, 6 May 2025 15:59:06 -0400 Subject: [PATCH] [Bugfix] Fix for the condition to accept empty encoder inputs for mllama (#17732) Signed-off-by: Gregory Shtrasberg --- vllm/engine/llm_engine.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/engine/llm_engine.py b/vllm/engine/llm_engine.py index 38a20a418e215..e0f57e0b450c6 100644 --- a/vllm/engine/llm_engine.py +++ b/vllm/engine/llm_engine.py @@ -2021,7 +2021,7 @@ class LLMEngine: if not prompt_ids: if prompt_type == "encoder" and model_config.is_multimodal_model: pass # Mllama may have empty encoder inputs for text-only data - if prompt_inputs["type"] == "embeds": + elif prompt_inputs["type"] == "embeds": pass else: raise ValueError(f"The {prompt_type} prompt cannot be empty")