From 366f6b3a4d92ee0b2df8e5620a88ddf57afc3681 Mon Sep 17 00:00:00 2001 From: Cyrus Leung Date: Wed, 30 Jul 2025 23:42:05 +0800 Subject: [PATCH] [Bugfix] Fix multi-api server not working for text models (#21933) Signed-off-by: DarkLight1337 --- vllm/config.py | 15 +-------------- 1 file changed, 1 insertion(+), 14 deletions(-) diff --git a/vllm/config.py b/vllm/config.py index 8e8c1198833c2..012a791a3c872 100644 --- a/vllm/config.py +++ b/vllm/config.py @@ -856,7 +856,7 @@ class ModelConfig: self.tokenizer = s3_tokenizer.dir def _init_multimodal_config(self) -> Optional["MultiModalConfig"]: - if self.registry.is_multimodal_model(self.architectures, self): + if self._model_info.supports_multimodal: return MultiModalConfig( limit_per_prompt=self.limit_mm_per_prompt, media_io_kwargs=self.media_io_kwargs, @@ -865,19 +865,6 @@ class ModelConfig: disable_mm_preprocessor_cache, interleave_mm_strings=self.interleave_mm_strings) - if self.limit_mm_per_prompt: - raise ValueError("`limit_mm_per_prompt` is only supported for " - "multimodal models.") - if self.mm_processor_kwargs: - raise ValueError("`mm_processor_kwargs` is only supported for " - "multimodal models.") - if self.disable_mm_preprocessor_cache: - raise ValueError("`disable_mm_preprocessor_cache` is only " - "supported for multimodal models.") - if self.interleave_mm_strings: - raise ValueError("`interleave_mm_strings` is only " - "supported for multimodal models.") - return None def _get_encoder_config(self):