mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-04-21 15:07:13 +08:00
fix(config): validate skip_tokenizer_init is not used with multimodal models
Add early validation to detect when skip_tokenizer_init=True is used with multimodal models like Gemma3. This combination is not supported because multimodal processors require a tokenizer for initialization. Previously, this would cause a confusing AttributeError: 'NoneType' object has no attribute 'image_token_id' deep in the transformers Gemma3Processor initialization. Now users get a clear error message explaining the incompatibility. Fixes #31123 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com> Signed-off-by: yurekami <yurekami@users.noreply.github.com>
This commit is contained in:
parent
7cd288a4b3
commit
16be10168c
@ -588,6 +588,19 @@ class ModelConfig:
|
||||
"repo name or path using the --tokenizer argument."
|
||||
)
|
||||
|
||||
# Multimodal models require a tokenizer for processor initialization
|
||||
# unless embedded inputs are enabled (enable_mm_embeds=True)
|
||||
if self.skip_tokenizer_init and self.is_multimodal_model:
|
||||
mm_config = getattr(self, "multimodal_config", None)
|
||||
if mm_config is None or not mm_config.enable_mm_embeds:
|
||||
raise ValueError(
|
||||
"Multimodal models require a tokenizer for processing. "
|
||||
"Please set skip_tokenizer_init=False when using multimodal "
|
||||
f"models like {self.model}. Alternatively, enable embedded "
|
||||
"inputs with enable_mm_embeds=True if your inputs are "
|
||||
"pre-embedded."
|
||||
)
|
||||
|
||||
if self.disable_sliding_window:
|
||||
# Set after get_and_verify_max_len to ensure that max_model_len
|
||||
# can be correctly capped to sliding window size
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user