From a8d2e326ecb70876aa73dce70dbe2434c64b710a Mon Sep 17 00:00:00 2001 From: Roger Wang Date: Mon, 27 Oct 2025 17:48:32 -0700 Subject: [PATCH] [Bugfix][CI] Fix config resolving logic with remote models (#27610) --- vllm/transformers_utils/config.py | 11 ++++++++--- 1 file changed, 8 insertions(+), 3 deletions(-) diff --git a/vllm/transformers_utils/config.py b/vllm/transformers_utils/config.py index 13de5939356e..34c0429a8067 100644 --- a/vllm/transformers_utils/config.py +++ b/vllm/transformers_utils/config.py @@ -622,9 +622,14 @@ def get_config( # Architecture mapping for models without explicit architectures field if not config.architectures: if config.model_type not in MODEL_MAPPING_NAMES: - raise ValueError(f"Cannot find architecture name for {config.model_type}") - model_type = MODEL_MAPPING_NAMES[config.model_type] - config.update({"architectures": [model_type]}) + logger.warning( + "Model config does not have a top-level 'architectures' field: " + "expecting `hf_overrides={'architectures': ['...']}` to be passed " + "in engine args." + ) + else: + model_type = MODEL_MAPPING_NAMES[config.model_type] + config.update({"architectures": [model_type]}) # ModelOpt 0.31.0 and after saves the quantization config in the model # config file.