From 0b343e32189bd22a893d0b2920914f0028a47f04 Mon Sep 17 00:00:00 2001 From: JJJYmmm <92386084+JJJYmmm@users.noreply.github.com> Date: Mon, 29 Sep 2025 01:56:12 +0800 Subject: [PATCH] [Bugfix] fix Qwen3VLMoe load when pp > 1 (#25838) Signed-off-by: liuye.hj Co-authored-by: liuye.hj Signed-off-by: yewentao256 --- vllm/model_executor/models/qwen3_vl_moe.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/vllm/model_executor/models/qwen3_vl_moe.py b/vllm/model_executor/models/qwen3_vl_moe.py index 52ea652b3765a..02cc5d6d66d15 100644 --- a/vllm/model_executor/models/qwen3_vl_moe.py +++ b/vllm/model_executor/models/qwen3_vl_moe.py @@ -212,6 +212,8 @@ class Qwen3MoeLLMModel(Qwen3MoeModel): # attempted to load as other weights later is_expert_weight = True name_mapped = name.replace(weight_name, param_name) + if is_pp_missing_parameter(name_mapped, self): + continue if is_fused_expert: loaded_weight = loaded_weight.transpose(-1, -2) # no bias @@ -230,8 +232,6 @@ class Qwen3MoeLLMModel(Qwen3MoeModel): name_mapped, params_dict, loaded_weight, shard_id, num_experts) else: - if is_pp_missing_parameter(name_mapped, self): - continue # Skip loading extra parameters for GPTQ/modelopt models if name_mapped.endswith( ignore_suffixes