From c5b8b5953a2e20e8358d0828aad11d259c073c50 Mon Sep 17 00:00:00 2001 From: Jee Jee Li Date: Thu, 17 Jul 2025 13:47:49 +0800 Subject: [PATCH] [Misc] Fix PhiMoE expert mapping (#21085) Signed-off-by: Jee Jee Li --- vllm/model_executor/models/phimoe.py | 7 +------ 1 file changed, 1 insertion(+), 6 deletions(-) diff --git a/vllm/model_executor/models/phimoe.py b/vllm/model_executor/models/phimoe.py index 0fc64e88a6b63..cfe0982204fa9 100644 --- a/vllm/model_executor/models/phimoe.py +++ b/vllm/model_executor/models/phimoe.py @@ -533,14 +533,9 @@ class PhiMoEModel(nn.Module): ("qkv_proj", "v_proj", "v"), ] - expert_params_mapping = FusedMoE.make_expert_params_mapping( - ckpt_gate_proj_name="w1", - ckpt_down_proj_name="w2", - ckpt_up_proj_name="w3", - num_experts=self.config.num_local_experts) - params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() + expert_params_mapping = self.get_expert_mapping() for name, loaded_weight in weights: if (self.quant_config is not None and (scale_name := self.quant_config.get_cache_scale(name))):