From 7e8d685775fe9e11c3cea79e84418a9f0bab4a5f Mon Sep 17 00:00:00 2001 From: Isotr0py Date: Sun, 10 Aug 2025 15:08:23 +0800 Subject: [PATCH] [Minor] Fix pre-commit error on main (#22579) Signed-off-by: Isotr0py <2037008807@qq.com> --- vllm/model_executor/layers/fused_moe/fused_moe.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/vllm/model_executor/layers/fused_moe/fused_moe.py b/vllm/model_executor/layers/fused_moe/fused_moe.py index 3ad5f5b7ad31d..86cc6e0e5dac8 100644 --- a/vllm/model_executor/layers/fused_moe/fused_moe.py +++ b/vllm/model_executor/layers/fused_moe/fused_moe.py @@ -1038,9 +1038,9 @@ def inplace_fused_experts( w2_zp: Optional[torch.Tensor] = None, a1_scale: Optional[torch.Tensor] = None, a2_scale: Optional[torch.Tensor] = None, - block_shape: Optional[List[int]] = None, + block_shape: Optional[List[int]] = None, #noqa: UP006 w1_bias: Optional[torch.Tensor] = None, - w2_bias: Optional[torch.Tensor] = None) -> None: #noqa: UP006 + w2_bias: Optional[torch.Tensor] = None) -> None: fused_experts_impl(hidden_states, w1, w2, topk_weights, topk_ids, True, activation, is_act_and_mul, apply_router_weight_on_input, use_fp8_w8a8,