From 9381b5cde09c542d74df835b2cba3225eff68d0e Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E6=9C=9D?= <108347991+BowTen@users.noreply.github.com> Date: Sun, 30 Nov 2025 14:29:13 +0800 Subject: [PATCH] [Doc]: Fix typo in fused_moe layer (#29731) Signed-off-by: BowTen --- vllm/model_executor/layers/fused_moe/layer.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/model_executor/layers/fused_moe/layer.py b/vllm/model_executor/layers/fused_moe/layer.py index 0ef3130b26333..e180b4f4ba23f 100644 --- a/vllm/model_executor/layers/fused_moe/layer.py +++ b/vllm/model_executor/layers/fused_moe/layer.py @@ -1422,7 +1422,7 @@ class FusedMoE(CustomOp): # do nothing. return p - # Do not update the layer paramater as the layer's MoE operations would + # Do not update the layer parameter as the layer's MoE operations would # expect the parameter's tensor to the same shape / stride. Instead, # make a new torch.nn.Parameter that is used just in the context of # EPLB.