diff --git a/vllm/attention/backends/mla/utils.py b/vllm/attention/backends/mla/utils.py index 28f00a2797dcf..9cdc31d2bd7d9 100644 --- a/vllm/attention/backends/mla/utils.py +++ b/vllm/attention/backends/mla/utils.py @@ -161,7 +161,7 @@ class MLAImplCommon(AttentionImpl): def _v_up_proj_and_o_proj(self, x): if envs.VLLM_MLA_PERFORM_MATRIX_ABSORPTION: - return self.o_proj_absored( + return self.o_proj_absorbed( x.reshape(-1, self.num_heads * self.kv_lora_rank))[0] else: x = torch.einsum("bnl,lnv->bnv", x, self.W_UV)