From aa19f297d2baa334b7836df71a64608e761fc0b6 Mon Sep 17 00:00:00 2001 From: Lucas Wilkinson Date: Wed, 29 Jan 2025 22:51:37 -0500 Subject: [PATCH] Update vllm/attention/backends/mla/utils.py Co-authored-by: Michael Goin --- vllm/attention/backends/mla/utils.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/attention/backends/mla/utils.py b/vllm/attention/backends/mla/utils.py index 28f00a2797dcf..9cdc31d2bd7d9 100644 --- a/vllm/attention/backends/mla/utils.py +++ b/vllm/attention/backends/mla/utils.py @@ -161,7 +161,7 @@ class MLAImplCommon(AttentionImpl): def _v_up_proj_and_o_proj(self, x): if envs.VLLM_MLA_PERFORM_MATRIX_ABSORPTION: - return self.o_proj_absored( + return self.o_proj_absorbed( x.reshape(-1, self.num_heads * self.kv_lora_rank))[0] else: x = torch.einsum("bnl,lnv->bnv", x, self.W_UV)