From 68d630a0c7273777c23a89e3abab75bd89eaa644 Mon Sep 17 00:00:00 2001 From: Gordon Wong Date: Sat, 22 Feb 2025 14:07:04 +0800 Subject: [PATCH] [ROCM] fix native attention function call (#13650) --- vllm/attention/backends/rocm_flash_attn.py | 1 - 1 file changed, 1 deletion(-) diff --git a/vllm/attention/backends/rocm_flash_attn.py b/vllm/attention/backends/rocm_flash_attn.py index f49b37842d9b1..e1a8d3d336139 100644 --- a/vllm/attention/backends/rocm_flash_attn.py +++ b/vllm/attention/backends/rocm_flash_attn.py @@ -717,7 +717,6 @@ class ROCmFlashAttentionImpl(AttentionImpl): self.num_heads, self.head_size, self.scale, - causal_mask, attn_masks, ) else: