diff --git a/vllm/attention/backends/rocm_flash_attn.py b/vllm/attention/backends/rocm_flash_attn.py index f49b37842d9b1..e1a8d3d336139 100644 --- a/vllm/attention/backends/rocm_flash_attn.py +++ b/vllm/attention/backends/rocm_flash_attn.py @@ -717,7 +717,6 @@ class ROCmFlashAttentionImpl(AttentionImpl): self.num_heads, self.head_size, self.scale, - causal_mask, attn_masks, ) else: