[Bugfix] Initialize attention bias on the same device as Query/Key/Value for QwenVL Series (#14031)

This commit is contained in:
Yang Liu 2025-02-28 23:36:08 +08:00 committed by GitHub
parent f7bee5c815
commit 9b61dd41e7
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
2 changed files with 4 additions and 2 deletions

View File

@ -323,7 +323,8 @@ class Qwen2_5_VisionAttention(nn.Module):
seqlens = (cu_seqlens[1:] - cu_seqlens[:-1]).tolist()
attn_bias = BlockDiagonalMask.from_seqlens(q_seqlen=seqlens,
kv_seqlen=None)
kv_seqlen=None,
device=q.device)
context_layer = xops.memory_efficient_attention_forward(
q, k, v, attn_bias=attn_bias, p=0, scale=None)

View File

@ -367,7 +367,8 @@ class Qwen2VisionAttention(nn.Module):
seqlens = (cu_seqlens[1:] - cu_seqlens[:-1]).tolist()
attn_bias = BlockDiagonalMask.from_seqlens(q_seqlen=seqlens,
kv_seqlen=None)
kv_seqlen=None,
device=q.device)
context_layer = xops.memory_efficient_attention_forward(
q, k, v, attn_bias=attn_bias, p=0, scale=None)