Add transformer_options to sageattn patch

This commit is contained in:
kijai 2025-09-13 11:04:17 +03:00
parent 36f6fdd7d4
commit 5dc5a15cc4

View File

@ -114,7 +114,7 @@ class BaseLoaderKJ:
sage_func = set_sage_func(sage_attention)
@torch.compiler.disable()
def attention_sage(q, k, v, heads, mask=None, attn_precision=None, skip_reshape=False, skip_output_reshape=False):
def attention_sage(q, k, v, heads, mask=None, attn_precision=None, skip_reshape=False, skip_output_reshape=False, transformer_options=None):
if skip_reshape:
b, _, _, dim_head = q.shape
tensor_layout="HND"