From 5dc5a15cc464afe58fb679de3af4a07ed612fdf4 Mon Sep 17 00:00:00 2001 From: kijai <40791699+kijai@users.noreply.github.com> Date: Sat, 13 Sep 2025 11:04:17 +0300 Subject: [PATCH] Add transformer_options to sageattn patch --- nodes/model_optimization_nodes.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/nodes/model_optimization_nodes.py b/nodes/model_optimization_nodes.py index c1f2e88..4821f15 100644 --- a/nodes/model_optimization_nodes.py +++ b/nodes/model_optimization_nodes.py @@ -114,7 +114,7 @@ class BaseLoaderKJ: sage_func = set_sage_func(sage_attention) @torch.compiler.disable() - def attention_sage(q, k, v, heads, mask=None, attn_precision=None, skip_reshape=False, skip_output_reshape=False): + def attention_sage(q, k, v, heads, mask=None, attn_precision=None, skip_reshape=False, skip_output_reshape=False, transformer_options=None): if skip_reshape: b, _, _, dim_head = q.shape tensor_layout="HND"