From e38739aefc2ef97a697d3bfe5c4bbd4e14787231 Mon Sep 17 00:00:00 2001 From: yuantao <2422264527@qq.com> Date: Sat, 15 Nov 2025 18:37:53 +0800 Subject: [PATCH] Add FLASH_SINK_ATTN to AttentionBackendEnum Signed-off-by: yuantao <2422264527@qq.com> --- vllm/attention/backends/registry.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/vllm/attention/backends/registry.py b/vllm/attention/backends/registry.py index f07a6059be377..0602607966720 100644 --- a/vllm/attention/backends/registry.py +++ b/vllm/attention/backends/registry.py @@ -42,6 +42,9 @@ class AttentionBackendEnum(enum.Enum, metaclass=_AttentionBackendEnumMeta): """ FLASH_ATTN = "vllm.v1.attention.backends.flash_attn.FlashAttentionBackend" + FLASH_SINK_ATTN = ( + "vllm.v1.attention.backends.flash_sink_attn.FlashSinkAttentionBackend" + ) TRITON_ATTN = "vllm.v1.attention.backends.triton_attn.TritonAttentionBackend" XFORMERS = "vllm.v1.attention.backends.xformers.XFormersAttentionBackend" ROCM_ATTN = "vllm.v1.attention.backends.rocm_attn.RocmAttentionBackend"