mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-24 15:25:01 +08:00
[Bugfix][Kernel] Give unique name to BlockSparseFlashAttention (#12040)
Signed-off-by: Chen Zhang <zhangch99@outlook.com>
This commit is contained in:
parent
2e0e017610
commit
a2d2acb4c8
@ -89,8 +89,7 @@ class BlocksparseFlashAttentionBackend(AttentionBackend):
|
||||
|
||||
@staticmethod
|
||||
def get_name() -> str:
|
||||
# For attention layer compatibility
|
||||
return "FLASH_ATTN"
|
||||
return "BLOCK_SPARSE_FLASH_ATTN"
|
||||
|
||||
@staticmethod
|
||||
def get_impl_cls() -> Type["BlocksparseFlashAttentionImpl"]:
|
||||
|
||||
@ -33,6 +33,7 @@ class _Backend(enum.Enum):
|
||||
HPU_ATTN = enum.auto()
|
||||
PALLAS = enum.auto()
|
||||
IPEX = enum.auto()
|
||||
BLOCK_SPARSE_FLASH_ATTN = enum.auto()
|
||||
NO_ATTENTION = enum.auto()
|
||||
|
||||
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user