mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-05-03 00:04:39 +08:00
[Bugfix][Kernel] Give unique name to BlockSparseFlashAttention (#12040)
Signed-off-by: Chen Zhang <zhangch99@outlook.com>
This commit is contained in:
parent
2e0e017610
commit
a2d2acb4c8
@ -89,8 +89,7 @@ class BlocksparseFlashAttentionBackend(AttentionBackend):
|
|||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def get_name() -> str:
|
def get_name() -> str:
|
||||||
# For attention layer compatibility
|
return "BLOCK_SPARSE_FLASH_ATTN"
|
||||||
return "FLASH_ATTN"
|
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def get_impl_cls() -> Type["BlocksparseFlashAttentionImpl"]:
|
def get_impl_cls() -> Type["BlocksparseFlashAttentionImpl"]:
|
||||||
|
|||||||
@ -33,6 +33,7 @@ class _Backend(enum.Enum):
|
|||||||
HPU_ATTN = enum.auto()
|
HPU_ATTN = enum.auto()
|
||||||
PALLAS = enum.auto()
|
PALLAS = enum.auto()
|
||||||
IPEX = enum.auto()
|
IPEX = enum.auto()
|
||||||
|
BLOCK_SPARSE_FLASH_ATTN = enum.auto()
|
||||||
NO_ATTENTION = enum.auto()
|
NO_ATTENTION = enum.auto()
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user