Skip to content

Commit

Permalink
[Bugfix][Kernel] Give unique name to BlockSparseFlashAttention (vllm-…
Browse files Browse the repository at this point in the history
…project#12040)

Signed-off-by: Chen Zhang <[email protected]>
Signed-off-by: hzh <[email protected]>
  • Loading branch information
heheda12345 authored and HwwwwwwwH committed Jan 22, 2025
1 parent 3a05c49 commit 87a687b
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 2 deletions.
3 changes: 1 addition & 2 deletions vllm/attention/backends/blocksparse_attn.py
Original file line number Diff line number Diff line change
Expand Up @@ -89,8 +89,7 @@ class BlocksparseFlashAttentionBackend(AttentionBackend):

@staticmethod
def get_name() -> str:
# For attention layer compatibility
return "FLASH_ATTN"
return "BLOCK_SPARSE_FLASH_ATTN"

@staticmethod
def get_impl_cls() -> Type["BlocksparseFlashAttentionImpl"]:
Expand Down
1 change: 1 addition & 0 deletions vllm/platforms/interface.py
Original file line number Diff line number Diff line change
Expand Up @@ -33,6 +33,7 @@ class _Backend(enum.Enum):
HPU_ATTN = enum.auto()
PALLAS = enum.auto()
IPEX = enum.auto()
BLOCK_SPARSE_FLASH_ATTN = enum.auto()
NO_ATTENTION = enum.auto()


Expand Down

0 comments on commit 87a687b

Please sign in to comment.