mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-10 22:28:21 +08:00
[BugFix] Revert ROCm Custom Paged Attention Env Flag Check (#17022)
Signed-off-by: vllmellm <vllm.ellm@embeddedllm.com>
This commit is contained in:
parent
1e013fa388
commit
7e081ba7ca
@ -118,6 +118,7 @@ def use_rocm_custom_paged_attention(qtype: torch.dtype, head_size: int,
|
|||||||
and (head_size == 64 or head_size == 128)
|
and (head_size == 64 or head_size == 128)
|
||||||
and (block_size == 16 or block_size == 32)
|
and (block_size == 16 or block_size == 32)
|
||||||
and (gqa_ratio >= 1 and gqa_ratio <= 16) and max_seq_len <= 32768
|
and (gqa_ratio >= 1 and gqa_ratio <= 16) and max_seq_len <= 32768
|
||||||
|
and (envs.VLLM_ROCM_CUSTOM_PAGED_ATTN)
|
||||||
and not (envs.VLLM_ROCM_USE_AITER_PAGED_ATTN
|
and not (envs.VLLM_ROCM_USE_AITER_PAGED_ATTN
|
||||||
and envs.VLLM_ROCM_USE_AITER))
|
and envs.VLLM_ROCM_USE_AITER))
|
||||||
|
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user