From eb447aff5609378ff63f935d86cce3b93fe62864 Mon Sep 17 00:00:00 2001 From: Juechen Liu Date: Sun, 28 Sep 2025 21:18:57 -0700 Subject: [PATCH] [Misc] fix tests failure by using current_platform (#25825) Signed-off-by: Juechen Liu Signed-off-by: yewentao256 --- vllm/attention/ops/triton_reshape_and_cache_flash.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/attention/ops/triton_reshape_and_cache_flash.py b/vllm/attention/ops/triton_reshape_and_cache_flash.py index cc9e1bb23b9b3..0d82935bb4185 100644 --- a/vllm/attention/ops/triton_reshape_and_cache_flash.py +++ b/vllm/attention/ops/triton_reshape_and_cache_flash.py @@ -137,7 +137,7 @@ def triton_reshape_and_cache_flash( # heuristics instead of autotuning TILE_SIZE = min(2048, triton.next_power_of_2(n)) - if torch.version.hip or torch.version.xpu: + if current_platform.is_rocm() or current_platform.is_xpu(): num_stages = 4 num_warps = 8 else: # cuda