From d2af67441ddf5965aaebf129802a0a9d38f0e225 Mon Sep 17 00:00:00 2001 From: Kunshang Ji Date: Thu, 25 Sep 2025 20:38:11 +0800 Subject: [PATCH] [XPU][Triton]add xpu config in triton_reshape_and_cache_flash (#25643) Signed-off-by: Kunshang Ji --- vllm/attention/ops/triton_reshape_and_cache_flash.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/attention/ops/triton_reshape_and_cache_flash.py b/vllm/attention/ops/triton_reshape_and_cache_flash.py index 883052cb46aab..cc9e1bb23b9b3 100644 --- a/vllm/attention/ops/triton_reshape_and_cache_flash.py +++ b/vllm/attention/ops/triton_reshape_and_cache_flash.py @@ -137,7 +137,7 @@ def triton_reshape_and_cache_flash( # heuristics instead of autotuning TILE_SIZE = min(2048, triton.next_power_of_2(n)) - if torch.version.hip: + if torch.version.hip or torch.version.xpu: num_stages = 4 num_warps = 8 else: # cuda