From 376e7eb838acb708eba60a6419ce1c0090a483c5 Mon Sep 17 00:00:00 2001 From: Sage Moore Date: Tue, 8 Jul 2025 19:12:45 +0000 Subject: [PATCH] minor change Signed-off-by: Sage Moore --- vllm/v1/attention/backends/flash_attn.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/v1/attention/backends/flash_attn.py b/vllm/v1/attention/backends/flash_attn.py index a82f1bb20fba3..527b31153410b 100755 --- a/vllm/v1/attention/backends/flash_attn.py +++ b/vllm/v1/attention/backends/flash_attn.py @@ -266,7 +266,7 @@ class FlashAttentionMetadataBuilder( use_cascade = common_prefix_len > 0 if use_cascade: - cu_prefix_query_lens = torch.tensor([0,num_actual_tokens], + cu_prefix_query_lens = torch.tensor([0, num_actual_tokens], dtype=torch.int32, device=self.runner.device) prefix_kv_lens = torch.tensor([common_prefix_len],