diff --git a/vllm/attention/ops/ipex_attn.py b/vllm/attention/ops/ipex_attn.py index b7e4ba4d7416..7207d0420a01 100644 --- a/vllm/attention/ops/ipex_attn.py +++ b/vllm/attention/ops/ipex_attn.py @@ -1,7 +1,7 @@ # SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project -from typing import Dict, List, Optional, Tuple +from typing import List, Optional, Tuple try: import intel_extension_for_pytorch.llm.modules as ipex_modules @@ -120,7 +120,7 @@ class _PagedAttention: @staticmethod def copy_blocks( kv_caches: List[torch.Tensor], - src_to_dists: Dict[int, List[int]], + src_to_dists: torch.Tensor, *args, ) -> None: key_caches = [kv_cache[0] for kv_cache in kv_caches]