diff --git a/vllm/attention/layers/cross_attention.py b/vllm/attention/layers/cross_attention.py index 5f814b23888b3..c24fa4e15f679 100644 --- a/vllm/attention/layers/cross_attention.py +++ b/vllm/attention/layers/cross_attention.py @@ -6,14 +6,13 @@ from typing import Optional import numpy as np import torch -from transformers import CacheConfig from vllm import envs from vllm.attention.backends.abstract import (AttentionBackend, AttentionMetadata, AttentionType) from vllm.attention.layer import Attention from vllm.attention.selector import get_attn_backend -from vllm.config import VllmConfig +from vllm.config import CacheConfig, VllmConfig from vllm.logger import init_logger from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.utils import cdiv