[bugfix] fix type[AttentionBackend] bug in kv_connector_base_v1 (#30051)

Signed-off-by: 01267596 <xiongkai123@cmbchina.com>
Co-authored-by: 01267596 <xiongkai123@cmbchina.com>
This commit is contained in:
kx 2025-12-06 15:11:31 +08:00 committed by GitHub
parent a238cbd89d
commit d6aeaddf4a
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -239,7 +239,7 @@ class KVConnectorBase_V1(ABC):
return
def register_cross_layers_kv_cache(
self, kv_cache: torch.Tensor, attn_backend: type[AttentionBackend]
self, kv_cache: torch.Tensor, attn_backend: type["AttentionBackend"]
):
"""
Initialize with a single KV cache tensor used by all layers.