diff --git a/vllm/v1/attention/backends/pallas.py b/vllm/v1/attention/backends/pallas.py index b7fc1ffeb65e6..52e12a1a506f5 100644 --- a/vllm/v1/attention/backends/pallas.py +++ b/vllm/v1/attention/backends/pallas.py @@ -167,10 +167,6 @@ class PallasAttentionBackendImpl(AttentionImpl): "are not implemented for " "PallasAttentionBackendImpl") - tpu_version = torch_xla.tpu.version() - if tpu_version < 4: - raise NotImplementedError("TPU version must be 4 or higher.") - def forward( self, layer: AttentionLayer,