diff --git a/examples/offline_inference/profiling_tpu/README.md b/examples/offline_inference/profiling_tpu/README.md index 6595efec43779..e0122c05cff1f 100644 --- a/examples/offline_inference/profiling_tpu/README.md +++ b/examples/offline_inference/profiling_tpu/README.md @@ -4,7 +4,7 @@ This script is used to profile the TPU performance of vLLM for specific prefill Note: an actual running server is a mix of both prefill of many shapes and decode of many shapes. -We assume you are on a TPU already (this was tested on TPU v6e) and have installed vLLM according to the [installation guide](https://docs.vllm.ai/en/latest/getting_started/installation/ai_accelerator/index.html). +We assume you are on a TPU already (this was tested on TPU v6e) and have installed vLLM according to the [Google TPU installation guide](https://docs.vllm.ai/en/latest/getting_started/installation/google_tpu.html). > In all examples below, we run several warmups before (so `--enforce-eager` is okay)