From 55e1c66da56a37150a85213092ed033665ec2bdb Mon Sep 17 00:00:00 2001 From: Kuntai Du Date: Mon, 14 Jul 2025 15:14:17 -0700 Subject: [PATCH] [Docs] remove outdated performance benchmark (#20935) Signed-off-by: Kuntai Du --- README.md | 2 -- 1 file changed, 2 deletions(-) diff --git a/README.md b/README.md index c4b1468552635..dc2f0afbe3538 100644 --- a/README.md +++ b/README.md @@ -63,8 +63,6 @@ vLLM is fast with: - Speculative decoding - Chunked prefill -**Performance benchmark**: We include a performance benchmark at the end of [our blog post](https://blog.vllm.ai/2024/09/05/perf-update.html). It compares the performance of vLLM against other LLM serving engines ([TensorRT-LLM](https://github.com/NVIDIA/TensorRT-LLM), [SGLang](https://github.com/sgl-project/sglang) and [LMDeploy](https://github.com/InternLM/lmdeploy)). The implementation is under [nightly-benchmarks folder](.buildkite/nightly-benchmarks/) and you can [reproduce](https://github.com/vllm-project/vllm/issues/8176) this benchmark using our one-click runnable script. - vLLM is flexible and easy to use with: - Seamless integration with popular Hugging Face models