From f257544709a8d9ccb8947e6f2c1779988c448ae7 Mon Sep 17 00:00:00 2001 From: Huy Do Date: Tue, 28 Oct 2025 19:39:15 -0700 Subject: [PATCH] Install pre-built xformers-0.0.32.post2 built with pt-2.9.0 (#27598) Signed-off-by: Huy Do Co-authored-by: Roger Wang --- docker/Dockerfile | 7 ------- requirements/cuda.txt | 2 +- 2 files changed, 1 insertion(+), 8 deletions(-) diff --git a/docker/Dockerfile b/docker/Dockerfile index eb1453126e6f..c5b729e03b17 100644 --- a/docker/Dockerfile +++ b/docker/Dockerfile @@ -361,13 +361,6 @@ RUN --mount=type=bind,from=build,src=/workspace/dist,target=/vllm-workspace/dist && uv pip install --system dist/*.whl --verbose \ --extra-index-url ${PYTORCH_CUDA_INDEX_BASE_URL}/cu$(echo $CUDA_VERSION | cut -d. -f1,2 | tr -d '.') -# TODO (huydhn): Remove this once xformers is released for 2.9.0 -RUN --mount=type=cache,target=/root/.cache/uv bash - <<'BASH' - . /etc/environment - export TORCH_CUDA_ARCH_LIST='7.5 8.0+PTX 9.0a' - uv pip install --system --no-build-isolation "git+https://github.com/facebookresearch/xformers@v0.0.32.post2" -BASH - # Install FlashInfer pre-compiled kernel cache and binaries # https://docs.flashinfer.ai/installation.html RUN --mount=type=cache,target=/root/.cache/uv \ diff --git a/requirements/cuda.txt b/requirements/cuda.txt index 7c5bc457d45b..9636e5b4b801 100644 --- a/requirements/cuda.txt +++ b/requirements/cuda.txt @@ -10,7 +10,7 @@ torchaudio==2.9.0 # These must be updated alongside torch torchvision==0.24.0 # Required for phi3v processor. See https://github.com/pytorch/vision?tab=readme-ov-file#installation for corresponding version # https://github.com/facebookresearch/xformers/releases/tag/v0.0.32.post1 -# xformers==0.0.32.post1; platform_system == 'Linux' and platform_machine == 'x86_64' # Requires PyTorch >= 2.8 +xformers==0.0.33+5d4b92a5.d20251026; platform_system == 'Linux' and platform_machine == 'x86_64' # Requires PyTorch >= 2.9 # FlashInfer should be updated together with the Dockerfile flashinfer-python==0.4.1 # Triton Kernels are needed for mxfp4 fused moe. (Should be updated alongside torch)