mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-10 14:05:39 +08:00
[XPU] upgrade torch & ipex 2.9 on XPU platform (#29307)
Signed-off-by: Kunshang Ji <kunshang.ji@intel.com>
This commit is contained in:
parent
5f9679a43b
commit
b8328b49fb
@ -1,4 +1,4 @@
|
|||||||
FROM intel/deep-learning-essentials:2025.1.3-0-devel-ubuntu24.04 AS vllm-base
|
FROM intel/deep-learning-essentials:2025.2.2-0-devel-ubuntu24.04 AS vllm-base
|
||||||
|
|
||||||
RUN wget -O- https://apt.repos.intel.com/intel-gpg-keys/GPG-PUB-KEY-INTEL-SW-PRODUCTS.PUB | gpg --dearmor | tee /usr/share/keyrings/oneapi-archive-keyring.gpg > /dev/null && \
|
RUN wget -O- https://apt.repos.intel.com/intel-gpg-keys/GPG-PUB-KEY-INTEL-SW-PRODUCTS.PUB | gpg --dearmor | tee /usr/share/keyrings/oneapi-archive-keyring.gpg > /dev/null && \
|
||||||
echo "deb [signed-by=/usr/share/keyrings/oneapi-archive-keyring.gpg] https://apt.repos.intel.com/oneapi all main" | tee /etc/apt/sources.list.d/oneAPI.list && \
|
echo "deb [signed-by=/usr/share/keyrings/oneapi-archive-keyring.gpg] https://apt.repos.intel.com/oneapi all main" | tee /etc/apt/sources.list.d/oneAPI.list && \
|
||||||
@ -25,10 +25,14 @@ RUN apt clean && apt-get update -y && \
|
|||||||
RUN update-alternatives --install /usr/bin/python3 python3 /usr/bin/python3.12 1
|
RUN update-alternatives --install /usr/bin/python3 python3 /usr/bin/python3.12 1
|
||||||
RUN update-alternatives --install /usr/bin/python python /usr/bin/python3.12 1
|
RUN update-alternatives --install /usr/bin/python python /usr/bin/python3.12 1
|
||||||
|
|
||||||
RUN apt install -y libze1 libze-dev libze-intel-gpu1 intel-opencl-icd libze-intel-gpu-raytracing
|
RUN apt install -y libze1 libze-dev libze-intel-gpu1 intel-opencl-icd libze-intel-gpu-raytracing intel-ocloc
|
||||||
|
|
||||||
|
# This oneccl contains the BMG support which is not the case for default version of oneapi 2025.2.
|
||||||
|
RUN wget https://github.com/uxlfoundation/oneCCL/releases/download/2021.15.6/intel-oneccl-2021.15.6.9_offline.sh
|
||||||
|
RUN bash intel-oneccl-2021.15.6.9_offline.sh -a --silent --eula accept && \
|
||||||
|
echo "source /opt/intel/oneapi/setvars.sh --force" >> /root/.bashrc && \
|
||||||
|
echo "source /opt/intel/oneapi/ccl/2021.15/env/vars.sh --force" >> /root/.bashrc
|
||||||
|
|
||||||
RUN wget https://github.com/uxlfoundation/oneCCL/releases/download/2021.15.4/intel-oneccl-2021.15.4.11_offline.sh
|
|
||||||
RUN bash intel-oneccl-2021.15.4.11_offline.sh -a --silent --eula accept && echo "source /opt/intel/oneapi/setvars.sh --force" >> /root/.bashrc
|
|
||||||
SHELL ["bash", "-c"]
|
SHELL ["bash", "-c"]
|
||||||
CMD ["bash", "-c", "source /root/.bashrc && exec bash"]
|
CMD ["bash", "-c", "source /root/.bashrc && exec bash"]
|
||||||
|
|
||||||
@ -72,6 +76,7 @@ RUN python3 -m pip install -e tests/vllm_test_utils
|
|||||||
ENV NIXL_VERSION=0.7.0
|
ENV NIXL_VERSION=0.7.0
|
||||||
RUN python3 /workspace/vllm/tools/install_nixl_from_source_ubuntu.py
|
RUN python3 /workspace/vllm/tools/install_nixl_from_source_ubuntu.py
|
||||||
|
|
||||||
|
# remove torch bundled oneccl to avoid conflicts
|
||||||
RUN --mount=type=cache,target=/root/.cache/pip \
|
RUN --mount=type=cache,target=/root/.cache/pip \
|
||||||
pip uninstall oneccl oneccl-devel -y
|
pip uninstall oneccl oneccl-devel -y
|
||||||
|
|
||||||
|
|||||||
@ -10,9 +10,9 @@ wheel
|
|||||||
jinja2>=3.1.6
|
jinja2>=3.1.6
|
||||||
datasets # for benchmark scripts
|
datasets # for benchmark scripts
|
||||||
numba == 0.61.2 # Required for N-gram speculative decoding
|
numba == 0.61.2 # Required for N-gram speculative decoding
|
||||||
torch==2.8.0+xpu
|
--extra-index-url=https://download.pytorch.org/whl/xpu
|
||||||
|
torch==2.9.0+xpu
|
||||||
torchaudio
|
torchaudio
|
||||||
torchvision
|
torchvision
|
||||||
--extra-index-url=https://download.pytorch.org/whl/xpu
|
|
||||||
|
|
||||||
intel-extension-for-pytorch @ https://intel-extension-for-pytorch.s3.us-east-1.amazonaws.com/ipex_dev/xpu/intel_extension_for_pytorch-2.8.10.post1%2Bxpu-cp312-cp312-linux_x86_64.whl
|
intel-extension-for-pytorch @ https://intel-extension-for-pytorch.s3.us-east-1.amazonaws.com/ipex_dev/xpu/intel_extension_for_pytorch-2.9.10.post0%2Bxpu-cp312-cp312-linux_x86_64.whl
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user