mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-11 15:06:13 +08:00
[XPU][CI] add v1/core test in xpu hardware ci (#20537)
Signed-off-by: Ma, Liangliang <liangliang.ma@intel.com>
This commit is contained in:
parent
2e610deb72
commit
2c5ebec064
@ -27,4 +27,6 @@ docker run \
|
|||||||
"${image_name}" \
|
"${image_name}" \
|
||||||
sh -c '
|
sh -c '
|
||||||
VLLM_USE_V1=1 python3 examples/offline_inference/basic/generate.py --model facebook/opt-125m --block-size 64 --enforce-eager
|
VLLM_USE_V1=1 python3 examples/offline_inference/basic/generate.py --model facebook/opt-125m --block-size 64 --enforce-eager
|
||||||
|
cd tests
|
||||||
|
pytest -v -s v1/core
|
||||||
'
|
'
|
||||||
|
|||||||
@ -47,7 +47,7 @@ FROM vllm-base AS vllm-openai
|
|||||||
|
|
||||||
# install additional dependencies for openai api server
|
# install additional dependencies for openai api server
|
||||||
RUN --mount=type=cache,target=/root/.cache/pip \
|
RUN --mount=type=cache,target=/root/.cache/pip \
|
||||||
pip install accelerate hf_transfer 'modelscope!=1.15.0'
|
pip install accelerate hf_transfer pytest 'modelscope!=1.15.0'
|
||||||
|
|
||||||
ENV VLLM_USAGE_SOURCE production-docker-image \
|
ENV VLLM_USAGE_SOURCE production-docker-image \
|
||||||
TRITON_XPU_PROFILE 1
|
TRITON_XPU_PROFILE 1
|
||||||
|
|||||||
@ -93,10 +93,6 @@ class XPUPlatform(Platform):
|
|||||||
"mode.")
|
"mode.")
|
||||||
model_config.enforce_eager = True
|
model_config.enforce_eager = True
|
||||||
|
|
||||||
if vllm_config.speculative_config is not None:
|
|
||||||
raise NotImplementedError(
|
|
||||||
"XPU does not support speculative decoding")
|
|
||||||
|
|
||||||
if vllm_config.device_config is not None:
|
if vllm_config.device_config is not None:
|
||||||
assert vllm_config.device_config.device_type == "xpu"
|
assert vllm_config.device_config.device_type == "xpu"
|
||||||
|
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user