mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-10 12:45:33 +08:00
[CI/Build] Skip prompt embeddings tests on V1-only CPU backend (#24721)
Signed-off-by: jiang1.li <jiang1.li@intel.com>
This commit is contained in:
parent
d21a36f5f9
commit
59d5d2c736
@ -119,6 +119,12 @@ def test_models(hf_runner, vllm_runner, example_prompts, model: str,
|
|||||||
# in parts of the operators
|
# in parts of the operators
|
||||||
pytest.skip(f"Skipping '{model}' model test with AITER kernel.")
|
pytest.skip(f"Skipping '{model}' model test with AITER kernel.")
|
||||||
|
|
||||||
|
# Note: can be removed when
|
||||||
|
# https://github.com/vllm-project/vllm/pull/24278 finished
|
||||||
|
if current_platform.is_cpu() and use_prompt_embeds:
|
||||||
|
pytest.skip("Skipping use_prompt_embeds=True with "
|
||||||
|
"V1-only CPU backend.")
|
||||||
|
|
||||||
with hf_runner(model) as hf_model:
|
with hf_runner(model) as hf_model:
|
||||||
hf_outputs = hf_model.generate_greedy_logprobs_limit(
|
hf_outputs = hf_model.generate_greedy_logprobs_limit(
|
||||||
example_prompts, max_tokens, num_logprobs)
|
example_prompts, max_tokens, num_logprobs)
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user