mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-13 02:25:36 +08:00
[CI/Build] Skip Mllama HF runner tests with Transformers v4.55.0 (#22659)
Signed-off-by: Isotr0py <2037008807@qq.com>
This commit is contained in:
parent
84cf78acee
commit
c90fb03df5
@ -6,6 +6,7 @@ from typing import Optional, overload
|
|||||||
import pytest
|
import pytest
|
||||||
import torch
|
import torch
|
||||||
from transformers import AutoConfig, AutoModelForImageTextToText, AutoTokenizer
|
from transformers import AutoConfig, AutoModelForImageTextToText, AutoTokenizer
|
||||||
|
from transformers import __version__ as TRANSFORMERS_VERSION
|
||||||
|
|
||||||
from vllm import LLM, SamplingParams
|
from vllm import LLM, SamplingParams
|
||||||
from vllm.attention.backends.flash_attn import FlashAttentionMetadata
|
from vllm.attention.backends.flash_attn import FlashAttentionMetadata
|
||||||
@ -285,6 +286,10 @@ def clear_cache():
|
|||||||
@pytest.mark.parametrize("max_tokens", [128])
|
@pytest.mark.parametrize("max_tokens", [128])
|
||||||
@pytest.mark.parametrize("num_logprobs", [5])
|
@pytest.mark.parametrize("num_logprobs", [5])
|
||||||
@pytest.mark.parametrize("attn_backend", LIST_ENC_DEC_SUPPORTED_BACKENDS)
|
@pytest.mark.parametrize("attn_backend", LIST_ENC_DEC_SUPPORTED_BACKENDS)
|
||||||
|
@pytest.mark.skipif(
|
||||||
|
TRANSFORMERS_VERSION == "4.55.0",
|
||||||
|
reason="Transformers v4.55.0 has a regression issue on mllama, "
|
||||||
|
"see: https://github.com/huggingface/transformers/pull/40083")
|
||||||
def test_models_single_leading_image(hf_runner, vllm_runner, image_assets,
|
def test_models_single_leading_image(hf_runner, vllm_runner, image_assets,
|
||||||
model, sizes, dtype, max_tokens,
|
model, sizes, dtype, max_tokens,
|
||||||
num_logprobs,
|
num_logprobs,
|
||||||
@ -313,6 +318,10 @@ def test_models_single_leading_image(hf_runner, vllm_runner, image_assets,
|
|||||||
@pytest.mark.parametrize("max_tokens", [128])
|
@pytest.mark.parametrize("max_tokens", [128])
|
||||||
@pytest.mark.parametrize("num_logprobs", [5])
|
@pytest.mark.parametrize("num_logprobs", [5])
|
||||||
@pytest.mark.parametrize("attn_backend", LIST_ENC_DEC_SUPPORTED_BACKENDS)
|
@pytest.mark.parametrize("attn_backend", LIST_ENC_DEC_SUPPORTED_BACKENDS)
|
||||||
|
@pytest.mark.skipif(
|
||||||
|
TRANSFORMERS_VERSION == "4.55.0",
|
||||||
|
reason="Transformers v4.55.0 has a regression issue on mllama, "
|
||||||
|
"see: https://github.com/huggingface/transformers/pull/40083")
|
||||||
def test_models_multi_leading_images(hf_runner, vllm_runner, image_assets,
|
def test_models_multi_leading_images(hf_runner, vllm_runner, image_assets,
|
||||||
model, dtype, max_tokens, num_logprobs,
|
model, dtype, max_tokens, num_logprobs,
|
||||||
attn_backend: _Backend) -> None:
|
attn_backend: _Backend) -> None:
|
||||||
@ -362,6 +371,10 @@ def test_models_multi_leading_images(hf_runner, vllm_runner, image_assets,
|
|||||||
@pytest.mark.parametrize("max_tokens", [128])
|
@pytest.mark.parametrize("max_tokens", [128])
|
||||||
@pytest.mark.parametrize("num_logprobs", [5])
|
@pytest.mark.parametrize("num_logprobs", [5])
|
||||||
@pytest.mark.parametrize("attn_backend", LIST_ENC_DEC_SUPPORTED_BACKENDS)
|
@pytest.mark.parametrize("attn_backend", LIST_ENC_DEC_SUPPORTED_BACKENDS)
|
||||||
|
@pytest.mark.skipif(
|
||||||
|
TRANSFORMERS_VERSION == "4.55.0",
|
||||||
|
reason="Transformers v4.55.0 has a regression issue on mllama, "
|
||||||
|
"see: https://github.com/huggingface/transformers/pull/40083")
|
||||||
def test_models_interleaved_images(hf_runner, vllm_runner, image_assets, model,
|
def test_models_interleaved_images(hf_runner, vllm_runner, image_assets, model,
|
||||||
dtype, max_tokens, num_logprobs,
|
dtype, max_tokens, num_logprobs,
|
||||||
attn_backend: _Backend) -> None:
|
attn_backend: _Backend) -> None:
|
||||||
@ -402,6 +415,10 @@ def test_models_interleaved_images(hf_runner, vllm_runner, image_assets, model,
|
|||||||
@pytest.mark.parametrize("dtype", ["bfloat16"])
|
@pytest.mark.parametrize("dtype", ["bfloat16"])
|
||||||
@pytest.mark.parametrize("max_tokens", [64])
|
@pytest.mark.parametrize("max_tokens", [64])
|
||||||
@pytest.mark.parametrize("num_logprobs", [5])
|
@pytest.mark.parametrize("num_logprobs", [5])
|
||||||
|
@pytest.mark.skipif(
|
||||||
|
TRANSFORMERS_VERSION == "4.55.0",
|
||||||
|
reason="Transformers v4.55.0 has a regression issue on mllama, "
|
||||||
|
"see: https://github.com/huggingface/transformers/pull/40083")
|
||||||
def test_models_distributed(
|
def test_models_distributed(
|
||||||
hf_runner,
|
hf_runner,
|
||||||
vllm_runner,
|
vllm_runner,
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user