mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-04-23 06:07:03 +08:00
[CI/Build] Update transformers to v4.55.2 (#23093)
Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
This commit is contained in:
parent
a70d0bd0a3
commit
d6a1a20973
@ -7,7 +7,7 @@ requests >= 2.26.0
|
|||||||
tqdm
|
tqdm
|
||||||
blake3
|
blake3
|
||||||
py-cpuinfo
|
py-cpuinfo
|
||||||
transformers >= 4.55.0
|
transformers >= 4.55.2
|
||||||
tokenizers >= 0.21.1 # Required for fast incremental detokenization.
|
tokenizers >= 0.21.1 # Required for fast incremental detokenization.
|
||||||
protobuf # Required by LlamaTokenizer.
|
protobuf # Required by LlamaTokenizer.
|
||||||
fastapi[standard] >= 0.115.0 # Required by FastAPI's form models in the OpenAI API server's audio transcriptions endpoint.
|
fastapi[standard] >= 0.115.0 # Required by FastAPI's form models in the OpenAI API server's audio transcriptions endpoint.
|
||||||
|
|||||||
@ -34,7 +34,7 @@ opencv-python-headless >= 4.11.0 # required for video test
|
|||||||
datamodel_code_generator # required for minicpm3 test
|
datamodel_code_generator # required for minicpm3 test
|
||||||
lm-eval[api]==0.4.8 # required for model evaluation test
|
lm-eval[api]==0.4.8 # required for model evaluation test
|
||||||
mteb[bm25s]>=1.38.11, <2 # required for mteb test
|
mteb[bm25s]>=1.38.11, <2 # required for mteb test
|
||||||
transformers==4.55.0
|
transformers==4.55.2
|
||||||
tokenizers==0.21.1
|
tokenizers==0.21.1
|
||||||
schemathesis>=3.39.15 # Required for openai schema test.
|
schemathesis>=3.39.15 # Required for openai schema test.
|
||||||
# quantization
|
# quantization
|
||||||
|
|||||||
@ -1139,7 +1139,7 @@ tqdm==4.66.6
|
|||||||
# transformers
|
# transformers
|
||||||
tqdm-multiprocess==0.0.11
|
tqdm-multiprocess==0.0.11
|
||||||
# via lm-eval
|
# via lm-eval
|
||||||
transformers==4.55.0
|
transformers==4.55.2
|
||||||
# via
|
# via
|
||||||
# -r requirements/test.in
|
# -r requirements/test.in
|
||||||
# genai-perf
|
# genai-perf
|
||||||
|
|||||||
@ -5,6 +5,7 @@ from typing import Optional, overload
|
|||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
import torch
|
import torch
|
||||||
|
from packaging.version import Version
|
||||||
from transformers import AutoConfig, AutoModelForImageTextToText, AutoTokenizer
|
from transformers import AutoConfig, AutoModelForImageTextToText, AutoTokenizer
|
||||||
from transformers import __version__ as TRANSFORMERS_VERSION
|
from transformers import __version__ as TRANSFORMERS_VERSION
|
||||||
|
|
||||||
@ -287,8 +288,8 @@ def clear_cache():
|
|||||||
@pytest.mark.parametrize("num_logprobs", [5])
|
@pytest.mark.parametrize("num_logprobs", [5])
|
||||||
@pytest.mark.parametrize("attn_backend", LIST_ENC_DEC_SUPPORTED_BACKENDS)
|
@pytest.mark.parametrize("attn_backend", LIST_ENC_DEC_SUPPORTED_BACKENDS)
|
||||||
@pytest.mark.skipif(
|
@pytest.mark.skipif(
|
||||||
TRANSFORMERS_VERSION == "4.55.0",
|
Version(TRANSFORMERS_VERSION) <= Version("4.55.2"),
|
||||||
reason="Transformers v4.55.0 has a regression issue on mllama, "
|
reason="Transformers v4.55 has a regression issue on mllama, "
|
||||||
"see: https://github.com/huggingface/transformers/pull/40083")
|
"see: https://github.com/huggingface/transformers/pull/40083")
|
||||||
def test_models_single_leading_image(hf_runner, vllm_runner, image_assets,
|
def test_models_single_leading_image(hf_runner, vllm_runner, image_assets,
|
||||||
model, sizes, dtype, max_tokens,
|
model, sizes, dtype, max_tokens,
|
||||||
@ -319,8 +320,8 @@ def test_models_single_leading_image(hf_runner, vllm_runner, image_assets,
|
|||||||
@pytest.mark.parametrize("num_logprobs", [5])
|
@pytest.mark.parametrize("num_logprobs", [5])
|
||||||
@pytest.mark.parametrize("attn_backend", LIST_ENC_DEC_SUPPORTED_BACKENDS)
|
@pytest.mark.parametrize("attn_backend", LIST_ENC_DEC_SUPPORTED_BACKENDS)
|
||||||
@pytest.mark.skipif(
|
@pytest.mark.skipif(
|
||||||
TRANSFORMERS_VERSION == "4.55.0",
|
Version(TRANSFORMERS_VERSION) <= Version("4.55.2"),
|
||||||
reason="Transformers v4.55.0 has a regression issue on mllama, "
|
reason="Transformers v4.55 has a regression issue on mllama, "
|
||||||
"see: https://github.com/huggingface/transformers/pull/40083")
|
"see: https://github.com/huggingface/transformers/pull/40083")
|
||||||
def test_models_multi_leading_images(hf_runner, vllm_runner, image_assets,
|
def test_models_multi_leading_images(hf_runner, vllm_runner, image_assets,
|
||||||
model, dtype, max_tokens, num_logprobs,
|
model, dtype, max_tokens, num_logprobs,
|
||||||
@ -372,8 +373,8 @@ def test_models_multi_leading_images(hf_runner, vllm_runner, image_assets,
|
|||||||
@pytest.mark.parametrize("num_logprobs", [5])
|
@pytest.mark.parametrize("num_logprobs", [5])
|
||||||
@pytest.mark.parametrize("attn_backend", LIST_ENC_DEC_SUPPORTED_BACKENDS)
|
@pytest.mark.parametrize("attn_backend", LIST_ENC_DEC_SUPPORTED_BACKENDS)
|
||||||
@pytest.mark.skipif(
|
@pytest.mark.skipif(
|
||||||
TRANSFORMERS_VERSION == "4.55.0",
|
Version(TRANSFORMERS_VERSION) <= Version("4.55.2"),
|
||||||
reason="Transformers v4.55.0 has a regression issue on mllama, "
|
reason="Transformers v4.55 has a regression issue on mllama, "
|
||||||
"see: https://github.com/huggingface/transformers/pull/40083")
|
"see: https://github.com/huggingface/transformers/pull/40083")
|
||||||
def test_models_interleaved_images(hf_runner, vllm_runner, image_assets, model,
|
def test_models_interleaved_images(hf_runner, vllm_runner, image_assets, model,
|
||||||
dtype, max_tokens, num_logprobs,
|
dtype, max_tokens, num_logprobs,
|
||||||
@ -416,8 +417,8 @@ def test_models_interleaved_images(hf_runner, vllm_runner, image_assets, model,
|
|||||||
@pytest.mark.parametrize("max_tokens", [64])
|
@pytest.mark.parametrize("max_tokens", [64])
|
||||||
@pytest.mark.parametrize("num_logprobs", [5])
|
@pytest.mark.parametrize("num_logprobs", [5])
|
||||||
@pytest.mark.skipif(
|
@pytest.mark.skipif(
|
||||||
TRANSFORMERS_VERSION == "4.55.0",
|
Version(TRANSFORMERS_VERSION) <= Version("4.55.2"),
|
||||||
reason="Transformers v4.55.0 has a regression issue on mllama, "
|
reason="Transformers v4.55 has a regression issue on mllama, "
|
||||||
"see: https://github.com/huggingface/transformers/pull/40083")
|
"see: https://github.com/huggingface/transformers/pull/40083")
|
||||||
def test_models_distributed(
|
def test_models_distributed(
|
||||||
hf_runner,
|
hf_runner,
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user