mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-10 08:14:54 +08:00
52 lines
2.2 KiB
Python
52 lines
2.2 KiB
Python
# SPDX-License-Identifier: Apache-2.0
|
|
# SPDX-FileCopyrightText: Copyright contributors to the vLLM project
|
|
import pytest
|
|
|
|
from tests.models.language.pooling.embed_utils import (
|
|
correctness_test_embed_models)
|
|
from tests.models.utils import CLSPoolingEmbedModelInfo, EmbedModelInfo
|
|
|
|
from .mteb_utils import mteb_test_embed_models
|
|
|
|
MODELS = [
|
|
########## BertModel
|
|
CLSPoolingEmbedModelInfo("intfloat/e5-small",
|
|
architecture="BertModel",
|
|
mteb_score=0.742285423,
|
|
enable_test=True),
|
|
CLSPoolingEmbedModelInfo("intfloat/e5-base",
|
|
architecture="BertModel",
|
|
enable_test=False),
|
|
CLSPoolingEmbedModelInfo("intfloat/e5-large",
|
|
architecture="BertModel",
|
|
enable_test=False),
|
|
CLSPoolingEmbedModelInfo("intfloat/multilingual-e5-small",
|
|
architecture="BertModel",
|
|
enable_test=False),
|
|
########## XLMRobertaModel
|
|
CLSPoolingEmbedModelInfo("intfloat/multilingual-e5-base",
|
|
architecture="XLMRobertaModel",
|
|
mteb_score=0.779325955,
|
|
enable_test=True),
|
|
CLSPoolingEmbedModelInfo("intfloat/multilingual-e5-large",
|
|
architecture="XLMRobertaModel",
|
|
enable_test=False),
|
|
CLSPoolingEmbedModelInfo("intfloat/multilingual-e5-large-instruct",
|
|
architecture="XLMRobertaModel",
|
|
enable_test=False),
|
|
]
|
|
|
|
|
|
@pytest.mark.parametrize("model_info", MODELS)
|
|
def test_embed_models_mteb(hf_runner, vllm_runner,
|
|
model_info: EmbedModelInfo) -> None:
|
|
mteb_test_embed_models(hf_runner, vllm_runner, model_info)
|
|
|
|
|
|
@pytest.mark.parametrize("model_info", MODELS)
|
|
def test_embed_models_correctness(hf_runner, vllm_runner,
|
|
model_info: EmbedModelInfo,
|
|
example_prompts) -> None:
|
|
correctness_test_embed_models(hf_runner, vllm_runner, model_info,
|
|
example_prompts)
|