mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-10 03:15:20 +08:00
41 lines
1.3 KiB
Python
41 lines
1.3 KiB
Python
# SPDX-License-Identifier: Apache-2.0
|
|
# SPDX-FileCopyrightText: Copyright contributors to the vLLM project
|
|
import pytest
|
|
|
|
from vllm.assets.image import ImageAsset
|
|
from vllm.config import ModelConfig
|
|
from vllm.multimodal import MULTIMODAL_REGISTRY
|
|
|
|
|
|
# yapf: disable
|
|
@pytest.mark.parametrize("model_id",
|
|
["llava-hf/llava-onevision-qwen2-0.5b-ov-hf"])
|
|
def test_multimodal_processor(model_id):
|
|
model_config = ModelConfig(
|
|
model=model_id,
|
|
model_impl="transformers",
|
|
)
|
|
|
|
mm_processor = MULTIMODAL_REGISTRY.create_processor(model_config)
|
|
|
|
image_pil = ImageAsset('cherry_blossom').pil_image
|
|
mm_data = {"image": image_pil}
|
|
str_prompt = "<|im_start|>user <image>\nWhat is the content of this image?<|im_end|><|im_start|>assistant\n" # noqa: E501
|
|
str_processed_inputs = mm_processor.apply(
|
|
prompt=str_prompt,
|
|
mm_data=mm_data,
|
|
hf_processor_mm_kwargs={},
|
|
)
|
|
|
|
ids_prompt = [
|
|
151644, 872, 220, 151646, 198, 3838, 374, 279, 2213, 315, 419, 2168,
|
|
30, 151645, 151644, 77091, 198
|
|
]
|
|
ids_processed_inputs = mm_processor.apply(
|
|
prompt=ids_prompt,
|
|
mm_data=mm_data,
|
|
hf_processor_mm_kwargs={},
|
|
)
|
|
|
|
assert str_processed_inputs["prompt"] == ids_processed_inputs["prompt"]
|