diff --git a/tests/entrypoints/openai/test_lora_resolvers.py b/tests/entrypoints/openai/test_lora_resolvers.py index b05fa379c69f..4856cafef44b 100644 --- a/tests/entrypoints/openai/test_lora_resolvers.py +++ b/tests/entrypoints/openai/test_lora_resolvers.py @@ -114,7 +114,7 @@ def mock_serving_setup(): mock_engine.add_lora.reset_mock() mock_engine.model_config = MockModelConfig() - mock_engine.processor = MagicMock() + mock_engine.input_processor = MagicMock() mock_engine.io_processor = MagicMock() models = OpenAIServingModels( diff --git a/tests/entrypoints/openai/test_serving_chat.py b/tests/entrypoints/openai/test_serving_chat.py index dd10384a7e8c..492e15fc82a6 100644 --- a/tests/entrypoints/openai/test_serving_chat.py +++ b/tests/entrypoints/openai/test_serving_chat.py @@ -429,7 +429,7 @@ async def test_serving_chat_returns_correct_model_name(): mock_engine.get_tokenizer.return_value = get_tokenizer(MODEL_NAME) mock_engine.errored = False mock_engine.model_config = MockModelConfig() - mock_engine.processor = MagicMock() + mock_engine.input_processor = MagicMock() mock_engine.io_processor = MagicMock() serving_chat = _build_serving_chat(mock_engine) @@ -459,7 +459,7 @@ async def test_serving_chat_should_set_correct_max_tokens(): mock_engine.get_tokenizer.return_value = get_tokenizer(MODEL_NAME) mock_engine.errored = False mock_engine.model_config = MockModelConfig() - mock_engine.processor = MagicMock() + mock_engine.input_processor = MagicMock() mock_engine.io_processor = MagicMock() serving_chat = _build_serving_chat(mock_engine) @@ -492,7 +492,7 @@ async def test_serving_chat_should_set_correct_max_tokens(): mock_engine.get_tokenizer.return_value = get_tokenizer(MODEL_NAME) mock_engine.errored = False mock_engine.model_config = mock_model_config - mock_engine.processor = MagicMock() + mock_engine.input_processor = MagicMock() mock_engine.io_processor = MagicMock() # Initialize the serving chat @@ -537,7 +537,7 @@ async def test_serving_chat_should_set_correct_max_tokens(): mock_engine.get_tokenizer.return_value = get_tokenizer(MODEL_NAME) mock_engine.errored = False mock_engine.model_config = mock_model_config - mock_engine.processor = MagicMock() + mock_engine.input_processor = MagicMock() mock_engine.io_processor = MagicMock() # Initialize the serving chat @@ -583,7 +583,7 @@ async def test_serving_chat_could_load_correct_generation_config(): mock_engine.get_tokenizer.return_value = get_tokenizer(MODEL_NAME) mock_engine.errored = False mock_engine.model_config = mock_model_config - mock_engine.processor = MagicMock() + mock_engine.input_processor = MagicMock() mock_engine.io_processor = MagicMock() # Initialize the serving chat @@ -629,7 +629,7 @@ async def test_serving_chat_did_set_correct_cache_salt(model_type): mock_engine.get_tokenizer.return_value = get_tokenizer(MODEL_NAME) mock_engine.errored = False mock_engine.model_config = mock_model_config - mock_engine.processor = MagicMock() + mock_engine.input_processor = MagicMock() mock_engine.io_processor = MagicMock() serving_chat = _build_serving_chat(mock_engine) @@ -662,7 +662,7 @@ async def test_serving_chat_data_parallel_rank_extraction(): mock_engine.get_tokenizer.return_value = get_tokenizer(MODEL_NAME) mock_engine.errored = False mock_engine.model_config = MockModelConfig() - mock_engine.processor = MagicMock() + mock_engine.input_processor = MagicMock() mock_engine.io_processor = MagicMock() # Mock the generate method to return an async generator diff --git a/tests/entrypoints/openai/test_serving_engine.py b/tests/entrypoints/openai/test_serving_engine.py index 46d8871441a7..29892d0bf38a 100644 --- a/tests/entrypoints/openai/test_serving_engine.py +++ b/tests/entrypoints/openai/test_serving_engine.py @@ -23,7 +23,7 @@ def serving() -> OpenAIServing: model_config.max_model_len = 32768 models = Mock(spec=OpenAIServingModels) models.model_config = model_config - models.processor = Mock() + models.input_processor = Mock() models.io_processor = Mock() serving = OpenAIServing( diff --git a/tests/entrypoints/openai/test_serving_models.py b/tests/entrypoints/openai/test_serving_models.py index 3c022870dba4..b585835a0667 100644 --- a/tests/entrypoints/openai/test_serving_models.py +++ b/tests/entrypoints/openai/test_serving_models.py @@ -30,7 +30,7 @@ async def _async_serving_models_init() -> OpenAIServingModels: mock_model_config = MagicMock(spec=ModelConfig) mock_model_config.max_model_len = 2048 mock_engine_client.model_config = mock_model_config - mock_engine_client.processor = MagicMock() + mock_engine_client.input_processor = MagicMock() mock_engine_client.io_processor = MagicMock() serving_models = OpenAIServingModels( diff --git a/tests/entrypoints/openai/test_serving_responses.py b/tests/entrypoints/openai/test_serving_responses.py index 93e11b61020c..6af32774cc5c 100644 --- a/tests/entrypoints/openai/test_serving_responses.py +++ b/tests/entrypoints/openai/test_serving_responses.py @@ -127,7 +127,7 @@ class TestInitializeToolSessions: model_config.get_diff_sampling_param.return_value = {} engine_client.model_config = model_config - engine_client.processor = MagicMock() + engine_client.input_processor = MagicMock() engine_client.io_processor = MagicMock() models = MagicMock() @@ -213,7 +213,7 @@ class TestValidateGeneratorInput: model_config.get_diff_sampling_param.return_value = {} engine_client.model_config = model_config - engine_client.processor = MagicMock() + engine_client.input_processor = MagicMock() engine_client.io_processor = MagicMock() models = MagicMock() diff --git a/tests/v1/engine/test_processor_multi_modal_uuids.py b/tests/v1/engine/test_process_multi_modal_uuids.py similarity index 87% rename from tests/v1/engine/test_processor_multi_modal_uuids.py rename to tests/v1/engine/test_process_multi_modal_uuids.py index cb6865e42ef8..1b11b8af49d1 100644 --- a/tests/v1/engine/test_processor_multi_modal_uuids.py +++ b/tests/v1/engine/test_process_multi_modal_uuids.py @@ -7,18 +7,17 @@ from vllm.assets.image import ImageAsset from vllm.assets.video import VideoAsset from vllm.config import CacheConfig, DeviceConfig, ModelConfig, VllmConfig from vllm.sampling_params import SamplingParams -from vllm.v1.engine import processor as processor_mod -from vllm.v1.engine.processor import Processor +from vllm.v1.engine import input_processor as input_processor_mod +from vllm.v1.engine.input_processor import InputProcessor cherry_pil_image = ImageAsset("cherry_blossom").pil_image stop_pil_image = ImageAsset("stop_sign").pil_image baby_reading_np_ndarrays = VideoAsset("baby_reading").np_ndarrays -# Mock processor for testing -def _mk_processor( +def _mock_input_processor( monkeypatch, *, mm_cache_gb: float = 4.0, enable_prefix_caching: bool = True -) -> Processor: +) -> InputProcessor: """ Create a Processor instance with minimal configuration suitable for unit tests without accessing external resources. @@ -36,7 +35,7 @@ def _mk_processor( raising=True, ) monkeypatch.setattr( - processor_mod, + input_processor_mod, "processor_cache_from_config", lambda vllm_config, mm_registry: None, raising=True, @@ -65,11 +64,11 @@ def _mk_processor( device_config=DeviceConfig(device="cpu"), ) - return Processor(vllm_config, tokenizer=None) + return InputProcessor(vllm_config, tokenizer=None) def test_multi_modal_uuids_length_mismatch_raises(monkeypatch): - processor = _mk_processor(monkeypatch) + input_processor = _mock_input_processor(monkeypatch) prompt = { "prompt": "USER: \nDescribe\nASSISTANT:", @@ -79,7 +78,7 @@ def test_multi_modal_uuids_length_mismatch_raises(monkeypatch): } with pytest.raises(ValueError, match="must have same length as data"): - processor.process_inputs( + input_processor.process_inputs( request_id="req-1", prompt=prompt, # type: ignore[arg-type] params=SamplingParams(), @@ -87,7 +86,7 @@ def test_multi_modal_uuids_length_mismatch_raises(monkeypatch): def test_multi_modal_uuids_missing_modality_raises(monkeypatch): - processor = _mk_processor(monkeypatch) + input_processor = _mock_input_processor(monkeypatch) prompt = { "prompt": "USER: