diff --git a/tests/v1/entrypoints/llm/test_struct_output_generate.py b/tests/v1/entrypoints/llm/test_struct_output_generate.py index 316e152e7395c..a00600b87eca1 100644 --- a/tests/v1/entrypoints/llm/test_struct_output_generate.py +++ b/tests/v1/entrypoints/llm/test_struct_output_generate.py @@ -121,7 +121,6 @@ def test_guided_decoding_deprecated(): assert sp1.structured_outputs == guided_decoding -@pytest.mark.skip_global_cleanup @pytest.mark.parametrize( "model_name, backend, tokenizer_mode, speculative_config", PARAMS_MODELS_BACKENDS_TOKENIZER_MODE, @@ -626,7 +625,6 @@ Make the response as short as possible. ) -@pytest.mark.skip_global_cleanup @pytest.mark.parametrize( "model_name, backend, tokenizer_mode, reasoning_parser, speculative_config", # noqa: E501 [ @@ -711,7 +709,6 @@ def test_structured_output_with_reasoning_matrices( jsonschema.validate(instance=output_json, schema=reasoning_schema) -@pytest.mark.skip_global_cleanup @pytest.mark.parametrize("model_name, tokenizer_mode", PARAMS_MODELS_TOKENIZER_MODE) def test_structured_output_auto_mode( unsupported_json_schema: dict[str, Any], @@ -758,7 +755,6 @@ def test_structured_output_auto_mode( assert isinstance(parsed_json, dict) -@pytest.mark.skip_global_cleanup def test_guidance_no_additional_properties(): llm = LLM( model="Qwen/Qwen2.5-1.5B-Instruct",