diff --git a/tests/kernels/quantization/test_silu_nvfp4_quant_fusion.py b/tests/kernels/quantization/test_silu_nvfp4_quant_fusion.py index 969f14cc3fe62..4325162ae94a9 100644 --- a/tests/kernels/quantization/test_silu_nvfp4_quant_fusion.py +++ b/tests/kernels/quantization/test_silu_nvfp4_quant_fusion.py @@ -8,7 +8,8 @@ from vllm.model_executor.layers.activation import SiluAndMul from vllm.platforms import current_platform from vllm.scalar_type import scalar_types -if not current_platform.has_device_capability(100): +if not (current_platform.has_device_capability(100) + and hasattr(torch.ops._C, "silu_and_mul_nvfp4_quant")): pytest.skip(reason="Nvfp4 Requires compute capability of 10 or above.", allow_module_level=True)