[Quantization] Skip Fp4 Test for compressed-tensors (#19217)

This commit is contained in:
Dipika Sikka 2025-06-05 14:21:53 -04:00 committed by GitHub
parent 9ef9173cfa
commit aa49f14832
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -651,6 +651,7 @@ def test_compressed_tensors_2of4_sparse_compressed(vllm_runner, args_2of4):
assert output
@pytest.mark.skip(reason="Skip until the model config is updated")
def test_compressed_tensors_nvfp4a16(vllm_runner):
# run weight only example
model = "nm-testing/TinyLlama-1.1B-Chat-v1.0-FP4"