diff --git a/tests/v1/engine/conftest.py b/tests/v1/engine/conftest.py index c5c5d35b83c3e..283a76dab6723 100644 --- a/tests/v1/engine/conftest.py +++ b/tests/v1/engine/conftest.py @@ -6,6 +6,7 @@ import torch from transformers import AutoTokenizer from tests.v1.engine.utils import ( + FULL_STRINGS, NUM_PROMPT_LOGPROBS_UNDER_TEST, NUM_SAMPLE_LOGPROBS_UNDER_TEST, PROMPT_LEN, @@ -18,8 +19,6 @@ from vllm.engine.arg_utils import EngineArgs from ...distributed.conftest import publisher_config, random_port # noqa: F401 -from tests.v1.engine.utils import FULL_STRINGS # isort: skip - EngineCoreSampleLogprobsType = list[tuple[torch.Tensor, torch.Tensor]] EngineCorePromptLogprobsType = tuple[torch.Tensor, torch.Tensor] diff --git a/tests/v1/tpu/test_topk_topp_sampler.py b/tests/v1/tpu/test_topk_topp_sampler.py index c2fc24442c7cd..c6634395bb167 100644 --- a/tests/v1/tpu/test_topk_topp_sampler.py +++ b/tests/v1/tpu/test_topk_topp_sampler.py @@ -8,10 +8,7 @@ import torch_xla from vllm.platforms import current_platform from vllm.v1.sample.ops.topk_topp_sampler import apply_top_k_top_p - -# isort: off from vllm.v1.sample.tpu.sampler import apply_top_k_top_p as apply_top_k_top_p_tpu -# isort: on if not current_platform.is_tpu(): pytest.skip("This test needs a TPU.", allow_module_level=True) diff --git a/vllm/model_executor/layers/quantization/compressed_tensors/schemes/__init__.py b/vllm/model_executor/layers/quantization/compressed_tensors/schemes/__init__.py index fc0634394ece3..ca286675ebd0c 100644 --- a/vllm/model_executor/layers/quantization/compressed_tensors/schemes/__init__.py +++ b/vllm/model_executor/layers/quantization/compressed_tensors/schemes/__init__.py @@ -15,6 +15,7 @@ from .compressed_tensors_w8a8_int8 import CompressedTensorsW8A8Int8 from .compressed_tensors_w8a16_fp8 import CompressedTensorsW8A16Fp8 from .compressed_tensors_wNa16 import WNA16_SUPPORTED_BITS, CompressedTensorsWNA16 +# This avoids circular import error from .compressed_tensors_24 import CompressedTensors24 # isort: skip __all__ = [ diff --git a/vllm/model_executor/models/qwen3_omni_moe_thinker.py b/vllm/model_executor/models/qwen3_omni_moe_thinker.py index d5a75e75aa43e..08bccee9e2d1a 100755 --- a/vllm/model_executor/models/qwen3_omni_moe_thinker.py +++ b/vllm/model_executor/models/qwen3_omni_moe_thinker.py @@ -80,17 +80,12 @@ from .interfaces import ( SupportsMultiModal, SupportsPP, ) - -# yapf conflicts with isort for this block -# yapf: disable from .qwen2_5_omni_thinker import ( Qwen2_5OmniConditionalGenerationMixin, Qwen2_5OmniThinkerDummyInputsBuilder, Qwen2_5OmniThinkerMultiModalProcessor, Qwen2_5OmniThinkerProcessingInfo, ) - -# yapf: enable from .qwen2_5_vl import ( Qwen2_5_VisionAttention, Qwen2_5_VisionRotaryEmbedding,