[Chore] Rename utils submodules (#26920)

Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
This commit is contained in:
Cyrus Leung 2025-10-16 11:58:13 +08:00 committed by GitHub
parent 509cdc0370
commit f6cdc9a02f
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
24 changed files with 28 additions and 24 deletions

View File

@ -12,7 +12,7 @@ from vllm.entrypoints.openai.api_server import (
from vllm.inputs import TextPrompt from vllm.inputs import TextPrompt
from vllm.lora.request import LoRARequest from vllm.lora.request import LoRARequest
from vllm.sampling_params import SamplingParams from vllm.sampling_params import SamplingParams
from vllm.utils.async_utils import merge_async_iterators from vllm.utils.asyncio import merge_async_iterators
MODEL_PATH = "zai-org/chatglm3-6b" MODEL_PATH = "zai-org/chatglm3-6b"
LORA_RANK = 64 LORA_RANK = 64

View File

@ -17,7 +17,7 @@ from transformers import (
) )
from vllm.platforms import current_platform from vllm.platforms import current_platform
from vllm.utils.func import identity from vllm.utils.functools import identity
from ....conftest import ( from ....conftest import (
IMAGE_ASSETS, IMAGE_ASSETS,

View File

@ -5,7 +5,7 @@ from collections.abc import AsyncIterator
import pytest import pytest
from vllm.utils.async_utils import merge_async_iterators from vllm.utils.asyncio import merge_async_iterators
async def _mock_async_iterator(idx: int): async def _mock_async_iterator(idx: int):

View File

@ -4,7 +4,7 @@
import pytest import pytest
from vllm.utils.func import deprecate_kwargs, supports_kw from vllm.utils.functools import deprecate_kwargs, supports_kw
from ..utils import error_on_warning from ..utils import error_on_warning

View File

@ -34,7 +34,7 @@ from vllm.inputs import TextPrompt, TokensPrompt
from vllm.lora.request import LoRARequest from vllm.lora.request import LoRARequest
from vllm.outputs import RequestOutput from vllm.outputs import RequestOutput
from vllm.sampling_params import BeamSearchParams from vllm.sampling_params import BeamSearchParams
from vllm.utils.async_utils import merge_async_iterators from vllm.utils.asyncio import merge_async_iterators
def run_vllm( def run_vllm(

View File

@ -51,7 +51,7 @@ from vllm.transformers_utils.chat_templates import get_chat_template_fallback_pa
from vllm.transformers_utils.processor import cached_get_processor from vllm.transformers_utils.processor import cached_get_processor
from vllm.transformers_utils.tokenizer import AnyTokenizer, MistralTokenizer from vllm.transformers_utils.tokenizer import AnyTokenizer, MistralTokenizer
from vllm.utils import random_uuid from vllm.utils import random_uuid
from vllm.utils.func import supports_kw from vllm.utils.functools import supports_kw
logger = init_logger(__name__) logger = init_logger(__name__)

View File

@ -35,7 +35,7 @@ from vllm.outputs import RequestOutput
from vllm.sampling_params import BeamSearchParams, SamplingParams from vllm.sampling_params import BeamSearchParams, SamplingParams
from vllm.transformers_utils.tokenizer import AnyTokenizer from vllm.transformers_utils.tokenizer import AnyTokenizer
from vllm.utils import as_list from vllm.utils import as_list
from vllm.utils.async_utils import merge_async_iterators from vllm.utils.asyncio import merge_async_iterators
logger = init_logger(__name__) logger = init_logger(__name__)

View File

@ -40,7 +40,7 @@ from vllm.outputs import (
) )
from vllm.pooling_params import PoolingParams from vllm.pooling_params import PoolingParams
from vllm.utils import chunk_list from vllm.utils import chunk_list
from vllm.utils.async_utils import merge_async_iterators from vllm.utils.asyncio import merge_async_iterators
logger = init_logger(__name__) logger = init_logger(__name__)

View File

@ -91,7 +91,7 @@ from vllm.tracing import (
) )
from vllm.transformers_utils.tokenizer import AnyTokenizer, MistralTokenizer from vllm.transformers_utils.tokenizer import AnyTokenizer, MistralTokenizer
from vllm.utils import is_list_of, random_uuid from vllm.utils import is_list_of, random_uuid
from vllm.utils.async_utils import ( from vllm.utils.asyncio import (
AsyncMicrobatchTokenizer, AsyncMicrobatchTokenizer,
collect_from_async_generator, collect_from_async_generator,
make_async, make_async,

View File

@ -36,7 +36,7 @@ from vllm.entrypoints.utils import _validate_truncation_size
from vllm.logger import init_logger from vllm.logger import init_logger
from vllm.outputs import PoolingOutput, PoolingRequestOutput from vllm.outputs import PoolingOutput, PoolingRequestOutput
from vllm.tasks import SupportedTask from vllm.tasks import SupportedTask
from vllm.utils.async_utils import merge_async_iterators from vllm.utils.asyncio import merge_async_iterators
logger = init_logger(__name__) logger = init_logger(__name__)

View File

@ -37,7 +37,7 @@ from vllm.logger import init_logger
from vllm.lora.request import LoRARequest from vllm.lora.request import LoRARequest
from vllm.outputs import PoolingRequestOutput, ScoringRequestOutput from vllm.outputs import PoolingRequestOutput, ScoringRequestOutput
from vllm.transformers_utils.tokenizer import AnyTokenizer, MistralTokenizer from vllm.transformers_utils.tokenizer import AnyTokenizer, MistralTokenizer
from vllm.utils.async_utils import make_async, merge_async_iterators from vllm.utils.asyncio import make_async, merge_async_iterators
logger = init_logger(__name__) logger = init_logger(__name__)

View File

@ -17,7 +17,7 @@ from vllm.inputs.data import TextPrompt as EngineTextPrompt
from vllm.inputs.data import TokensPrompt as EngineTokensPrompt from vllm.inputs.data import TokensPrompt as EngineTokensPrompt
from vllm.inputs.parse import get_prompt_components, parse_raw_prompts from vllm.inputs.parse import get_prompt_components, parse_raw_prompts
from vllm.transformers_utils.tokenizer import AnyTokenizer from vllm.transformers_utils.tokenizer import AnyTokenizer
from vllm.utils.async_utils import AsyncMicrobatchTokenizer from vllm.utils.asyncio import AsyncMicrobatchTokenizer
@dataclass(frozen=True) @dataclass(frozen=True)

View File

@ -17,7 +17,7 @@ from vllm.logger import init_logger
from vllm.lora.request import LoRARequest from vllm.lora.request import LoRARequest
from vllm.sequence import ExecuteModelRequest from vllm.sequence import ExecuteModelRequest
from vllm.tasks import SupportedTask from vllm.tasks import SupportedTask
from vllm.utils.async_utils import make_async from vllm.utils.asyncio import make_async
from vllm.v1.outputs import SamplerOutput from vllm.v1.outputs import SamplerOutput
from vllm.v1.worker.worker_base import WorkerBase from vllm.v1.worker.worker_base import WorkerBase

View File

@ -24,7 +24,7 @@ from vllm.utils import (
get_ip, get_ip,
get_open_port, get_open_port,
) )
from vllm.utils.async_utils import make_async from vllm.utils.asyncio import make_async
from vllm.v1.outputs import SamplerOutput from vllm.v1.outputs import SamplerOutput
if ray is not None: if ray is not None:

View File

@ -29,7 +29,7 @@ from vllm.model_executor.layers.quantization.utils.fp8_utils import (
) )
from vllm.utils import has_deep_gemm from vllm.utils import has_deep_gemm
from vllm.utils.deep_gemm import m_grouped_fp8_gemm_nt_contiguous from vllm.utils.deep_gemm import m_grouped_fp8_gemm_nt_contiguous
from vllm.utils.func import run_once from vllm.utils.functools import run_once
logger = init_logger(__name__) logger = init_logger(__name__)

View File

@ -24,7 +24,7 @@ from vllm.inputs import TokensPrompt
from vllm.inputs.data import PromptType from vllm.inputs.data import PromptType
from vllm.logger import init_logger from vllm.logger import init_logger
from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.layers.quantization import QuantizationConfig
from vllm.utils.func import supports_kw from vllm.utils.functools import supports_kw
from .interfaces_base import VllmModel, is_pooling_model from .interfaces_base import VllmModel, is_pooling_model

View File

@ -15,7 +15,7 @@ import torch.nn as nn
from typing_extensions import TypeIs, TypeVar from typing_extensions import TypeIs, TypeVar
from vllm.logger import init_logger from vllm.logger import init_logger
from vllm.utils.func import supports_kw from vllm.utils.functools import supports_kw
if TYPE_CHECKING: if TYPE_CHECKING:
from vllm.config import VllmConfig from vllm.config import VllmConfig

View File

@ -26,7 +26,7 @@ from vllm.logger import init_logger
from vllm.transformers_utils.processor import cached_processor_from_config from vllm.transformers_utils.processor import cached_processor_from_config
from vllm.transformers_utils.tokenizer import AnyTokenizer, decode_tokens, encode_tokens from vllm.transformers_utils.tokenizer import AnyTokenizer, decode_tokens, encode_tokens
from vllm.utils import flatten_2d_lists, full_groupby from vllm.utils import flatten_2d_lists, full_groupby
from vllm.utils.func import get_allowed_kwarg_only_overrides from vllm.utils.functools import get_allowed_kwarg_only_overrides
from vllm.utils.jsontree import JSONTree, json_map_leaves from vllm.utils.jsontree import JSONTree, json_map_leaves
from .hasher import MultiModalHasher from .hasher import MultiModalHasher

View File

@ -5,7 +5,7 @@ import os
from collections.abc import Mapping from collections.abc import Mapping
from vllm.logger import init_logger from vllm.logger import init_logger
from vllm.utils.func import run_once from vllm.utils.functools import run_once
TRACE_HEADERS = ["traceparent", "tracestate"] TRACE_HEADERS = ["traceparent", "tracestate"]

View File

@ -16,7 +16,7 @@ from transformers.processing_utils import ProcessorMixin
from transformers.video_processing_utils import BaseVideoProcessor from transformers.video_processing_utils import BaseVideoProcessor
from typing_extensions import TypeVar from typing_extensions import TypeVar
from vllm.utils.func import get_allowed_kwarg_only_overrides from vllm.utils.functools import get_allowed_kwarg_only_overrides
if TYPE_CHECKING: if TYPE_CHECKING:
from vllm.config import ModelConfig from vllm.config import ModelConfig

View File

@ -1,6 +1,10 @@
# SPDX-License-Identifier: Apache-2.0 # SPDX-License-Identifier: Apache-2.0
# SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Copyright contributors to the vLLM project
"""Contains helpers related to asynchronous code.""" """
Contains helpers related to asynchronous code.
This is similar in concept to the `asyncio` module.
"""
import asyncio import asyncio
import contextlib import contextlib

View File

@ -30,8 +30,8 @@ from vllm.transformers_utils.config import maybe_register_config_serialize_by_va
from vllm.transformers_utils.tokenizer import AnyTokenizer, init_tokenizer_from_configs from vllm.transformers_utils.tokenizer import AnyTokenizer, init_tokenizer_from_configs
from vllm.usage.usage_lib import UsageContext from vllm.usage.usage_lib import UsageContext
from vllm.utils import Device, as_list, cdiv from vllm.utils import Device, as_list, cdiv
from vllm.utils.async_utils import cancel_task_threadsafe from vllm.utils.asyncio import cancel_task_threadsafe
from vllm.utils.func import deprecate_kwargs from vllm.utils.functools import deprecate_kwargs
from vllm.v1.engine import EngineCoreRequest from vllm.v1.engine import EngineCoreRequest
from vllm.v1.engine.core_client import EngineCoreClient from vllm.v1.engine.core_client import EngineCoreClient
from vllm.v1.engine.exceptions import EngineDeadError, EngineGenerateError from vllm.v1.engine.exceptions import EngineDeadError, EngineGenerateError

View File

@ -29,7 +29,7 @@ from vllm.utils import (
get_open_zmq_inproc_path, get_open_zmq_inproc_path,
make_zmq_socket, make_zmq_socket,
) )
from vllm.utils.async_utils import in_loop from vllm.utils.asyncio import in_loop
from vllm.v1.engine import ( from vllm.v1.engine import (
EngineCoreOutputs, EngineCoreOutputs,
EngineCoreRequest, EngineCoreRequest,