mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-14 02:24:58 +08:00
[Chore] Rename utils submodules (#26920)
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
This commit is contained in:
parent
509cdc0370
commit
f6cdc9a02f
@ -12,7 +12,7 @@ from vllm.entrypoints.openai.api_server import (
|
|||||||
from vllm.inputs import TextPrompt
|
from vllm.inputs import TextPrompt
|
||||||
from vllm.lora.request import LoRARequest
|
from vllm.lora.request import LoRARequest
|
||||||
from vllm.sampling_params import SamplingParams
|
from vllm.sampling_params import SamplingParams
|
||||||
from vllm.utils.async_utils import merge_async_iterators
|
from vllm.utils.asyncio import merge_async_iterators
|
||||||
|
|
||||||
MODEL_PATH = "zai-org/chatglm3-6b"
|
MODEL_PATH = "zai-org/chatglm3-6b"
|
||||||
LORA_RANK = 64
|
LORA_RANK = 64
|
||||||
|
|||||||
@ -17,7 +17,7 @@ from transformers import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
from vllm.platforms import current_platform
|
from vllm.platforms import current_platform
|
||||||
from vllm.utils.func import identity
|
from vllm.utils.functools import identity
|
||||||
|
|
||||||
from ....conftest import (
|
from ....conftest import (
|
||||||
IMAGE_ASSETS,
|
IMAGE_ASSETS,
|
||||||
|
|||||||
@ -5,7 +5,7 @@ from collections.abc import AsyncIterator
|
|||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from vllm.utils.async_utils import merge_async_iterators
|
from vllm.utils.asyncio import merge_async_iterators
|
||||||
|
|
||||||
|
|
||||||
async def _mock_async_iterator(idx: int):
|
async def _mock_async_iterator(idx: int):
|
||||||
|
|||||||
@ -4,7 +4,7 @@
|
|||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from vllm.utils.func import deprecate_kwargs, supports_kw
|
from vllm.utils.functools import deprecate_kwargs, supports_kw
|
||||||
|
|
||||||
from ..utils import error_on_warning
|
from ..utils import error_on_warning
|
||||||
|
|
||||||
|
|||||||
@ -34,7 +34,7 @@ from vllm.inputs import TextPrompt, TokensPrompt
|
|||||||
from vllm.lora.request import LoRARequest
|
from vllm.lora.request import LoRARequest
|
||||||
from vllm.outputs import RequestOutput
|
from vllm.outputs import RequestOutput
|
||||||
from vllm.sampling_params import BeamSearchParams
|
from vllm.sampling_params import BeamSearchParams
|
||||||
from vllm.utils.async_utils import merge_async_iterators
|
from vllm.utils.asyncio import merge_async_iterators
|
||||||
|
|
||||||
|
|
||||||
def run_vllm(
|
def run_vllm(
|
||||||
|
|||||||
@ -51,7 +51,7 @@ from vllm.transformers_utils.chat_templates import get_chat_template_fallback_pa
|
|||||||
from vllm.transformers_utils.processor import cached_get_processor
|
from vllm.transformers_utils.processor import cached_get_processor
|
||||||
from vllm.transformers_utils.tokenizer import AnyTokenizer, MistralTokenizer
|
from vllm.transformers_utils.tokenizer import AnyTokenizer, MistralTokenizer
|
||||||
from vllm.utils import random_uuid
|
from vllm.utils import random_uuid
|
||||||
from vllm.utils.func import supports_kw
|
from vllm.utils.functools import supports_kw
|
||||||
|
|
||||||
logger = init_logger(__name__)
|
logger = init_logger(__name__)
|
||||||
|
|
||||||
|
|||||||
@ -35,7 +35,7 @@ from vllm.outputs import RequestOutput
|
|||||||
from vllm.sampling_params import BeamSearchParams, SamplingParams
|
from vllm.sampling_params import BeamSearchParams, SamplingParams
|
||||||
from vllm.transformers_utils.tokenizer import AnyTokenizer
|
from vllm.transformers_utils.tokenizer import AnyTokenizer
|
||||||
from vllm.utils import as_list
|
from vllm.utils import as_list
|
||||||
from vllm.utils.async_utils import merge_async_iterators
|
from vllm.utils.asyncio import merge_async_iterators
|
||||||
|
|
||||||
logger = init_logger(__name__)
|
logger = init_logger(__name__)
|
||||||
|
|
||||||
|
|||||||
@ -40,7 +40,7 @@ from vllm.outputs import (
|
|||||||
)
|
)
|
||||||
from vllm.pooling_params import PoolingParams
|
from vllm.pooling_params import PoolingParams
|
||||||
from vllm.utils import chunk_list
|
from vllm.utils import chunk_list
|
||||||
from vllm.utils.async_utils import merge_async_iterators
|
from vllm.utils.asyncio import merge_async_iterators
|
||||||
|
|
||||||
logger = init_logger(__name__)
|
logger = init_logger(__name__)
|
||||||
|
|
||||||
|
|||||||
@ -91,7 +91,7 @@ from vllm.tracing import (
|
|||||||
)
|
)
|
||||||
from vllm.transformers_utils.tokenizer import AnyTokenizer, MistralTokenizer
|
from vllm.transformers_utils.tokenizer import AnyTokenizer, MistralTokenizer
|
||||||
from vllm.utils import is_list_of, random_uuid
|
from vllm.utils import is_list_of, random_uuid
|
||||||
from vllm.utils.async_utils import (
|
from vllm.utils.asyncio import (
|
||||||
AsyncMicrobatchTokenizer,
|
AsyncMicrobatchTokenizer,
|
||||||
collect_from_async_generator,
|
collect_from_async_generator,
|
||||||
make_async,
|
make_async,
|
||||||
|
|||||||
@ -36,7 +36,7 @@ from vllm.entrypoints.utils import _validate_truncation_size
|
|||||||
from vllm.logger import init_logger
|
from vllm.logger import init_logger
|
||||||
from vllm.outputs import PoolingOutput, PoolingRequestOutput
|
from vllm.outputs import PoolingOutput, PoolingRequestOutput
|
||||||
from vllm.tasks import SupportedTask
|
from vllm.tasks import SupportedTask
|
||||||
from vllm.utils.async_utils import merge_async_iterators
|
from vllm.utils.asyncio import merge_async_iterators
|
||||||
|
|
||||||
logger = init_logger(__name__)
|
logger = init_logger(__name__)
|
||||||
|
|
||||||
|
|||||||
@ -37,7 +37,7 @@ from vllm.logger import init_logger
|
|||||||
from vllm.lora.request import LoRARequest
|
from vllm.lora.request import LoRARequest
|
||||||
from vllm.outputs import PoolingRequestOutput, ScoringRequestOutput
|
from vllm.outputs import PoolingRequestOutput, ScoringRequestOutput
|
||||||
from vllm.transformers_utils.tokenizer import AnyTokenizer, MistralTokenizer
|
from vllm.transformers_utils.tokenizer import AnyTokenizer, MistralTokenizer
|
||||||
from vllm.utils.async_utils import make_async, merge_async_iterators
|
from vllm.utils.asyncio import make_async, merge_async_iterators
|
||||||
|
|
||||||
logger = init_logger(__name__)
|
logger = init_logger(__name__)
|
||||||
|
|
||||||
|
|||||||
@ -17,7 +17,7 @@ from vllm.inputs.data import TextPrompt as EngineTextPrompt
|
|||||||
from vllm.inputs.data import TokensPrompt as EngineTokensPrompt
|
from vllm.inputs.data import TokensPrompt as EngineTokensPrompt
|
||||||
from vllm.inputs.parse import get_prompt_components, parse_raw_prompts
|
from vllm.inputs.parse import get_prompt_components, parse_raw_prompts
|
||||||
from vllm.transformers_utils.tokenizer import AnyTokenizer
|
from vllm.transformers_utils.tokenizer import AnyTokenizer
|
||||||
from vllm.utils.async_utils import AsyncMicrobatchTokenizer
|
from vllm.utils.asyncio import AsyncMicrobatchTokenizer
|
||||||
|
|
||||||
|
|
||||||
@dataclass(frozen=True)
|
@dataclass(frozen=True)
|
||||||
|
|||||||
@ -17,7 +17,7 @@ from vllm.logger import init_logger
|
|||||||
from vllm.lora.request import LoRARequest
|
from vllm.lora.request import LoRARequest
|
||||||
from vllm.sequence import ExecuteModelRequest
|
from vllm.sequence import ExecuteModelRequest
|
||||||
from vllm.tasks import SupportedTask
|
from vllm.tasks import SupportedTask
|
||||||
from vllm.utils.async_utils import make_async
|
from vllm.utils.asyncio import make_async
|
||||||
from vllm.v1.outputs import SamplerOutput
|
from vllm.v1.outputs import SamplerOutput
|
||||||
from vllm.v1.worker.worker_base import WorkerBase
|
from vllm.v1.worker.worker_base import WorkerBase
|
||||||
|
|
||||||
|
|||||||
@ -24,7 +24,7 @@ from vllm.utils import (
|
|||||||
get_ip,
|
get_ip,
|
||||||
get_open_port,
|
get_open_port,
|
||||||
)
|
)
|
||||||
from vllm.utils.async_utils import make_async
|
from vllm.utils.asyncio import make_async
|
||||||
from vllm.v1.outputs import SamplerOutput
|
from vllm.v1.outputs import SamplerOutput
|
||||||
|
|
||||||
if ray is not None:
|
if ray is not None:
|
||||||
|
|||||||
@ -29,7 +29,7 @@ from vllm.model_executor.layers.quantization.utils.fp8_utils import (
|
|||||||
)
|
)
|
||||||
from vllm.utils import has_deep_gemm
|
from vllm.utils import has_deep_gemm
|
||||||
from vllm.utils.deep_gemm import m_grouped_fp8_gemm_nt_contiguous
|
from vllm.utils.deep_gemm import m_grouped_fp8_gemm_nt_contiguous
|
||||||
from vllm.utils.func import run_once
|
from vllm.utils.functools import run_once
|
||||||
|
|
||||||
logger = init_logger(__name__)
|
logger = init_logger(__name__)
|
||||||
|
|
||||||
|
|||||||
@ -24,7 +24,7 @@ from vllm.inputs import TokensPrompt
|
|||||||
from vllm.inputs.data import PromptType
|
from vllm.inputs.data import PromptType
|
||||||
from vllm.logger import init_logger
|
from vllm.logger import init_logger
|
||||||
from vllm.model_executor.layers.quantization import QuantizationConfig
|
from vllm.model_executor.layers.quantization import QuantizationConfig
|
||||||
from vllm.utils.func import supports_kw
|
from vllm.utils.functools import supports_kw
|
||||||
|
|
||||||
from .interfaces_base import VllmModel, is_pooling_model
|
from .interfaces_base import VllmModel, is_pooling_model
|
||||||
|
|
||||||
|
|||||||
@ -15,7 +15,7 @@ import torch.nn as nn
|
|||||||
from typing_extensions import TypeIs, TypeVar
|
from typing_extensions import TypeIs, TypeVar
|
||||||
|
|
||||||
from vllm.logger import init_logger
|
from vllm.logger import init_logger
|
||||||
from vllm.utils.func import supports_kw
|
from vllm.utils.functools import supports_kw
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
if TYPE_CHECKING:
|
||||||
from vllm.config import VllmConfig
|
from vllm.config import VllmConfig
|
||||||
|
|||||||
@ -26,7 +26,7 @@ from vllm.logger import init_logger
|
|||||||
from vllm.transformers_utils.processor import cached_processor_from_config
|
from vllm.transformers_utils.processor import cached_processor_from_config
|
||||||
from vllm.transformers_utils.tokenizer import AnyTokenizer, decode_tokens, encode_tokens
|
from vllm.transformers_utils.tokenizer import AnyTokenizer, decode_tokens, encode_tokens
|
||||||
from vllm.utils import flatten_2d_lists, full_groupby
|
from vllm.utils import flatten_2d_lists, full_groupby
|
||||||
from vllm.utils.func import get_allowed_kwarg_only_overrides
|
from vllm.utils.functools import get_allowed_kwarg_only_overrides
|
||||||
from vllm.utils.jsontree import JSONTree, json_map_leaves
|
from vllm.utils.jsontree import JSONTree, json_map_leaves
|
||||||
|
|
||||||
from .hasher import MultiModalHasher
|
from .hasher import MultiModalHasher
|
||||||
|
|||||||
@ -5,7 +5,7 @@ import os
|
|||||||
from collections.abc import Mapping
|
from collections.abc import Mapping
|
||||||
|
|
||||||
from vllm.logger import init_logger
|
from vllm.logger import init_logger
|
||||||
from vllm.utils.func import run_once
|
from vllm.utils.functools import run_once
|
||||||
|
|
||||||
TRACE_HEADERS = ["traceparent", "tracestate"]
|
TRACE_HEADERS = ["traceparent", "tracestate"]
|
||||||
|
|
||||||
|
|||||||
@ -16,7 +16,7 @@ from transformers.processing_utils import ProcessorMixin
|
|||||||
from transformers.video_processing_utils import BaseVideoProcessor
|
from transformers.video_processing_utils import BaseVideoProcessor
|
||||||
from typing_extensions import TypeVar
|
from typing_extensions import TypeVar
|
||||||
|
|
||||||
from vllm.utils.func import get_allowed_kwarg_only_overrides
|
from vllm.utils.functools import get_allowed_kwarg_only_overrides
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
if TYPE_CHECKING:
|
||||||
from vllm.config import ModelConfig
|
from vllm.config import ModelConfig
|
||||||
|
|||||||
@ -1,6 +1,10 @@
|
|||||||
# SPDX-License-Identifier: Apache-2.0
|
# SPDX-License-Identifier: Apache-2.0
|
||||||
# SPDX-FileCopyrightText: Copyright contributors to the vLLM project
|
# SPDX-FileCopyrightText: Copyright contributors to the vLLM project
|
||||||
"""Contains helpers related to asynchronous code."""
|
"""
|
||||||
|
Contains helpers related to asynchronous code.
|
||||||
|
|
||||||
|
This is similar in concept to the `asyncio` module.
|
||||||
|
"""
|
||||||
|
|
||||||
import asyncio
|
import asyncio
|
||||||
import contextlib
|
import contextlib
|
||||||
@ -30,8 +30,8 @@ from vllm.transformers_utils.config import maybe_register_config_serialize_by_va
|
|||||||
from vllm.transformers_utils.tokenizer import AnyTokenizer, init_tokenizer_from_configs
|
from vllm.transformers_utils.tokenizer import AnyTokenizer, init_tokenizer_from_configs
|
||||||
from vllm.usage.usage_lib import UsageContext
|
from vllm.usage.usage_lib import UsageContext
|
||||||
from vllm.utils import Device, as_list, cdiv
|
from vllm.utils import Device, as_list, cdiv
|
||||||
from vllm.utils.async_utils import cancel_task_threadsafe
|
from vllm.utils.asyncio import cancel_task_threadsafe
|
||||||
from vllm.utils.func import deprecate_kwargs
|
from vllm.utils.functools import deprecate_kwargs
|
||||||
from vllm.v1.engine import EngineCoreRequest
|
from vllm.v1.engine import EngineCoreRequest
|
||||||
from vllm.v1.engine.core_client import EngineCoreClient
|
from vllm.v1.engine.core_client import EngineCoreClient
|
||||||
from vllm.v1.engine.exceptions import EngineDeadError, EngineGenerateError
|
from vllm.v1.engine.exceptions import EngineDeadError, EngineGenerateError
|
||||||
|
|||||||
@ -29,7 +29,7 @@ from vllm.utils import (
|
|||||||
get_open_zmq_inproc_path,
|
get_open_zmq_inproc_path,
|
||||||
make_zmq_socket,
|
make_zmq_socket,
|
||||||
)
|
)
|
||||||
from vllm.utils.async_utils import in_loop
|
from vllm.utils.asyncio import in_loop
|
||||||
from vllm.v1.engine import (
|
from vllm.v1.engine import (
|
||||||
EngineCoreOutputs,
|
EngineCoreOutputs,
|
||||||
EngineCoreRequest,
|
EngineCoreRequest,
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user