mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-05-09 01:16:53 +08:00
[ez] move harmony utils to parser folder (#30117)
Signed-off-by: Andrew Xia <axia@fb.com> Co-authored-by: Andrew Xia <axia@fb.com>
This commit is contained in:
parent
671427efbf
commit
421125d03a
0
tests/entrypoints/openai/parser/__init__.py
Normal file
0
tests/entrypoints/openai/parser/__init__.py
Normal file
@ -4,7 +4,7 @@
|
|||||||
from openai.types.responses import ResponseFunctionToolCall, ResponseReasoningItem
|
from openai.types.responses import ResponseFunctionToolCall, ResponseReasoningItem
|
||||||
from openai_harmony import Author, Message, Role, TextContent
|
from openai_harmony import Author, Message, Role, TextContent
|
||||||
|
|
||||||
from vllm.entrypoints.harmony_utils import (
|
from vllm.entrypoints.openai.parser.harmony_utils import (
|
||||||
has_custom_tools,
|
has_custom_tools,
|
||||||
parse_input_to_harmony_message,
|
parse_input_to_harmony_message,
|
||||||
parse_output_message,
|
parse_output_message,
|
||||||
@ -726,7 +726,7 @@ async def test_function_calling_required(client: OpenAI, model_name: str):
|
|||||||
@pytest.mark.asyncio
|
@pytest.mark.asyncio
|
||||||
@pytest.mark.parametrize("model_name", [MODEL_NAME])
|
@pytest.mark.parametrize("model_name", [MODEL_NAME])
|
||||||
async def test_system_message_with_tools(client: OpenAI, model_name: str):
|
async def test_system_message_with_tools(client: OpenAI, model_name: str):
|
||||||
from vllm.entrypoints.harmony_utils import get_system_message
|
from vllm.entrypoints.openai.parser.harmony_utils import get_system_message
|
||||||
|
|
||||||
# Test with custom tools enabled - commentary channel should be available
|
# Test with custom tools enabled - commentary channel should be available
|
||||||
sys_msg = get_system_message(with_custom_tools=True)
|
sys_msg = get_system_message(with_custom_tools=True)
|
||||||
|
|||||||
@ -19,7 +19,7 @@ from vllm import envs
|
|||||||
from vllm.entrypoints.chat_utils import (
|
from vllm.entrypoints.chat_utils import (
|
||||||
ChatTemplateContentFormatOption,
|
ChatTemplateContentFormatOption,
|
||||||
)
|
)
|
||||||
from vllm.entrypoints.harmony_utils import (
|
from vllm.entrypoints.openai.parser.harmony_utils import (
|
||||||
get_encoding,
|
get_encoding,
|
||||||
get_streamable_parser_for_assistant,
|
get_streamable_parser_for_assistant,
|
||||||
render_for_completion,
|
render_for_completion,
|
||||||
|
|||||||
@ -21,7 +21,8 @@ from vllm.entrypoints.chat_utils import (
|
|||||||
get_history_tool_calls_cnt,
|
get_history_tool_calls_cnt,
|
||||||
make_tool_call_id,
|
make_tool_call_id,
|
||||||
)
|
)
|
||||||
from vllm.entrypoints.harmony_utils import (
|
from vllm.entrypoints.logger import RequestLogger
|
||||||
|
from vllm.entrypoints.openai.parser.harmony_utils import (
|
||||||
get_developer_message,
|
get_developer_message,
|
||||||
get_stop_tokens_for_assistant_actions,
|
get_stop_tokens_for_assistant_actions,
|
||||||
get_streamable_parser_for_assistant,
|
get_streamable_parser_for_assistant,
|
||||||
@ -30,7 +31,6 @@ from vllm.entrypoints.harmony_utils import (
|
|||||||
parse_input_to_harmony_message,
|
parse_input_to_harmony_message,
|
||||||
render_for_completion,
|
render_for_completion,
|
||||||
)
|
)
|
||||||
from vllm.entrypoints.logger import RequestLogger
|
|
||||||
from vllm.entrypoints.openai.protocol import (
|
from vllm.entrypoints.openai.protocol import (
|
||||||
ChatCompletionLogProb,
|
ChatCompletionLogProb,
|
||||||
ChatCompletionLogProbs,
|
ChatCompletionLogProbs,
|
||||||
|
|||||||
@ -64,7 +64,8 @@ from vllm.entrypoints.context import (
|
|||||||
SimpleContext,
|
SimpleContext,
|
||||||
StreamingHarmonyContext,
|
StreamingHarmonyContext,
|
||||||
)
|
)
|
||||||
from vllm.entrypoints.harmony_utils import (
|
from vllm.entrypoints.logger import RequestLogger
|
||||||
|
from vllm.entrypoints.openai.parser.harmony_utils import (
|
||||||
construct_harmony_previous_input_messages,
|
construct_harmony_previous_input_messages,
|
||||||
get_developer_message,
|
get_developer_message,
|
||||||
get_stop_tokens_for_assistant_actions,
|
get_stop_tokens_for_assistant_actions,
|
||||||
@ -76,7 +77,6 @@ from vllm.entrypoints.harmony_utils import (
|
|||||||
parse_response_input,
|
parse_response_input,
|
||||||
render_for_completion,
|
render_for_completion,
|
||||||
)
|
)
|
||||||
from vllm.entrypoints.logger import RequestLogger
|
|
||||||
from vllm.entrypoints.openai.protocol import (
|
from vllm.entrypoints.openai.protocol import (
|
||||||
DeltaMessage,
|
DeltaMessage,
|
||||||
ErrorResponse,
|
ErrorResponse,
|
||||||
|
|||||||
@ -4,7 +4,7 @@ import json
|
|||||||
from collections.abc import Sequence
|
from collections.abc import Sequence
|
||||||
from typing import TYPE_CHECKING
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
from vllm.entrypoints.harmony_utils import parse_output_into_messages
|
from vllm.entrypoints.openai.parser.harmony_utils import parse_output_into_messages
|
||||||
from vllm.entrypoints.openai.protocol import (
|
from vllm.entrypoints.openai.protocol import (
|
||||||
ChatCompletionRequest,
|
ChatCompletionRequest,
|
||||||
DeltaMessage,
|
DeltaMessage,
|
||||||
|
|||||||
@ -5,7 +5,7 @@ from collections.abc import Sequence
|
|||||||
|
|
||||||
from transformers import PreTrainedTokenizerBase
|
from transformers import PreTrainedTokenizerBase
|
||||||
|
|
||||||
from vllm.entrypoints.harmony_utils import parse_chat_output
|
from vllm.entrypoints.openai.parser.harmony_utils import parse_chat_output
|
||||||
from vllm.entrypoints.openai.protocol import ChatCompletionRequest, DeltaMessage
|
from vllm.entrypoints.openai.protocol import ChatCompletionRequest, DeltaMessage
|
||||||
from vllm.entrypoints.tool_server import ToolServer
|
from vllm.entrypoints.tool_server import ToolServer
|
||||||
from vllm.logger import init_logger
|
from vllm.logger import init_logger
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user