mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-01-01 19:16:33 +08:00
[Bugfix] Add mm_processor_kwargs to chat-related protocols (#13644)
This commit is contained in:
parent
c7b07a95a6
commit
a30c093502
@ -974,6 +974,10 @@ class EmbeddingChatRequest(OpenAIBaseModel):
|
||||
description=("Additional kwargs to pass to the template renderer. "
|
||||
"Will be accessible by the chat template."),
|
||||
)
|
||||
mm_processor_kwargs: Optional[Dict[str, Any]] = Field(
|
||||
default=None,
|
||||
description=("Additional kwargs to pass to the HF processor."),
|
||||
)
|
||||
priority: int = Field(
|
||||
default=0,
|
||||
description=(
|
||||
@ -1394,6 +1398,10 @@ class TokenizeChatRequest(OpenAIBaseModel):
|
||||
description=("Additional kwargs to pass to the template renderer. "
|
||||
"Will be accessible by the chat template."),
|
||||
)
|
||||
mm_processor_kwargs: Optional[Dict[str, Any]] = Field(
|
||||
default=None,
|
||||
description=("Additional kwargs to pass to the HF processor."),
|
||||
)
|
||||
|
||||
@model_validator(mode="before")
|
||||
@classmethod
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user