[Bugfix] Incorrect another MM data format in vllm bench throughput (#26462)

Signed-off-by: Huy Do <huydhn@gmail.com>
This commit is contained in:
Huy Do 2025-10-08 22:58:46 -07:00 committed by GitHub
parent bb6d8c21f9
commit 8bd696fa53
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -59,16 +59,16 @@ def run_vllm(
prompts: list[Union[TextPrompt, TokensPrompt]] = [] prompts: list[Union[TextPrompt, TokensPrompt]] = []
sampling_params: list[SamplingParams] = [] sampling_params: list[SamplingParams] = []
for request in requests: for request in requests:
prompts.append( prompt = (
TokensPrompt( TokensPrompt(prompt_token_ids=request.prompt["prompt_token_ids"])
prompt_token_ids=request.prompt["prompt_token_ids"],
multi_modal_data=request.multi_modal_data,
)
if "prompt_token_ids" in request.prompt if "prompt_token_ids" in request.prompt
else TextPrompt( else TextPrompt(prompt=request.prompt)
prompt=request.prompt, multi_modal_data=request.multi_modal_data
)
) )
if request.multi_modal_data:
assert isinstance(request.multi_modal_data, dict)
prompt["multi_modal_data"] = request.multi_modal_data
prompts.append(prompt)
sampling_params.append( sampling_params.append(
SamplingParams( SamplingParams(
n=n, n=n,