From 299252ea8243db5c783fbaa4162d611ad10cf2f4 Mon Sep 17 00:00:00 2001 From: Wentao Ye <44945378+yewentao256@users.noreply.github.com> Date: Thu, 10 Jul 2025 15:48:13 -0400 Subject: [PATCH] [CI] Fix pre commit issue (#20782) Signed-off-by: yewentao256 --- vllm/entrypoints/openai/serving_score.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/vllm/entrypoints/openai/serving_score.py b/vllm/entrypoints/openai/serving_score.py index b4fdbfcc7f60..8d47a417f9cd 100644 --- a/vllm/entrypoints/openai/serving_score.py +++ b/vllm/entrypoints/openai/serving_score.py @@ -216,8 +216,8 @@ class ServingScores(OpenAIServing): # cross_encoder models defaults to using pad_token. tokenized_prompts = await asyncio.gather(*( tokenize_async( - text=t1, # type: ignore[arg-type] - text_pair=t2, # type: ignore[arg-type] + text=t1, # type: ignore[arg-type] + text_pair=t2, # type: ignore[arg-type] **tokenization_kwargs) for t1, t2 in input_pairs)) else: # `llm as reranker` models defaults to not using pad_token.