diff --git a/vllm/entrypoints/openai/serving_score.py b/vllm/entrypoints/openai/serving_score.py index b4fdbfcc7f60..8d47a417f9cd 100644 --- a/vllm/entrypoints/openai/serving_score.py +++ b/vllm/entrypoints/openai/serving_score.py @@ -216,8 +216,8 @@ class ServingScores(OpenAIServing): # cross_encoder models defaults to using pad_token. tokenized_prompts = await asyncio.gather(*( tokenize_async( - text=t1, # type: ignore[arg-type] - text_pair=t2, # type: ignore[arg-type] + text=t1, # type: ignore[arg-type] + text_pair=t2, # type: ignore[arg-type] **tokenization_kwargs) for t1, t2 in input_pairs)) else: # `llm as reranker` models defaults to not using pad_token.