[Bugfix][benchmarks] Fix input token calculation for rerank benchmark metrics (#30596)

Signed-off-by: vensen <vensenmu@gmail.com>
This commit is contained in:
Vensen 2025-12-14 22:57:15 +08:00 committed by GitHub
parent ae88aada38
commit add4b0ca44
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
3 changed files with 7 additions and 2 deletions

View File

@ -235,7 +235,9 @@ async def get_request(
def calculate_metrics_for_embeddings(
outputs: list[RequestFuncOutput], dur_s: float, selected_percentiles: list[float]
outputs: list[RequestFuncOutput],
dur_s: float,
selected_percentiles: list[float],
) -> EmbedBenchmarkMetrics:
"""Calculate the metrics for the embedding requests.

View File

@ -120,6 +120,7 @@ class RerankResult(BaseModel):
class RerankUsage(BaseModel):
prompt_tokens: int
total_tokens: int

View File

@ -502,5 +502,7 @@ class ServingScores(OpenAIServing):
id=request_id,
model=model_name,
results=results,
usage=RerankUsage(total_tokens=num_prompt_tokens),
usage=RerankUsage(
total_tokens=num_prompt_tokens, prompt_tokens=num_prompt_tokens
),
)