From b5f93d063126dd94518a7773fed74c5b07823719 Mon Sep 17 00:00:00 2001 From: Lukas Kreussel <65088241+LLukas22@users.noreply.github.com> Date: Fri, 15 Sep 2023 02:33:01 +0200 Subject: [PATCH] Only fail if logit_bias has actual values (#1045) --- vllm/entrypoints/openai/api_server.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/vllm/entrypoints/openai/api_server.py b/vllm/entrypoints/openai/api_server.py index 0f1a09b7a5fe4..bc827b3441954 100644 --- a/vllm/entrypoints/openai/api_server.py +++ b/vllm/entrypoints/openai/api_server.py @@ -196,7 +196,7 @@ async def create_chat_completion(request: ChatCompletionRequest, if error_check_ret is not None: return error_check_ret - if request.logit_bias is not None: + if request.logit_bias is not None and len(request.logit_bias) > 0: # TODO: support logit_bias in vLLM engine. return create_error_response(HTTPStatus.BAD_REQUEST, "logit_bias is not currently supported") @@ -379,7 +379,7 @@ async def create_completion(request: CompletionRequest, raw_request: Request): return create_error_response(HTTPStatus.BAD_REQUEST, "suffix is not currently supported") - if request.logit_bias is not None: + if request.logit_bias is not None and len(request.logit_bias) > 0: # TODO: support logit_bias in vLLM engine. return create_error_response(HTTPStatus.BAD_REQUEST, "logit_bias is not currently supported")