diff --git a/vllm/entrypoints/anthropic/serving_messages.py b/vllm/entrypoints/anthropic/serving_messages.py index 11c96adf332f..340dabf0e711 100644 --- a/vllm/entrypoints/anthropic/serving_messages.py +++ b/vllm/entrypoints/anthropic/serving_messages.py @@ -231,9 +231,11 @@ class AnthropicServingMessages(OpenAIServingChat): See https://docs.anthropic.com/en/api/messages for the API specification. This API mimics the Anthropic messages API. """ - logger.debug("Received messages request %s", request.model_dump_json()) + if logger.isEnabledFor(logging.DEBUG): + logger.debug("Received messages request %s", request.model_dump_json()) chat_req = self._convert_anthropic_to_openai_request(request) - logger.debug("Convert to OpenAI request %s", request.model_dump_json()) + if logger.isEnabledFor(logging.DEBUG): + logger.debug("Convert to OpenAI request %s", chat_req.model_dump_json()) generator = await self.create_chat_completion(chat_req, raw_request) if isinstance(generator, ErrorResponse): diff --git a/vllm/entrypoints/openai/api_server.py b/vllm/entrypoints/openai/api_server.py index aba961ac6919..cd55993ae4ba 100644 --- a/vllm/entrypoints/openai/api_server.py +++ b/vllm/entrypoints/openai/api_server.py @@ -648,10 +648,9 @@ async def create_messages(request: AnthropicMessagesRequest, raw_request: Reques return translate_error_response(generator) elif isinstance(generator, AnthropicMessagesResponse): - logger.debug( - "Anthropic Messages Response: %s", generator.model_dump(exclude_none=True) - ) - return JSONResponse(content=generator.model_dump(exclude_none=True)) + resp = generator.model_dump(exclude_none=True) + logger.debug("Anthropic Messages Response: %s", resp) + return JSONResponse(content=resp) return StreamingResponse(content=generator, media_type="text/event-stream")