From 89a0315f4cdf34da40d7634a069068c09082b7dc Mon Sep 17 00:00:00 2001 From: Michael Yao Date: Fri, 9 May 2025 16:20:55 +0800 Subject: [PATCH] [Doc] Update several links in reasoning_outputs.md (#17846) Signed-off-by: windsonsea --- docs/source/features/reasoning_outputs.md | 14 ++++++++------ 1 file changed, 8 insertions(+), 6 deletions(-) diff --git a/docs/source/features/reasoning_outputs.md b/docs/source/features/reasoning_outputs.md index a079eb8b77e77..4759d0c26c35b 100644 --- a/docs/source/features/reasoning_outputs.md +++ b/docs/source/features/reasoning_outputs.md @@ -17,7 +17,9 @@ vLLM currently supports the following reasoning models: | [IBM Granite 3.2 language models](https://huggingface.co/collections/ibm-granite/granite-32-language-models-67b3bc8c13508f6d064cff9a) | `granite` | ❌ | ❌ | | [Qwen3 series](https://huggingface.co/collections/Qwen/qwen3-67dd247413f0e2e4f653967f) | `qwen3` | `guided_json`, `guided_regex` | ✅ | -- IBM Granite 3.2 reasoning is disabled by default; to enable it, you must also pass `thinking=True` in your `chat_template_kwargs`. +:::{note} +IBM Granite 3.2 reasoning is disabled by default; to enable it, you must also pass `thinking=True` in your `chat_template_kwargs`. +::: ## Quickstart @@ -83,7 +85,7 @@ Streaming chat completions are also supported for reasoning models. The `reasoni } ``` -OpenAI Python client library does not officially support `reasoning_content` attribute for streaming output. But the client support extra attributes in the response. You can use `hasattr` to check if the `reasoning_content` attribute is present in the response. For example: +OpenAI Python client library does not officially support `reasoning_content` attribute for streaming output. But the client supports extra attributes in the response. You can use `hasattr` to check if the `reasoning_content` attribute is present in the response. For example: ```python from openai import OpenAI @@ -221,7 +223,7 @@ print(f"Function called: {tool_call.name}") print(f"Arguments: {tool_call.arguments}") ``` -For more examples, please refer to . +For more examples, please refer to . ## Limitations @@ -229,7 +231,7 @@ For more examples, please refer to . ```python # import the required packages @@ -286,7 +288,7 @@ class ExampleParser(ReasoningParser): """ ``` -Additionally, to enable structured output, you'll need to create a new `Reasoner` similar to the one in `vllm/model_executor/guided_decoding/reasoner/deepseek_reasoner.py`. +Additionally, to enable structured output, you'll need to create a new `Reasoner` similar to the one in . ```python @dataclass @@ -312,7 +314,7 @@ class DeepSeekReasoner(Reasoner): ... ``` -The structured output engine like `xgrammar` will use `end_token_id` to check if the reasoning content is present in the model output and skip the structured output if it is the case. +The structured output engine like [xgrammar](https://github.com/mlc-ai/xgrammar) will use `end_token_id` to check if the reasoning content is present in the model output and skip the structured output if it is the case. Finally, you can enable reasoning for the model by using the `--reasoning-parser` flags.