diff --git a/vllm/reasoning/deepseek_v3_reasoning_parser.py b/vllm/reasoning/deepseek_v3_reasoning_parser.py index 6604f70badbcf..4e6758586bf42 100644 --- a/vllm/reasoning/deepseek_v3_reasoning_parser.py +++ b/vllm/reasoning/deepseek_v3_reasoning_parser.py @@ -26,6 +26,8 @@ class DeepSeekV3ReasoningParser(ReasoningParser): chat_kwargs = kwargs.pop("chat_template_kwargs", {}) or {} thinking = bool(chat_kwargs.pop("thinking", False)) + enable_thinking = bool(chat_kwargs.pop("enable_thinking", False)) + thinking = thinking or enable_thinking if thinking: self._parser = DeepSeekR1ReasoningParser(tokenizer, *args, **kwargs) diff --git a/vllm/tokenizers/deepseek_v32.py b/vllm/tokenizers/deepseek_v32.py index bf279a5cf67c5..d519b61ddb76d 100644 --- a/vllm/tokenizers/deepseek_v32.py +++ b/vllm/tokenizers/deepseek_v32.py @@ -50,6 +50,8 @@ class DeepseekV32Tokenizer(CachedHfTokenizer): **kwargs, ) -> str | list[int]: thinking = kwargs.get("thinking", False) + enable_thinking = kwargs.get("enable_thinking", False) + thinking = thinking or enable_thinking thinking_mode = "thinking" if not thinking: thinking_mode = "chat"