mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-01-28 13:27:15 +08:00
[gpt-oss][Bugfix]Fix streamableparser for missing handling of certain token_ids (#24306)
Signed-off-by: chaunceyjiang <chaunceyjiang@gmail.com>
This commit is contained in:
parent
7812bcf278
commit
23a6c5280e
@ -238,11 +238,11 @@ class StreamingHarmonyContext(HarmonyContext):
|
||||
# (finished=True), then the next token processed will mark the
|
||||
# beginning of a new message
|
||||
self.first_tok_of_message = output.finished
|
||||
tok = output.outputs[0].token_ids[0]
|
||||
self.parser.process(tok)
|
||||
for tok in output.outputs[0].token_ids:
|
||||
self.parser.process(tok)
|
||||
self._update_num_output_tokens(output.outputs[0].token_ids)
|
||||
# Check if the current token is part of reasoning content
|
||||
self._update_num_reasoning_tokens([tok])
|
||||
self._update_num_reasoning_tokens(output.outputs[0].token_ids)
|
||||
self.last_tok = tok
|
||||
else:
|
||||
# Handle the case of tool output in direct message format
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user