mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-03-20 15:16:54 +08:00
fix: don't skip first special token. (#1497)
This commit is contained in:
parent
28b47d1e49
commit
beac8dd461
@ -120,7 +120,11 @@ def detokenize_incrementally(
|
||||
# tokenizers (bigger = more conservative).
|
||||
# Subtract 1 extra to account for the generated token.
|
||||
prefix_offset = max(len(output_tokens) - 6, 0)
|
||||
read_offset = max(len(output_tokens) - 1, 0)
|
||||
# If the first new token is a special token, we can't skip 1 extra token
|
||||
if skip_special_tokens and new_token_id in tokenizer.all_special_ids:
|
||||
read_offset = max(len(output_tokens), 0)
|
||||
else:
|
||||
read_offset = max(len(output_tokens) - 1, 0)
|
||||
else:
|
||||
# Put new_token_id in a list so skip_special_tokens is respected
|
||||
new_tokens = tokenizer.convert_ids_to_tokens(
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user