mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-10 05:04:58 +08:00
[Fix] Fix the condition of max_seq_len (#477)
This commit is contained in:
parent
58df2883cb
commit
2bdea7ac11
@ -190,7 +190,7 @@ class Scheduler:
|
||||
break
|
||||
|
||||
num_prompt_tokens = seq_group.get_seqs()[0].get_len()
|
||||
if num_prompt_tokens >= self.scheduler_config.max_seq_len:
|
||||
if num_prompt_tokens > self.scheduler_config.max_seq_len:
|
||||
logger.warning(
|
||||
f"Input prompt ({num_prompt_tokens} tokens) is too long"
|
||||
" and exceeds limit of "
|
||||
|
||||
@ -300,7 +300,7 @@ class LLMEngine:
|
||||
continue
|
||||
|
||||
# Check if the sequence has reached max_seq_len.
|
||||
if (seq.get_len() >=
|
||||
if (seq.get_len() >
|
||||
self.scheduler.scheduler_config.max_seq_len):
|
||||
self.scheduler.free_seq(
|
||||
seq, SequenceStatus.FINISHED_LENGTH_CAPPED)
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user