diff --git a/vllm/core/scheduler.py b/vllm/core/scheduler.py index 2b37a3af2629..bff357748596 100644 --- a/vllm/core/scheduler.py +++ b/vllm/core/scheduler.py @@ -196,7 +196,7 @@ class Scheduler: logger.warning( f"Input prompt ({num_prompt_tokens} tokens) is too long" " and exceeds limit of " - f"{self.scheduler_config.max_seq_len}") + f"{self.scheduler_config.max_model_len}") for seq in seq_group.get_seqs(): seq.status = SequenceStatus.FINISHED_IGNORED ignored_seq_groups.append(seq_group)