From 328d231c17a6c414a40d6abf31c4deb410a750f8 Mon Sep 17 00:00:00 2001 From: MoeedDar Date: Tue, 18 Jul 2023 16:47:59 +0100 Subject: [PATCH] Fixed old name reference for max_seq_len --- vllm/core/scheduler.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/core/scheduler.py b/vllm/core/scheduler.py index 2b37a3af26295..bff3577485969 100644 --- a/vllm/core/scheduler.py +++ b/vllm/core/scheduler.py @@ -196,7 +196,7 @@ class Scheduler: logger.warning( f"Input prompt ({num_prompt_tokens} tokens) is too long" " and exceeds limit of " - f"{self.scheduler_config.max_seq_len}") + f"{self.scheduler_config.max_model_len}") for seq in seq_group.get_seqs(): seq.status = SequenceStatus.FINISHED_IGNORED ignored_seq_groups.append(seq_group)