mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-14 07:25:01 +08:00
[Misc] Fix estimated max model len msg (#18966)
Signed-off-by: Yong Hoon Shin <yhshin@meta.com>
This commit is contained in:
parent
dff80b0e42
commit
1e123529d7
@ -544,8 +544,9 @@ def check_enough_kv_cache_memory(vllm_config: VllmConfig,
|
|||||||
available_memory)
|
available_memory)
|
||||||
estimated_msg = ""
|
estimated_msg = ""
|
||||||
if estimated_max_len > 0:
|
if estimated_max_len > 0:
|
||||||
estimated_msg = " Based on the available memory,"
|
estimated_msg = (
|
||||||
f" the estimated maximum model length is {estimated_max_len}."
|
"Based on the available memory, "
|
||||||
|
f"the estimated maximum model length is {estimated_max_len}.")
|
||||||
|
|
||||||
raise ValueError(
|
raise ValueError(
|
||||||
f"To serve at least one request with the models's max seq len "
|
f"To serve at least one request with the models's max seq len "
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user