diff --git a/vllm/config.py b/vllm/config.py index 84aa14b7c860..57b9df236477 100644 --- a/vllm/config.py +++ b/vllm/config.py @@ -1484,7 +1484,7 @@ class CacheConfig: sizes up to 32 are supported. On HPU devices, block size defaults to 128. This config has no static default. If left unspecified by the user, it will - be set in `Platform.check_and_update_configs()` based on the current + be set in `Platform.check_and_update_config()` based on the current platform.""" gpu_memory_utilization: float = 0.9 """The fraction of GPU memory to be used for the model executor, which can