mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-03-26 06:03:40 +08:00
chore: Improve GPU cleanup in tests
Centralizes the GPU memory cleanup logic into a single static method to prevent flaky test failures from OOM errors. Signed-off-by: AzizCode92 <azizbenothman76@gmail.com>
This commit is contained in:
parent
276ec32ba9
commit
2c9ed5b738
@ -170,14 +170,18 @@ class RemoteOpenAIServer:
|
||||
except subprocess.TimeoutExpired:
|
||||
# force kill if needed
|
||||
self.proc.kill()
|
||||
# GPU memory cleanup
|
||||
self.__cleanup_gpu_memory()
|
||||
|
||||
@staticmethod
|
||||
def __cleanup_gpu_memory():
|
||||
try:
|
||||
if current_platform.is_cuda() or current_platform.is_rocm():
|
||||
if current_platform.is_cuda_alike():
|
||||
num_devices = cuda_device_count_stateless()
|
||||
if num_devices > 0:
|
||||
wait_for_gpu_memory_to_clear(devices=list(
|
||||
range(num_devices)),
|
||||
threshold_ratio=0.05)
|
||||
threshold_ratio=0.05,
|
||||
timeout_s=60)
|
||||
except Exception as e:
|
||||
print(f"GPU cleanup warning: {e}")
|
||||
|
||||
@ -276,16 +280,7 @@ class RemoteOpenAIServerCustom(RemoteOpenAIServer):
|
||||
# force kill if needed
|
||||
self.proc.kill()
|
||||
|
||||
# GPU memory cleaning
|
||||
try:
|
||||
if current_platform.is_cuda() or current_platform.is_rocm():
|
||||
num_devices = cuda_device_count_stateless()
|
||||
if num_devices > 0:
|
||||
wait_for_gpu_memory_to_clear(devices=list(
|
||||
range(num_devices)),
|
||||
threshold_ratio=0.05)
|
||||
except Exception as e:
|
||||
print(f"GPU cleanup warning: {e}")
|
||||
self.__cleanup_gpu_memory()
|
||||
|
||||
|
||||
def _test_completion(
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user