mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-03-16 10:17:13 +08:00
Fix precommit
Signed-off-by: Tyler Michael Smith <tysmith@redhat.com>
This commit is contained in:
parent
feeb17303d
commit
2f86f710dd
@ -222,13 +222,15 @@ class GroupCoordinator:
|
||||
|
||||
for ranks in group_ranks:
|
||||
device_group = torch.distributed.new_group(
|
||||
ranks, backend=torch_distributed_backend,
|
||||
ranks,
|
||||
backend=torch_distributed_backend,
|
||||
timeout=envs.VLLM_DISTRIBUTED_INIT_TIMEOUT_SECONDS)
|
||||
# a group with `gloo` backend, to allow direct coordination between
|
||||
# processes through the CPU.
|
||||
cpu_group = torch.distributed.new_group(ranks,
|
||||
backend="gloo",
|
||||
timeout=envs.VLLM_DISTRIBUTED_INIT_TIMEOUT_SECONDS)
|
||||
cpu_group = torch.distributed.new_group(
|
||||
ranks,
|
||||
backend="gloo",
|
||||
timeout=envs.VLLM_DISTRIBUTED_INIT_TIMEOUT_SECONDS)
|
||||
if self.rank in ranks:
|
||||
self.ranks = ranks
|
||||
self.world_size = len(ranks)
|
||||
|
||||
@ -507,7 +507,7 @@ environment_variables: dict[str, Callable[[], Any]] = {
|
||||
lambda: int(os.getenv("VLLM_IMAGE_FETCH_TIMEOUT", "5")),
|
||||
|
||||
# Timeout for torch distributed calls
|
||||
"VLLM_DISTRIBUTED_INIT_TIMEOUT_SECONDS":
|
||||
"VLLM_DISTRIBUTED_INIT_TIMEOUT_SECONDS":
|
||||
lambda: maybe_convert_int(os.getenv("VLLM_IMAGE_FETCH_TIMEOUT", None)),
|
||||
|
||||
# Timeout for fetching videos when serving multimodal models
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user