mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-12 08:25:01 +08:00
[XPU] Make pp group initilized for pipeline-parallelism (#11648)
Signed-off-by: yisheng <yi.sheng@intel.com>
This commit is contained in:
parent
d0169e1b0f
commit
d93d2d74fd
@ -11,6 +11,7 @@ import torch.distributed
|
|||||||
from vllm.config import VllmConfig
|
from vllm.config import VllmConfig
|
||||||
from vllm.distributed import (ensure_model_parallel_initialized,
|
from vllm.distributed import (ensure_model_parallel_initialized,
|
||||||
init_distributed_environment)
|
init_distributed_environment)
|
||||||
|
from vllm.distributed.parallel_state import get_pp_group
|
||||||
from vllm.logger import init_logger
|
from vllm.logger import init_logger
|
||||||
from vllm.model_executor import set_random_seed
|
from vllm.model_executor import set_random_seed
|
||||||
from vllm.platforms import current_platform
|
from vllm.platforms import current_platform
|
||||||
@ -176,3 +177,8 @@ class XPUWorker(LoraNotSupportedWorkerBase, Worker):
|
|||||||
parallel_config.pipeline_parallel_size)
|
parallel_config.pipeline_parallel_size)
|
||||||
# global all_reduce needed for overall oneccl warm up
|
# global all_reduce needed for overall oneccl warm up
|
||||||
torch.distributed.all_reduce(torch.zeros(1).xpu())
|
torch.distributed.all_reduce(torch.zeros(1).xpu())
|
||||||
|
|
||||||
|
if parallel_config.pipeline_parallel_size > 1:
|
||||||
|
# Add pp group init to avoid
|
||||||
|
# p2p communication as the first call
|
||||||
|
get_pp_group().all_reduce(torch.zeros(1).xpu())
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user