mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-01-05 04:17:28 +08:00
[V1] Add RayExecutor support for AsyncLLM (api server) (#11712)
This commit is contained in:
parent
d1d49397e7
commit
fbf2564554
@ -22,6 +22,7 @@ from vllm.v1.engine.core_client import EngineCoreClient
|
||||
from vllm.v1.engine.detokenizer import Detokenizer
|
||||
from vllm.v1.engine.processor import Processor
|
||||
from vllm.v1.executor.abstract import Executor
|
||||
from vllm.v1.executor.ray_utils import initialize_ray_cluster
|
||||
|
||||
logger = init_logger(__name__)
|
||||
|
||||
@ -131,7 +132,11 @@ class AsyncLLM(EngineClient):
|
||||
executor_class: Type[Executor]
|
||||
distributed_executor_backend = (
|
||||
vllm_config.parallel_config.distributed_executor_backend)
|
||||
if distributed_executor_backend == "mp":
|
||||
if distributed_executor_backend == "ray":
|
||||
initialize_ray_cluster(vllm_config.parallel_config)
|
||||
from vllm.v1.executor.ray_executor import RayExecutor
|
||||
executor_class = RayExecutor
|
||||
elif distributed_executor_backend == "mp":
|
||||
from vllm.v1.executor.multiproc_executor import MultiprocExecutor
|
||||
executor_class = MultiprocExecutor
|
||||
else:
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user