mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-01-02 16:07:30 +08:00
[misc][distributed] improve tests (#6488)
This commit is contained in:
parent
09c2eb85dd
commit
7f62077af5
@ -4,12 +4,14 @@ from ..utils import RemoteOpenAIServer
|
||||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
"TP_SIZE, PP_SIZE, EAGER_MODE, CHUNKED_PREFILL, MODEL_NAME", [
|
||||
"TP_SIZE, PP_SIZE, EAGER_MODE, CHUNKED_PREFILL, MODEL_NAME",
|
||||
[
|
||||
(2, 2, 0, 1, "meta-llama/Meta-Llama-3-8B"),
|
||||
(2, 2, 1, 0, "meta-llama/Meta-Llama-3-8B"),
|
||||
(1, 3, 0, 0, "meta-llama/Meta-Llama-3-8B"),
|
||||
(1, 4, 0, 1, "meta-llama/Meta-Llama-3-8B"),
|
||||
(1, 4, 1, 0, "meta-llama/Meta-Llama-3-8B"),
|
||||
# TODO: figure out why PP=4 tests are flaky
|
||||
# (1, 4, 0, 1, "meta-llama/Meta-Llama-3-8B"),
|
||||
# (1, 4, 1, 0, "meta-llama/Meta-Llama-3-8B"),
|
||||
])
|
||||
def test_compare_tp(TP_SIZE, PP_SIZE, EAGER_MODE, CHUNKED_PREFILL, MODEL_NAME):
|
||||
pp_args = [
|
||||
|
||||
@ -170,7 +170,7 @@ class MessageQueue:
|
||||
self.n_remote_reader = n_remote_reader
|
||||
|
||||
if connect_ip is None:
|
||||
connect_ip = get_ip()
|
||||
connect_ip = get_ip() if n_remote_reader > 0 else "127.0.0.1"
|
||||
|
||||
context = Context()
|
||||
|
||||
@ -230,6 +230,8 @@ class MessageQueue:
|
||||
remote_sync_port=remote_sync_port,
|
||||
)
|
||||
|
||||
logger.info("vLLM message queue communication handle: %s", self.handle)
|
||||
|
||||
def export_handle(self) -> Handle:
|
||||
return self.handle
|
||||
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user