From 60ae00303381e956d2d3dbd4c2a99fff398f4e32 Mon Sep 17 00:00:00 2001 From: yewentao256 Date: Fri, 17 Oct 2025 08:24:52 -0700 Subject: [PATCH] fix mypy Signed-off-by: yewentao256 --- vllm/v1/engine/async_llm.py | 2 +- vllm/v1/engine/llm_engine.py | 11 ++--------- 2 files changed, 3 insertions(+), 10 deletions(-) diff --git a/vllm/v1/engine/async_llm.py b/vllm/v1/engine/async_llm.py index 4e5c4cd6a4b67..c487475a72eea 100644 --- a/vllm/v1/engine/async_llm.py +++ b/vllm/v1/engine/async_llm.py @@ -649,7 +649,7 @@ class AsyncLLM(EngineClient): return self.tokenizer async def is_tracing_enabled(self) -> bool: - return self.observability_config.otlp_traces_endpoint is not None + return self.observability_config.otlp_traces_endpoint is not None # type: ignore async def do_log_stats(self) -> None: if self.logger_manager: diff --git a/vllm/v1/engine/llm_engine.py b/vllm/v1/engine/llm_engine.py index 3646b986da71c..c2572ab07debd 100644 --- a/vllm/v1/engine/llm_engine.py +++ b/vllm/v1/engine/llm_engine.py @@ -287,7 +287,7 @@ class LLMEngine: # Add the request to EngineCore. self.engine_core.add_request(child_request) - def step(self) -> list[RequestOutput] | list[PoolingRequestOutput]: + def step(self) -> list[RequestOutput | PoolingRequestOutput]: if self.should_execute_dummy_batch: self.should_execute_dummy_batch = False self.engine_core.execute_dummy_batch() @@ -318,14 +318,7 @@ class LLMEngine: ) self.do_log_stats_with_interval() - ro = processed_outputs.request_outputs - if not ro: - return [] - first = ro[0] - if isinstance(first, RequestOutput): - return [x for x in ro if isinstance(x, RequestOutput)] - else: - return [x for x in ro if isinstance(x, PoolingRequestOutput)] + return processed_outputs.request_outputs def start_profile(self): self.engine_core.profile(True)