From 2731e8cbcb74c1fb2abfa4f8c460f9949a2c8297 Mon Sep 17 00:00:00 2001 From: Sage Moore Date: Mon, 2 Jun 2025 19:30:01 +0000 Subject: [PATCH] temporarily remove enable_microbatching Signed-off-by: Sage Moore --- vllm/config.py | 9 --------- 1 file changed, 9 deletions(-) diff --git a/vllm/config.py b/vllm/config.py index 33afedbacecda..5382e9a16829d 100644 --- a/vllm/config.py +++ b/vllm/config.py @@ -1741,15 +1741,6 @@ class ParallelConfig: rank: int = 0 """Global rank in distributed setup.""" - always_microbatch_if_enabled: bool = True - """Always microbatch if microbatching is enabled. Easier to sync between - dp workers.""" - - microbatching_token_threshold: int = 4 - """The threshold for microbatching. If the number of tokens in the - request is greater than this threshold, microbatching will be used. - Otherwise, the request will be processed in a single batch.""" - @property def world_size_across_dp(self) -> int: """world_size_across_dp is TPxPPxDP, it is the size of the world