From 65e403d335159a0e4829c83ddbc44c612b66c49e Mon Sep 17 00:00:00 2001 From: bk-201 Date: Thu, 11 Dec 2025 16:00:16 +0000 Subject: [PATCH] remove outdated comment Signed-off-by: bk-201 --- vllm/lora/layers/row_parallel_linear.py | 4 ---- 1 file changed, 4 deletions(-) diff --git a/vllm/lora/layers/row_parallel_linear.py b/vllm/lora/layers/row_parallel_linear.py index d860359877377..958aa6af36746 100644 --- a/vllm/lora/layers/row_parallel_linear.py +++ b/vllm/lora/layers/row_parallel_linear.py @@ -63,10 +63,6 @@ class RowParallelLinearWithLoRA(BaseLinearLayerWithLoRA): input_parallel = splitted_input[self.tp_rank].contiguous() # Matrix multiply. - # Only fuse bias add into GEMM for rank 0 (matches base - # RowParallelLinear behavior). This ensures bias will not get - # added more than once in TP>1 case and matches the numerical - # behavior of the unwrapped layer bias_ = ( None if (self.tp_rank > 0 or self.base_layer.skip_bias_add)