From a9e15c21efbbc5b4a7a1e69e40378fdfe1acdcb7 Mon Sep 17 00:00:00 2001 From: Asaf Joseph Gardin <39553475+Josephasafg@users.noreply.github.com> Date: Wed, 17 Dec 2025 10:48:53 +0200 Subject: [PATCH] [Mamba] Removed disable cascade attn in MambaModelConfig (#30712) Signed-off-by: asafg <39553475+Josephasafg@users.noreply.github.com> --- vllm/model_executor/models/config.py | 6 ------ 1 file changed, 6 deletions(-) diff --git a/vllm/model_executor/models/config.py b/vllm/model_executor/models/config.py index 4b08472538db4..a3624b1cfa5f2 100644 --- a/vllm/model_executor/models/config.py +++ b/vllm/model_executor/models/config.py @@ -308,12 +308,6 @@ class MambaModelConfig(VerifyAndUpdateConfig): if cache_config.mamba_block_size is None: cache_config.mamba_block_size = model_config.max_model_len - # TODO(tdoublep): remove once cascade attention is supported - logger.info( - "Disabling cascade attention since it is not supported for hybrid models." - ) - model_config.disable_cascade_attn = True - class HybridAttentionMambaModelConfig(VerifyAndUpdateConfig): @classmethod