mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-23 09:45:58 +08:00
[Mamba] Removed disable cascade attn in MambaModelConfig (#30712)
Signed-off-by: asafg <39553475+Josephasafg@users.noreply.github.com>
This commit is contained in:
parent
20fda43151
commit
a9e15c21ef
@ -308,12 +308,6 @@ class MambaModelConfig(VerifyAndUpdateConfig):
|
||||
if cache_config.mamba_block_size is None:
|
||||
cache_config.mamba_block_size = model_config.max_model_len
|
||||
|
||||
# TODO(tdoublep): remove once cascade attention is supported
|
||||
logger.info(
|
||||
"Disabling cascade attention since it is not supported for hybrid models."
|
||||
)
|
||||
model_config.disable_cascade_attn = True
|
||||
|
||||
|
||||
class HybridAttentionMambaModelConfig(VerifyAndUpdateConfig):
|
||||
@classmethod
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user