mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-23 19:55:02 +08:00
[Mamba] Removed disable cascade attn in MambaModelConfig (#30712)
Signed-off-by: asafg <39553475+Josephasafg@users.noreply.github.com>
This commit is contained in:
parent
20fda43151
commit
a9e15c21ef
@ -308,12 +308,6 @@ class MambaModelConfig(VerifyAndUpdateConfig):
|
|||||||
if cache_config.mamba_block_size is None:
|
if cache_config.mamba_block_size is None:
|
||||||
cache_config.mamba_block_size = model_config.max_model_len
|
cache_config.mamba_block_size = model_config.max_model_len
|
||||||
|
|
||||||
# TODO(tdoublep): remove once cascade attention is supported
|
|
||||||
logger.info(
|
|
||||||
"Disabling cascade attention since it is not supported for hybrid models."
|
|
||||||
)
|
|
||||||
model_config.disable_cascade_attn = True
|
|
||||||
|
|
||||||
|
|
||||||
class HybridAttentionMambaModelConfig(VerifyAndUpdateConfig):
|
class HybridAttentionMambaModelConfig(VerifyAndUpdateConfig):
|
||||||
@classmethod
|
@classmethod
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user