mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-03-16 14:17:16 +08:00
[CI][AMD] spec_decode:eagle skip FLASH_ATTN for deepseek on ROCm (#29827)
Signed-off-by: Divakar Verma <divakar.verma@amd.com>
This commit is contained in:
parent
a690fb5bd6
commit
e2fbfc955e
@ -414,7 +414,10 @@ def test_eagle_correctness(
|
|||||||
)
|
)
|
||||||
|
|
||||||
if attn_backend == "FLASH_ATTN" and current_platform.is_rocm():
|
if attn_backend == "FLASH_ATTN" and current_platform.is_rocm():
|
||||||
m.setenv("VLLM_ROCM_USE_AITER", "1")
|
if "deepseek" in model_setup[1].lower():
|
||||||
|
pytest.skip("FLASH_ATTN for deepseek not supported on ROCm platform")
|
||||||
|
else:
|
||||||
|
m.setenv("VLLM_ROCM_USE_AITER", "1")
|
||||||
|
|
||||||
method, model_name, spec_model_name, tp_size = model_setup
|
method, model_name, spec_model_name, tp_size = model_setup
|
||||||
max_model_len = 2048
|
max_model_len = 2048
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user