Fix test_memory_usage_no_spec (#17754)

Signed-off-by: Yong Hoon Shin <yhshin@meta.com>
This commit is contained in:
Yong Hoon Shin 2025-05-07 00:10:33 -07:00 committed by GitHub
parent 8a15c2603a
commit 324a3119b0
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -42,12 +42,12 @@ we can ensure we go through the _no_spec codepath for most of our engine steps.
def test_memory_usage_no_spec():
previous_memory_allocated = None
llm = vllm.LLM(
model=MAIN_MODEL,
speculative_model=SPEC_MODEL,
num_speculative_tokens=3,
speculative_disable_by_batch_size=SPEC_DISABLE_BATCH_SIZE,
)
llm = vllm.LLM(model=MAIN_MODEL,
speculative_config={
"model": SPEC_MODEL,
"num_speculative_tokens": 3,
"disable_by_batch_size": SPEC_DISABLE_BATCH_SIZE,
})
batch_sequences = set()
engine = llm.llm_engine