mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-27 05:55:15 +08:00
[Bugfix] DeepSeek V3.2 MTP metadata & CUDA graph issues (#26779)
Signed-off-by: xiaohajiayou <923390377@qq.com>
This commit is contained in:
parent
30a14b034f
commit
d811b442d3
@ -109,6 +109,7 @@ class EagleProposer:
|
||||
else []
|
||||
)
|
||||
|
||||
self.use_cuda_graph = self.use_cuda_graph and bool(self.cudagraph_batch_sizes)
|
||||
# persistent buffers for cuda graph
|
||||
self.input_ids = torch.zeros(
|
||||
self.max_num_tokens, dtype=torch.int32, device=device
|
||||
@ -939,7 +940,7 @@ class EagleProposer:
|
||||
self.vllm_config, DeepseekV32IndexerCache
|
||||
)
|
||||
draft_indexer_layer_names = indexer_layers.keys() - target_indexer_layer_names
|
||||
self.attn_layer_names = list(draft_attn_layer_names)
|
||||
self.attn_layer_names = list(draft_attn_layer_names - draft_indexer_layer_names)
|
||||
self.indexer_layer_names = list(draft_indexer_layer_names)
|
||||
|
||||
if self.indexer_layer_names:
|
||||
@ -1050,16 +1051,18 @@ class EagleProposer:
|
||||
num_tokens: int,
|
||||
use_cudagraphs=True,
|
||||
) -> None:
|
||||
if use_cudagraphs and num_tokens <= self.cudagraph_batch_sizes[-1]:
|
||||
# Determine if CUDA graphs should be used for this run.
|
||||
cudagraphs_enabled = use_cudagraphs and self.use_cuda_graph
|
||||
if cudagraphs_enabled and num_tokens <= self.cudagraph_batch_sizes[-1]:
|
||||
num_tokens = self.vllm_config.pad_for_cudagraph(num_tokens)
|
||||
|
||||
with set_forward_context(
|
||||
None,
|
||||
self.vllm_config,
|
||||
num_tokens=num_tokens,
|
||||
cudagraph_runtime_mode=CUDAGraphMode.PIECEWISE
|
||||
if use_cudagraphs
|
||||
else CUDAGraphMode.NONE,
|
||||
cudagraph_runtime_mode=(
|
||||
CUDAGraphMode.PIECEWISE if cudagraphs_enabled else CUDAGraphMode.NONE
|
||||
),
|
||||
):
|
||||
if self.supports_mm_inputs:
|
||||
input_ids = None
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user