mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-13 17:25:38 +08:00
[BugFix] Fix seeded random sampling with encoder-decoder models (#8870)
Co-authored-by: Roger Wang <ywang@roblox.com>
This commit is contained in:
parent
3d49776bbb
commit
31f46a0d35
@ -268,11 +268,13 @@ class EncoderDecoderModelRunner(GPUModelRunnerBase[EncoderDecoderModelInput]):
|
|||||||
encoder_input_positions=encoder_input_positions_tensor,
|
encoder_input_positions=encoder_input_positions_tensor,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
generators = self.get_generators(finished_requests_ids)
|
||||||
sampling_metadata = SamplingMetadata.prepare(seq_group_metadata_list,
|
sampling_metadata = SamplingMetadata.prepare(seq_group_metadata_list,
|
||||||
model_input.seq_lens,
|
model_input.seq_lens,
|
||||||
model_input.query_lens,
|
model_input.query_lens,
|
||||||
self.device,
|
self.device,
|
||||||
self.pin_memory)
|
self.pin_memory,
|
||||||
|
generators=generators)
|
||||||
is_prompt = (seq_group_metadata_list[0].is_prompt
|
is_prompt = (seq_group_metadata_list[0].is_prompt
|
||||||
if seq_group_metadata_list else None)
|
if seq_group_metadata_list else None)
|
||||||
return dataclasses.replace(model_input,
|
return dataclasses.replace(model_input,
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user