From d8f094a92aad5b102a798de7d6da1454972620c7 Mon Sep 17 00:00:00 2001 From: Reid <61492567+reidliu41@users.noreply.github.com> Date: Sun, 6 Apr 2025 10:57:18 +0800 Subject: [PATCH] [Misc] format output for encoder_decoder.py (#16095) Signed-off-by: reidliu41 Co-authored-by: reidliu41 --- examples/offline_inference/encoder_decoder.py | 11 ++++++----- 1 file changed, 6 insertions(+), 5 deletions(-) diff --git a/examples/offline_inference/encoder_decoder.py b/examples/offline_inference/encoder_decoder.py index 8765d1812cc53..c6ccfd42ec85b 100644 --- a/examples/offline_inference/encoder_decoder.py +++ b/examples/offline_inference/encoder_decoder.py @@ -75,8 +75,6 @@ prompts = [ enc_dec_prompt1, enc_dec_prompt2, enc_dec_prompt3 ] + zipped_prompt_list -print(prompts) - # Create a sampling params object. sampling_params = SamplingParams( temperature=0, @@ -91,10 +89,13 @@ sampling_params = SamplingParams( outputs = llm.generate(prompts, sampling_params) # Print the outputs. -for output in outputs: +print("-" * 50) +for i, output in enumerate(outputs): prompt = output.prompt encoder_prompt = output.encoder_prompt generated_text = output.outputs[0].text - print(f"Encoder prompt: {encoder_prompt!r}, " - f"Decoder prompt: {prompt!r}, " + print(f"Output {i+1}:") + print(f"Encoder prompt: {encoder_prompt!r}\n" + f"Decoder prompt: {prompt!r}\n" f"Generated text: {generated_text!r}") + print("-" * 50)