From 4ec2cee000af209a9499e0696993834af4f45035 Mon Sep 17 00:00:00 2001 From: Reid <61492567+reidliu41@users.noreply.github.com> Date: Wed, 26 Mar 2025 18:12:47 +0800 Subject: [PATCH] [Misc] improve example script output (#15528) Signed-off-by: reidliu41 Co-authored-by: reidliu41 --- examples/offline_inference/basic/basic.py | 5 ++++- examples/offline_inference/basic/chat.py | 5 +++-- examples/offline_inference/basic/classify.py | 4 +++- examples/offline_inference/basic/embed.py | 4 +++- examples/offline_inference/basic/score.py | 4 +++- 5 files changed, 16 insertions(+), 6 deletions(-) diff --git a/examples/offline_inference/basic/basic.py b/examples/offline_inference/basic/basic.py index a6e96c0bb4339..2ba5ec1192b19 100644 --- a/examples/offline_inference/basic/basic.py +++ b/examples/offline_inference/basic/basic.py @@ -18,7 +18,10 @@ llm = LLM(model="facebook/opt-125m") # that contain the prompt, generated text, and other information. outputs = llm.generate(prompts, sampling_params) # Print the outputs. +print("\nGenerated Outputs:\n" + "-" * 60) for output in outputs: prompt = output.prompt generated_text = output.outputs[0].text - print(f"Prompt: {prompt!r}, Generated text: {generated_text!r}") \ No newline at end of file + print(f"Prompt: {prompt!r}") + print(f"Output: {generated_text!r}") + print("-" * 60) \ No newline at end of file diff --git a/examples/offline_inference/basic/chat.py b/examples/offline_inference/basic/chat.py index b2523e533a40a..2dea45f843cf3 100644 --- a/examples/offline_inference/basic/chat.py +++ b/examples/offline_inference/basic/chat.py @@ -27,12 +27,13 @@ def main(args: dict): sampling_params.top_k = top_k def print_outputs(outputs): + print("\nGenerated Outputs:\n" + "-" * 80) for output in outputs: prompt = output.prompt generated_text = output.outputs[0].text - print(f"Prompt: {prompt!r}") + print(f"Prompt: {prompt!r}\n") print(f"Generated text: {generated_text!r}") - print("-" * 80) + print("-" * 80) print("=" * 80) diff --git a/examples/offline_inference/basic/classify.py b/examples/offline_inference/basic/classify.py index 4ef949b4784de..72c29e4c77c30 100644 --- a/examples/offline_inference/basic/classify.py +++ b/examples/offline_inference/basic/classify.py @@ -23,12 +23,14 @@ def main(args: Namespace): outputs = model.classify(prompts) # Print the outputs. + print("\nGenerated Outputs:\n" + "-" * 60) for prompt, output in zip(prompts, outputs): probs = output.outputs.probs probs_trimmed = ((str(probs[:16])[:-1] + ", ...]") if len(probs) > 16 else probs) - print(f"Prompt: {prompt!r} | " + print(f"Prompt: {prompt!r} \n" f"Class Probabilities: {probs_trimmed} (size={len(probs)})") + print("-" * 60) if __name__ == "__main__": diff --git a/examples/offline_inference/basic/embed.py b/examples/offline_inference/basic/embed.py index f1655b6dbe111..0283909a2a84a 100644 --- a/examples/offline_inference/basic/embed.py +++ b/examples/offline_inference/basic/embed.py @@ -23,12 +23,14 @@ def main(args: Namespace): outputs = model.embed(prompts) # Print the outputs. + print("\nGenerated Outputs:\n" + "-" * 60) for prompt, output in zip(prompts, outputs): embeds = output.outputs.embedding embeds_trimmed = ((str(embeds[:16])[:-1] + ", ...]") if len(embeds) > 16 else embeds) - print(f"Prompt: {prompt!r} | " + print(f"Prompt: {prompt!r} \n" f"Embeddings: {embeds_trimmed} (size={len(embeds)})") + print("-" * 60) if __name__ == "__main__": diff --git a/examples/offline_inference/basic/score.py b/examples/offline_inference/basic/score.py index 2d21f1f0e3971..83b8253f4e257 100644 --- a/examples/offline_inference/basic/score.py +++ b/examples/offline_inference/basic/score.py @@ -22,9 +22,11 @@ def main(args: Namespace): outputs = model.score(text_1, texts_2) # Print the outputs. + print("\nGenerated Outputs:\n" + "-" * 60) for text_2, output in zip(texts_2, outputs): score = output.outputs.score - print(f"Pair: {[text_1, text_2]!r} | Score: {score}") + print(f"Pair: {[text_1, text_2]!r} \nScore: {score}") + print("-" * 60) if __name__ == "__main__":