From f71b00a19e25abebdff13a5ae9acc93749125e11 Mon Sep 17 00:00:00 2001 From: Isotr0py Date: Wed, 5 Mar 2025 23:57:10 +0800 Subject: [PATCH] [Bugfix] Fix broken vision language example (#14292) Signed-off-by: Isotr0py <2037008807@qq.com> --- examples/offline_inference/vision_language.py | 40 +++++++++---------- 1 file changed, 18 insertions(+), 22 deletions(-) diff --git a/examples/offline_inference/vision_language.py b/examples/offline_inference/vision_language.py index a0a71f18ed949..270c0f59cc58e 100644 --- a/examples/offline_inference/vision_language.py +++ b/examples/offline_inference/vision_language.py @@ -152,15 +152,13 @@ def run_h2ovl(questions: list[str], modality: str): tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True) - prompts = [ - tokenizer.apply_chat_template([{ - 'role': 'user', - 'content': f"\n{question}" - }], - tokenize=False, - add_generation_prompt=True) - for question in questions - ] + messages = [[{ + 'role': 'user', + 'content': f"\n{question}" + }] for question in questions] + prompts = tokenizer.apply_chat_template(messages, + tokenize=False, + add_generation_prompt=True) # Stop tokens for H2OVL-Mississippi # https://huggingface.co/h2oai/h2ovl-mississippi-800m @@ -209,15 +207,13 @@ def run_internvl(questions: list[str], modality: str): tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True) - prompts = [ - tokenizer.apply_chat_template([{ - 'role': 'user', - 'content': f"\n{question}" - }], - tokenize=False, - add_generation_prompt=True) - for question in questions - ] + messages = [[{ + 'role': 'user', + 'content': f"\n{question}" + }] for question in questions] + prompts = tokenizer.apply_chat_template(messages, + tokenize=False, + add_generation_prompt=True) # Stop tokens for InternVL # models variants may have different stop tokens @@ -399,7 +395,7 @@ def run_mllama(questions: list[str], modality: str): ) tokenizer = AutoTokenizer.from_pretrained(model_name) - messages = [{ + messages = [[{ "role": "user", "content": [{ @@ -408,7 +404,7 @@ def run_mllama(questions: list[str], modality: str): "type": "text", "text": f"{question}" }] - } for question in questions] + }] for question in questions] prompts = tokenizer.apply_chat_template(messages, add_generation_prompt=True, tokenize=False) @@ -454,10 +450,10 @@ def run_nvlm_d(questions: list[str], modality: str): tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True) - messages = [{ + messages = [[{ 'role': 'user', 'content': f"\n{question}" - } for question in questions] + }] for question in questions] prompts = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)