mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-24 23:15:01 +08:00
[Doc] fix mutltimodal_inputs.md gh examples link (#20497)
Signed-off-by: Guy Stone <guys@spotify.com>
This commit is contained in:
parent
c108781c85
commit
d3f05c9248
@ -228,7 +228,7 @@ Our OpenAI-compatible server accepts multi-modal data via the [Chat Completions
|
||||
If no default chat template is available, we will first look for a built-in fallback in <gh-file:vllm/transformers_utils/chat_templates/registry.py>.
|
||||
If no fallback is available, an error is raised and you have to provide the chat template manually via the `--chat-template` argument.
|
||||
|
||||
For certain models, we provide alternative chat templates inside <gh-dir:vllm/examples>.
|
||||
For certain models, we provide alternative chat templates inside <gh-dir:examples>.
|
||||
For example, VLM2Vec uses <gh-file:examples/template_vlm2vec.jinja> which is different from the default one for Phi-3-Vision.
|
||||
|
||||
### Image Inputs
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user