mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-10 08:25:01 +08:00
[doc] fold long code block (#20795)
Signed-off-by: reidliu41 <reid201711@gmail.com>
This commit is contained in:
parent
5d09152ff1
commit
6a9e6b2abf
@ -279,7 +279,7 @@ Some models, e.g., [Granite Speech](https://huggingface.co/ibm-granite/granite-s
|
|||||||
|
|
||||||
To this end, we allow registration of default multimodal LoRAs to handle this automatically, where users can map each modality to a LoRA adapter to automatically apply it when the corresponding inputs are present. Note that currently, we only allow one LoRA per prompt; if several modalities are provided, each of which are registered to a given modality, none of them will be applied.
|
To this end, we allow registration of default multimodal LoRAs to handle this automatically, where users can map each modality to a LoRA adapter to automatically apply it when the corresponding inputs are present. Note that currently, we only allow one LoRA per prompt; if several modalities are provided, each of which are registered to a given modality, none of them will be applied.
|
||||||
|
|
||||||
Example usage for offline inference:
|
??? code "Example usage for offline inference"
|
||||||
|
|
||||||
```python
|
```python
|
||||||
from transformers import AutoTokenizer
|
from transformers import AutoTokenizer
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user