mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-11 04:54:59 +08:00
[Docs] add __init__.py to vllm/model_executor/layers/quantization/compressed_tensors/transform (#24974)
Signed-off-by: samzong <samzong.lu@gmail.com>
This commit is contained in:
parent
2506ce5189
commit
138f0d1e75
@ -680,7 +680,7 @@ class Blip2ForConditionalGeneration(nn.Module, SupportsMultiModal, SupportsPP,
|
|||||||
batch.
|
batch.
|
||||||
|
|
||||||
Info:
|
Info:
|
||||||
[Blip2ImageInputs][]
|
[`Blip2ImageInputs`][vllm.model_executor.models.blip2.Blip2ImageInputs]
|
||||||
"""
|
"""
|
||||||
|
|
||||||
if intermediate_tensors is not None:
|
if intermediate_tensors is not None:
|
||||||
|
|||||||
@ -737,7 +737,7 @@ class LlavaForConditionalGeneration(nn.Module, SupportsMultiModal, SupportsPP):
|
|||||||
inputs_embeds: Optional tensor of input embeddings.
|
inputs_embeds: Optional tensor of input embeddings.
|
||||||
|
|
||||||
Info:
|
Info:
|
||||||
[LlavaImageInputs][]
|
[`LlavaImageInputs`][vllm.model_executor.models.llava.LlavaImageInputs]
|
||||||
"""
|
"""
|
||||||
if intermediate_tensors is not None:
|
if intermediate_tensors is not None:
|
||||||
inputs_embeds = None
|
inputs_embeds = None
|
||||||
|
|||||||
@ -527,7 +527,8 @@ class LlavaNextForConditionalGeneration(nn.Module, SupportsMultiModal,
|
|||||||
Unlike in LLaVA-1.5, the number of image tokens inputted to the language
|
Unlike in LLaVA-1.5, the number of image tokens inputted to the language
|
||||||
model depends on the original size of the input image. Including the
|
model depends on the original size of the input image. Including the
|
||||||
original image token in the input, the required number of image tokens
|
original image token in the input, the required number of image tokens
|
||||||
is given by [get_llava_next_image_feature_size][].
|
is given by [`LlavaNextProcessingInfo.get_num_image_tokens`][vllm.\
|
||||||
|
model_executor.models.llava_next.LlavaNextProcessingInfo.get_num_image_tokens].
|
||||||
|
|
||||||
This way, the `positions` and `attn_metadata` are consistent
|
This way, the `positions` and `attn_metadata` are consistent
|
||||||
with the `input_ids`.
|
with the `input_ids`.
|
||||||
@ -540,7 +541,7 @@ class LlavaNextForConditionalGeneration(nn.Module, SupportsMultiModal,
|
|||||||
inputs_embeds: Optional tensor of input embeddings.
|
inputs_embeds: Optional tensor of input embeddings.
|
||||||
|
|
||||||
Info:
|
Info:
|
||||||
[LlavaNextImageInputs][]
|
[`LlavaNextImageInputs`][vllm.model_executor.models.llava_next.LlavaNextImageInputs]
|
||||||
"""
|
"""
|
||||||
if intermediate_tensors is not None:
|
if intermediate_tensors is not None:
|
||||||
inputs_embeds = None
|
inputs_embeds = None
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user