[Bugfix] Support compile for Transformers multimodal (#23095)

Signed-off-by: raushan <raushan@huggingface.co>
This commit is contained in:
Raushan Turganbay 2025-08-18 15:35:48 +02:00 committed by GitHub
parent 569aefd134
commit 0e3bb543f0
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -709,6 +709,13 @@ def flatten_and_concat(x: list[torch.Tensor]) -> torch.Tensor:
MultiModalProcessor,
info=MultiModalProcessingInfo,
dummy_inputs=MultiModalDummyInputsBuilder)
@support_torch_compile(
dynamic_arg_dims={
"input_ids": 0,
"positions": -1,
"intermediate_tensors": 0,
"inputs_embeds": 0,
}) # set `positions` to last dim to support Qwen-mrope
class TransformersForMultimodalLM(TransformersForCausalLM, SupportsMultiModal):
# Backwards compatibility for prev released models. State dicts back then
# had different formats and cannot be loaded with `AutoModel` mapping as is