mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-17 03:35:54 +08:00
[Bugfix] Fix GLM-4.1-V video prompt update (#20635)
Signed-off-by: Isotr0py <2037008807@qq.com>
This commit is contained in:
parent
32dffc2772
commit
b9fca83256
@ -65,7 +65,7 @@ from vllm.multimodal.parse import (ImageSize, MultiModalDataItems,
|
||||
MultiModalDataParser)
|
||||
from vllm.multimodal.processing import (BaseMultiModalProcessor,
|
||||
BaseProcessingInfo, PromptReplacement,
|
||||
PromptUpdate)
|
||||
PromptUpdate, PromptUpdateDetails)
|
||||
from vllm.multimodal.profiling import BaseDummyInputsBuilder
|
||||
from vllm.platforms import _Backend
|
||||
from vllm.sequence import IntermediateTensors
|
||||
@ -1213,7 +1213,10 @@ class Glm4vMultiModalProcessor(BaseMultiModalProcessor[Glm4vProcessingInfo]):
|
||||
placeholder.append(eoi_token_id)
|
||||
placeholder.extend(frame_idx)
|
||||
placeholder.append(eov_token_id)
|
||||
return placeholder
|
||||
return PromptUpdateDetails.select_token_id(
|
||||
placeholder,
|
||||
embed_token_id=hf_processor.video_token_id,
|
||||
)
|
||||
|
||||
return [
|
||||
PromptReplacement(
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user