From 69bff9bc893475fbd64c47633cb8ece46cd462c7 Mon Sep 17 00:00:00 2001 From: Zaida Zhou <58739961+zhouzaida@users.noreply.github.com> Date: Fri, 25 Apr 2025 11:16:32 +0800 Subject: [PATCH] fix float16 support for kimi-vl (#17156) Co-authored-by: zhouzaida --- vllm/model_executor/models/kimi_vl.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/vllm/model_executor/models/kimi_vl.py b/vllm/model_executor/models/kimi_vl.py index 5c39907d79e4..8cb8bc22fc0e 100644 --- a/vllm/model_executor/models/kimi_vl.py +++ b/vllm/model_executor/models/kimi_vl.py @@ -340,8 +340,7 @@ class KimiVLForConditionalGeneration(nn.Module, SupportsMultiModal): else: pixel_values = pixel_values.reshape(-1, num_channels, patch_size, patch_size) - # fp32 -> bf16 - pixel_values = pixel_values.to(torch.bfloat16) + pixel_values = pixel_values.to(self.vision_tower.dtype) # image_grid_hws.shape = (N, 2) assert image_grid_hws.ndim == 2, f"unexpected shape for image_grid_hws: {image_grid_hws.shape}"