From 1d9b737e054ecbbf6250e0541b74d4e8ca668406 Mon Sep 17 00:00:00 2001 From: dancingpipi Date: Wed, 6 Dec 2023 02:52:48 +0800 Subject: [PATCH] Support ChatGLMForConditionalGeneration (#1932) Co-authored-by: shujunhua1 --- vllm/model_executor/model_loader.py | 1 + 1 file changed, 1 insertion(+) diff --git a/vllm/model_executor/model_loader.py b/vllm/model_executor/model_loader.py index fe6ff36edb882..14739d95e1132 100644 --- a/vllm/model_executor/model_loader.py +++ b/vllm/model_executor/model_loader.py @@ -19,6 +19,7 @@ _MODEL_REGISTRY = { "BaichuanForCausalLM": BaichuanForCausalLM, # baichuan-13b "BloomForCausalLM": BloomForCausalLM, "ChatGLMModel": ChatGLMForCausalLM, + "ChatGLMForConditionalGeneration": ChatGLMForCausalLM, "FalconForCausalLM": FalconForCausalLM, "GPT2LMHeadModel": GPT2LMHeadModel, "GPTBigCodeForCausalLM": GPTBigCodeForCausalLM,