From 415d1095278d896135728a54b4307dde3bc83243 Mon Sep 17 00:00:00 2001 From: Zhuohan Li Date: Thu, 16 Nov 2023 14:47:26 -0800 Subject: [PATCH] [Fix] Update Supported Models List (#1690) --- README.md | 2 ++ docs/source/models/supported_models.rst | 9 +++++++++ 2 files changed, 11 insertions(+) diff --git a/README.md b/README.md index 1d8aba50d853..7586bb36ea35 100644 --- a/README.md +++ b/README.md @@ -49,6 +49,7 @@ vLLM seamlessly supports many Hugging Face models, including the following archi - Aquila & Aquila2 (`BAAI/AquilaChat2-7B`, `BAAI/AquilaChat2-34B`, `BAAI/Aquila-7B`, `BAAI/AquilaChat-7B`, etc.) - Baichuan (`baichuan-inc/Baichuan-7B`, `baichuan-inc/Baichuan-13B-Chat`, etc.) - BLOOM (`bigscience/bloom`, `bigscience/bloomz`, etc.) +- ChatGLM (`THUDM/chatglm2-6b`, `THUDM/chatglm3-6b`, etc.) - Falcon (`tiiuae/falcon-7b`, `tiiuae/falcon-40b`, `tiiuae/falcon-rw-7b`, etc.) - GPT-2 (`gpt2`, `gpt2-xl`, etc.) - GPT BigCode (`bigcode/starcoder`, `bigcode/gpt_bigcode-santacoder`, etc.) @@ -61,6 +62,7 @@ vLLM seamlessly supports many Hugging Face models, including the following archi - OPT (`facebook/opt-66b`, `facebook/opt-iml-max-30b`, etc.) - Phi-1.5 (`microsoft/phi-1_5`, etc.) - Qwen (`Qwen/Qwen-7B`, `Qwen/Qwen-7B-Chat`, etc.) +- Yi (`01-ai/Yi-6B`, `01-ai/Yi-34B`, etc.) Install vLLM with pip or [from source](https://vllm.readthedocs.io/en/latest/getting_started/installation.html#build-from-source): diff --git a/docs/source/models/supported_models.rst b/docs/source/models/supported_models.rst index 17f5379ddafe..1d69d6fd6afd 100644 --- a/docs/source/models/supported_models.rst +++ b/docs/source/models/supported_models.rst @@ -20,6 +20,9 @@ Alongside each architecture, we include some popular models that use it. * - :code:`BaiChuanForCausalLM` - Baichuan - :code:`baichuan-inc/Baichuan-7B`, :code:`baichuan-inc/Baichuan-13B-Chat`, etc. + * - :code:`ChatGLMModel` + - ChatGLM + - :code:`THUDM/chatglm2-6b`, :code:`THUDM/chatglm3-6b`, etc. * - :code:`BloomForCausalLM` - BLOOM, BLOOMZ, BLOOMChat - :code:`bigscience/bloom`, :code:`bigscience/bloomz`, etc. @@ -53,9 +56,15 @@ Alongside each architecture, we include some popular models that use it. * - :code:`OPTForCausalLM` - OPT, OPT-IML - :code:`facebook/opt-66b`, :code:`facebook/opt-iml-max-30b`, etc. + * - :code:`PhiForCausalLM` + - Phi-1.5 + - :code:`microsoft/phi-1_5`, etc. * - :code:`QWenLMHeadModel` - Qwen - :code:`Qwen/Qwen-7B`, :code:`Qwen/Qwen-7B-Chat`, etc. + * - :code:`YiForCausalLM` + - Yi + - :code:`01-ai/Yi-6B`, :code:`01-ai/Yi-34B`, etc. If your model uses one of the above model architectures, you can seamlessly run your model with vLLM. Otherwise, please refer to :ref:`Adding a New Model ` for instructions on how to implement support for your model.