diff --git a/model-list.json b/model-list.json index 47425c8a..872b2e7d 100644 --- a/model-list.json +++ b/model-list.json @@ -691,7 +691,7 @@ "url": "https://huggingface.co/openai/clip-vit-large-patch14/resolve/main/pytorch_model.bin" }, { - "name": "CLIPVision model (IP-Adapter)", + "name": "CLIPVision model (IP-Adapter) 1.5", "type": "clip_vision", "base": "SD1.5", "save_path": "clip_vision/SD1.5", @@ -701,7 +701,7 @@ "url": "https://huggingface.co/h94/IP-Adapter/resolve/main/models/image_encoder/pytorch_model.bin" }, { - "name": "CLIPVision model (IP-Adapter)", + "name": "CLIPVision model (IP-Adapter) XL", "type": "clip_vision", "base": "SDXL", "save_path": "clip_vision/SDXL", diff --git a/node_db/new/model-list.json b/node_db/new/model-list.json index fc2981c8..8f46d3f0 100644 --- a/node_db/new/model-list.json +++ b/node_db/new/model-list.json @@ -658,7 +658,7 @@ }, { - "name": "CLIPVision model (IP-Adapter)", + "name": "CLIPVision model (IP-Adapter) 1.5", "type": "clip_vision", "base": "SD1.5", "save_path": "clip_vision/SD1.5", @@ -668,7 +668,7 @@ "url": "https://huggingface.co/h94/IP-Adapter/resolve/main/models/image_encoder/pytorch_model.bin" }, { - "name": "CLIPVision model (IP-Adapter)", + "name": "CLIPVision model (IP-Adapter) XL", "type": "clip_vision", "base": "SDXL", "save_path": "clip_vision/SDXL",