mirror of
https://github.com/invoke-ai/InvokeAI.git
synced 2026-04-23 03:00:31 -04:00
feat(mm): fix clip vision starter model bases, add ref to actual models
This commit is contained in:
@@ -37,16 +37,20 @@ cyberrealistic_negative = StarterModel(
|
||||
)
|
||||
|
||||
# region CLIP Image Encoders
|
||||
|
||||
# This is CLIP-ViT-H-14-laion2B-s32B-b79K
|
||||
ip_adapter_sd_image_encoder = StarterModel(
|
||||
name="IP Adapter SD1.5 Image Encoder",
|
||||
base=BaseModelType.StableDiffusion1,
|
||||
base=BaseModelType.Any,
|
||||
source="InvokeAI/ip_adapter_sd_image_encoder",
|
||||
description="IP Adapter SD Image Encoder",
|
||||
type=ModelType.CLIPVision,
|
||||
)
|
||||
|
||||
# This is CLIP-ViT-bigG-14-laion2B-39B-b160k
|
||||
ip_adapter_sdxl_image_encoder = StarterModel(
|
||||
name="IP Adapter SDXL Image Encoder",
|
||||
base=BaseModelType.StableDiffusionXL,
|
||||
base=BaseModelType.Any,
|
||||
source="InvokeAI/ip_adapter_sdxl_image_encoder",
|
||||
description="IP Adapter SDXL Image Encoder",
|
||||
type=ModelType.CLIPVision,
|
||||
|
||||
Reference in New Issue
Block a user