fix(mm): loader for clip embed

This commit is contained in:
psychedelicious
2025-09-23 14:26:03 +10:00
parent f852c03ba5
commit 4ae20f4876

View File

@@ -37,7 +37,6 @@ from invokeai.backend.flux.util import get_flux_ae_params, get_flux_transformers
from invokeai.backend.model_manager.config import (
AnyModelConfig,
CheckpointConfigBase,
CLIPEmbedCheckpointConfig,
CLIPEmbedDiffusersConfig,
ControlNetCheckpointConfig,
ControlNetDiffusersConfig,
@@ -131,27 +130,6 @@ class CLIPDiffusersLoader(ModelLoader):
)
@ModelLoaderRegistry.register(base=BaseModelType.Any, type=ModelType.CLIPEmbed, format=ModelFormat.Checkpoint)
class CLIPCheckpointLoader(ModelLoader):
"""Class to load main models."""
def _load_model(
self,
config: AnyModelConfig,
submodel_type: Optional[SubModelType] = None,
) -> AnyModel:
if not isinstance(config, CLIPEmbedCheckpointConfig):
raise ValueError("Only CLIPEmbedCheckpointConfig models are currently supported here.")
match submodel_type:
case SubModelType.TextEncoder:
return CLIPTextModel.from_pretrained(Path(config.path), use_safetensors=True)
case _:
raise ValueError(
f"Only TextEncoder submodels are currently supported. Received: {submodel_type.value if submodel_type else 'None'}"
)
@ModelLoaderRegistry.register(base=BaseModelType.Any, type=ModelType.T5Encoder, format=ModelFormat.BnbQuantizedLlmInt8b)
class BnbQuantizedLlmInt8bCheckpointModel(ModelLoader):
"""Class to load main models."""