From 71972c370922e8f7e16bcc9e0fb6ff7d64368c14 Mon Sep 17 00:00:00 2001 From: Jordan Date: Thu, 23 Feb 2023 01:44:13 -0700 Subject: [PATCH] re-enable load attn procs support (no multiplier) --- ldm/modules/lora_manager.py | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/ldm/modules/lora_manager.py b/ldm/modules/lora_manager.py index b62feae2fb..e9dfe688ec 100644 --- a/ldm/modules/lora_manager.py +++ b/ldm/modules/lora_manager.py @@ -334,9 +334,8 @@ class LoraManager: file = Path(path, "pytorch_lora_weights.bin") if path.is_dir() and file.is_file(): - print(f"Diffusers lora is currently disabled: {path}") - # print(f"loading lora: {path}") - # self.unet.load_attn_procs(path.absolute().as_posix()) + print(f"loading lora: {path}") + self.unet.load_attn_procs(path.absolute().as_posix()) else: path_file = Path(self.lora_path, f'{name}.ckpt') if Path(self.lora_path, f'{name}.safetensors').exists():