diff --git a/modules/sd_models_xl.py b/modules/sd_models_xl.py index 1418348a..89a55635 100644 --- a/modules/sd_models_xl.py +++ b/modules/sd_models_xl.py @@ -12,7 +12,7 @@ import ldm_patched.modules.model_management as model_management def get_learned_conditioning(self: sgm.models.diffusion.DiffusionEngine, batch: prompt_parser.SdConditioning | list[str]): - model_management.load_model_gpu(self.clip_patcher) + model_management.load_model_gpu(self.forge_objects.clip.patcher) for embedder in self.conditioner.embedders: embedder.ucg_rate = 0.0 diff --git a/modules_forge/forge_clip.py b/modules_forge/forge_clip.py index 65fc2480..b71afd82 100644 --- a/modules_forge/forge_clip.py +++ b/modules_forge/forge_clip.py @@ -5,7 +5,7 @@ from modules.shared import opts class CLIP_SD_15_L(FrozenCLIPEmbedderWithCustomWords): def encode_with_transformers(self, tokens): - model_management.load_models_gpu([self.patcher.patcher]) + model_management.load_model_gpu(self.forge_objects.clip.patcher) outputs = self.wrapped.transformer(input_ids=tokens, output_hidden_states=-opts.CLIP_stop_at_last_layers) if opts.CLIP_stop_at_last_layers > 1: @@ -30,7 +30,7 @@ class CLIP_SD_21_H(FrozenCLIPEmbedderWithCustomWords): self.id_pad = 0 def encode_with_transformers(self, tokens): - model_management.load_models_gpu([self.patcher.patcher]) + model_management.load_model_gpu(self.forge_objects.clip.patcher) outputs = self.wrapped.transformer(tokens, output_hidden_states=self.wrapped.layer == "hidden") if self.wrapped.layer == "last":