From 388b70134b10a71d8666a6c1c2749d7ad896467f Mon Sep 17 00:00:00 2001 From: layerdiffusion <19834515+lllyasviel@users.noreply.github.com> Date: Sun, 25 Aug 2024 20:28:40 -0700 Subject: [PATCH] fix offline loras --- backend/patcher/lora.py | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/backend/patcher/lora.py b/backend/patcher/lora.py index fdd6f67d..cb18c871 100644 --- a/backend/patcher/lora.py +++ b/backend/patcher/lora.py @@ -421,13 +421,15 @@ class LoraLoader: if gguf_cls is not None: from backend.operations_gguf import ParameterGGUF weight = gguf_cls.quantize_pytorch(weight, gguf_real_shape) - utils.set_attr_raw(self.model, key, ParameterGGUF.make( + weight = ParameterGGUF.make( data=weight, gguf_type=gguf_type, gguf_cls=gguf_cls, gguf_real_shape=gguf_real_shape, parent=parent_layer - )) + ) + gguf_cls.bake_layer(parent_layer, weight, gguf_cls.computation_dtype) + utils.set_attr_raw(self.model, key, weight) continue utils.set_attr_raw(self.model, key, torch.nn.Parameter(weight, requires_grad=False))