diff --git a/backend/patcher/lora.py b/backend/patcher/lora.py index fdd6f67d..cb18c871 100644 --- a/backend/patcher/lora.py +++ b/backend/patcher/lora.py @@ -421,13 +421,15 @@ class LoraLoader: if gguf_cls is not None: from backend.operations_gguf import ParameterGGUF weight = gguf_cls.quantize_pytorch(weight, gguf_real_shape) - utils.set_attr_raw(self.model, key, ParameterGGUF.make( + weight = ParameterGGUF.make( data=weight, gguf_type=gguf_type, gguf_cls=gguf_cls, gguf_real_shape=gguf_real_shape, parent=parent_layer - )) + ) + gguf_cls.bake_layer(parent_layer, weight, gguf_cls.computation_dtype) + utils.set_attr_raw(self.model, key, weight) continue utils.set_attr_raw(self.model, key, torch.nn.Parameter(weight, requires_grad=False))