From 79bb9be92b27832032f50b68133d390593e52d14 Mon Sep 17 00:00:00 2001 From: Jaret Burkett Date: Wed, 28 May 2025 07:42:30 -0600 Subject: [PATCH] Fix issue with saving chroma full finetune. --- extensions_built_in/diffusion_models/chroma/chroma_model.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/extensions_built_in/diffusion_models/chroma/chroma_model.py b/extensions_built_in/diffusion_models/chroma/chroma_model.py index 31eb6986..3888909a 100644 --- a/extensions_built_in/diffusion_models/chroma/chroma_model.py +++ b/extensions_built_in/diffusion_models/chroma/chroma_model.py @@ -411,6 +411,8 @@ class ChromaModel(BaseModel): return self.text_encoder[1].encoder.block[0].layer[0].SelfAttention.q.weight.requires_grad def save_model(self, output_path, meta, save_dtype): + if not output_path.endswith(".safetensors"): + output_path = output_path + ".safetensors" # only save the unet transformer: Chroma = unwrap_model(self.model) state_dict = transformer.state_dict()