Experimental features and bug fixes

This commit is contained in:
Jaret Burkett
2025-02-04 13:36:34 -07:00
parent e6180d1e1d
commit 216ab164ce
6 changed files with 26 additions and 16 deletions

View File

@@ -66,6 +66,7 @@ from optimum.quanto import freeze, qfloat8, quantize, QTensor, qint4
from toolkit.accelerator import get_accelerator, unwrap_model
from typing import TYPE_CHECKING
from toolkit.print import print_acc
from diffusers import FluxFillPipeline
if TYPE_CHECKING:
from toolkit.lora_special import LoRASpecialNetwork
@@ -560,7 +561,10 @@ class StableDiffusion:
)
# hack in model gpu splitter
if self.model_config.split_model_over_gpus:
add_model_gpu_splitter_to_flux(transformer)
add_model_gpu_splitter_to_flux(
transformer,
other_module_param_count_scale=self.model_config.split_model_other_module_param_count_scale
)
if not self.low_vram:
# for low v ram, we leave it on the cpu. Quantizes slower, but allows training on primary gpu