* Fix Checkpoint Merging #1359,#1095
- checkpoint_list[] contains the CheckpointInfo.title which is "checkpointname.safetensor [hash]"
when a checkpoint is selected to be loaded during merge, we try to match it with just "checkpointname.safetensor".
-> use checkpoint_aliases[] which already contains the checkpoint key in all possible variants.
- replaced removed sd_models.read_state_dict() with sd_models.load_torch_file()
- replaced removed sd_vae.load_vae_dict() with sd_vae.load_torch_file()
- uncommented create_config() for now, since it calls a removed method: sd_models_config.find_checkpoint_config_near_filename()
* Follow up merge fix for #1359#1095
- read_state_dict() does nothing, replaced 2 occurrences with load_torch_file()
- now merging actually merges again
- checkpoint_list[] contains the CheckpointInfo.title which is "checkpointname.safetensor [hash]"
when a checkpoint is selected to be loaded during merge, we try to match it with just "checkpointname.safetensor".
-> use checkpoint_aliases[] which already contains the checkpoint key in all possible variants.
- replaced removed sd_models.read_state_dict() with sd_models.load_torch_file()
- replaced removed sd_vae.load_vae_dict() with sd_vae.load_torch_file()
- uncommented create_config() for now, since it calls a removed method: sd_models_config.find_checkpoint_config_near_filename()
do not save config if it's the default for checkpoint merger tab
change file naming scheme for checkpoint merger tab
allow just saving A without any merging for checkpoint merger tab
some stylistic changes for UI in checkpoint merger tab
maximum of 2 models loaded at once. delete unneeded model before next
step. fix 'teritary' -> 'tertiary'. gracefully fail when "add
difference" is selected without a tertiary model
There is no need to re-hash the input image each iteration of the loop.
This also reverts PR #4026 as it was determined the cache hits it avoids
were actually valid.