* fix GFPGAN to work with visibility < 1
* fix codeformer to work with visibility < 1
* try harder to download GFPGAN model. Old method would download only if there were no .pth models in the GFPGAN directory. If codeformer was used before GFPGAN, the supporting models are already downloaded into the GFPGAN directory.
* Fix Checkpoint Merging #1359,#1095
- checkpoint_list[] contains the CheckpointInfo.title which is "checkpointname.safetensor [hash]"
when a checkpoint is selected to be loaded during merge, we try to match it with just "checkpointname.safetensor".
-> use checkpoint_aliases[] which already contains the checkpoint key in all possible variants.
- replaced removed sd_models.read_state_dict() with sd_models.load_torch_file()
- replaced removed sd_vae.load_vae_dict() with sd_vae.load_torch_file()
- uncommented create_config() for now, since it calls a removed method: sd_models_config.find_checkpoint_config_near_filename()
* Follow up merge fix for #1359#1095
- read_state_dict() does nothing, replaced 2 occurrences with load_torch_file()
- now merging actually merges again
hide the color/opacity/softness controls on the inpaint tab:
consistent with high contrast option +
they don't offer real control anyway (color irrelevant to mask, mask gets thresholded to binary)
1. fix several problems related to layerdiffuse not unloaded
2. fix several problems related to Fooocus inpaint
3. Slightly speed up on-the-fly LoRAs by precomputing them to computation dtype
In *Settings -> Disregard fields from pasted infotext* there is a very long list of things that can optionally be ignored when parsing infotext. Now it is a slightly longer list, and includes `Lora hashes`.
main issue: Upscaling would fail on single image when controlnet used.
minor issues: On the way to fixing my oversight of not accounting for control images in the gallery, I found that attempting to upscale a control image would fail due to trying to access infotext that doesn't exist. Then I handled a case previously caught by an assert more gracefully. Unhandled, these minor issues would lose the current gallery, so these extra fixes are good QoL.
Then I found another related minor issue if grids are not displayed.
- checkpoint_list[] contains the CheckpointInfo.title which is "checkpointname.safetensor [hash]"
when a checkpoint is selected to be loaded during merge, we try to match it with just "checkpointname.safetensor".
-> use checkpoint_aliases[] which already contains the checkpoint key in all possible variants.
- replaced removed sd_models.read_state_dict() with sd_models.load_torch_file()
- replaced removed sd_vae.load_vae_dict() with sd_vae.load_torch_file()
- uncommented create_config() for now, since it calls a removed method: sd_models_config.find_checkpoint_config_near_filename()
* update ui.js - correct index for extras tab
one character change
* Update postprocessing.py
fix missing attribute orig_name by using name instead
avoid duplication of postprocessing text. Previously written twice, to png sections postprocessing and extras.
* Update postprocessing.py
unnecessary line
underlying gallery object changed with gradio update, old code broken, new code more simple
added check for attempt to upscale grid
removed redundant check already covered by second assert
1. Add an option to allow users to use UNet in fp8/gguf but lora in fp16.
2. All FP16 loras do not need patch. Others will only patch again when lora weight change.
3. FP8 unet + fp16 lora are available (somewhat only available) in Forge now. This also solves some “LoRA too subtle” problems.
4. Significantly speed up all gguf models (in Async mode) by using independent thread (CUDA stream) to compute and dequant at the same time, even when low-bit weights are already on GPU.
5. View “online lora” as a module similar to ControlLoRA so that it is moved to GPU together with model when sampling, achieving significant speedup and perfect low VRAM management simultaneously.