Commit Graph

8350 Commits

Author SHA1 Message Date
layerdiffusion
31bed671ac change some dtype behaviors based on community feedbacks
only influence old devices like 1080/70/60/50.
please remove cmd flags if you are on 1080/70/60/50 and previously used many cmd flags to tune performance
2024-08-21 08:46:52 -07:00
layerdiffusion
2b1e7366a7 Update requirements.txt 2024-08-21 03:31:24 -07:00
lllyasviel
c73ff3724c update some codes related to win32 2024-08-21 03:23:34 -07:00
layerdiffusion
3e253012b5 huggingface-hub==0.24.6 2024-08-20 23:44:30 -07:00
layerdiffusion
49435de094 add removal hint for space 2024-08-20 23:23:43 -07:00
layerdiffusion
579ff49225 revise space 2024-08-20 23:05:25 -07:00
layerdiffusion
389d011fee revise space 2024-08-20 23:01:34 -07:00
layerdiffusion
8b90bd88b5 fix #1367 2024-08-20 22:35:25 -07:00
layerdiffusion
7d99a193e9 space use better delete logics 2024-08-20 22:24:05 -07:00
layerdiffusion
69f238ea38 revise hints 2024-08-20 22:01:31 -07:00
layerdiffusion
8752bfc1b0 revise space 2024-08-20 21:53:02 -07:00
layerdiffusion
250ae27749 The First IDM-VTON that pass 4GB VRAM
100% reproduce official results
2024-08-20 21:30:47 -07:00
layerdiffusion
25e97a8895 revise space 2024-08-20 21:27:16 -07:00
layerdiffusion
1096c708cc revise swap module name 2024-08-20 21:18:53 -07:00
layerdiffusion
4ef29f9546 revise space 2024-08-20 20:40:56 -07:00
DenOfEquity
a46cfa6a1d minor fixes related to Extras tab (#1312)
* update ui.js - correct index for extras tab

one character change

* Update postprocessing.py

fix missing attribute orig_name by using name instead
avoid duplication of postprocessing text. Previously written twice, to png sections postprocessing and extras.

* Update postprocessing.py

unnecessary line
2024-08-20 20:28:42 -07:00
altoiddealer
8bf98cee93 Update README.md (#1346) 2024-08-20 20:28:22 -07:00
DenOfEquity
b2353a4911 fix hires-fix button (#1360)
underlying gallery object changed with gradio update, old code broken, new code more simple
added check for attempt to upscale grid
removed redundant check already covered by second assert
2024-08-20 20:28:07 -07:00
layerdiffusion
cb783405bb revise space 2024-08-20 20:23:33 -07:00
layerdiffusion
7d9f1350f2 revise space 2024-08-20 19:03:17 -07:00
layerdiffusion
56740824e2 add hints and reduce prints to only release slider 2024-08-20 19:02:30 -07:00
layerdiffusion
e750407053 Update README.md 2024-08-20 18:23:12 -07:00
layerdiffusion
0252ad86be missing print 2024-08-20 08:47:06 -07:00
layerdiffusion
74aacc5d4b make "GPU weights" also available to SDXL 2024-08-20 08:19:44 -07:00
layerdiffusion
8fd889dcad fix #1336 2024-08-20 08:04:09 -07:00
layerdiffusion
5452bc6ac3 All Forge Spaces Now Pass 4GB VRAM
and they all 100% reproduce author results
2024-08-20 08:01:10 -07:00
Panchovix
f136f86fee Merge pull request #1340 from DenOfEquity/fix-for-new-samplers
Fix for new samplers
2024-08-20 10:12:37 -04:00
DenOfEquity
c127e60cf0 Update sd_samplers_kdiffusion.py
add new samplers here
2024-08-20 15:01:58 +01:00
DenOfEquity
8c7db614ba Update alter_samplers.py
move new samplers from here
2024-08-20 15:00:56 +01:00
layerdiffusion
14ac95f908 fix 2024-08-20 01:37:01 -07:00
layerdiffusion
6f411a4940 fix loras on nf4 models when activate "loras in fp16" 2024-08-20 01:29:52 -07:00
layerdiffusion
65ec461f8a revise space 2024-08-19 22:43:09 -07:00
layerdiffusion
fef6df29d9 Update README.md 2024-08-19 22:33:20 -07:00
layerdiffusion
6c7c85628e change News to Quick List 2024-08-19 22:30:56 -07:00
layerdiffusion
5ecc525664 fix #1322 2024-08-19 20:19:13 -07:00
layerdiffusion
475524496d revise 2024-08-19 18:54:54 -07:00
Panchovix
8eeeace725 Merge pull request #1316 from lllyasviel/more_samplers1
Add samplers: HeunPP2, IPDNM, IPNDM_V, DEIS
2024-08-19 20:49:37 -04:00
Panchovix
2fc1708a59 Add samplers: HeunPP2, IPDNM, IPNDM_V, DEIS
Pending: CFG++ Samplers, ODE Samplers
The latter is probably easy to implement, the former needs modifications in sd_samplers_cfg_denoiser.py
2024-08-19 20:48:41 -04:00
Panchovix
9bc2d04ca9 Merge pull request #1310 from lllyasviel/more_schedulers
Add Align Your Steps GITS, AYS 11 Steps and AYS 32 Steps Schedulers.
2024-08-19 16:58:52 -04:00
Panchovix
9f5a27ca4e Add Align Your Steps GITS, AYS 11 Steps and AYS 32 Steps Schedulers. 2024-08-19 16:57:58 -04:00
layerdiffusion
d7151b4dcd add low vram warning 2024-08-19 11:08:01 -07:00
layerdiffusion
2f1d04759f avoid some mysteries problems when using lots of python local delegations 2024-08-19 09:47:04 -07:00
layerdiffusion
0b70b7287c gradio 2024-08-19 09:12:38 -07:00
layerdiffusion
584b6c998e #1294 2024-08-19 09:09:22 -07:00
layerdiffusion
054a3416f1 revise space logics 2024-08-19 08:06:24 -07:00
layerdiffusion
96f264ec6a add a way to save models 2024-08-19 06:30:49 -07:00
layerdiffusion
4e8ba14dd0 info 2024-08-19 05:13:28 -07:00
layerdiffusion
d03fc5c2b1 speed up a bit 2024-08-19 05:06:46 -07:00
layerdiffusion
d38e560e42 Implement some rethinking about LoRA system
1. Add an option to allow users to use UNet in fp8/gguf but lora in fp16.
2. All FP16 loras do not need patch. Others will only patch again when lora weight change.
3. FP8 unet + fp16 lora are available (somewhat only available) in Forge now. This also solves some “LoRA too subtle” problems.
4. Significantly speed up all gguf models (in Async mode) by using independent thread (CUDA stream) to compute and dequant at the same time, even when low-bit weights are already on GPU.
5. View “online lora” as a module similar to ControlLoRA so that it is moved to GPU together with model when sampling, achieving significant speedup and perfect low VRAM management simultaneously.
2024-08-19 04:31:59 -07:00
layerdiffusion
e5f213c21e upload some GGUF supports 2024-08-19 01:09:50 -07:00