Commit Graph

35 Commits

Author SHA1 Message Date
Jaret Burkett
bb6db3d635 Added support for caching text embeddings. This is just initial support and will probably fail for some models. Still needs to be ompimized 2025-08-07 10:27:55 -06:00
Jaret Burkett
4c4a10d439 Remove vision model from qwen text encoder as it is not needed for image generation currently 2025-08-06 11:40:02 -06:00
Jaret Burkett
14ccf2f3ce Refactor qwen5b model code to be qwen 5b specific 2025-08-06 10:54:56 -06:00
Jaret Burkett
5d8922fca2 Add ability to designate a dataset as i2v or t2v for models that support it 2025-08-06 09:29:47 -06:00
Jaret Burkett
93202c7a2b Training working for Qwen Image 2025-08-04 21:14:30 +00:00
Jaret Burkett
9da8b5408e Initial but untested support for qwen_image 2025-08-04 13:29:37 -06:00
Jaret Burkett
a558d5b68f Move transformer back to device on aggresive wan 2.2 pipeline after generation. 2025-07-29 09:13:47 -06:00
Jaret Burkett
1d1199b15b Fix bug that prevented training wan 2.2 with batch size greater than 1 2025-07-29 09:06:25 -06:00
Jaret Burkett
ca7c5c950b Add support for Wan2.2 5B 2025-07-29 05:31:54 -06:00
Jaret Burkett
cefa2ca5fe Added initial support for Hidream E1 training 2025-07-27 15:12:56 -06:00
Daniel Verdu
a77ba5a089 fix: Guidance incorrect shape 2025-07-18 12:49:18 +02:00
Jaret Burkett
611969ec1f Allow control image for omnigen training and sampling 2025-07-09 13:54:55 -06:00
Jaret Burkett
bbb57de6ec Speed up omnigen TE loading 2025-07-05 09:32:00 -06:00
Jaret Burkett
5906a76666 Fixed issue with flux kontext forcing generation image sizes 2025-06-29 05:38:20 -06:00
Jaret Burkett
57a81bc0db Update base model version for kontext meta 2025-06-28 14:48:36 -06:00
Jaret Burkett
01a3c8a9b1 Fix device issue 2025-06-26 19:14:25 -06:00
Jaret Burkett
4f91cb7148 Fix issue with gradient checkpointing and flux kontext 2025-06-26 19:03:12 -06:00
Jaret Burkett
446b0b6989 Remove revision for kontext 2025-06-26 16:46:58 -06:00
Jaret Burkett
60ef2f1df7 Added support for FLUX.1-Kontext-dev 2025-06-26 15:24:37 -06:00
Jaret Burkett
8d9c47316a Work on mean flow. Minor bug fixes. Omnigen improvements 2025-06-26 13:46:20 -06:00
Jaret Burkett
19ea8ecc38 Added support for finetuning OmniGen2. 2025-06-25 13:58:16 -06:00
Jaret Burkett
ffaf2f154a Fix issue with the way chroma handled gradient checkpointing. 2025-05-28 08:41:47 -06:00
Jaret Burkett
79bb9be92b Fix issue with saving chroma full finetune. 2025-05-28 07:42:30 -06:00
Jaret Burkett
79499fa795 Allow fine tuning pruned versions of chroma. Allow flash attention 2 for chroma if it is installed. 2025-05-21 07:02:50 -06:00
Jaret Burkett
6174ba474e Fixed issue with chroma sampling 2025-05-10 18:30:23 +00:00
Jaret Burkett
43cb5603ad Added chroma model to the ui. Added logic to easily pull latest, use local, or use a specific version of chroma. Allow ustom name or path in the ui for custom models 2025-05-07 12:06:30 -06:00
Jaret Burkett
d9700bdb99 Added initial support for f-lite model 2025-05-01 11:15:18 -06:00
Jaret Burkett
add83df5cc Fixed issue with training hidream when batch size is larger than 1 2025-04-21 17:26:29 +00:00
Jaret Burkett
77001ee77f Upodate model tag on loras 2025-04-19 10:41:27 -06:00
Jaret Burkett
0f99fce004 Adjust hidream lora names to work with comfy 2025-04-16 09:24:23 -06:00
Jaret Burkett
524bd2edfc Make flash attn optional. Handle larger batch sizes. 2025-04-14 14:34:46 +00:00
Jaret Burkett
3a5ea2c742 Remove some moe stuff for finetuning. Drastically reduces vram usage 2025-04-14 00:57:34 +00:00
Jaret Burkett
f80cf99f40 Hidream is training, but has a memory leak 2025-04-13 23:28:18 +00:00
Jaret Burkett
594e166ca3 Initial support for hidream. Still a WIP 2025-04-13 13:50:11 -06:00
Jaret Burkett
7c21eac1b3 Added support for Lodestone Rock's Chroma model 2025-04-05 13:21:36 -06:00