"macapp/webpack.plugins.ts" did not exist on "c5bafaff5424230fce761e37456da50e738f1b1a"
- 27 Jan, 2025 1 commit
-
-
hlky authored
* [training] Convert to ImageFolder script * make
-
- 24 Jan, 2025 1 commit
-
-
Leo Jiang authored
* NPU Adaption for Sanna --------- Co-authored-by:
J石页 <jiangshuo9@h-partners.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
- 21 Jan, 2025 3 commits
-
-
Muyang Li authored
Remove the FP32 Wrapper Co-authored-by:Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com>
-
jiqing-feng authored
* enable dreambooth_lora on other devices Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * enable xpu Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * check cuda device before empty cache Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix comment Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * import free_memory Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> --------- Signed-off-by:
jiqing-feng <jiqing.feng@intel.com>
-
Sayak Paul authored
change licensing to 2025 from 2024.
-
- 15 Jan, 2025 1 commit
-
-
Leo Jiang authored
Co-authored-by:J石页 <jiangshuo9@h-partners.com>
-
- 30 Dec, 2024 1 commit
-
-
Sayak Paul authored
* add ds support to lora sd3. Co-authored-by:
leisuzz <jiangshuonb@gmail.com> * style. --------- Co-authored-by:
leisuzz <jiangshuonb@gmail.com> Co-authored-by:
Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com>
-
- 23 Dec, 2024 2 commits
-
-
Sayak Paul authored
* post release 0.32.0 * stylew
-
Sayak Paul authored
* sana lora training tests and misc. * remove push to hub * Update examples/dreambooth/train_dreambooth_lora_sana.py Co-authored-by:
Aryan <aryan@huggingface.co> --------- Co-authored-by:
Aryan <aryan@huggingface.co>
-
- 19 Dec, 2024 1 commit
-
-
Sayak Paul authored
Update README_sana.md to update the default model
-
- 18 Dec, 2024 2 commits
-
-
Sayak Paul authored
fix: reamde -> readme
-
Sayak Paul authored
* feat: lora support for SANA. * make fix-copies * rename test class. * attention_kwargs -> cross_attention_kwargs. * Revert "attention_kwargs -> cross_attention_kwargs." This reverts commit 23433bf9bccc12e0f2f55df26bae58a894e8b43b. * exhaust 119 max line limit * sana lora fine-tuning script. * readme * add a note about the supported models. * Apply suggestions from code review Co-authored-by:
Aryan <aryan@huggingface.co> * style * docs for attention_kwargs. * remove lora_scale from pag pipeline. * copy fix --------- Co-authored-by:
Aryan <aryan@huggingface.co>
-
- 12 Dec, 2024 1 commit
-
-
Ethan Smith authored
* fix min-snr implementation https://github.com/kohya-ss/sd-scripts/blob/main/library/custom_train_functions.py#L66 * Update train_dreambooth.py fix variable name mse_loss_weights * fix divisor * make style --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com>
-
- 25 Nov, 2024 1 commit
-
-
SkyCol authored
Add files via upload
-
- 24 Nov, 2024 1 commit
-
-
Linoy Tsaban authored
* smol change to fix checkpoint saving & resuming (as done in train_dreambooth_sd3.py) * style * modify comment to explain reasoning behind hidden size check
-
- 19 Nov, 2024 1 commit
-
-
Linoy Tsaban authored
* memory improvement as done here: https://github.com/huggingface/diffusers/pull/9829 * fix bug * fix bug * style --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
- 08 Nov, 2024 1 commit
-
-
SahilCarterr authored
* fix use_dora * fix style and quality * fix use_dora with peft version --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
- 06 Nov, 2024 1 commit
-
-
SahilCarterr authored
* updated encode prompt and clip encod prompt --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
- 01 Nov, 2024 3 commits
-
-
Leo Jiang authored
* Improve NPU performance * Improve NPU performance * Improve NPU performance * Improve NPU performance * [bugfix] bugfix for npu free memory * [bugfix] bugfix for npu free memory * [bugfix] bugfix for npu free memory * Reduce memory cost for flux training process --------- Co-authored-by:
蒋硕 <jiangshuo9@h-partners.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
Boseong Jeon authored
Handling mixed precision and add unwarp Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com>
-
Leo Jiang authored
* NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX --------- Co-authored-by:蒋硕 <jiangshuo9@h-partners.com>
-
- 31 Oct, 2024 1 commit
-
-
Sayak Paul authored
* use the lr when using 8bit adam. * remove lr as we pack it in params_to_optimize. --------- Co-authored-by:Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com>
-
- 28 Oct, 2024 3 commits
-
-
Linoy Tsaban authored
* make lora target modules configurable and change the default * style * make lora target modules configurable and change the default * fix bug when using prodigy and training te * fix mixed precision training as proposed in https://github.com/huggingface/diffusers/pull/9565 for full dreambooth as well * add test and notes * style * address sayaks comments * style * fix test --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
Linoy Tsaban authored
* configurable layers * configurable layers * update README * style * add test * style * add layer test, update readme, add nargs * readme * test style * remove print, change nargs * test arg change * style * revert nargs 2/2 * address sayaks comments * style * address sayaks comments
-
Biswaroop authored
[Fix] remove setting lr for T5 text encoder when using prodigy in flux dreambooth lora script (#9473) * fix: removed setting of text encoder lr for T5 as it's not being tuned * fix: removed setting of text encoder lr for T5 as it's not being tuned --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com>
-
- 25 Oct, 2024 1 commit
-
-
Ina authored
* flux pipline: readability enhancement.
-
- 23 Oct, 2024 1 commit
-
-
Linoy Tsaban authored
* improve readme * style --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
- 22 Oct, 2024 1 commit
-
-
Sayak Paul authored
* post-release * style
-
- 16 Oct, 2024 1 commit
-
-
Linoy Tsaban authored
* add latent caching + smol updates * update license * replace with free_memory * add --upcast_before_saving to allow saving transformer weights in lower precision * fix models to accumulate * fix mixed precision issue as proposed in https://github.com/huggingface/diffusers/pull/9565 * smol update to readme * style * fix caching latents * style * add tests for latent caching * style * fix latent caching --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
- 15 Oct, 2024 1 commit
-
-
0x名無し authored
* fixed issue #9350, Tensor is deprecated * ran make style
-
- 28 Sep, 2024 1 commit
-
-
Sayak Paul authored
* fix: retain memory utility. * fix * quality * free_memory.
-
- 16 Sep, 2024 1 commit
-
-
suzukimain authored
* [docs] Replace runwayml/stable-diffusion-v1-5 with Lykon/dreamshaper-8 Updated documentation as runwayml/stable-diffusion-v1-5 has been removed from Huggingface. * Update docs/source/en/using-diffusers/inpaint.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Replace with stable-diffusion-v1-5/stable-diffusion-v1-5 * Update inpaint.md --------- Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
- 15 Sep, 2024 1 commit
-
-
Linoy Tsaban authored
* add ostris trainer to README & add cache latents of vae * add ostris trainer to README & add cache latents of vae * style * readme * add test for latent caching * add ostris noise scheduler https://github.com/ostris/ai-toolkit/blob/9ee1ef2a0a2a9a02b92d114a95f21312e5906e54/toolkit/samplers/custom_flowmatch_sampler.py#L95 * style * fix import * style * fix tests * style * --change upcasting of transformer? * update readme according to main * keep only latent caching * add configurable param for final saving of trained layers- --upcast_before_saving * style * Update examples/dreambooth/README_flux.md Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * Update examples/dreambooth/README_flux.md Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * use clear_objs_and_retain_memory from utilities * style --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
- 14 Sep, 2024 1 commit
-
-
Leo Jiang authored
* Fix dtype error * [bugfix] Fixed the issue on sd3 dreambooth training * [bugfix] Fixed the issue on sd3 dreambooth training --------- Co-authored-by:
蒋硕 <jiangshuo9@h-partners.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
- 11 Sep, 2024 1 commit
-
-
Sayak Paul authored
fix some fast gpu tests.
-
- 05 Sep, 2024 1 commit
-
-
Linoy Tsaban authored
* improve readme * improve readme * improve readme * improve readme
-
- 03 Sep, 2024 1 commit
-
-
Sayak Paul authored
-
- 26 Aug, 2024 1 commit
-
-
Linoy Tsaban authored
* fix shape * fix prompt encoding * style * fix device * add comment
-
- 19 Aug, 2024 1 commit
-
-
townwish4git authored
Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com>
-
- 18 Aug, 2024 1 commit
-
-
Tolga Cangöz authored
* chore: Fix cpu offloading usage * Trim trailing white space * docs: update Kolors model link in kolors.md
-