1. 14 Sep, 2024 1 commit
  2. 14 Aug, 2024 1 commit
  3. 26 Jul, 2024 1 commit
    • Sayak Paul's avatar
      [Chore] add `LoraLoaderMixin` to the inits (#8981) · d87fe95f
      Sayak Paul authored
      
      
      * introduce  to promote reusability.
      
      * up
      
      * add more tests
      
      * up
      
      * remove comments.
      
      * fix fuse_nan test
      
      * clarify the scope of fuse_lora and unfuse_lora
      
      * remove space
      
      * rewrite fuse_lora a bit.
      
      * feedback
      
      * copy over load_lora_into_text_encoder.
      
      * address dhruv's feedback.
      
      * fix-copies
      
      * fix issubclass.
      
      * num_fused_loras
      
      * fix
      
      * fix
      
      * remove mapping
      
      * up
      
      * fix
      
      * style
      
      * fix-copies
      
      * change to SD3TransformerLoRALoadersMixin
      
      * Apply suggestions from code review
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      * up
      
      * handle wuerstchen
      
      * up
      
      * move lora to lora_pipeline.py
      
      * up
      
      * fix-copies
      
      * fix documentation.
      
      * comment set_adapters().
      
      * fix-copies
      
      * fix set_adapters() at the model level.
      
      * fix?
      
      * fix
      
      * loraloadermixin.
      
      ---------
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      d87fe95f
  4. 25 Jul, 2024 2 commits
    • YiYi Xu's avatar
      Revert "[LoRA] introduce LoraBaseMixin to promote reusability." (#8976) · 62863bb1
      YiYi Xu authored
      Revert "[LoRA] introduce LoraBaseMixin to promote reusability. (#8774)"
      
      This reverts commit 527430d0.
      62863bb1
    • Sayak Paul's avatar
      [LoRA] introduce LoraBaseMixin to promote reusability. (#8774) · 527430d0
      Sayak Paul authored
      
      
      * introduce  to promote reusability.
      
      * up
      
      * add more tests
      
      * up
      
      * remove comments.
      
      * fix fuse_nan test
      
      * clarify the scope of fuse_lora and unfuse_lora
      
      * remove space
      
      * rewrite fuse_lora a bit.
      
      * feedback
      
      * copy over load_lora_into_text_encoder.
      
      * address dhruv's feedback.
      
      * fix-copies
      
      * fix issubclass.
      
      * num_fused_loras
      
      * fix
      
      * fix
      
      * remove mapping
      
      * up
      
      * fix
      
      * style
      
      * fix-copies
      
      * change to SD3TransformerLoRALoadersMixin
      
      * Apply suggestions from code review
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      * up
      
      * handle wuerstchen
      
      * up
      
      * move lora to lora_pipeline.py
      
      * up
      
      * fix-copies
      
      * fix documentation.
      
      * comment set_adapters().
      
      * fix-copies
      
      * fix set_adapters() at the model level.
      
      * fix?
      
      * fix
      
      ---------
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      527430d0
  5. 13 Jun, 2024 1 commit
  6. 29 May, 2024 1 commit
  7. 20 May, 2024 1 commit
  8. 02 Apr, 2024 1 commit
    • Bagheera's avatar
      7529 do not disable autocast for cuda devices (#7530) · 8e963d1c
      Bagheera authored
      
      
      * 7529 do not disable autocast for cuda devices
      
      * Remove typecasting error check for non-mps platforms, as a correct autocast implementation makes it a non-issue
      
      * add autocast fix to other training examples
      
      * disable native_amp for dreambooth (sdxl)
      
      * disable native_amp for pix2pix (sdxl)
      
      * remove tests from remaining files
      
      * disable native_amp on huggingface accelerator for every training example that uses it
      
      * convert more usages of autocast to nullcontext, make style fixes
      
      * make style fixes
      
      * style.
      
      * Empty-Commit
      
      ---------
      Co-authored-by: default avatarbghira <bghira@users.github.com>
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      8e963d1c
  9. 18 Mar, 2024 1 commit
  10. 13 Mar, 2024 1 commit
  11. 04 Mar, 2024 1 commit
    • Linoy Tsaban's avatar
      [training scripts] add tags of diffusers-training (#7206) · 8da360aa
      Linoy Tsaban authored
      * add tags for diffusers training
      
      * add tags for diffusers training
      
      * add tags for diffusers training
      
      * add tags for diffusers training
      
      * add tags for diffusers training
      
      * add tags for diffusers training
      
      * add dora tags for drambooth lora scripts
      
      * style
      8da360aa
  12. 26 Feb, 2024 1 commit
  13. 09 Feb, 2024 2 commits
  14. 08 Feb, 2024 2 commits
  15. 07 Feb, 2024 1 commit
    • Sayak Paul's avatar
      [Model Card] standardize dreambooth model card (#6729) · 76696dca
      Sayak Paul authored
      * feat: standarize model card creation for dreambooth training.
      
      * correct 'inference
      
      * remove comments.
      
      * take component out of kwargs
      
      * style
      
      * add: card template to have a leaner description.
      
      * widget support.
      
      * propagate changes to train_dreambooth_lora
      
      * propagate changes to custom diffusion
      
      * make widget properly type-annotated
      76696dca
  16. 16 Jan, 2024 1 commit
  17. 12 Jan, 2024 1 commit
  18. 05 Jan, 2024 1 commit
  19. 28 Dec, 2023 1 commit
  20. 24 Dec, 2023 1 commit
  21. 18 Dec, 2023 1 commit
  22. 07 Dec, 2023 1 commit
    • Younes Belkada's avatar
      [`PEFT`] Adapt example scripts to use PEFT (#5388) · c2717317
      Younes Belkada authored
      
      
      * adapt example scripts to use PEFT
      
      * Update examples/text_to_image/train_text_to_image_lora.py
      
      * fix
      
      * add for SDXL
      
      * oops
      
      * make sure to install peft
      
      * fix
      
      * fix
      
      * fix dreambooth and lora
      
      * more fixes
      
      * add peft to requirements.txt
      
      * fix
      
      * final fix
      
      * add peft version in requirements
      
      * remove comment
      
      * change variable names
      
      * add few lines in readme
      
      * add to reqs
      
      * style
      
      * fix issues
      
      * fix lora dreambooth xl tests
      
      * init_lora_weights to gaussian and add out proj where missing
      
      * ammend requirements.
      
      * ammend requirements.txt
      
      * add correct peft versions
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      c2717317
  23. 01 Dec, 2023 1 commit
  24. 27 Nov, 2023 1 commit
  25. 21 Nov, 2023 1 commit
    • Patrick von Platen's avatar
      [Lora] Seperate logic (#5809) · 13d73d93
      Patrick von Platen authored
      * [Lora] Seperate logic
      
      * [Lora] Seperate logic
      
      * [Lora] Seperate logic
      
      * add comments to explain the code better
      
      * add comments to explain the code better
      13d73d93
  26. 17 Nov, 2023 1 commit
  27. 14 Nov, 2023 1 commit
    • Sayak Paul's avatar
      [Refactor] refactor `loaders.py` to make it cleaner and leaner. (#5771) · ded93f79
      Sayak Paul authored
      
      
      * refactor loaders.py to make it cleaner and leaner.
      
      * refactor loaders init
      
      * inits.
      
      * textual inversion to the init.
      
      * inits.
      
      * remove certain modules from the main init.
      
      * AttnProcsLayers
      
      * fix imports
      
      * avoid circular import.
      
      * fix circular import pt 2.
      
      * address PR comments
      
      * imports
      
      * fix: imports.
      
      * remove from main init for avoiding circular deps.
      
      * remove spurious deps.
      
      * fix-copies.
      
      * fix imports.
      
      * more debug
      
      * more debug
      
      * Apply suggestions from code review
      
      * Apply suggestions from code review
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      ded93f79
  28. 10 Nov, 2023 1 commit
  29. 06 Nov, 2023 1 commit
  30. 16 Oct, 2023 1 commit
  31. 11 Oct, 2023 1 commit
  32. 08 Oct, 2023 1 commit
  33. 02 Oct, 2023 1 commit
    • Sayak Paul's avatar
      fix: how print training resume logs. (#5117) · d56825e4
      Sayak Paul authored
      
      
      * fix: how print training resume logs.
      
      * propagate changes to text-to-image scripts.
      
      * propagate changes to instructpix2pix.
      
      * propagate changes to dreambooth
      
      * propagate changes to custom diffusion and instructpix2pix
      
      * propagate changes to kandinsky
      
      * propagate changes to textual inv.
      
      * debug
      
      * fix: checkpointing.
      
      * debug
      
      * debug
      
      * debug
      
      * back to the square
      
      * debug
      
      * debug
      
      * change condition order.
      
      * debug
      
      * debug
      
      * debug
      
      * debug
      
      * revert to original
      
      * clean
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      d56825e4
  34. 14 Sep, 2023 1 commit
  35. 08 Sep, 2023 1 commit
  36. 25 Aug, 2023 1 commit
  37. 17 Aug, 2023 1 commit