1. 26 Jul, 2024 1 commit
    • Sayak Paul's avatar
      [Chore] add `LoraLoaderMixin` to the inits (#8981) · d87fe95f
      Sayak Paul authored
      
      
      * introduce  to promote reusability.
      
      * up
      
      * add more tests
      
      * up
      
      * remove comments.
      
      * fix fuse_nan test
      
      * clarify the scope of fuse_lora and unfuse_lora
      
      * remove space
      
      * rewrite fuse_lora a bit.
      
      * feedback
      
      * copy over load_lora_into_text_encoder.
      
      * address dhruv's feedback.
      
      * fix-copies
      
      * fix issubclass.
      
      * num_fused_loras
      
      * fix
      
      * fix
      
      * remove mapping
      
      * up
      
      * fix
      
      * style
      
      * fix-copies
      
      * change to SD3TransformerLoRALoadersMixin
      
      * Apply suggestions from code review
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      * up
      
      * handle wuerstchen
      
      * up
      
      * move lora to lora_pipeline.py
      
      * up
      
      * fix-copies
      
      * fix documentation.
      
      * comment set_adapters().
      
      * fix-copies
      
      * fix set_adapters() at the model level.
      
      * fix?
      
      * fix
      
      * loraloadermixin.
      
      ---------
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      d87fe95f
  2. 25 Jul, 2024 2 commits
    • YiYi Xu's avatar
      Revert "[LoRA] introduce LoraBaseMixin to promote reusability." (#8976) · 62863bb1
      YiYi Xu authored
      Revert "[LoRA] introduce LoraBaseMixin to promote reusability. (#8774)"
      
      This reverts commit 527430d0.
      62863bb1
    • Sayak Paul's avatar
      [LoRA] introduce LoraBaseMixin to promote reusability. (#8774) · 527430d0
      Sayak Paul authored
      
      
      * introduce  to promote reusability.
      
      * up
      
      * add more tests
      
      * up
      
      * remove comments.
      
      * fix fuse_nan test
      
      * clarify the scope of fuse_lora and unfuse_lora
      
      * remove space
      
      * rewrite fuse_lora a bit.
      
      * feedback
      
      * copy over load_lora_into_text_encoder.
      
      * address dhruv's feedback.
      
      * fix-copies
      
      * fix issubclass.
      
      * num_fused_loras
      
      * fix
      
      * fix
      
      * remove mapping
      
      * up
      
      * fix
      
      * style
      
      * fix-copies
      
      * change to SD3TransformerLoRALoadersMixin
      
      * Apply suggestions from code review
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      * up
      
      * handle wuerstchen
      
      * up
      
      * move lora to lora_pipeline.py
      
      * up
      
      * fix-copies
      
      * fix documentation.
      
      * comment set_adapters().
      
      * fix-copies
      
      * fix set_adapters() at the model level.
      
      * fix?
      
      * fix
      
      ---------
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      527430d0
  3. 18 Jul, 2024 1 commit
  4. 03 Jul, 2024 2 commits
  5. 26 Jun, 2024 1 commit
  6. 25 Jun, 2024 1 commit
  7. 22 Jun, 2024 1 commit
  8. 21 Jun, 2024 1 commit
  9. 20 Jun, 2024 1 commit
  10. 18 Jun, 2024 1 commit
    • Gæros's avatar
      [LoRA] text encoder: read the ranks for all the attn modules (#8324) · 298ce679
      Gæros authored
      
      
      * [LoRA] text encoder: read the ranks for all the attn modules
      
       * In addition to out_proj, read the ranks of adapters for q_proj, k_proj, and  v_proj
      
       * Allow missing adapters (UNet already supports this)
      
      * ruff format loaders.lora
      
      * [LoRA] add tests for partial text encoders LoRAs
      
      * [LoRA] update test_simple_inference_with_partial_text_lora to be deterministic
      
      * [LoRA] comment justifying test_simple_inference_with_partial_text_lora
      
      * style
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      298ce679
  11. 12 Jun, 2024 1 commit
  12. 05 Jun, 2024 1 commit
    • Sayak Paul's avatar
      [LoRA] Remove legacy LoRA code and related adjustments (#8316) · a0542c19
      Sayak Paul authored
      * remove legacy code from load_attn_procs.
      
      * finish first draft
      
      * fix more.
      
      * fix more
      
      * add test
      
      * add serialization support.
      
      * fix-copies
      
      * require peft backend for lora tests
      
      * style
      
      * fix test
      
      * fix loading.
      
      * empty
      
      * address benjamin's feedback.
      a0542c19
  13. 12 May, 2024 1 commit
  14. 09 May, 2024 1 commit
  15. 08 May, 2024 1 commit
  16. 03 May, 2024 1 commit
  17. 01 May, 2024 1 commit
  18. 25 Apr, 2024 1 commit
  19. 12 Apr, 2024 1 commit
  20. 30 Mar, 2024 1 commit
  21. 29 Mar, 2024 1 commit
    • UmerHA's avatar
      Implements Blockwise lora (#7352) · 03024468
      UmerHA authored
      
      
      * Initial commit
      
      * Implemented block lora
      
      - implemented block lora
      - updated docs
      - added tests
      
      * Finishing up
      
      * Reverted unrelated changes made by make style
      
      * Fixed typo
      
      * Fixed bug + Made text_encoder_2 scalable
      
      * Integrated some review feedback
      
      * Incorporated review feedback
      
      * Fix tests
      
      * Made every module configurable
      
      * Adapter to new lora test structure
      
      * Final cleanup
      
      * Some more final fixes
      
      - Included examples in `using_peft_for_inference.md`
      - Added hint that only attns are scaled
      - Removed NoneTypes
      - Added test to check mismatching lens of adapter names / weights raise error
      
      * Update using_peft_for_inference.md
      
      * Update using_peft_for_inference.md
      
      * Make style, quality, fix-copies
      
      * Updated tutorial;Warning if scale/adapter mismatch
      
      * floats are forwarded as-is; changed tutorial scale
      
      * make style, quality, fix-copies
      
      * Fixed typo in tutorial
      
      * Moved some warnings into `lora_loader_utils.py`
      
      * Moved scale/lora mismatch warnings back
      
      * Integrated final review suggestions
      
      * Empty commit to trigger CI
      
      * Reverted emoty commit to trigger CI
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      03024468
  22. 26 Mar, 2024 1 commit
    • Sayak Paul's avatar
      feat: support DoRA LoRA from community (#7371) · 699dfb08
      Sayak Paul authored
      * feat: support dora loras from community
      
      * safe-guard dora operations under peft version.
      
      * pop use_dora when False
      
      * make dora lora from kohya work.
      
      * fix: kohya conversion utils.
      
      * add a fast test for DoRA compatibility..
      
      * add a nightly test.
      699dfb08
  23. 19 Mar, 2024 1 commit
  24. 13 Mar, 2024 1 commit
  25. 27 Feb, 2024 1 commit
  26. 23 Feb, 2024 1 commit
  27. 09 Feb, 2024 1 commit
  28. 08 Feb, 2024 1 commit
  29. 16 Jan, 2024 1 commit
  30. 12 Jan, 2024 1 commit
  31. 03 Jan, 2024 1 commit
  32. 26 Dec, 2023 2 commits
  33. 21 Dec, 2023 1 commit
    • Will Berman's avatar
      open muse (#5437) · 40398152
      Will Berman authored
      
      
      amused
      
      rename
      
      Update docs/source/en/api/pipelines/amused.md
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      AdaLayerNormContinuous default values
      
      custom micro conditioning
      
      micro conditioning docs
      
      put lookup from codebook in constructor
      
      fix conversion script
      
      remove manual fused flash attn kernel
      
      add training script
      
      temp remove training script
      
      add dummy gradient checkpointing func
      
      clarify temperatures is an instance variable by setting it
      
      remove additional SkipFF block args
      
      hardcode norm args
      
      rename tests folder
      
      fix paths and samples
      
      fix tests
      
      add training script
      
      training readme
      
      lora saving and loading
      
      non-lora saving/loading
      
      some readme fixes
      
      guards
      
      Update docs/source/en/api/pipelines/amused.md
      Co-authored-by: default avatarSuraj Patil <surajp815@gmail.com>
      
      Update examples/amused/README.md
      Co-authored-by: default avatarSuraj Patil <surajp815@gmail.com>
      
      Update examples/amused/train_amused.py
      Co-authored-by: default avatarSuraj Patil <surajp815@gmail.com>
      
      vae upcasting
      
      add fp16 integration tests
      
      use tuple for micro cond
      
      copyrights
      
      remove casts
      
      delegate to torch.nn.LayerNorm
      
      move temperature to pipeline call
      
      upsampling/downsampling changes
      40398152
  34. 16 Dec, 2023 1 commit
  35. 06 Dec, 2023 1 commit
  36. 02 Dec, 2023 1 commit
  37. 01 Dec, 2023 1 commit