"vscode:/vscode.git/clone" did not exist on "14cc5c0599c01d1406a1a5edcab944ad436cfaba"
  1. 11 Jul, 2023 1 commit
    • Sayak Paul's avatar
      [Examples] Add a training script for SDXL DreamBooth LoRA (#4016) · 3d74dc2a
      Sayak Paul authored
      
      
      * add dreambooth lora script for SDXL incorporating latest changes.
      
      * remove use_auth_token=True.
      
      * add: documentation
      
      * remove unneeded cli.
      
      * increase the number of training steps in the readme.
      
      * add LoraLoaderMixin to the subclassing mix.
      
      * add sdxl lora dreambooth test.
      
      * add: inference code sample.
      
      * add: refiner output.
      
      * add LoraLoaderMixin to the mix of classes of StableDiffusionXLImg2ImgPipeline.
      
      * change default resolution of DreamBoothDataset.
      
      * better sdxl report path.
      
      * Apply suggestions from code review
      Co-authored-by: default avatarPedro Cuenca <pedro@huggingface.co>
      
      ---------
      Co-authored-by: default avatarPedro Cuenca <pedro@huggingface.co>
      3d74dc2a
  2. 09 Jul, 2023 2 commits
    • Patrick von Platen's avatar
      make style · 4a3e5748
      Patrick von Platen authored
      4a3e5748
    • Will Berman's avatar
      Refactor LoRA (#3778) · c2a28c34
      Will Berman authored
      
      
      * refactor to support patching LoRA into T5
      
      instantiate the lora linear layer on the same device as the regular linear layer
      
      get lora rank from state dict
      
      tests
      
      fmt
      
      can create lora layer in float32 even when rest of model is float16
      
      fix loading model hook
      
      remove load_lora_weights_ and T5 dispatching
      
      remove Unet#attn_processors_state_dict
      
      docstrings
      
      * text encoder monkeypatch class method
      
      * fix test
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      c2a28c34
  3. 07 Jul, 2023 1 commit
  4. 15 Jun, 2023 2 commits
  5. 08 Jun, 2023 2 commits
  6. 06 Jun, 2023 1 commit
    • Sayak Paul's avatar
      [LoRA] feat: add lora attention processor for pt 2.0. (#3594) · 8669e831
      Sayak Paul authored
      * feat: add lora attention processor for pt 2.0.
      
      * explicit context manager for SDPA.
      
      * switch to flash attention
      
      * make shapes compatible to work optimally with SDPA.
      
      * fix: circular import problem.
      
      * explicitly specify the flash attention kernel in sdpa
      
      * fall back to efficient attention context manager.
      
      * remove explicit dispatch.
      
      * fix: removed processor.
      
      * fix: remove optional from type annotation.
      
      * feat: make changes regarding LoRAAttnProcessor2_0.
      
      * remove confusing warning.
      
      * formatting.
      
      * relax tolerance for PT 2.0
      
      * fix: loading message.
      
      * remove unnecessary logging.
      
      * add: entry to the docs.
      
      * add: network_alpha argument.
      
      * relax tolerance.
      8669e831
  7. 05 Jun, 2023 2 commits
  8. 02 Jun, 2023 2 commits
    • Will Berman's avatar
      dreambooth if docs - stage II, more info (#3628) · 5911a3aa
      Will Berman authored
      
      
      * dreambooth if docs - stage II, more info
      
      * Update docs/source/en/training/dreambooth.mdx
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update docs/source/en/training/dreambooth.mdx
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update docs/source/en/training/dreambooth.mdx
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      * download instructions for downsized images
      
      * update source README to match docs
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      5911a3aa
    • Takuma Mori's avatar
      Support Kohya-ss style LoRA file format (in a limited capacity) (#3437) · 8e552bb4
      Takuma Mori authored
      
      
      * add _convert_kohya_lora_to_diffusers
      
      * make style
      
      * add scaffold
      
      * match result: unet attention only
      
      * fix monkey-patch for text_encoder
      
      * with CLIPAttention
      
      While the terrible images are no longer produced,
      the results do not match those from the hook ver.
      This may be due to not setting the network_alpha value.
      
      * add to support network_alpha
      
      * generate diff image
      
      * fix monkey-patch for text_encoder
      
      * add test_text_encoder_lora_monkey_patch()
      
      * verify that it's okay to release the attn_procs
      
      * fix closure version
      
      * add comment
      
      * Revert "fix monkey-patch for text_encoder"
      
      This reverts commit bb9c61e6faecc1935c9c4319c77065837655d616.
      
      * Fix to reuse utility functions
      
      * make LoRAAttnProcessor targets to self_attn
      
      * fix LoRAAttnProcessor target
      
      * make style
      
      * fix split key
      
      * Update src/diffusers/loaders.py
      
      * remove TEXT_ENCODER_TARGET_MODULES loop
      
      * add print memory usage
      
      * remove test_kohya_loras_scaffold.py
      
      * add: doc on LoRA civitai
      
      * remove print statement and refactor in the doc.
      
      * fix state_dict test for kohya-ss style lora
      
      * Apply suggestions from code review
      Co-authored-by: default avatarTakuma Mori <takuma104@gmail.com>
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      8e552bb4
  9. 31 May, 2023 2 commits
  10. 26 May, 2023 1 commit
  11. 24 May, 2023 1 commit
  12. 22 May, 2023 2 commits
    • Patrick von Platen's avatar
      make style · 2b56e8ca
      Patrick von Platen authored
      2b56e8ca
    • Ambrosiussen's avatar
      DataLoader respecting EXIF data in Training Images (#3465) · b8b5daae
      Ambrosiussen authored
      * DataLoader will now bake in any transforms or image manipulations contained in the EXIF
      
      Images may have rotations stored in EXIF. Training using such images will cause those transforms to be ignored while training and thus produce unexpected results
      
      * Fixed the Dataloading EXIF issue in main DreamBooth training as well
      
      * Run make style (black & isort)
      b8b5daae
  13. 19 May, 2023 1 commit
  14. 17 May, 2023 3 commits
  15. 11 May, 2023 1 commit
  16. 09 May, 2023 1 commit
    • Will Berman's avatar
      if dreambooth lora (#3360) · a757b2db
      Will Berman authored
      * update IF stage I pipelines
      
      add fixed variance schedulers and lora loading
      
      * added kv lora attn processor
      
      * allow loading into alternative lora attn processor
      
      * make vae optional
      
      * throw away predicted variance
      
      * allow loading into added kv lora layer
      
      * allow load T5
      
      * allow pre compute text embeddings
      
      * set new variance type in schedulers
      
      * fix copies
      
      * refactor all prompt embedding code
      
      class prompts are now included in pre-encoding code
      max tokenizer length is now configurable
      embedding attention mask is now configurable
      
      * fix for when variance type is not defined on scheduler
      
      * do not pre compute validation prompt if not present
      
      * add example test for if lora dreambooth
      
      * add check for train text encoder and pre compute text embeddings
      a757b2db
  17. 03 May, 2023 1 commit
  18. 28 Apr, 2023 2 commits
  19. 26 Apr, 2023 2 commits
  20. 22 Apr, 2023 1 commit
  21. 20 Apr, 2023 1 commit
  22. 12 Apr, 2023 2 commits
  23. 11 Apr, 2023 1 commit
  24. 04 Apr, 2023 1 commit
  25. 29 Mar, 2023 1 commit
  26. 23 Mar, 2023 1 commit
  27. 15 Mar, 2023 1 commit
  28. 14 Mar, 2023 1 commit