1. 11 Sep, 2023 3 commits
    • Patrick von Platen's avatar
      Refactor model offload (#4514) · 93579650
      Patrick von Platen authored
      
      
      * [Draft] Refactor model offload
      
      * [Draft] Refactor model offload
      
      * Apply suggestions from code review
      
      * cpu offlaod updates
      
      * remove model cpu offload from individual pipelines
      
      * add hook to offload models to cpu
      
      * clean up
      
      * model offload
      
      * add model cpu offload string
      
      * make style
      
      * clean up
      
      * fixes for offload issues
      
      * fix tests issues
      
      * resolve merge conflicts
      
      * update src/diffusers/pipelines/pipeline_utils.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * make style
      
      * Update src/diffusers/pipelines/latent_diffusion/pipeline_latent_diffusion.py
      
      ---------
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      93579650
    • Dhruv Nair's avatar
      Lazy Import for Diffusers (#4829) · b6e0b016
      Dhruv Nair authored
      
      
      * initial commit
      
      * move modules to import struct
      
      * add dummy objects and _LazyModule
      
      * add lazy import to schedulers
      
      * clean up unused imports
      
      * lazy import on models module
      
      * lazy import for schedulers module
      
      * add lazy import to pipelines module
      
      * lazy import altdiffusion
      
      * lazy import audio diffusion
      
      * lazy import audioldm
      
      * lazy import consistency model
      
      * lazy import controlnet
      
      * lazy import dance diffusion ddim ddpm
      
      * lazy import deepfloyd
      
      * lazy import kandinksy
      
      * lazy imports
      
      * lazy import semantic diffusion
      
      * lazy imports
      
      * lazy import stable diffusion
      
      * move sd output to its own module
      
      * clean up
      
      * lazy import t2iadapter
      
      * lazy import unclip
      
      * lazy import versatile and vq diffsuion
      
      * lazy import vq diffusion
      
      * helper to fetch objects from modules
      
      * lazy import sdxl
      
      * lazy import txt2vid
      
      * lazy import stochastic karras
      
      * fix model imports
      
      * fix bug
      
      * lazy import
      
      * clean up
      
      * clean up
      
      * fixes for tests
      
      * fixes for tests
      
      * clean up
      
      * remove import of torch_utils from utils module
      
      * clean up
      
      * clean up
      
      * fix mistake import statement
      
      * dedicated modules for exporting and loading
      
      * remove testing utils from utils module
      
      * fixes from  merge conflicts
      
      * Update src/diffusers/pipelines/kandinsky2_2/__init__.py
      
      * fix docs
      
      * fix alt diffusion copied from
      
      * fix check dummies
      
      * fix more docs
      
      * remove accelerate import from utils module
      
      * add type checking
      
      * make style
      
      * fix check dummies
      
      * remove torch import from xformers check
      
      * clean up error message
      
      * fixes after upstream merges
      
      * dummy objects fix
      
      * fix tests
      
      * remove unused module import
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      b6e0b016
    • Will Berman's avatar
      Revert revert and install accelerate main (#4963) · 4191ddee
      Will Berman authored
      * Revert "Temp Revert "[Core] better support offloading when side loading is enabled… (#4927)"
      
      This reverts commit 2ab17049.
      
      * tests: install accelerate from main
      4191ddee
  2. 09 Sep, 2023 1 commit
  3. 05 Sep, 2023 2 commits
  4. 04 Sep, 2023 2 commits
    • Sayak Paul's avatar
      [Core] LoRA improvements pt. 3 (#4842) · c81a88b2
      Sayak Paul authored
      
      
      * throw warning when more than one lora is attempted to be fused.
      
      * introduce support of lora scale during fusion.
      
      * change test name
      
      * changes
      
      * change to _lora_scale
      
      * lora_scale to call whenever applicable.
      
      * debugging
      
      * lora_scale additional.
      
      * cross_attention_kwargs
      
      * lora_scale -> scale.
      
      * lora_scale fix
      
      * lora_scale in patched projection.
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * styling.
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * remove unneeded prints.
      
      * remove unneeded prints.
      
      * assign cross_attention_kwargs.
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * clean up.
      
      * refactor scale retrieval logic a bit.
      
      * fix nonetypw
      
      * fix: tests
      
      * add more tests
      
      * more fixes.
      
      * figure out a way to pass lora_scale.
      
      * Apply suggestions from code review
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * unify the retrieval logic of lora_scale.
      
      * move adjust_lora_scale_text_encoder to lora.py.
      
      * introduce dynamic adjustment lora scale support to sd
      
      * fix up copies
      
      * Empty-Commit
      
      * add: test to check fusion equivalence on different scales.
      
      * handle lora fusion warning.
      
      * make lora smaller
      
      * make lora smaller
      
      * make lora smaller
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      c81a88b2
    • Erwann Millon's avatar
      check for unet_lora_layers in sdxl pipeline's save_lora_weights method (#4821) · 2fa4b3ff
      Erwann Millon authored
      run make fix-copies and make style
      2fa4b3ff
  5. 02 Sep, 2023 1 commit
  6. 01 Sep, 2023 2 commits
  7. 31 Aug, 2023 3 commits
  8. 30 Aug, 2023 2 commits
  9. 28 Aug, 2023 2 commits
  10. 26 Aug, 2023 1 commit
    • Sayak Paul's avatar
      [Core] Support negative conditions in SDXL (#4774) · 3be0ff90
      Sayak Paul authored
      * add: support negative conditions.
      
      * fix: key
      
      * add: tests
      
      * address PR feedback.
      
      * add documentation
      
      * add img2img support.
      
      * add inpainting support.
      
      * ad controlnet support
      
      * Apply suggestions from code review
      
      * modify wording in the doc.
      3be0ff90
  11. 25 Aug, 2023 2 commits
  12. 22 Aug, 2023 1 commit
  13. 17 Aug, 2023 1 commit
  14. 16 Aug, 2023 1 commit
  15. 12 Aug, 2023 1 commit
  16. 11 Aug, 2023 1 commit
  17. 10 Aug, 2023 2 commits
  18. 08 Aug, 2023 1 commit
  19. 03 Aug, 2023 6 commits
  20. 30 Jul, 2023 1 commit
  21. 28 Jul, 2023 1 commit
  22. 26 Jul, 2023 1 commit
  23. 25 Jul, 2023 2 commits