1. 04 Nov, 2022 1 commit
  2. 03 Nov, 2022 1 commit
    • Pedro Cuenca's avatar
      Continuation of #1035 (#1120) · 269109db
      Pedro Cuenca authored
      
      
      * remove batch size from repeat
      
      * repeat empty string if uncond_tokens is none
      
      * fix inpaint pipes
      
      * return back whitespace to pass code quality
      
      * Apply suggestions from code review
      
      * Fix typos.
      Co-authored-by: default avatarHad <had-95@yandex.ru>
      269109db
  3. 02 Nov, 2022 1 commit
    • MatthieuTPHR's avatar
      Up to 2x speedup on GPUs using memory efficient attention (#532) · 98c42134
      MatthieuTPHR authored
      
      
      * 2x speedup using memory efficient attention
      
      * remove einops dependency
      
      * Swap K, M in op instantiation
      
      * Simplify code, remove unnecessary maybe_init call and function, remove unused self.scale parameter
      
      * make xformers a soft dependency
      
      * remove one-liner functions
      
      * change one letter variable to appropriate names
      
      * Remove Env variable dependency, remove MemoryEfficientCrossAttention class and use enable_xformers_memory_efficient_attention method
      
      * Add memory efficient attention toggle to img2img and inpaint pipelines
      
      * Clearer management of xformers' availability
      
      * update optimizations markdown to add info about memory efficient attention
      
      * add benchmarks for TITAN RTX
      
      * More detailed explanation of how the mem eff benchmark were ran
      
      * Removing autocast from optimization markdown
      
      * import_utils: import torch only if is available
      Co-authored-by: default avatarNouamane Tazi <nouamane98@gmail.com>
      98c42134
  4. 31 Oct, 2022 4 commits
  5. 30 Oct, 2022 1 commit
  6. 28 Oct, 2022 1 commit
  7. 27 Oct, 2022 2 commits
  8. 26 Oct, 2022 2 commits
    • Hu Ye's avatar
      bd06dd02
    • Pi Esposito's avatar
      minimal stable diffusion GPU memory usage with accelerate hooks (#850) · b2e2d141
      Pi Esposito authored
      * add method to enable cuda with minimal gpu usage to stable diffusion
      
      * add test to minimal cuda memory usage
      
      * ensure all models but unet are onn torch.float32
      
      * move to cpu_offload along with minor internal changes to make it work
      
      * make it test against accelerate master branch
      
      * coming back, its official: I don't know how to make it test againt the master branch from accelerate
      
      * make it install accelerate from master on tests
      
      * go back to accelerate>=0.11
      
      * undo prettier formatting on yml files
      
      * undo prettier formatting on yml files againn
      b2e2d141
  9. 25 Oct, 2022 2 commits
  10. 24 Oct, 2022 1 commit
  11. 19 Oct, 2022 4 commits
  12. 18 Oct, 2022 2 commits
  13. 14 Oct, 2022 1 commit
  14. 13 Oct, 2022 7 commits
  15. 12 Oct, 2022 1 commit
  16. 11 Oct, 2022 2 commits
  17. 10 Oct, 2022 1 commit
    • Patrick von Platen's avatar
      [Low CPU memory] + device map (#772) · fab17528
      Patrick von Platen authored
      
      
      * add accelerate to load models with smaller memory footprint
      
      * remove low_cpu_mem_usage as it is reduntant
      
      * move accelerate init weights context to modelling utils
      
      * add test to ensure results are the same when loading with accelerate
      
      * add tests to ensure ram usage gets lower when using accelerate
      
      * move accelerate logic to single snippet under modelling utils and remove it from configuration utils
      
      * format code using to pass quality check
      
      * fix imports with isor
      
      * add accelerate to test extra deps
      
      * only import accelerate if device_map is set to auto
      
      * move accelerate availability check to diffusers import utils
      
      * format code
      
      * add device map to pipeline abstraction
      
      * lint it to pass PR quality check
      
      * fix class check to use accelerate when using diffusers ModelMixin subclasses
      
      * use low_cpu_mem_usage in transformers if device_map is not available
      
      * NoModuleLayer
      
      * comment out tests
      
      * up
      
      * uP
      
      * finish
      
      * Update src/diffusers/pipelines/stable_diffusion/safety_checker.py
      
      * finish
      
      * uP
      
      * make style
      Co-authored-by: default avatarPi Esposito <piero.skywalker@gmail.com>
      fab17528
  18. 07 Oct, 2022 1 commit
    • Suraj Patil's avatar
      [img2img, inpainting] fix fp16 inference (#769) · 92d70863
      Suraj Patil authored
      * handle dtype in vae and image2image pipeline
      
      * fix inpaint in fp16
      
      * dtype should be handled in add_noise
      
      * style
      
      * address review comments
      
      * add simple fast tests to check fp16
      
      * fix test name
      
      * put mask in fp16
      92d70863
  19. 06 Oct, 2022 3 commits
  20. 05 Oct, 2022 2 commits