1. 27 Oct, 2022 1 commit
    • Pedro Cuenca's avatar
      Continuation of #942: additional float64 failure (#996) · 1d04e1b4
      Pedro Cuenca authored
      * Add failing test for #940.
      
      * Do not use torch.float64 in mps.
      
      * style
      
      * Temporarily skip add_noise for IPNDMScheduler.
      
      Until #990 is addressed.
      
      * Fix additional float64 error in mps.
      
      * Improve add_noise test
      
      * Slight edit – I think it's clearer this way.
      1d04e1b4
  2. 26 Oct, 2022 2 commits
    • Pi Esposito's avatar
      minimal stable diffusion GPU memory usage with accelerate hooks (#850) · b2e2d141
      Pi Esposito authored
      * add method to enable cuda with minimal gpu usage to stable diffusion
      
      * add test to minimal cuda memory usage
      
      * ensure all models but unet are onn torch.float32
      
      * move to cpu_offload along with minor internal changes to make it work
      
      * make it test against accelerate master branch
      
      * coming back, its official: I don't know how to make it test againt the master branch from accelerate
      
      * make it install accelerate from master on tests
      
      * go back to accelerate>=0.11
      
      * undo prettier formatting on yml files
      
      * undo prettier formatting on yml files againn
      b2e2d141
    • Pedro Cuenca's avatar
      Do not use torch.float64 on the mps device (#942) · 0343d8f5
      Pedro Cuenca authored
      * Add failing test for #940.
      
      * Do not use torch.float64 in mps.
      
      * style
      
      * Temporarily skip add_noise for IPNDMScheduler.
      
      Until #990 is addressed.
      0343d8f5
  3. 25 Oct, 2022 4 commits
  4. 24 Oct, 2022 1 commit
  5. 22 Oct, 2022 1 commit
  6. 21 Oct, 2022 1 commit
  7. 20 Oct, 2022 4 commits
  8. 19 Oct, 2022 4 commits
  9. 18 Oct, 2022 2 commits
  10. 17 Oct, 2022 2 commits
    • Pedro Cuenca's avatar
      Fix autoencoder test (#886) · 100e094c
      Pedro Cuenca authored
      Fix autoencoder test.
      100e094c
    • Anton Lozhkov's avatar
      Add Apple M1 tests (#796) · cca59ce3
      Anton Lozhkov authored
      
      
      * [CI] Add Apple M1 tests
      
      * setup-python
      
      * python build
      
      * conda install
      
      * remove branch
      
      * only 3.8 is built for osx-arm
      
      * try fetching prebuilt tokenizers
      
      * use user cache
      
      * update shells
      
      * Reports and cleanup
      
      * -> MPS
      
      * Disable parallel tests
      
      * Better naming
      
      * investigate worker crash
      
      * return xdist
      
      * restart
      
      * num_workers=2
      
      * still crashing?
      
      * faulthandler for segfaults
      
      * faulthandler for segfaults
      
      * remove restarts, stop on segfault
      
      * torch version
      
      * change installation order
      
      * Use pre-RC version of PyTorch.
      
      To be updated when it is released.
      
      * Skip crashing test on MPS, add new one that works.
      
      * Skip cuda tests in mps device.
      
      * Actually use generator in test.
      
      I think this was a typo.
      
      * make style
      Co-authored-by: default avatarPedro Cuenca <pedro@huggingface.co>
      cca59ce3
  11. 14 Oct, 2022 2 commits
  12. 13 Oct, 2022 5 commits
  13. 12 Oct, 2022 2 commits
  14. 11 Oct, 2022 1 commit
    • Akash Pannu's avatar
      Flax: Trickle down `norm_num_groups` (#789) · a1242044
      Akash Pannu authored
      * pass norm_num_groups param and add tests
      
      * set resnet_groups for FlaxUNetMidBlock2D
      
      * fixed docstrings
      
      * fixed typo
      
      * using is_flax_available util and created require_flax decorator
      a1242044
  15. 10 Oct, 2022 2 commits
    • Patrick von Platen's avatar
      Fix gradient checkpointing test (#797) · 22963ed8
      Patrick von Platen authored
      * Fix gradient checkpointing test
      
      * more tsets
      22963ed8
    • Patrick von Platen's avatar
      [Low CPU memory] + device map (#772) · fab17528
      Patrick von Platen authored
      
      
      * add accelerate to load models with smaller memory footprint
      
      * remove low_cpu_mem_usage as it is reduntant
      
      * move accelerate init weights context to modelling utils
      
      * add test to ensure results are the same when loading with accelerate
      
      * add tests to ensure ram usage gets lower when using accelerate
      
      * move accelerate logic to single snippet under modelling utils and remove it from configuration utils
      
      * format code using to pass quality check
      
      * fix imports with isor
      
      * add accelerate to test extra deps
      
      * only import accelerate if device_map is set to auto
      
      * move accelerate availability check to diffusers import utils
      
      * format code
      
      * add device map to pipeline abstraction
      
      * lint it to pass PR quality check
      
      * fix class check to use accelerate when using diffusers ModelMixin subclasses
      
      * use low_cpu_mem_usage in transformers if device_map is not available
      
      * NoModuleLayer
      
      * comment out tests
      
      * up
      
      * uP
      
      * finish
      
      * Update src/diffusers/pipelines/stable_diffusion/safety_checker.py
      
      * finish
      
      * uP
      
      * make style
      Co-authored-by: default avatarPi Esposito <piero.skywalker@gmail.com>
      fab17528
  16. 07 Oct, 2022 4 commits
  17. 06 Oct, 2022 2 commits