1. 11 Oct, 2022 6 commits
    • Pedro Cuenca's avatar
      `mps`: Alternative implementation for `repeat_interleave` (#766) · 24b8b5cf
      Pedro Cuenca authored
      
      
      * mps: alt. implementation for repeat_interleave
      
      * style
      
      * Bump mps version of PyTorch in the documentation.
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSuraj Patil <surajp815@gmail.com>
      
      * Simplify: do not check for device.
      
      * style
      
      * Fix repeat dimensions:
      
      - The unconditional embeddings are always created from a single prompt.
      - I was shadowing the batch_size var.
      
      * Split long lines as suggested by Suraj.
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarSuraj Patil <surajp815@gmail.com>
      24b8b5cf
    • Omar Sanseviero's avatar
      Fix indentation in the code example (#802) · 757babfc
      Omar Sanseviero authored
      Update custom_pipelines.mdx
      757babfc
    • spezialspezial's avatar
      Eventually preserve this typo? :) (#804) · e8959528
      spezialspezial authored
      e8959528
    • Akash Pannu's avatar
      Flax: Trickle down `norm_num_groups` (#789) · a1242044
      Akash Pannu authored
      * pass norm_num_groups param and add tests
      
      * set resnet_groups for FlaxUNetMidBlock2D
      
      * fixed docstrings
      
      * fixed typo
      
      * using is_flax_available util and created require_flax decorator
      a1242044
    • Suraj Patil's avatar
      stable diffusion fine-tuning (#356) · 66a5279a
      Suraj Patil authored
      
      
      * begin text2image script
      
      * loading the datasets, preprocessing & transforms
      
      * handle input features correctly
      
      * add gradient checkpointing support
      
      * fix output names
      
      * run unet in train mode not text encoder
      
      * use no_grad instead of freezing params
      
      * default max steps None
      
      * pad to longest
      
      * don't pad when tokenizing
      
      * fix encode on multi gpu
      
      * fix stupid bug
      
      * add random flip
      
      * add ema
      
      * fix ema
      
      * put ema on cpu
      
      * improve EMA model
      
      * contiguous_format
      
      * don't warp vae and text encode in accelerate
      
      * remove no_grad
      
      * use randn_like
      
      * fix resize
      
      * improve few things
      
      * log epoch loss
      
      * set log level
      
      * don't log each step
      
      * remove max_length from collate
      
      * style
      
      * add report_to option
      
      * make scale_lr false by default
      
      * add grad clipping
      
      * add an option to use 8bit adam
      
      * fix logging in multi-gpu, log every step
      
      * more comments
      
      * remove eval for now
      
      * adress review comments
      
      * add requirements file
      
      * begin readme
      
      * begin readme
      
      * fix typo
      
      * fix push to hub
      
      * populate readme
      
      * update readme
      
      * remove use_auth_token from the script
      
      * address some review comments
      
      * better mixed precision support
      
      * remove redundant to
      
      * create ema model early
      
      * Apply suggestions from code review
      Co-authored-by: default avatarPedro Cuenca <pedro@huggingface.co>
      
      * better description for train_data_dir
      
      * add diffusers in requirements
      
      * update dataset_name_mapping
      
      * update readme
      
      * add inference example
      Co-authored-by: default avataranton-l <anton@huggingface.co>
      Co-authored-by: default avatarPedro Cuenca <pedro@huggingface.co>
      66a5279a
    • Suraj Patil's avatar
      support bf16 for stable diffusion (#792) · 797b290e
      Suraj Patil authored
      * support bf16 for stable diffusion
      
      * fix typo
      
      * address review comments
      797b290e
  2. 10 Oct, 2022 7 commits
    • Henrik Forstén's avatar
      DreamBooth DeepSpeed support for under 8 GB VRAM training (#735) · 81bdbb5e
      Henrik Forstén authored
      * Support deepspeed
      
      * Dreambooth DeepSpeed documentation
      
      * Remove unnecessary casts, documentation
      
      Due to recent commits some casts to half precision are not necessary
      anymore.
      
      Mention that DeepSpeed's version of Adam is about 2x faster.
      
      * Review comments
      81bdbb5e
    • Nathan Lambert's avatar
      fix typo docstring in unet2d (#798) · 71ca10c6
      Nathan Lambert authored
      fix typo docstring
      71ca10c6
    • Patrick von Platen's avatar
      Fix gradient checkpointing test (#797) · 22963ed8
      Patrick von Platen authored
      * Fix gradient checkpointing test
      
      * more tsets
      22963ed8
    • Patrick von Platen's avatar
      [Low CPU memory] + device map (#772) · fab17528
      Patrick von Platen authored
      
      
      * add accelerate to load models with smaller memory footprint
      
      * remove low_cpu_mem_usage as it is reduntant
      
      * move accelerate init weights context to modelling utils
      
      * add test to ensure results are the same when loading with accelerate
      
      * add tests to ensure ram usage gets lower when using accelerate
      
      * move accelerate logic to single snippet under modelling utils and remove it from configuration utils
      
      * format code using to pass quality check
      
      * fix imports with isor
      
      * add accelerate to test extra deps
      
      * only import accelerate if device_map is set to auto
      
      * move accelerate availability check to diffusers import utils
      
      * format code
      
      * add device map to pipeline abstraction
      
      * lint it to pass PR quality check
      
      * fix class check to use accelerate when using diffusers ModelMixin subclasses
      
      * use low_cpu_mem_usage in transformers if device_map is not available
      
      * NoModuleLayer
      
      * comment out tests
      
      * up
      
      * uP
      
      * finish
      
      * Update src/diffusers/pipelines/stable_diffusion/safety_checker.py
      
      * finish
      
      * uP
      
      * make style
      Co-authored-by: default avatarPi Esposito <piero.skywalker@gmail.com>
      fab17528
    • Nathan Lambert's avatar
      add sigmoid betas (#777) · feaa7324
      Nathan Lambert authored
      * add sigmoid betas
      
      * convert to torch
      
      * add comment on source
      feaa7324
    • Nathan Lambert's avatar
      Clean up resnet.py file (#780) · a73f8b72
      Nathan Lambert authored
      * clean up resnet.py
      
      * make style and quality
      
      * minor formatting
      a73f8b72
    • lowinli's avatar
      debug an exception (#638) · 5af6eed9
      lowinli authored
      
      
      * debug an exception
      
      if dst_path is not a file, it will raise Exception in the function src_path.samefile:
      FileNotFoundError: [Errno 2] No such file or directory: '/home/lilongwei/notebook/onnx_diffusion/vae_decoder/model.onnx'
      
      * Update src/diffusers/onnx_utils.py
      Co-authored-by: default avatarAnton Lozhkov <aglozhkov@gmail.com>
      5af6eed9
  3. 07 Oct, 2022 12 commits
  4. 06 Oct, 2022 15 commits