1. 23 Aug, 2023 1 commit
    • Ollin Boer Bohan's avatar
      Fix AutoencoderTiny encoder scaling convention (#4682) · 052bf328
      Ollin Boer Bohan authored
      * Fix AutoencoderTiny encoder scaling convention
      
        * Add [-1, 1] -> [0, 1] rescaling to EncoderTiny
      
        * Move [0, 1] -> [-1, 1] rescaling from AutoencoderTiny.decode to DecoderTiny
          (i.e. immediately after the final conv, as early as possible)
      
        * Fix missing [0, 255] -> [0, 1] rescaling in AutoencoderTiny.forward
      
        * Update AutoencoderTinyIntegrationTests to protect against scaling issues.
          The new test constructs a simple image, round-trips it through AutoencoderTiny,
          and confirms the decoded result is approximately equal to the source image.
          This test checks behavior with and without tiling enabled.
          This test will fail if new AutoencoderTiny scaling issues are introduced.
      
        * Context: Raw TAESD weights expect images in [0, 1], but diffusers'
          convention represents images with zero-centered values in [-1, 1],
          so AutoencoderTiny needs to scale / unscale images at the start of
          encoding and at the end of decoding in order to work with diffusers.
      
      * Re-add existing AutoencoderTiny test, update golden values
      
      * Add comments to AutoencoderTiny.forward
      052bf328
  2. 22 Aug, 2023 2 commits
  3. 21 Aug, 2023 1 commit
    • Sanchit Gandhi's avatar
      Add AudioLDM 2 (#4549) · 7a24977c
      Sanchit Gandhi authored
      
      
      * from audioldm
      
      * unet down + mid
      
      * vae, clap, flan-t5
      
      * start sequence audio mae
      
      * iterate on audioldm encoder
      
      * finish encoder
      
      * finish weight conversion
      
      * text pre-processing
      
      * gpt2 pre-processing
      
      * fix projection model
      
      * working
      
      * unet equivalence
      
      * finish in base
      
      * add unet cond
      
      * finish unet
      
      * finish custom unet
      
      * start clean-up
      
      * revert base unet changes
      
      * refactor pre-processing
      
      * tests: from audioldm
      
      * fix some tests
      
      * more fixes
      
      * iterate on tests
      
      * make fix copies
      
      * harden fast tests
      
      * slow integration tests
      
      * finish tests
      
      * update checkpoint
      
      * update copyright
      
      * docs
      
      * remove outdated method
      
      * add docstring
      
      * make style
      
      * remove decode latents
      
      * enable cpu offload
      
      * (text_encoder_1, tokenizer_1) -> (text_encoder, tokenizer)
      
      * more clean up
      
      * more refactor
      
      * build pr docs
      
      * Update docs/source/en/api/pipelines/audioldm2.md
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      * small clean
      
      * tidy conversion
      
      * update for large checkpoint
      
      * generate -> generate_language_model
      
      * full clap model
      
      * shrink clap-audio in tests
      
      * fix large integration test
      
      * fix fast tests
      
      * use generation config
      
      * make style
      
      * update docs
      
      * finish docs
      
      * finish doc
      
      * update tests
      
      * fix last test
      
      * syntax
      
      * finalise tests
      
      * refactor projection model in prep for TTS
      
      * fix fast tests
      
      * style
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      7a24977c
  4. 18 Aug, 2023 1 commit
  5. 17 Aug, 2023 5 commits
  6. 16 Aug, 2023 3 commits
    • Sayak Paul's avatar
      [Core] feat: MultiControlNet support for SDXL ControlNet pipeline (#4597) · 50495991
      Sayak Paul authored
      * core: add multicontrolnet support to sdxl controlnet
      
      * modify checks.
      
      * fix: original_size determination
      
      * add: tests for multi controlnet sdxl.
      
      * remove unnecessary prints.
      50495991
    • Dirk Morris's avatar
      Fix unipc use_karras_sigmas exception - fixes huggingface/diffusers#4580 (#4581) · a7de9650
      Dirk Morris authored
      * Fix unipc karras sigmas exception - fixes huggingface/diffusers#4580
      
      * Add unipc scheduler tests for karras sigmas
      a7de9650
    • nikhil-masterful's avatar
      Add GLIGEN implementation (#4441) · da5ab51d
      nikhil-masterful authored
      * Add GLIGEN implementation
      
      * GLIGEN: Fix code quality check failures
      
      * GLIGEN: Fix Import block un-sorted or un-formatted failures
      
      * GLIGEN: Fix check_repository_consistency failures
      
      * GLIGEN: Add 'PositionNet' to versatile_diffusion/modeling_text_unet.py
      
      * GLIGEN: check_repository_consistency: fix 'copy does not match' error
      
      * GLIGEN: Fix review comments (1)
      
      * GLIGEN: Fix E721 Do not compare types, use `isinstance()` failures
      
      * GLIGEN : Ensure _encode_prompt() copy matches to StableDiffusionPipeline
      
      * GLIGEN: Fix ruff E721 failure in unidiffuser/test_unidiffuser.py
      
      * GLIGEN: doc_builder: restyle pipeline_stable_diffusion_gligen.py
      
      * GIGLEN: reset files unrelated to gligen
      
      * GLIGEN: Fix documentation comments (1)
      
      * GLIGEN: Fix review comments (2)
      
      * GLIGEN: Added FastTest
      
      * GLIGEN: Fix review comments (3)
      da5ab51d
  7. 15 Aug, 2023 1 commit
  8. 11 Aug, 2023 1 commit
  9. 09 Aug, 2023 3 commits
  10. 07 Aug, 2023 1 commit
  11. 04 Aug, 2023 4 commits
  12. 03 Aug, 2023 2 commits
  13. 02 Aug, 2023 2 commits
    • Sayak Paul's avatar
      [Feat] add tiny Autoencoder for (almost) instant decoding (#4384) · 18fc40c1
      Sayak Paul authored
      
      
      * add: model implementation of tiny autoencoder.
      
      * add: inits.
      
      * push the latest devs.
      
      * add: conversion script and finish.
      
      * add: scaling factor args.
      
      * debugging
      
      * fix denormalization.
      
      * fix: positional argument.
      
      * handle use_torch_2_0_or_xformers.
      
      * handle post_quant_conv
      
      * handle dtype
      
      * fix: sdxl image processor for tiny ae.
      
      * fix: sdxl image processor for tiny ae.
      
      * unify upcasting logic.
      
      * copied from madness.
      
      * remove trailing whitespace.
      
      * set is_tiny_vae = False
      
      * address PR comments.
      
      * change to AutoencoderTiny
      
      * make act_fn an str throughout
      
      * fix: apply_forward_hook decorator call
      
      * get rid of the special is_tiny_vae flag.
      
      * directly scale the output.
      
      * fix dummies?
      
      * fix: act_fn.
      
      * get rid of the Clamp() layer.
      
      * bring back copied from.
      
      * movement of the blocks to appropriate modules.
      
      * add: docstrings to AutoencoderTiny
      
      * add: documentation.
      
      * changes to the conversion script.
      
      * add doc entry.
      
      * settle tests.
      
      * style
      
      * add one slow test.
      
      * fix
      
      * fix 2
      
      * fix 2
      
      * fix: 4
      
      * fix: 5
      
      * finish integration tests
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * style
      
      ---------
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      18fc40c1
    • Sayak Paul's avatar
      [LoRA] Fix SDXL text encoder LoRAs (#4371) · 816ca004
      Sayak Paul authored
      
      
      * temporarily disable text encoder loras.
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debbuging.
      
      * modify doc.
      
      * rename tests.
      
      * print slices.
      
      * fix: assertions
      
      * Apply suggestions from code review
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      816ca004
  14. 01 Aug, 2023 2 commits
  15. 28 Jul, 2023 3 commits
    • Sayak Paul's avatar
      [Feat] Support SDXL Kohya-style LoRA (#4287) · 4a4cdd6b
      Sayak Paul authored
      
      
      * sdxl lora changes.
      
      * better name replacement.
      
      * better replacement.
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * remove print.
      
      * print state dict keys.
      
      * print
      
      * distingisuih better
      
      * debuggable.
      
      * fxi: tyests
      
      * fix: arg from training script.
      
      * access from class.
      
      * run style
      
      * debug
      
      * save intermediate
      
      * some simplifications for SDXL LoRA
      
      * styling
      
      * unet config is not needed in diffusers format.
      
      * fix: dynamic SGM block mapping for SDXL kohya loras (#4322)
      
      * Use lora compatible layers for linear proj_in/proj_out (#4323)
      
      * improve condition for using the sgm_diffusers mapping
      
      * informative comment.
      
      * load compatible keys and embedding layer maaping.
      
      * Get SDXL 1.0 example lora to load
      
      * simplify
      
      * specif ranks and hidden sizes.
      
      * better handling of k rank and hidden
      
      * debug
      
      * debug
      
      * debug
      
      * debug
      
      * debug
      
      * fix: alpha keys
      
      * add check for handling LoRAAttnAddedKVProcessor
      
      * sanity comment
      
      * modifications for text encoder SDXL
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * denugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * up
      
      * up
      
      * up
      
      * up
      
      * up
      
      * up
      
      * unneeded comments.
      
      * unneeded comments.
      
      * kwargs for the other attention processors.
      
      * kwargs for the other attention processors.
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * improve
      
      * debugging
      
      * debugging
      
      * more print
      
      * Fix alphas
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * clean up
      
      * clean up.
      
      * debugging
      
      * fix: text
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarBatuhan Taskaya <batuhan@python.org>
      4a4cdd6b
    • Patrick von Platen's avatar
      [ONNX] Don't download ONNX model by default (#4338) · 306a7bd0
      Patrick von Platen authored
      * [Download] Don't download ONNX weights by default
      
      * [Download] Don't download ONNX weights by default
      
      * [Download] Don't download ONNX weights by default
      
      * fix more
      
      * finish
      
      * finish
      
      * finish
      306a7bd0
    • Patrick von Platen's avatar
      [SDXL Refiner] Fix refiner forward pass for batched input (#4327) · 18b018c8
      Patrick von Platen authored
      * fix_batch_xl
      
      * Fix other pipelines as well
      
      * up
      
      * up
      
      * Update tests/pipelines/stable_diffusion_xl/test_stable_diffusion_xl_inpaint.py
      
      * sort
      
      * up
      
      * Finish it all up Co-authored-by: Bagheera <bghira@users.github.com>
      
      * Co-authored-by: Bagheera bghira@users.github.com
      
      * Co-authored-by: Bagheera <bghira@users.github.com>
      
      * Finish it all up Co-authored-by: Bagheera <bghira@users.github.com>
      18b018c8
  16. 27 Jul, 2023 3 commits
  17. 26 Jul, 2023 1 commit
  18. 25 Jul, 2023 4 commits
    • Batuhan Taskaya's avatar
      Load Kohya-ss style LoRAs with auxilary states (#4147) · ff8f5808
      Batuhan Taskaya authored
      
      
      * Support to load Kohya-ss style LoRA file format (without restrictions)
      Co-Authored-By: default avatarTakuma Mori <takuma104@gmail.com>
      Co-Authored-By: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      * tmp: add sdxl to mlp_modules
      
      ---------
      Co-authored-by: default avatarTakuma Mori <takuma104@gmail.com>
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      ff8f5808
    • Harutatsu Akiyama's avatar
      [SDXL and IP2P]: instruction pix2pix XL training and pipeline (#4079) · 428dbfec
      Harutatsu Akiyama authored
      
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * [Community] Implementation of the IADB community pipeline (#3996)
      
      * community pipeline: implementation of iadb
      
      * iadb.py: reformat using black
      
      * iadb.py: linting update
      
      * add kandinsky to readme table (#4081)
      Co-authored-by: default avataryiyixuxu <yixu310@gmail,com>
      
      * [From Single File] Force accelerate to be installed (#4078)
      
      force accelerate to be installed
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Support instruction pix2pix sdxl
      
      * Clean up IP2P SDXL code
      
      * Clean up IP2P SDXL code
      
      * [IP2P and SDXL] clean up code
      
      * [IP2P and SDXL] clean up code
      
      * [IP2P and SDXL] clean up code
      
      * [IP2P SDXL] Address code reviews
      
      * [IP2P SDXL] Address code reviews, add docs, tests
      
      * [IP2P SDXL] Address code reviews, add docs, tests
      
      * [IP2P SDXL] Address code reviews, add docs, tests
      
      * [IP2P SDXL] Address code reviews, add docs, tests
      
      * [IP2P SDXL] Address code reviews, add docs, tests
      
      * [IP2P SDXL] Address code reviews, add docs, tests
      
      * [IP2P SDXL] Address code reviews, add docs, tests
      
      * [IP2P SDXL] Address code reviews, add docs, tests
      
      * [IP2P SDXL] Address code reviews, add docs, tests
      
      * [IP2P SDXL] Address code reviews, add docs, tests
      
      * [IP2P SDXL] Address code reviews, add docs, tests
      
      * [IP2P SDXL] Address code reviews, add docs, tests
      
      * [IP2P SDXL] Address code reviews
      
      * [IP2P SDXL] Address code reviews
      
      * [IP2P SDXL] Add README_SDXL
      
      * [IP2P SDXL] Address code reviews
      
      * [IP2P SDXL] Address code reviews
      
      * [IP2P SDXL] Fix the copy problems
      
      * [IP2P SDXL] Add license
      
      * [IP2P SDXL] Add license
      
      * [IP2P SDXL] Add license
      
      * [IP2P SDXL] Address code reivew for selecting VAE andd others
      
      * [IP2P SDXL] Update README_sdxl
      
      * [IP2P SDXL] Update __init__
      
      * [IP2P SDXL] Update dummy_torch_and_transformers_and_invisible_watermark_objects
      
      * address patrick's comments and some additions to readmes.
      
      ---------
      Co-authored-by: default avatarHarutatsu Akiyama <kf.zy.qin@gmail.com>
      Co-authored-by: default avatarThomas Chambon <36728882+tchambon@users.noreply.github.com>
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      Co-authored-by: default avataryiyixuxu <yixu310@gmail,com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      428dbfec
    • YiYi Xu's avatar
      [draft v2] AutoPipeline (#4138) · c11d11d6
      YiYi Xu authored
      
      
      * initial
      
      * style
      
      * from ...pipelines -> from ..pipeline_util
      
      * make style
      
      * fix-copies
      
      * fix value_guided_sampling oops
      
      * style
      
      * add test
      
      * Show failing test
      
      * update from_pipe
      
      * fix
      
      * add controlnet, additional test and register unused original config
      
      * update for controlnet
      
      * Apply suggestions from code review
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * store unused config as private attribute and pass if can
      
      * add doc
      
      * kandinsky inpaint pipeline does not work with decoder checkpoint
      
      * update doc
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      * style
      
      * Apply suggestions from code review
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * fix
      
      * Apply suggestions from code review
      
      ---------
      Co-authored-by: default avataryiyixuxu <yixu310@gmail,com>
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      c11d11d6
    • Sayak Paul's avatar
      [SDXL DreamBooth LoRA] add support for text encoder fine-tuning (#4097) · 365e8461
      Sayak Paul authored
      
      
      * Allow low precision sd xl
      
      * finish
      
      * finish
      
      * feat: initial draft for supporting text encoder lora finetuning for SDXL DreamBooth
      
      * fix: variable assignments.
      
      * add: autocast block.
      
      * add debugging
      
      * vae dtype hell
      
      * fix: vae dtype hell.
      
      * fix: vae dtype hell 3.
      
      * clean up
      
      * lora text encoder loader.
      
      * fix: unwrapping models.
      
      * add: tests.
      
      * docs.
      
      * handle unexpected keys.
      
      * fix vae dtype in the final inference.
      
      * fix scope problem.
      
      * fix: save_model_card args.
      
      * initialize: prefix to None.
      
      * fix: dtype issues.
      
      * apply gixes.
      
      * debgging.
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * add: fast tests.
      
      * pre-tokenize.
      
      * address: will's comments.
      
      * fix: loader and tests.
      
      * fix: dataloader.
      
      * simplify dataloader.
      
      * length.
      
      * simplification.
      
      * make style && make quality
      
      * simplify state_dict munging
      
      * fix: tests.
      
      * fix: state_dict packing.
      
      * Apply suggestions from code review
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      365e8461