1. 09 Jan, 2026 1 commit
  2. 03 Dec, 2025 1 commit
  3. 10 Nov, 2025 1 commit
  4. 15 Oct, 2025 1 commit
  5. 30 Sep, 2025 1 commit
  6. 16 Sep, 2025 1 commit
  7. 26 Aug, 2025 1 commit
  8. 22 Jul, 2025 1 commit
  9. 26 Jun, 2025 1 commit
  10. 19 Jun, 2025 1 commit
  11. 19 May, 2025 1 commit
  12. 01 May, 2025 1 commit
  13. 12 Mar, 2025 1 commit
  14. 03 Mar, 2025 1 commit
  15. 28 Jan, 2025 1 commit
    • Aryan's avatar
      Refactor gradient checkpointing (#10611) · c4d4ac21
      Aryan authored
      * update
      
      * remove unused fn
      
      * apply suggestions based on review
      
      * update + cleanup 🧹
      
      * more cleanup 🧹
      
      * make fix-copies
      
      * update test
      c4d4ac21
  16. 22 Jan, 2025 1 commit
    • Aryan's avatar
      [core] Layerwise Upcasting (#10347) · beacaa55
      Aryan authored
      
      
      * update
      
      * update
      
      * make style
      
      * remove dynamo disable
      
      * add coauthor
      Co-Authored-By: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      * update
      
      * update
      
      * update
      
      * update mixin
      
      * add some basic tests
      
      * update
      
      * update
      
      * non_blocking
      
      * improvements
      
      * update
      
      * norm.* -> norm
      
      * apply suggestions from review
      
      * add example
      
      * update hook implementation to the latest changes from pyramid attention broadcast
      
      * deinitialize should raise an error
      
      * update doc page
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * update docs
      
      * update
      
      * refactor
      
      * fix _always_upcast_modules for asym ae and vq_model
      
      * fix lumina embedding forward to not depend on weight dtype
      
      * refactor tests
      
      * add simple lora inference tests
      
      * _always_upcast_modules -> _precision_sensitive_module_patterns
      
      * remove todo comments about review; revert changes to self.dtype in unets because .dtype on ModelMixin should be able to handle fp8 weight case
      
      * check layer dtypes in lora test
      
      * fix UNet1DModelTests::test_layerwise_upcasting_inference
      
      * _precision_sensitive_module_patterns -> _skip_layerwise_casting_patterns based on feedback
      
      * skip test in NCSNppModelTests
      
      * skip tests for AutoencoderTinyTests
      
      * skip tests for AutoencoderOobleckTests
      
      * skip tests for UNet1DModelTests - unsupported pytorch operations
      
      * layerwise_upcasting -> layerwise_casting
      
      * skip tests for UNetRLModelTests; needs next pytorch release for currently unimplemented operation support
      
      * add layerwise fp8 pipeline test
      
      * use xfail
      
      * Apply suggestions from code review
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      * add assertion with fp32 comparison; add tolerance to fp8-fp32 vs fp32-fp32 comparison (required for a few models' test to pass)
      
      * add note about memory consumption on tesla CI runner for failing test
      
      ---------
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      beacaa55
  17. 21 Jan, 2025 1 commit
  18. 20 Jan, 2025 1 commit
  19. 08 Jan, 2025 1 commit
  20. 20 Dec, 2024 1 commit
  21. 19 Dec, 2024 1 commit
  22. 18 Dec, 2024 1 commit
  23. 16 Dec, 2024 1 commit
  24. 13 Dec, 2024 1 commit
  25. 08 Nov, 2024 1 commit
  26. 06 Sep, 2024 1 commit
    • Aryan's avatar
      [core] Freenoise memory improvements (#9262) · 6dfa4996
      Aryan authored
      * update
      
      * implement prompt interpolation
      
      * make style
      
      * resnet memory optimizations
      
      * more memory optimizations; todo: refactor
      
      * update
      
      * update animatediff controlnet with latest changes
      
      * refactor chunked inference changes
      
      * remove print statements
      
      * update
      
      * chunk -> split
      
      * remove changes from incorrect conflict resolution
      
      * remove changes from incorrect conflict resolution
      
      * add explanation of SplitInferenceModule
      
      * update docs
      
      * Revert "update docs"
      
      This reverts commit c55a50a271b2cefa8fe340a4f2a3ab9b9d374ec0.
      
      * update docstring for freenoise split inference
      
      * apply suggestions from review
      
      * add tests
      
      * apply suggestions from review
      6dfa4996
  27. 04 Sep, 2024 1 commit
  28. 28 Aug, 2024 1 commit
    • Aryan's avatar
      AnimateDiff prompt travel (#9231) · cbc2ec8f
      Aryan authored
      * update
      
      * implement prompt interpolation
      
      * make style
      
      * resnet memory optimizations
      
      * more memory optimizations; todo: refactor
      
      * update
      
      * update animatediff controlnet with latest changes
      
      * refactor chunked inference changes
      
      * remove print statements
      
      * undo memory optimization changes
      
      * update docstrings
      
      * fix tests
      
      * fix pia tests
      
      * apply suggestions from review
      
      * add tests
      
      * update comment
      cbc2ec8f
  29. 07 Aug, 2024 2 commits
    • Aryan's avatar
      [feat] allow sparsectrl to be loaded from single file (#9073) · f6df2244
      Aryan authored
      
      
      * allow sparsectrl to be loaded with single file
      
      * update
      
      ---------
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      f6df2244
    • Aryan's avatar
      [core] FreeNoise (#8948) · 16a93f1a
      Aryan authored
      
      
      * initial work draft for freenoise; needs massive cleanup
      
      * fix freeinit bug
      
      * add animatediff controlnet implementation
      
      * revert attention changes
      
      * add freenoise
      
      * remove old helper functions
      
      * add decode batch size param to all pipelines
      
      * make style
      
      * fix copied from comments
      
      * make fix-copies
      
      * make style
      
      * copy animatediff controlnet implementation from #8972
      
      * add experimental support for num_frames not perfectly fitting context length, ocntext stride
      
      * make unet motion model lora work again based on #8995
      
      * copy load video utils from #8972
      
      * copied from AnimateDiff::prepare_latents
      
      * address the case where last batch of frames does not match length of indices in prepare latents
      
      * decode_batch_size->vae_batch_size; batch vae encode support in animatediff vid2vid
      
      * revert sparsectrl and sdxl freenoise changes
      
      * revert pia
      
      * add freenoise tests
      
      * make fix-copies
      
      * improve docstrings
      
      * add freenoise tests to animatediff controlnet
      
      * update tests
      
      * Update src/diffusers/models/unets/unet_motion_model.py
      
      * add freenoise to animatediff pag
      
      * address review comments
      
      * make style
      
      * update tests
      
      * make fix-copies
      
      * fix error message
      
      * remove copied from comment
      
      * fix imports in tests
      
      * update
      
      ---------
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      16a93f1a
  30. 03 Aug, 2024 1 commit
    • Aryan's avatar
      [refactor] create modeling blocks specific to AnimateDiff (#8979) · fbe29c62
      Aryan authored
      
      
      * animatediff specific transformer model
      
      * make style
      
      * make fix-copies
      
      * move blocks to unet motion model
      
      * make style
      
      * remove dummy object
      
      * fix incorrectly passed param causing test failures
      
      * rename model and output class
      
      * fix sparsectrl imports
      
      * remove todo comments
      
      * remove temporal double self attn param from controlnet sparsectrl
      
      * add deprecated versions of blocks
      
      * apply suggestions from review
      
      * update
      
      ---------
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      fbe29c62
  31. 30 Jul, 2024 1 commit
  32. 26 Jul, 2024 2 commits
    • Álvaro Somoza's avatar
      [Kolors] Add IP Adapter (#8901) · 73acebb8
      Álvaro Somoza authored
      * initial draft
      
      * apply suggestions
      
      * fix failing test
      
      * added ipa to img2img
      
      * add docs
      
      * apply suggestions
      73acebb8
    • Aryan's avatar
      [core] AnimateDiff SparseCtrl (#8897) · 5c53ca5e
      Aryan authored
      * initial sparse control model draft
      
      * remove unnecessary implementation
      
      * copy animatediff pipeline
      
      * remove deprecated callbacks
      
      * update
      
      * update pipeline implementation progress
      
      * make style
      
      * make fix-copies
      
      * update progress
      
      * add partially working pipeline
      
      * remove debug prints
      
      * add model docs
      
      * dummy objects
      
      * improve motion lora conversion script
      
      * fix bugs
      
      * update docstrings
      
      * remove unnecessary model params; docs
      
      * address review comment
      
      * add copied from to zero_module
      
      * copy animatediff test
      
      * add fast tests
      
      * update docs
      
      * update
      
      * update pipeline docs
      
      * fix expected slice values
      
      * fix license
      
      * remove get_down_block usage
      
      * remove temporal_double_self_attention from get_down_block
      
      * update
      
      * update docs with org and documentation images
      
      * make from_unet work in sparsecontrolnetmodel
      
      * add latest freeinit test from #8969
      
      * make fix-copies
      
      * LoraLoaderMixin -> StableDiffsuionLoraLoaderMixin
      5c53ca5e
  33. 25 Jul, 2024 1 commit
    • Aryan's avatar
      [tests] speed up animatediff tests (#8846) · 3ae0ee88
      Aryan authored
      
      
      * speed up animatediff tests
      
      * fix pia test_ip_adapter_single
      
      * fix tests/pipelines/pia/test_pia.py::PIAPipelineFastTests::test_dict_tuple_outputs_equivalent
      
      * update
      
      * fix ip adapter tests
      
      * skip test_from_pipe_consistent_config tests
      
      * fix prompt_embeds test
      
      * update test_from_pipe_consistent_config tests
      
      * fix expected_slice values
      
      * remove temporal_norm_num_groups from UpBlockMotion
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      3ae0ee88
  34. 24 Jul, 2024 1 commit
    • Sayak Paul's avatar
      [Core] fix QKV fusion for attention (#8829) · 50d21f7c
      Sayak Paul authored
      * start debugging the problem,
      
      * start
      
      * fix
      
      * fix
      
      * fix imports.
      
      * handle hunyuan
      
      * remove residuals.
      
      * add a check for making sure there's appropriate procs.
      
      * add more rigor to the tests.
      
      * fix test
      
      * remove redundant check
      
      * fix-copies
      
      * move check_qkv_fusion_matches_attn_procs_length and check_qkv_fusion_processors_exist.
      50d21f7c
  35. 12 Jul, 2024 1 commit
  36. 08 Jul, 2024 1 commit
  37. 02 Jul, 2024 1 commit
  38. 27 Jun, 2024 1 commit
    • Mathis Koroglu's avatar
      Motion Model / Adapter versatility (#8301) · 3e0d128d
      Mathis Koroglu authored
      * Motion Model / Adapter versatility
      
      - allow to use a different number of layers per block
      - allow to use a different number of transformer per layers per block
      - allow a different number of motion attention head per block
      - use dropout argument in get_down/up_block in 3d blocks
      
      * Motion Model added arguments renamed & refactoring
      
      * Add test for asymmetric UNetMotionModel
      3e0d128d