1. 14 Apr, 2025 1 commit
  2. 13 Apr, 2025 4 commits
  3. 12 Apr, 2025 1 commit
  4. 11 Apr, 2025 4 commits
  5. 10 Apr, 2025 9 commits
  6. 09 Apr, 2025 13 commits
  7. 08 Apr, 2025 8 commits
    • Sayak Paul's avatar
      [bistandbytes] improve replacement warnings for bnb (#11132) · 1a048124
      Sayak Paul authored
      * improve replacement warnings for bnb
      
      * updates to docs.
      1a048124
    • Sayak Paul's avatar
      [feat] implement `record_stream` when using CUDA streams during group offloading (#11081) · 4b27c4a4
      Sayak Paul authored
      
      
      * implement record_stream for better performance.
      
      * fix
      
      * style.
      
      * merge #11097
      
      * Update src/diffusers/hooks/group_offloading.py
      Co-authored-by: default avatarAryan <aryan@huggingface.co>
      
      * fixes
      
      * docstring.
      
      * remaining todos in low_cpu_mem_usage
      
      * tests
      
      * updates to docs.
      
      ---------
      Co-authored-by: default avatarAryan <aryan@huggingface.co>
      4b27c4a4
    • hlky's avatar
      Flux quantized with lora (#10990) · 5d49b3e8
      hlky authored
      
      
      * Flux quantized with lora
      
      * fix
      
      * changes
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      * Apply style fixes
      
      * enable model cpu offload()
      
      * Update src/diffusers/loaders/lora_pipeline.py
      Co-authored-by: default avatarhlky <hlky@hlky.ac>
      
      * update
      
      * Apply suggestions from code review
      
      * update
      
      * add peft as an additional dependency for gguf
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      Co-authored-by: default avatargithub-actions[bot] <github-actions[bot]@users.noreply.github.com>
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      5d49b3e8
    • Linoy Tsaban's avatar
      [Flux LoRA] fix issues in flux lora scripts (#11111) · 71f34fc5
      Linoy Tsaban authored
      
      
      * remove custom scheduler
      
      * update requirements.txt
      
      * log_validation with mixed precision
      
      * add intermediate embeddings saving when checkpointing is enabled
      
      * remove comment
      
      * fix validation
      
      * add unwrap_model for accelerator, torch.no_grad context for validation, fix accelerator.accumulate call in advanced script
      
      * revert unwrap_model change temp
      
      * add .module to address distributed training bug + replace accelerator.unwrap_model with unwrap model
      
      * changes to align advanced script with canonical script
      
      * make changes for distributed training + unify unwrap_model calls in advanced script
      
      * add module.dtype fix to dreambooth script
      
      * unify unwrap_model calls in dreambooth script
      
      * fix condition in validation run
      
      * mixed precision
      
      * Update examples/advanced_diffusion_training/train_dreambooth_lora_flux_advanced.py
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      * smol style change
      
      * change autocast
      
      * Apply style fixes
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      Co-authored-by: default avatargithub-actions[bot] <github-actions[bot]@users.noreply.github.com>
      71f34fc5
    • Yao Matrix's avatar
      introduce compute arch specific expectations and fix test_sd3_img2img_inference failure (#11227) · c51b6bd8
      Yao Matrix authored
      
      
      * add arch specfic expectations support, to support different arch's numerical characteristics
      Signed-off-by: default avatarYAO Matrix <matrix.yao@intel.com>
      
      * fix typo
      Signed-off-by: default avatarYAO Matrix <matrix.yao@intel.com>
      
      * Apply suggestions from code review
      
      * Apply style fixes
      
      * Update src/diffusers/utils/testing_utils.py
      
      ---------
      Signed-off-by: default avatarYAO Matrix <matrix.yao@intel.com>
      Co-authored-by: default avatarhlky <hlky@hlky.ac>
      Co-authored-by: default avatargithub-actions[bot] <github-actions[bot]@users.noreply.github.com>
      c51b6bd8
    • Benjamin Bossan's avatar
      [LoRA] Implement hot-swapping of LoRA (#9453) · fb544996
      Benjamin Bossan authored
      * [WIP][LoRA] Implement hot-swapping of LoRA
      
      This PR adds the possibility to hot-swap LoRA adapters. It is WIP.
      
      Description
      
      As of now, users can already load multiple LoRA adapters. They can
      offload existing adapters or they can unload them (i.e. delete them).
      However, they cannot "hotswap" adapters yet, i.e. substitute the weights
      from one LoRA adapter with the weights of another, without the need to
      create a separate LoRA adapter.
      
      Generally, hot-swapping may not appear not super useful but when the
      model is compiled, it is necessary to prevent recompilation. See #9279
      for more context.
      
      Caveats
      
      To hot-swap a LoRA adapter for another, these two adapters should target
      exactly the same layers and the "hyper-parameters" of the two adapters
      should be identical. For instance, the LoRA alpha has to be the same:
      Given that we keep the alpha from the first adapter, the LoRA scaling
      would be incorrect for the second adapter otherwise.
      
      Theoretically, we could override the scaling dict with the alpha values
      derived from the second adapter's config, but changing the dict will
      trigger a guard for recompilation, defeating the main purpose of the
      feature.
      
      I also found that compilation flags can have an impact on whether this
      works or not. E.g. when passing "reduce-overhead", there will be errors
      of the type:
      
      > input name: arg861_1. data pointer changed from 139647332027392 to
      139647331054592
      
      I don't know enough about compilation to determine whether this is
      problematic or not.
      
      Current state
      
      This is obviously WIP right now to collect feedback and discuss which
      direction to take this. If this PR turns out to be useful, the
      hot-swapping functions will be added to PEFT itself and can be imported
      here (or there is a separate copy in diffusers to avoid the need for a
      min PEFT version to use this feature).
      
      Moreover, more tests need to be added to better cover this feature,
      although we don't necessarily need tests for the hot-swapping
      functionality itself, since those tests will be added to PEFT.
      
      Furthermore, as of now, this is only implemented for the unet. Other
      pipeline components have yet to implement this feature.
      
      Finally, it should be properly documented.
      
      I would like to collect feedback on the current state of the PR before
      putting more time into finalizing it.
      
      * Reviewer feedback
      
      * Reviewer feedback, adjust test
      
      * Fix, doc
      
      * Make fix
      
      * Fix for possible g++ error
      
      * Add test for recompilation w/o hotswapping
      
      * Make hotswap work
      
      Requires https://github.com/huggingface/peft/pull/2366
      
      More changes to make hotswapping work. Together with the mentioned PEFT
      PR, the tests pass for me locally.
      
      List of changes:
      
      - docstring for hotswap
      - remove code copied from PEFT, import from PEFT now
      - adjustments to PeftAdapterMixin.load_lora_adapter (unfortunately, some
        state dict renaming was necessary, LMK if there is a better solution)
      - adjustments to UNet2DConditionLoadersMixin._process_lora: LMK if this
        is even necessary or not, I'm unsure what the overall relationship is
        between this and PeftAdapterMixin.load_lora_adapter
      - also in UNet2DConditionLoadersMixin._process_lora, I saw that there is
        no LoRA unloading when loading the adapter fails, so I added it
        there (in line with what happens in PeftAdapterMixin.load_lora_adapter)
      - rewritten tests to avoid shelling out, make the test more precise by
        making sure that the outputs align, parametrize it
      - also checked the pipeline code mentioned in this comment:
        https://github.com/huggingface/diffusers/pull/9453#issuecomment-2418508871;
      
      
        when running this inside the with
        torch._dynamo.config.patch(error_on_recompile=True) context, there is
        no error, so I think hotswapping is now working with pipelines.
      
      * Address reviewer feedback:
      
      - Revert deprecated method
      - Fix PEFT doc link to main
      - Don't use private function
      - Clarify magic numbers
      - Add pipeline test
      
      Moreover:
      - Extend docstrings
      - Extend existing test for outputs != 0
      - Extend existing test for wrong adapter name
      
      * Change order of test decorators
      
      parameterized.expand seems to ignore skip decorators if added in last
      place (i.e. innermost decorator).
      
      * Split model and pipeline tests
      
      Also increase test coverage by also targeting conv2d layers (support of
      which was added recently on the PEFT PR).
      
      * Reviewer feedback: Move decorator to test classes
      
      ... instead of having them on each test method.
      
      * Apply suggestions from code review
      Co-authored-by: default avatarhlky <hlky@hlky.ac>
      
      * Reviewer feedback: version check, TODO comment
      
      * Add enable_lora_hotswap method
      
      * Reviewer feedback: check _lora_loadable_modules
      
      * Revert changes in unet.py
      
      * Add possibility to ignore enabled at wrong time
      
      * Fix docstrings
      
      * Log possible PEFT error, test
      
      * Raise helpful error if hotswap not supported
      
      I.e. for the text encoder
      
      * Formatting
      
      * More linter
      
      * More ruff
      
      * Doc-builder complaint
      
      * Update docstring:
      
      - mention no text encoder support yet
      - make it clear that LoRA is meant
      - mention that same adapter name should be passed
      
      * Fix error in docstring
      
      * Update more methods with hotswap argument
      
      - SDXL
      - SD3
      - Flux
      
      No changes were made to load_lora_into_transformer.
      
      * Add hotswap argument to load_lora_into_transformer
      
      For SD3 and Flux. Use shorter docstring for brevity.
      
      * Extend docstrings
      
      * Add version guards to tests
      
      * Formatting
      
      * Fix LoRA loading call to add prefix=None
      
      See:
      https://github.com/huggingface/diffusers/pull/10187#issuecomment-2717571064
      
      
      
      * Run make fix-copies
      
      * Add hot swap documentation to the docs
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      Co-authored-by: default avatarhlky <hlky@hlky.ac>
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      fb544996
    • Álvaro Somoza's avatar
      [Training] Better image interpolation in training scripts (#11206) · 723dbdd3
      Álvaro Somoza authored
      
      
      * initial
      
      * Update examples/dreambooth/train_dreambooth_lora_sdxl.py
      Co-authored-by: default avatarhlky <hlky@hlky.ac>
      
      * update
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      Co-authored-by: default avatarhlky <hlky@hlky.ac>
      723dbdd3
    • Bhavay Malhotra's avatar
      [train_controlnet.py] Fix the LR schedulers when num_train_epochs is passed in... · fbf61f46
      Bhavay Malhotra authored
      
      [train_controlnet.py] Fix the LR schedulers when num_train_epochs is passed in a distributed training env (#8461)
      
      * Create diffusers.yml
      
      * fix num_train_epochs
      
      * Delete diffusers.yml
      
      * Fixed Changes
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      fbf61f46