1. 25 Sep, 2025 1 commit
  2. 12 Sep, 2025 1 commit
  3. 11 Sep, 2025 1 commit
  4. 10 Sep, 2025 1 commit
  5. 28 Aug, 2025 1 commit
  6. 14 Aug, 2025 1 commit
  7. 29 Jul, 2025 1 commit
  8. 10 Jul, 2025 2 commits
  9. 23 Jun, 2025 1 commit
  10. 20 Jun, 2025 1 commit
  11. 11 Jun, 2025 1 commit
    • Sayak Paul's avatar
      [tests] model-level `device_map` clarifications (#11681) · 91545666
      Sayak Paul authored
      * add clarity in documentation for device_map
      
      * docs
      
      * fix how compiler tester mixins are used.
      
      * propagate
      
      * more
      
      * typo.
      
      * fix tests
      
      * fix order of decroators.
      
      * clarify more.
      
      * more test cases.
      
      * fix doc
      
      * fix device_map docstring in pipeline_utils.
      
      * more examples
      
      * more
      
      * update
      
      * remove code for stuff that is already supported.
      
      * fix stuff.
      91545666
  12. 09 Jun, 2025 1 commit
  13. 21 May, 2025 1 commit
  14. 19 May, 2025 1 commit
  15. 13 May, 2025 1 commit
  16. 09 May, 2025 1 commit
    • Sayak Paul's avatar
      feat: pipeline-level quantization config (#11130) · 599c8871
      Sayak Paul authored
      
      
      * feat: pipeline-level quant config.
      Co-authored-by: default avatarSunMarc <marc.sun@hotmail.fr>
      
      condition better.
      
      support mapping.
      
      improvements.
      
      [Quantization] Add Quanto backend (#10756)
      
      * update
      
      * updaet
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * Update docs/source/en/quantization/quanto.md
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * Update src/diffusers/quantizers/quanto/utils.py
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      * update
      
      * update
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      [Single File] Add single file loading for SANA Transformer (#10947)
      
      * added support for from_single_file
      
      * added diffusers mapping script
      
      * added testcase
      
      * bug fix
      
      * updated tests
      
      * corrected code quality
      
      * corrected code quality
      
      ---------
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      [LoRA] Improve warning messages when LoRA loading becomes a no-op (#10187)
      
      * updates
      
      * updates
      
      * updates
      
      * updates
      
      * notebooks revert
      
      * fix-copies.
      
      * seeing
      
      * fix
      
      * revert
      
      * fixes
      
      * fixes
      
      * fixes
      
      * remove print
      
      * fix
      
      * conflicts ii.
      
      * updates
      
      * fixes
      
      * better filtering of prefix.
      
      ---------
      Co-authored-by: default avatarhlky <hlky@hlky.ac>
      
      [LoRA] CogView4 (#10981)
      
      * update
      
      * make fix-copies
      
      * update
      
      [Tests] improve quantization tests by additionally measuring the inference memory savings (#11021)
      
      * memory usage tests
      
      * fixes
      
      * gguf
      
      [`Research Project`] Add AnyText: Multilingual Visual Text Generation And Editing (#8998)
      
      * Add initial template
      
      * Second template
      
      * feat: Add TextEmbeddingModule to AnyTextPipeline
      
      * feat: Add AuxiliaryLatentModule template to AnyTextPipeline
      
      * Add bert tokenizer from the anytext repo for now
      
      * feat: Update AnyTextPipeline's modify_prompt method
      
      This commit adds improvements to the modify_prompt method in the AnyTextPipeline class. The method now handles special characters and replaces selected string prompts with a placeholder. Additionally, it includes a check for Chinese text and translation using the trans_pipe.
      
      * Fill in the `forward` pass of `AuxiliaryLatentModule`
      
      * `make style && make quality`
      
      * `chore: Update bert_tokenizer.py with a TODO comment suggesting the use of the transformers library`
      
      * Update error handling to raise and logging
      
      * Add `create_glyph_lines` function into `TextEmbeddingModule`
      
      * make style
      
      * Up
      
      * Up
      
      * Up
      
      * Up
      
      * Remove several comments
      
      * refactor: Remove ControlNetConditioningEmbedding and update code accordingly
      
      * Up
      
      * Up
      
      * up
      
      * refactor: Update AnyTextPipeline to include new optional parameters
      
      * up
      
      * feat: Add OCR model and its components
      
      * chore: Update `TextEmbeddingModule` to include OCR model components and dependencies
      
      * chore: Update `AuxiliaryLatentModule` to include VAE model and its dependencies for masked image in the editing task
      
      * `make style`
      
      * refactor: Update `AnyTextPipeline`'s docstring
      
      * Update `AuxiliaryLatentModule` to include info dictionary so that text processing is done once
      
      * simplify
      
      * `make style`
      
      * Converting `TextEmbeddingModule` to ordinary `encode_prompt()` function
      
      * Simplify for now
      
      * `make style`
      
      * Up
      
      * feat: Add scripts to convert AnyText controlnet to diffusers
      
      * `make style`
      
      * Fix: Move glyph rendering to `TextEmbeddingModule` from `AuxiliaryLatentModule`
      
      * make style
      
      * Up
      
      * Simplify
      
      * Up
      
      * feat: Add safetensors module for loading model file
      
      * Fix device issues
      
      * Up
      
      * Up
      
      * refactor: Simplify
      
      * refactor: Simplify code for loading models and handling data types
      
      * `make style`
      
      * refactor: Update to() method in FrozenCLIPEmbedderT3 and TextEmbeddingModule
      
      * refactor: Update dtype in embedding_manager.py to match proj.weight
      
      * Up
      
      * Add attribution and adaptation information to pipeline_anytext.py
      
      * Update usage example
      
      * Will refactor `controlnet_cond_embedding` initialization
      
      * Add `AnyTextControlNetConditioningEmbedding` template
      
      * Refactor organization
      
      * style
      
      * style
      
      * Move custom blocks from `AuxiliaryLatentModule` to `AnyTextControlNetConditioningEmbedding`
      
      * Follow one-file policy
      
      * style
      
      * [Docs] Update README and pipeline_anytext.py to use AnyTextControlNetModel
      
      * [Docs] Update import statement for AnyTextControlNetModel in pipeline_anytext.py
      
      * [Fix] Update import path for ControlNetModel, ControlNetOutput in anytext_controlnet.py
      
      * Refactor AnyTextControlNet to use configurable conditioning embedding channels
      
      * Complete control net conditioning embedding in AnyTextControlNetModel
      
      * up
      
      * [FIX] Ensure embeddings use correct device in AnyTextControlNetModel
      
      * up
      
      * up
      
      * style
      
      * [UPDATE] Revise README and example code for AnyTextPipeline integration with DiffusionPipeline
      
      * [UPDATE] Update example code in anytext.py to use correct font file and improve clarity
      
      * down
      
      * [UPDATE] Refactor BasicTokenizer usage to a new Checker class for text processing
      
      * update pillow
      
      * [UPDATE] Remove commented-out code and unnecessary docstring in anytext.py and anytext_controlnet.py for improved clarity
      
      * [REMOVE] Delete frozen_clip_embedder_t3.py as it is in the anytext.py file
      
      * [UPDATE] Replace edict with dict for configuration in anytext.py and RecModel.py for consistency
      
      * 🆙
      
      
      
      * style
      
      * [UPDATE] Revise README.md for clarity, remove unused imports in anytext.py, and add author credits in anytext_controlnet.py
      
      * style
      
      * Update examples/research_projects/anytext/README.md
      Co-authored-by: default avatarAryan <contact.aryanvs@gmail.com>
      
      * Remove commented-out image preparation code in AnyTextPipeline
      
      * Remove unnecessary blank line in README.md
      
      [Quantization] Allow loading TorchAO serialized Tensor objects with torch>=2.6  (#11018)
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      fix: mixture tiling sdxl pipeline - adjust gerating time_ids & embeddings  (#11012)
      
      small fix on generating time_ids & embeddings
      
      [LoRA] support wan i2v loras from the world. (#11025)
      
      * support wan i2v loras from the world.
      
      * remove copied from.
      
      * upates
      
      * add lora.
      
      Fix SD3 IPAdapter feature extractor (#11027)
      
      chore: fix help messages in advanced diffusion examples (#10923)
      
      Fix missing **kwargs in lora_pipeline.py (#11011)
      
      * Update lora_pipeline.py
      
      * Apply style fixes
      
      * fix-copies
      
      ---------
      Co-authored-by: default avatarhlky <hlky@hlky.ac>
      Co-authored-by: default avatargithub-actions[bot] <github-actions[bot]@users.noreply.github.com>
      
      Fix for multi-GPU WAN inference (#10997)
      
      Ensure that hidden_state and shift/scale are on the same device when running with multiple GPUs
      
      Co-authored-by: Jimmy <39@🇺🇸.com>
      
      [Refactor] Clean up import utils boilerplate (#11026)
      
      * update
      
      * update
      
      * update
      
      Use `output_size` in `repeat_interleave` (#11030)
      
      [hybrid inference 🍯🐝] Add VAE encode (#11017)
      
      * [hybrid inference 🍯🐝
      
      ] Add VAE encode
      
      * _toctree: add vae encode
      
      * Add endpoints, tests
      
      * vae_encode docs
      
      * vae encode benchmarks
      
      * api reference
      
      * changelog
      
      * Update docs/source/en/hybrid_inference/overview.md
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      * update
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      Wan Pipeline scaling fix, type hint warning, multi generator fix (#11007)
      
      * Wan Pipeline scaling fix, type hint warning, multi generator fix
      
      * Apply suggestions from code review
      
      [LoRA] change to warning from info when notifying the users about a LoRA no-op (#11044)
      
      * move to warning.
      
      * test related changes.
      
      Rename Lumina(2)Text2ImgPipeline -> Lumina(2)Pipeline (#10827)
      
      * Rename Lumina(2)Text2ImgPipeline -> Lumina(2)Pipeline
      
      ---------
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      
      making ```formatted_images``` initialization compact (#10801)
      
      compact writing
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      
      Fix aclnnRepeatInterleaveIntWithDim error on NPU for get_1d_rotary_pos_embed (#10820)
      
      * get_1d_rotary_pos_embed support npu
      
      * Update src/diffusers/models/embeddings.py
      
      ---------
      Co-authored-by: default avatarKai zheng <kaizheng@KaideMacBook-Pro.local>
      Co-authored-by: default avatarhlky <hlky@hlky.ac>
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      
      [Tests] restrict memory tests for quanto for certain schemes. (#11052)
      
      * restrict memory tests for quanto for certain schemes.
      
      * Apply suggestions from code review
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      * fixes
      
      * style
      
      ---------
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      [LoRA] feat: support non-diffusers wan t2v loras. (#11059)
      
      feat: support non-diffusers wan t2v loras.
      
      [examples/controlnet/train_controlnet_sd3.py] Fixes #11050 - Cast prompt_embeds and pooled_prompt_embeds to weight_dtype to prevent dtype mismatch (#11051)
      
      Fix: dtype mismatch of prompt embeddings in sd3 controlnet training
      Co-authored-by: default avatarAndreas Jörg <andreasjoerg@MacBook-Pro-von-Andreas-2.fritz.box>
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      reverts accidental change that removes attn_mask in attn. Improves fl… (#11065)
      
      reverts accidental change that removes attn_mask in attn. Improves flux ptxla by using flash block sizes. Moves encoding outside the for loop.
      Co-authored-by: default avatarJuan Acevedo <jfacevedo@google.com>
      
      Fix deterministic issue when getting pipeline dtype and device (#10696)
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      [Tests] add requires peft decorator. (#11037)
      
      * add requires peft decorator.
      
      * install peft conditionally.
      
      * conditional deps.
      Co-authored-by: default avatarDN6 <dhruv.nair@gmail.com>
      
      ---------
      Co-authored-by: default avatarDN6 <dhruv.nair@gmail.com>
      
      CogView4 Control Block (#10809)
      
      * cogview4 control training
      
      ---------
      Co-authored-by: default avatarOleehyO <leehy0357@gmail.com>
      Co-authored-by: default avataryiyixuxu <yixu310@gmail.com>
      
      [CI] pin transformers version for benchmarking. (#11067)
      
      pin transformers version for benchmarking.
      
      updates
      
      Fix Wan I2V Quality (#11087)
      
      * fix_wan_i2v_quality
      
      * Update src/diffusers/pipelines/wan/pipeline_wan_i2v.py
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      
      * Update src/diffusers/pipelines/wan/pipeline_wan_i2v.py
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      
      * Update src/diffusers/pipelines/wan/pipeline_wan_i2v.py
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      
      * Update pipeline_wan_i2v.py
      
      ---------
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      Co-authored-by: default avatarhlky <hlky@hlky.ac>
      
      LTX 0.9.5 (#10968)
      
      * update
      
      ---------
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      Co-authored-by: default avatarhlky <hlky@hlky.ac>
      
      make PR GPU tests conditioned on styling. (#11099)
      
      Group offloading improvements (#11094)
      
      update
      
      Fix pipeline_flux_controlnet.py (#11095)
      
      * Fix pipeline_flux_controlnet.py
      
      * Fix style
      
      update readme instructions. (#11096)
      Co-authored-by: default avatarJuan Acevedo <jfacevedo@google.com>
      
      Resolve stride mismatch in UNet's ResNet to support Torch DDP (#11098)
      
      Modify UNet's ResNet implementation to resolve stride mismatch in Torch's DDP
      
      Fix Group offloading behaviour when using streams (#11097)
      
      * update
      
      * update
      
      Quality options in `export_to_video` (#11090)
      
      * Quality options in `export_to_video`
      
      * make style
      
      improve more.
      
      add placeholders for docstrings.
      
      formatting.
      
      smol fix.
      
      solidify validation and annotation
      
      * Revert "feat: pipeline-level quant config."
      
      This reverts commit 316ff46b7648bfa24525ac02c284afcf440404aa.
      
      * feat: implement pipeline-level quantization config
      Co-authored-by: default avatarSunMarc <marc@huggingface.co>
      
      * update
      
      * fixes
      
      * fix validation.
      
      * add tests and other improvements.
      
      * add tests
      
      * import quality
      
      * remove prints.
      
      * add docs.
      
      * fixes to docs.
      
      * doc fixes.
      
      * doc fixes.
      
      * add validation to the input quantization_config.
      
      * clarify recommendations.
      
      * docs
      
      * add to ci.
      
      * todo.
      
      ---------
      Co-authored-by: default avatarSunMarc <marc@huggingface.co>
      599c8871
  17. 30 Apr, 2025 1 commit
  18. 18 Apr, 2025 1 commit
  19. 15 Apr, 2025 1 commit
  20. 02 Apr, 2025 2 commits
  21. 21 Mar, 2025 1 commit
  22. 19 Mar, 2025 1 commit
  23. 15 Mar, 2025 1 commit
  24. 10 Mar, 2025 1 commit
  25. 25 Feb, 2025 1 commit
  26. 24 Feb, 2025 1 commit
  27. 22 Feb, 2025 1 commit
    • Daniel Regado's avatar
      Comprehensive type checking for `from_pretrained` kwargs (#10758) · 9c7e2051
      Daniel Regado authored
      
      
      * More robust from_pretrained init_kwargs type checking
      
      * Corrected for Python 3.10
      
      * Type checks subclasses and fixed type warnings
      
      * More type corrections and skip tokenizer type checking
      
      * make style && make quality
      
      * Updated docs and types for Lumina pipelines
      
      * Fixed check for empty signature
      
      * changed location of helper functions
      
      * make style
      
      ---------
      Co-authored-by: default avatarhlky <hlky@hlky.ac>
      9c7e2051
  28. 20 Feb, 2025 1 commit
  29. 19 Feb, 2025 2 commits
  30. 14 Feb, 2025 1 commit
    • Aryan's avatar
      Module Group Offloading (#10503) · 9a147b82
      Aryan authored
      
      
      * update
      
      * fix
      
      * non_blocking; handle parameters and buffers
      
      * update
      
      * Group offloading with cuda stream prefetching (#10516)
      
      * cuda stream prefetch
      
      * remove breakpoints
      
      * update
      
      * copy model hook implementation from pab
      
      * update; ~very workaround based implementation but it seems to work as expected; needs cleanup and rewrite
      
      * more workarounds to make it actually work
      
      * cleanup
      
      * rewrite
      
      * update
      
      * make sure to sync current stream before overwriting with pinned params
      
      not doing so will lead to erroneous computations on the GPU and cause bad results
      
      * better check
      
      * update
      
      * remove hook implementation to not deal with merge conflict
      
      * re-add hook changes
      
      * why use more memory when less memory do trick
      
      * why still use slightly more memory when less memory do trick
      
      * optimise
      
      * add model tests
      
      * add pipeline tests
      
      * update docs
      
      * add layernorm and groupnorm
      
      * address review comments
      
      * improve tests; add docs
      
      * improve docs
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * apply suggestions from code review
      
      * update tests
      
      * apply suggestions from review
      
      * enable_group_offloading -> enable_group_offload for naming consistency
      
      * raise errors if multiple offloading strategies used; add relevant tests
      
      * handle .to() when group offload applied
      
      * refactor some repeated code
      
      * remove unintentional change from merge conflict
      
      * handle .cuda()
      
      ---------
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      9a147b82
  31. 05 Feb, 2025 1 commit
  32. 04 Feb, 2025 1 commit
  33. 27 Jan, 2025 1 commit
  34. 21 Jan, 2025 1 commit
  35. 14 Jan, 2025 1 commit
    • Marc Sun's avatar
      [FEAT] DDUF format (#10037) · fbff43ac
      Marc Sun authored
      
      
      * load and save dduf archive
      
      * style
      
      * switch to zip uncompressed
      
      * updates
      
      * Update src/diffusers/pipelines/pipeline_utils.py
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      * Update src/diffusers/pipelines/pipeline_utils.py
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      * first draft
      
      * remove print
      
      * switch to dduf_file for consistency
      
      * switch to huggingface hub api
      
      * fix log
      
      * add a basic test
      
      * Update src/diffusers/configuration_utils.py
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      * Update src/diffusers/pipelines/pipeline_utils.py
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      * Update src/diffusers/pipelines/pipeline_utils.py
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      
      * fix
      
      * fix variant
      
      * change saving logic
      
      * DDUF - Load transformers components manually (#10171)
      
      * update hfh version
      
      * Load transformers components manually
      
      * load encoder from_pretrained with state_dict
      
      * working version with transformers and tokenizer !
      
      * add generation_config case
      
      * fix tests
      
      * remove saving for now
      
      * typing
      
      * need next version from transformers
      
      * Update src/diffusers/configuration_utils.py
      Co-authored-by: default avatarLucain <lucain@huggingface.co>
      
      * check path corectly
      
      * Apply suggestions from code review
      Co-authored-by: default avatarLucain <lucain@huggingface.co>
      
      * udapte
      
      * typing
      
      * remove check for subfolder
      
      * quality
      
      * revert setup changes
      
      * oups
      
      * more readable condition
      
      * add loading from the hub test
      
      * add basic docs.
      
      * Apply suggestions from code review
      Co-authored-by: default avatarLucain <lucain@huggingface.co>
      
      * add example
      
      * add
      
      * make functions private
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * minor.
      
      * fixes
      
      * fix
      
      * change the precdence of parameterized.
      
      * error out when custom pipeline is passed with dduf_file.
      
      * updates
      
      * fix
      
      * updates
      
      * fixes
      
      * updates
      
      * fix xfail condition.
      
      * fix xfail
      
      * fixes
      
      * sharded checkpoint compat
      
      * add test for sharded checkpoint
      
      * add suggestions
      
      * Update src/diffusers/models/model_loading_utils.py
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      
      * from suggestions
      
      * add class attributes to flag dduf tests
      
      * last one
      
      * fix logic
      
      * remove comment
      
      * revert changes
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      Co-authored-by: default avatarLucain <lucain@huggingface.co>
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      fbff43ac
  36. 08 Jan, 2025 1 commit
  37. 06 Jan, 2025 1 commit
    • Ameer Azam's avatar
      Regarding the RunwayML path for V1.5 did change to... · 4f5e3e35
      Ameer Azam authored
      Regarding the RunwayML path for V1.5 did change to stable-diffusion-v1-5/[stable-diffusion-v1-5/ stable-diffusion-inpainting] (#10476)
      
      * Update pipeline_controlnet.py
      
      * Update pipeline_controlnet_img2img.py
      
      runwayml Take-down so change all from to this
      stable-diffusion-v1-5/stable-diffusion-v1-5
      
      * Update pipeline_controlnet_inpaint.py
      
      * runwayml take-down make change to sd-legacy
      
      * runwayml take-down make change to sd-legacy
      
      * runwayml take-down make change to sd-legacy
      
      * runwayml take-down make change to sd-legacy
      
      * Update convert_blipdiffusion_to_diffusers.py
      
      style change
      4f5e3e35