1. 08 Dec, 2025 2 commits
    • sayakpaul's avatar
      Release: v0.36.0-release · 9380e588
      sayakpaul authored
      9380e588
    • CalamitousFelicitousness's avatar
      Add ZImageImg2ImgPipeline (#12751) · 2246d2c7
      CalamitousFelicitousness authored
      
      
      * Add ZImageImg2ImgPipeline
      
      Updated the pipeline structure to include ZImageImg2ImgPipeline
          alongside ZImagePipeline.
      Implemented the ZImageImg2ImgPipeline class for image-to-image
          transformations, including necessary methods for
          encoding prompts, preparing latents, and denoising.
      Enhanced the auto_pipeline to map the new ZImageImg2ImgPipeline
          for image generation tasks.
      Added unit tests for ZImageImg2ImgPipeline to ensure
          functionality and performance.
      Updated dummy objects to include ZImageImg2ImgPipeline for
          testing purposes.
      
      * Address review comments for ZImageImg2ImgPipeline
      
      - Add `# Copied from` annotations to encode_prompt and _encode_prompt
      - Add ZImagePipeline to auto_pipeline.py for AutoPipeline support
      
      * Add ZImage pipeline documentation
      
      ---------
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      Co-authored-by: default avatarÁlvaro Somoza <asomoza@users.noreply.github.com>
      2246d2c7
  2. 06 Dec, 2025 1 commit
    • Tran Thanh Luan's avatar
      [Feat] TaylorSeer Cache (#12648) · 6290fdfd
      Tran Thanh Luan authored
      
      
      * init taylor_seer cache
      
      * make compatible with any tuple size returned
      
      * use logger for printing, add warmup feature
      
      * still update in warmup steps
      
      * refractor, add docs
      
      * add configurable cache, skip compute module
      
      * allow special cache ids only
      
      * add stop_predicts (cooldown)
      
      * update docs
      
      * apply ruff
      
      * update to handle multple calls per timestep
      
      * refractor to use state manager
      
      * fix format & doc
      
      * chores: naming, remove redundancy
      
      * add docs
      
      * quality & style
      
      * fix taylor precision
      
      * Apply style fixes
      
      * add tests
      
      * Apply style fixes
      
      * Remove TaylorSeerCacheTesterMixin from flux2 tests
      
      * rename identifiers, use more expressive taylor predict loop
      
      * torch compile compatible
      
      * Apply style fixes
      
      * Update src/diffusers/hooks/taylorseer_cache.py
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      * update docs
      
      * make fix-copies
      
      * fix example usage.
      
      * remove tests on flux kontext
      
      ---------
      Co-authored-by: default avatartoilaluan <toilaluan@github.com>
      Co-authored-by: default avatargithub-actions[bot] <github-actions[bot]@users.noreply.github.com>
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      6290fdfd
  3. 03 Dec, 2025 1 commit
  4. 02 Dec, 2025 1 commit
  5. 01 Dec, 2025 1 commit
  6. 29 Nov, 2025 1 commit
  7. 25 Nov, 2025 2 commits
    • Sayak Paul's avatar
      let's go Flux2 🚀 (#12711) · 5ffb73d4
      Sayak Paul authored
      
      
      * add vae
      
      * Initial commit for Flux 2 Transformer implementation
      
      * add pipeline part
      
      * small edits to the pipeline and conversion
      
      * update conversion script
      
      * fix
      
      * up up
      
      * finish pipeline
      
      * Remove Flux IP Adapter logic for now
      
      * Remove deprecated 3D id logic
      
      * Remove ControlNet logic for now
      
      * Add link to ViT-22B paper as reference for parallel transformer blocks such as the Flux 2 single stream block
      
      * update pipeline
      
      * Don't use biases for input projs and output AdaNorm
      
      * up
      
      * Remove bias for double stream block text QKV projections
      
      * Add script to convert Flux 2 transformer to diffusers
      
      * make style and make quality
      
      * fix a few things.
      
      * allow sft files to go.
      
      * fix image processor
      
      * fix batch
      
      * style a bit
      
      * Fix some bugs in Flux 2 transformer implementation
      
      * Fix dummy input preparation and fix some test bugs
      
      * fix dtype casting in timestep guidance module.
      
      * resolve conflicts.,
      
      * remove ip adapter stuff.
      
      * Fix Flux 2 transformer consistency test
      
      * Fix bug in Flux2TransformerBlock (double stream block)
      
      * Get remaining Flux 2 transformer tests passing
      
      * make style; make quality; make fix-copies
      
      * remove stuff.
      
      * fix type annotaton.
      
      * remove unneeded stuff from tests
      
      * tests
      
      * up
      
      * up
      
      * add sf support
      
      * Remove unused IP Adapter and ControlNet logic from transformer (#9)
      
      * copied from
      
      * Apply suggestions from code review
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      Co-authored-by: default avatarapolinário <joaopaulo.passos@gmail.com>
      
      * up
      
      * up
      
      * up
      
      * up
      
      * up
      
      * Refactor Flux2Attention into separate classes for double stream and single stream attention
      
      * Add _supports_qkv_fusion to AttentionModuleMixin to allow subclasses to disable QKV fusion
      
      * Have Flux2ParallelSelfAttention inherit from AttentionModuleMixin with _supports_qkv_fusion=False
      
      * Log debug message when calling fuse_projections on a AttentionModuleMixin subclass that does not support QKV fusion
      
      * Address review comments
      
      * Update src/diffusers/pipelines/flux2/pipeline_flux2.py
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      
      * up
      
      * Remove maybe_allow_in_graph decorators for Flux 2 transformer blocks (#12)
      
      * up
      
      * support ostris loras. (#13)
      
      * up
      
      * update schdule
      
      * up
      
      * up (#17)
      
      * add training scripts (#16)
      
      * add training scripts
      Co-authored-by: default avatarLinoy Tsaban <linoytsaban@gmail.com>
      
      * model cpu offload in validation.
      
      * add flux.2 readme
      
      * add img2img and tests
      
      * cpu offload in log validation
      
      * Apply suggestions from code review
      
      * fix
      
      * up
      
      * fixes
      
      * remove i2i training tests for now.
      
      ---------
      Co-authored-by: default avatarLinoy Tsaban <linoytsaban@gmail.com>
      Co-authored-by: default avatarlinoytsaban <linoy@huggingface.co>
      
      * up
      
      ---------
      Co-authored-by: default avataryiyixuxu <yixu310@gmail.com>
      Co-authored-by: default avatarDaniel Gu <dgu8957@gmail.com>
      Co-authored-by: default avataryiyi@huggingface.co <yiyi@ip-10-53-87-203.ec2.internal>
      Co-authored-by: default avatardg845 <58458699+dg845@users.noreply.github.com>
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      Co-authored-by: default avatarapolinário <joaopaulo.passos@gmail.com>
      Co-authored-by: default avataryiyi@huggingface.co <yiyi@ip-26-0-160-103.ec2.internal>
      Co-authored-by: default avatarLinoy Tsaban <linoytsaban@gmail.com>
      Co-authored-by: default avatarlinoytsaban <linoy@huggingface.co>
      5ffb73d4
    • Jerry Wu's avatar
      Add Support for Z-Image Series (#12703) · 4088e8a8
      Jerry Wu authored
      
      
      * Add Support for Z-Image.
      
      * Reformatting with make style, black & isort.
      
      * Remove init, Modify import utils, Merge forward in transformers block, Remove once func in pipeline.
      
      * modified main model forward, freqs_cis left
      
      * refactored to add B dim
      
      * fixed stack issue
      
      * fixed modulation bug
      
      * fixed modulation bug
      
      * fix bug
      
      * remove value_from_time_aware_config
      
      * styling
      
      * Fix neg embed and devide / bug; Reuse pad zero tensor; Turn cat -> repeat; Add hint for attn processor.
      
      * Replace padding with pad_sequence; Add gradient checkpointing.
      
      * Fix flash_attn3 in dispatch attn backend by _flash_attn_forward, replace its origin implement; Add DocString in pipeline for that.
      
      * Fix Docstring and Make Style.
      
      * Revert "Fix flash_attn3 in dispatch attn backend by _flash_attn_forward, replace its origin implement; Add DocString in pipeline for that."
      
      This reverts commit fbf26b7ed11d55146103c97740bad4a5f91744e0.
      
      * update z-image docstring
      
      * Revert attention dispatcher
      
      * update z-image docstring
      
      * styling
      
      * Recover attention_dispatch.py with its origin impl, later would special commit for fa3 compatibility.
      
      * Fix prev bug, and support for prompt_embeds pass in args after prompt pre-encode as List of torch Tensor.
      
      * Remove einop dependency.
      
      * remove redundant imports & make fix-copies
      
      * fix import
      
      ---------
      Co-authored-by: default avatarliudongyang <liudongyang0114@gmail.com>
      4088e8a8
  8. 17 Nov, 2025 1 commit
  9. 13 Nov, 2025 1 commit
  10. 10 Nov, 2025 2 commits
    • YiYi Xu's avatar
      [modular] wan! (#12611) · b455dc94
      YiYi Xu authored
      * update, remove intermediaate_inputs
      
      * support image2video
      
      * revert dynamic steps to simplify
      
      * refactor vae encoder block
      
      * support flf2video!
      
      * add support for wan2.2 14B
      
      * style
      
      * Apply suggestions from code review
      
      * input dynamic step -> additiional input step
      
      * up
      
      * fix init
      
      * update dtype
      b455dc94
    • Jay Wu's avatar
      add ChronoEdit (#12593) · 04f9d2bf
      Jay Wu authored
      
      
      * add ChronoEdit
      
      * add ref to  original function & remove wan2.2 logics
      
      * Update src/diffusers/pipelines/chronoedit/pipeline_chronoedit.py
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      
      * Update src/diffusers/pipelines/chronoedit/pipeline_chronoedit.py
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      
      * add ChronoeEdit test
      
      * add docs
      
      * add docs
      
      * make fix-copies
      
      * fix chronoedit test
      
      ---------
      Co-authored-by: default avatarwjay <wjay@nvidia.com>
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      04f9d2bf
  11. 06 Nov, 2025 1 commit
  12. 28 Oct, 2025 1 commit
    • galbria's avatar
      Bria fibo (#12545) · 84e16575
      galbria authored
      
      
      * Bria FIBO pipeline
      
      * style fixs
      
      * fix CR
      
      * Refactor BriaFibo classes and update pipeline parameters
      
      - Updated BriaFiboAttnProcessor and BriaFiboAttention classes to reflect changes from Flux equivalents.
      - Modified the _unpack_latents method in BriaFiboPipeline to improve clarity.
      - Increased the default max_sequence_length to 3000 and added a new optional parameter do_patching.
      - Cleaned up test_pipeline_bria_fibo.py by removing unused imports and skipping unsupported tests.
      
      * edit the docs of FIBO
      
      * Remove unused BriaFibo imports and update CPU offload method in BriaFiboPipeline
      
      * Refactor FIBO classes to BriaFibo naming convention
      
      - Updated class names from FIBO to BriaFibo for consistency across the module.
      - Modified instances of FIBOEmbedND, FIBOTimesteps, TextProjection, and TimestepProjEmbeddings to reflect the new naming.
      - Ensured all references in the BriaFiboTransformer2DModel are updated accordingly.
      
      * Add BriaFiboTransformer2DModel import to transformers module
      
      * Remove unused BriaFibo imports from modular pipelines and add BriaFiboTransformer2DModel and BriaFiboPipeline classes to dummy objects for enhanced compatibility with torch and transformers.
      
      * Update BriaFibo classes with copied documentation and fix import typo in pipeline module
      
      - Added documentation comments indicating the source of copied code in BriaFiboTransformerBlock and _pack_latents methods.
      - Corrected the import statement for BriaFiboPipeline in the pipelines module.
      
      * Remove unused BriaFibo imports from __init__.py to streamline modular pipelines.
      
      * Refactor documentation comments in BriaFibo classes to indicate inspiration from existing implementations
      
      - Updated comments in BriaFiboAttnProcessor, BriaFiboAttention, and BriaFiboPipeline to reflect that the code is inspired by other modules rather than copied.
      - Enhanced clarity on the origins of the methods to maintain proper attribution.
      
      * change Inspired by to Based on
      
      * add reference link and fix trailing whitespace
      
      * Add BriaFiboTransformer2DModel documentation and update comments in BriaFibo classes
      
      - Introduced a new documentation file for BriaFiboTransformer2DModel.
      - Updated comments in BriaFiboAttnProcessor, BriaFiboAttention, and BriaFiboPipeline to clarify the origins of the code, indicating copied sources for better attribution.
      
      ---------
      Co-authored-by: default avatarsayakpaul <spsayakpaul@gmail.com>
      84e16575
  13. 24 Oct, 2025 1 commit
  14. 22 Oct, 2025 1 commit
    • David Bertoin's avatar
      Prx (#12525) · dd07b19e
      David Bertoin authored
      * rename photon to prx
      
      * rename photon into prx
      
      * Revert .gitignore to state before commit b7fb0fe9d63bf766bbe3c42ac154a043796dd370
      
      * rename photon to prx
      
      * rename photon into prx
      
      * Revert .gitignore to state before commit b7fb0fe9d63bf766bbe3c42ac154a043796dd370
      
      * make fix-copies
      dd07b19e
  15. 21 Oct, 2025 1 commit
  16. 18 Oct, 2025 1 commit
  17. 10 Oct, 2025 1 commit
  18. 05 Oct, 2025 1 commit
  19. 24 Sep, 2025 1 commit
    • Aryan's avatar
      Context Parallel w/ Ring & Ulysses & Unified Attention (#11941) · dcb6dd9b
      Aryan authored
      
      
      * update
      
      * update
      
      * add coauthor
      Co-Authored-By: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      * improve test
      
      * handle ip adapter params correctly
      
      * fix chroma qkv fusion test
      
      * fix fastercache implementation
      
      * fix more tests
      
      * fight more tests
      
      * add back set_attention_backend
      
      * update
      
      * update
      
      * make style
      
      * make fix-copies
      
      * make ip adapter processor compatible with attention dispatcher
      
      * refactor chroma as well
      
      * remove rmsnorm assert
      
      * minify and deprecate npu/xla processors
      
      * update
      
      * refactor
      
      * refactor; support flash attention 2 with cp
      
      * fix
      
      * support sage attention with cp
      
      * make torch compile compatible
      
      * update
      
      * refactor
      
      * update
      
      * refactor
      
      * refactor
      
      * add ulysses backward
      
      * try to make dreambooth script work; accelerator backward not playing well
      
      * Revert "try to make dreambooth script work; accelerator backward not playing well"
      
      This reverts commit 768d0ea6fa6a305d12df1feda2afae3ec80aa449.
      
      * workaround compilation problems with triton when doing all-to-all
      
      * support wan
      
      * handle backward correctly
      
      * support qwen
      
      * support ltx
      
      * make fix-copies
      
      * Update src/diffusers/models/modeling_utils.py
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      * apply review suggestions
      
      * update docs
      
      * add explanation
      
      * make fix-copies
      
      * add docstrings
      
      * support passing parallel_config to from_pretrained
      
      * apply review suggestions
      
      * make style
      
      * update
      
      * Update docs/source/en/api/parallel.md
      Co-authored-by: default avatarAryan <aryan@huggingface.co>
      
      * up
      
      ---------
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      Co-authored-by: default avatarsayakpaul <spsayakpaul@gmail.com>
      dcb6dd9b
  20. 21 Sep, 2025 1 commit
  21. 16 Sep, 2025 1 commit
  22. 09 Sep, 2025 1 commit
  23. 08 Sep, 2025 1 commit
  24. 03 Sep, 2025 1 commit
    • Ishan Modi's avatar
      [Quantization] Add TRT-ModelOpt as a Backend (#11173) · 4acbfbf1
      Ishan Modi authored
      
      
      * initial commit
      
      * update
      
      * updates
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * addressed PR comments
      
      * update
      
      * addressed PR comments
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * updates
      
      * update
      
      * update
      
      * addressed PR comments
      
      * updates
      
      * code formatting
      
      * update
      
      * addressed PR comments
      
      * addressed PR comments
      
      * addressed PR comments
      
      * addressed PR comments
      
      * fix docs and dependencies
      
      * fixed dependency test
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      4acbfbf1
  25. 31 Aug, 2025 1 commit
  26. 22 Aug, 2025 1 commit
  27. 20 Aug, 2025 1 commit
    • galbria's avatar
      Bria 3 2 pipeline (#12010) · 7993be9e
      galbria authored
      
      
      * Add Bria model and pipeline to diffusers
      
      - Introduced `BriaTransformer2DModel` and `BriaPipeline` for enhanced image generation capabilities.
      - Updated import structures across various modules to include the new Bria components.
      - Added utility functions and output classes specific to the Bria pipeline.
      - Implemented tests for the Bria pipeline to ensure functionality and output integrity.
      
      * with working tests
      
      * style and quality pass
      
      * adding docs
      
      * add to overview
      
      * fixes from "make fix-copies"
      
      * Refactor transformer_bria.py and pipeline_bria.py: Introduce new EmbedND class for rotary position embedding, and enhance Timestep and TimestepProjEmbeddings classes. Add utility functions for handling negative prompts and generating original sigmas in pipeline_bria.py.
      
      * remove redundent and duplicates tests and fix bf16
      slow test
      
      * style fixes
      
      * small doc update
      
      * Enhance Bria 3.2 documentation and implementation
      
      - Updated the GitHub repository link for Bria 3.2.
      - Added usage instructions for the gated model access.
      - Introduced the BriaTransformerBlock and BriaAttention classes to the model architecture.
      - Refactored existing classes to integrate Bria-specific components, including BriaEmbedND and BriaPipeline.
      - Updated the pipeline output class to reflect Bria-specific functionality.
      - Adjusted test cases to align with the new Bria model structure.
      
      * Refactor Bria model components and update documentation
      
      - Removed outdated inference example from Bria 3.2 documentation.
      - Introduced the BriaTransformerBlock class to enhance model architecture.
      - Updated attention handling to use `attention_kwargs` instead of `joint_attention_kwargs`.
      - Improved import structure in the Bria pipeline to handle optional dependencies.
      - Adjusted test cases to reflect changes in model dtype assertions.
      
      * Update Bria model reference in documentation to reflect new file naming convention
      
      * Update docs/source/en/_toctree.yml
      
      * Refactor BriaPipeline to inherit from DiffusionPipeline instead of FluxPipeline, updating imports accordingly.
      
      * move the __call__ func to the end of file
      
      * Update BriaPipeline example to use bfloat16 for precision sensitivity for better result
      
      * make style && make quality &&  make fix-copiessource
      
      ---------
      Co-authored-by: default avatarLinoy Tsaban <57615435+linoytsaban@users.noreply.github.com>
      Co-authored-by: default avatarAryan <contact.aryanvs@gmail.com>
      7993be9e
  28. 19 Aug, 2025 1 commit
  29. 18 Aug, 2025 1 commit
  30. 17 Aug, 2025 1 commit
    • naykun's avatar
      Qwen Image Edit Support (#12164) · e682af20
      naykun authored
      * feat(qwen-image):
      add qwen-image-edit support
      
      * fix(qwen image):
      - compatible with torch.compile in new rope setting
      - fix init import
      - add prompt truncation in img2img and inpaint pipe
      - remove unused logic and comment
      - add copy statement
      - guard logic for rope video shape tuple
      
      * fix(qwen image):
      - make fix-copies
      - update doc
      e682af20
  31. 13 Aug, 2025 1 commit
  32. 12 Aug, 2025 1 commit
  33. 07 Aug, 2025 1 commit
    • dg845's avatar
      Implement Frequency-Decoupled Guidance (FDG) as a Guider (#11976) · d45199a2
      dg845 authored
      
      
      * Initial commit implementing frequency-decoupled guidance (FDG) as a guider
      
      * Update FrequencyDecoupledGuidance docstring to describe FDG
      
      * Update project so that it accepts any number of non-batch dims
      
      * Change guidance_scale and other params to accept a list of params for each freq level
      
      * Add comment with Laplacian pyramid shapes
      
      * Add function to import_utils to check if the kornia package is available
      
      * Only import from kornia if package is available
      
      * Fix bug: use pred_cond/uncond in freq space rather than data space
      
      * Allow guidance rescaling to be done in data space or frequency space (speculative)
      
      * Add kornia install instructions to kornia import error message
      
      * Add config to control whether operations are upcast to fp64
      
      * Add parallel_weights recommended values to docstring
      
      * Apply style fixes
      
      * make fix-copies
      
      ---------
      Co-authored-by: default avatargithub-actions[bot] <github-actions[bot]@users.noreply.github.com>
      Co-authored-by: default avatarAryan <aryan@huggingface.co>
      d45199a2
  34. 03 Aug, 2025 1 commit
    • naykun's avatar
      Qwen-Image (#12055) · 8e53cd95
      naykun authored
      
      
      * (feat): qwen-image integration
      
      * fix(qwen-image):
      - remove unused logics related to controlnet/ip-adapter
      
      * fix(qwen-image):
      - compatible with attention dispatcher
      - cond cache support
      
      * fix(qwen-image):
      - cond cache registry
      - attention backend argument
      - fix copies
      
      * fix(qwen-image):
      - remove local test
      
      * Update src/diffusers/models/transformers/transformer_qwenimage.py
      
      ---------
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      8e53cd95
  35. 29 Jul, 2025 1 commit
  36. 23 Jul, 2025 1 commit
  37. 17 Jul, 2025 1 commit