1. 03 Dec, 2025 1 commit
  2. 01 Dec, 2025 1 commit
  3. 25 Nov, 2025 1 commit
    • Sayak Paul's avatar
      let's go Flux2 🚀 (#12711) · 5ffb73d4
      Sayak Paul authored
      
      
      * add vae
      
      * Initial commit for Flux 2 Transformer implementation
      
      * add pipeline part
      
      * small edits to the pipeline and conversion
      
      * update conversion script
      
      * fix
      
      * up up
      
      * finish pipeline
      
      * Remove Flux IP Adapter logic for now
      
      * Remove deprecated 3D id logic
      
      * Remove ControlNet logic for now
      
      * Add link to ViT-22B paper as reference for parallel transformer blocks such as the Flux 2 single stream block
      
      * update pipeline
      
      * Don't use biases for input projs and output AdaNorm
      
      * up
      
      * Remove bias for double stream block text QKV projections
      
      * Add script to convert Flux 2 transformer to diffusers
      
      * make style and make quality
      
      * fix a few things.
      
      * allow sft files to go.
      
      * fix image processor
      
      * fix batch
      
      * style a bit
      
      * Fix some bugs in Flux 2 transformer implementation
      
      * Fix dummy input preparation and fix some test bugs
      
      * fix dtype casting in timestep guidance module.
      
      * resolve conflicts.,
      
      * remove ip adapter stuff.
      
      * Fix Flux 2 transformer consistency test
      
      * Fix bug in Flux2TransformerBlock (double stream block)
      
      * Get remaining Flux 2 transformer tests passing
      
      * make style; make quality; make fix-copies
      
      * remove stuff.
      
      * fix type annotaton.
      
      * remove unneeded stuff from tests
      
      * tests
      
      * up
      
      * up
      
      * add sf support
      
      * Remove unused IP Adapter and ControlNet logic from transformer (#9)
      
      * copied from
      
      * Apply suggestions from code review
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      Co-authored-by: default avatarapolinário <joaopaulo.passos@gmail.com>
      
      * up
      
      * up
      
      * up
      
      * up
      
      * up
      
      * Refactor Flux2Attention into separate classes for double stream and single stream attention
      
      * Add _supports_qkv_fusion to AttentionModuleMixin to allow subclasses to disable QKV fusion
      
      * Have Flux2ParallelSelfAttention inherit from AttentionModuleMixin with _supports_qkv_fusion=False
      
      * Log debug message when calling fuse_projections on a AttentionModuleMixin subclass that does not support QKV fusion
      
      * Address review comments
      
      * Update src/diffusers/pipelines/flux2/pipeline_flux2.py
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      
      * up
      
      * Remove maybe_allow_in_graph decorators for Flux 2 transformer blocks (#12)
      
      * up
      
      * support ostris loras. (#13)
      
      * up
      
      * update schdule
      
      * up
      
      * up (#17)
      
      * add training scripts (#16)
      
      * add training scripts
      Co-authored-by: default avatarLinoy Tsaban <linoytsaban@gmail.com>
      
      * model cpu offload in validation.
      
      * add flux.2 readme
      
      * add img2img and tests
      
      * cpu offload in log validation
      
      * Apply suggestions from code review
      
      * fix
      
      * up
      
      * fixes
      
      * remove i2i training tests for now.
      
      ---------
      Co-authored-by: default avatarLinoy Tsaban <linoytsaban@gmail.com>
      Co-authored-by: default avatarlinoytsaban <linoy@huggingface.co>
      
      * up
      
      ---------
      Co-authored-by: default avataryiyixuxu <yixu310@gmail.com>
      Co-authored-by: default avatarDaniel Gu <dgu8957@gmail.com>
      Co-authored-by: default avataryiyi@huggingface.co <yiyi@ip-10-53-87-203.ec2.internal>
      Co-authored-by: default avatardg845 <58458699+dg845@users.noreply.github.com>
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      Co-authored-by: default avatarapolinário <joaopaulo.passos@gmail.com>
      Co-authored-by: default avataryiyi@huggingface.co <yiyi@ip-26-0-160-103.ec2.internal>
      Co-authored-by: default avatarLinoy Tsaban <linoytsaban@gmail.com>
      Co-authored-by: default avatarlinoytsaban <linoy@huggingface.co>
      5ffb73d4
  4. 17 Nov, 2025 1 commit
  5. 12 Nov, 2025 1 commit
    • Quentin Gallouédec's avatar
      ArXiv -> HF Papers (#12583) · f3db38c1
      Quentin Gallouédec authored
      * Update pipeline_skyreels_v2_i2v.py
      
      * Update README.md
      
      * Update torch_utils.py
      
      * Update torch_utils.py
      
      * Update guider_utils.py
      
      * Update pipeline_ltx.py
      
      * Update pipeline_bria.py
      
      * Apply suggestion from @qgallouedec
      
      * Update autoencoder_kl_qwenimage.py
      
      * Update pipeline_prx.py
      
      * Update pipeline_wan_vace.py
      
      * Update pipeline_skyreels_v2.py
      
      * Update pipeline_skyreels_v2_diffusion_forcing.py
      
      * Update pipeline_bria_fibo.py
      
      * Update pipeline_skyreels_v2_diffusion_forcing_i2v.py
      
      * Update pipeline_ltx_condition.py
      
      * Update pipeline_ltx_image2video.py
      
      * Update regional_prompting_stable_diffusion.py
      
      * make style
      
      * style
      
      * style
      f3db38c1
  6. 10 Nov, 2025 1 commit
  7. 30 Oct, 2025 1 commit
  8. 28 Oct, 2025 1 commit
  9. 24 Oct, 2025 1 commit
  10. 22 Oct, 2025 2 commits
  11. 15 Oct, 2025 1 commit
  12. 30 Sep, 2025 1 commit
  13. 22 Sep, 2025 1 commit
  14. 16 Sep, 2025 1 commit
    • Zijian Zhou's avatar
      Fix autoencoder_kl_wan.py bugs for Wan2.2 VAE (#12335) · d06750a5
      Zijian Zhou authored
      * Update autoencoder_kl_wan.py
      
      When using the Wan2.2 VAE, the spatial compression ratio calculated here is incorrect. It should be 16 instead of 8. Pass it in directly via the config to ensure it’s correct here.
      
      * Update autoencoder_kl_wan.py
      d06750a5
  15. 18 Aug, 2025 1 commit
  16. 04 Aug, 2025 2 commits
  17. 03 Aug, 2025 1 commit
    • naykun's avatar
      Qwen-Image (#12055) · 8e53cd95
      naykun authored
      
      
      * (feat): qwen-image integration
      
      * fix(qwen-image):
      - remove unused logics related to controlnet/ip-adapter
      
      * fix(qwen-image):
      - compatible with attention dispatcher
      - cond cache support
      
      * fix(qwen-image):
      - cond cache registry
      - attention backend argument
      - fix copies
      
      * fix(qwen-image):
      - remove local test
      
      * Update src/diffusers/models/transformers/transformer_qwenimage.py
      
      ---------
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      8e53cd95
  18. 02 Aug, 2025 2 commits
  19. 01 Aug, 2025 1 commit
  20. 28 Jul, 2025 1 commit
  21. 24 Jun, 2025 1 commit
  22. 19 Jun, 2025 1 commit
  23. 18 Jun, 2025 1 commit
  24. 30 May, 2025 1 commit
  25. 19 May, 2025 1 commit
  26. 07 May, 2025 1 commit
    • Aryan's avatar
      Cosmos (#10660) · 7b904941
      Aryan authored
      
      
      * begin transformer conversion
      
      * refactor
      
      * refactor
      
      * refactor
      
      * refactor
      
      * refactor
      
      * refactor
      
      * update
      
      * add conversion script
      
      * add pipeline
      
      * make fix-copies
      
      * remove einops
      
      * update docs
      
      * gradient checkpointing
      
      * add transformer test
      
      * update
      
      * debug
      
      * remove prints
      
      * match sigmas
      
      * add vae pt. 1
      
      * finish CV* vae
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * make fix-copies
      
      * update
      
      * make fix-copies
      
      * fix
      
      * update
      
      * update
      
      * make fix-copies
      
      * update
      
      * update tests
      
      * handle device and dtype for safety checker; required in latest diffusers
      
      * remove enable_gqa and use repeat_interleave instead
      
      * enforce safety checker; use dummy checker in fast tests
      
      * add review suggestion for ONNX export
      Co-Authored-By: default avatarAsfiya Baig <asfiyab@nvidia.com>
      
      * fix safety_checker issues when not passed explicitly
      
      We could either do what's done in this commit, or update the Cosmos examples to explicitly pass the safety checker
      
      * use cosmos guardrail package
      
      * auto format docs
      
      * update conversion script to support 14B models
      
      * update name CosmosPipeline -> CosmosTextToWorldPipeline
      
      * update docs
      
      * fix docs
      
      * fix group offload test failing for vae
      
      ---------
      Co-authored-by: default avatarAsfiya Baig <asfiyab@nvidia.com>
      7b904941
  27. 05 May, 2025 1 commit
  28. 15 Apr, 2025 1 commit
  29. 13 Apr, 2025 1 commit
  30. 11 Apr, 2025 1 commit
  31. 05 Apr, 2025 1 commit
  32. 02 Apr, 2025 1 commit
    • Bruno Magalhaes's avatar
      remove unnecessary call to `F.pad` (#10620) · fe2b3974
      Bruno Magalhaes authored
      * rewrite memory count without implicitly using dimensions by @ic-synth
      
      * replace F.pad by built-in padding in Conv3D
      
      * in-place sums to reduce memory allocations
      
      * fixed trailing whitespace
      
      * file reformatted
      
      * in-place sums
      
      * simpler in-place expressions
      
      * removed in-place sum, may affect backward propagation logic
      
      * removed in-place sum, may affect backward propagation logic
      
      * removed in-place sum, may affect backward propagation logic
      
      * reverted change
      fe2b3974
  33. 18 Mar, 2025 1 commit
  34. 12 Mar, 2025 1 commit
  35. 07 Mar, 2025 2 commits
  36. 03 Mar, 2025 1 commit