You need to sign in or sign up before continuing.
  1. 04 Dec, 2025 4 commits
  2. 03 Dec, 2025 8 commits
  3. 02 Dec, 2025 3 commits
    • Jerry Wu's avatar
      Fix TPU (torch_xla) compatibility Error about tensor repeat func along with empty dim. (#12770) · 9379b239
      Jerry Wu authored
      
      
      * Refactor image padding logic to pervent zero tensor in transformer_z_image.py
      
      * Apply style fixes
      
      * Add more support to fix repeat bug on tpu devices.
      
      * Fix for dynamo compile error for multi if-branches.
      
      ---------
      Co-authored-by: default avatarMingjia Li <mingjiali@tju.edu.cn>
      Co-authored-by: default avatarMingjia Li <mail@mingjia.li>
      Co-authored-by: default avatargithub-actions[bot] <github-actions[bot]@users.noreply.github.com>
      9379b239
    • Guo-Hua Wang's avatar
      Add support for Ovis-Image (#12740) · 4f136f84
      Guo-Hua Wang authored
      
      
      * add ovis_image
      
      * fix code quality
      
      * optimize pipeline_ovis_image.py according to the feedbacks
      
      * optimize imports
      
      * add docs
      
      * make style
      
      * make style
      
      * add ovis to toctree
      
      * oops
      
      ---------
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      4f136f84
    • CalamitousFelicitousness's avatar
      Add ZImage LoRA support and integrate into ZImagePipeline (#12750) · edf36f51
      CalamitousFelicitousness authored
      
      
      * Add ZImage LoRA support and integrate into ZImagePipeline
      
      * Add LoRA test for Z-Image
      
      * Move the LoRA test
      
      * Fix ZImage LoRA scale support and test configuration
      
      * Add ZImage LoRA test overrides for architecture differences
      
      - Override test_lora_fuse_nan to use ZImage's 'layers' attribute
        instead of 'transformer_blocks'
      - Skip block-level LoRA scaling test (not supported in ZImage)
      - Add required imports: numpy, torch_device, check_if_lora_correctly_set
      
      * Add ZImageLoraLoaderMixin to LoRA documentation
      
      * Use conditional import for peft.LoraConfig in ZImage tests
      
      * Override test_correct_lora_configs_with_different_ranks for ZImage
      
      ZImage uses 'attention.to_k' naming convention instead of 'attn.to_k',
      so the base test's module name search loop never finds a match. This
      override uses the correct naming pattern for ZImage architecture.
      
      * Add is_flaky decorator to ZImage LoRA tests initialise padding tokens
      
      * Skip ZImage LoRA test class entirely
      
      Skip the entire ZImageLoRATests class due to non-deterministic behavior
      from complex64 RoPE operations and torch.empty padding tokens.
      LoRA functionality works correctly with real models.
      
      Clean up removed:
      - Individual @unittest.skip decorators
      - @is_flaky decorator overrides for inherited methods
      - Custom test method overrides
      - Global torch deterministic settings
      - Unused imports (numpy, is_flaky, check_if_lora_correctly_set)
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      Co-authored-by: default avatarÁlvaro Somoza <asomoza@users.noreply.github.com>
      edf36f51
  4. 01 Dec, 2025 8 commits
  5. 29 Nov, 2025 1 commit
  6. 28 Nov, 2025 2 commits
  7. 27 Nov, 2025 2 commits
  8. 26 Nov, 2025 6 commits
    • Jerry Wu's avatar
      Support unittest for Z-image ️ (#12715) · e6d46123
      Jerry Wu authored
      
      
      * Add Support for Z-Image.
      
      * Reformatting with make style, black & isort.
      
      * Remove init, Modify import utils, Merge forward in transformers block, Remove once func in pipeline.
      
      * modified main model forward, freqs_cis left
      
      * refactored to add B dim
      
      * fixed stack issue
      
      * fixed modulation bug
      
      * fixed modulation bug
      
      * fix bug
      
      * remove value_from_time_aware_config
      
      * styling
      
      * Fix neg embed and devide / bug; Reuse pad zero tensor; Turn cat -> repeat; Add hint for attn processor.
      
      * Replace padding with pad_sequence; Add gradient checkpointing.
      
      * Fix flash_attn3 in dispatch attn backend by _flash_attn_forward, replace its origin implement; Add DocString in pipeline for that.
      
      * Fix Docstring and Make Style.
      
      * Revert "Fix flash_attn3 in dispatch attn backend by _flash_attn_forward, replace its origin implement; Add DocString in pipeline for that."
      
      This reverts commit fbf26b7ed11d55146103c97740bad4a5f91744e0.
      
      * update z-image docstring
      
      * Revert attention dispatcher
      
      * update z-image docstring
      
      * styling
      
      * Recover attention_dispatch.py with its origin impl, later would special commit for fa3 compatibility.
      
      * Fix prev bug, and support for prompt_embeds pass in args after prompt pre-encode as List of torch Tensor.
      
      * Remove einop dependency.
      
      * remove redundant imports & make fix-copies
      
      * fix import
      
      * Support for num_images_per_prompt>1; Remove redundant unquote variables.
      
      * Fix bugs for num_images_per_prompt with actual batch.
      
      * Add unit tests for Z-Image.
      
      * Refine unitest and skip for cases needed separate test env; Fix compatibility with unitest in model, mostly precision formating.
      
      * Add clean env for test_save_load_float16 separ test; Add Note; Styling.
      
      * Update dtype mentioned by yiyi.
      
      ---------
      Co-authored-by: default avatarliudongyang <liudongyang0114@gmail.com>
      e6d46123
    • David El Malih's avatar
      Improve docstrings and type hints in scheduling_dpmsolver_multistep.py (#12710) · a88a7b4f
      David El Malih authored
      * Improve docstrings and type hints in multiple diffusion schedulers
      
      * docs: update Imagen Video paper link to Hugging Face Papers.
      a88a7b4f
    • Sayak Paul's avatar
      [docs] put autopipeline after overview and hunyuanimage in images (#12548) · c8656ed7
      Sayak Paul authored
      put autopipeline after overview and hunyuanimage in images
      c8656ed7
    • Sayak Paul's avatar
      [docs] Correct flux2 links (#12716) · 94c9613f
      Sayak Paul authored
      * fix links
      
      * up
      94c9613f
    • Sayak Paul's avatar
      [lora]: Fix Flux2 LoRA NaN test (#12714) · b91e8c0d
      Sayak Paul authored
      
      
      * up
      
      * Update tests/lora/test_lora_layers_flux2.py
      Co-authored-by: default avatardg845 <58458699+dg845@users.noreply.github.com>
      
      ---------
      Co-authored-by: default avatardg845 <58458699+dg845@users.noreply.github.com>
      b91e8c0d
    • Andrei Filatov's avatar
  9. 25 Nov, 2025 3 commits
    • Sayak Paul's avatar
      let's go Flux2 🚀 (#12711) · 5ffb73d4
      Sayak Paul authored
      
      
      * add vae
      
      * Initial commit for Flux 2 Transformer implementation
      
      * add pipeline part
      
      * small edits to the pipeline and conversion
      
      * update conversion script
      
      * fix
      
      * up up
      
      * finish pipeline
      
      * Remove Flux IP Adapter logic for now
      
      * Remove deprecated 3D id logic
      
      * Remove ControlNet logic for now
      
      * Add link to ViT-22B paper as reference for parallel transformer blocks such as the Flux 2 single stream block
      
      * update pipeline
      
      * Don't use biases for input projs and output AdaNorm
      
      * up
      
      * Remove bias for double stream block text QKV projections
      
      * Add script to convert Flux 2 transformer to diffusers
      
      * make style and make quality
      
      * fix a few things.
      
      * allow sft files to go.
      
      * fix image processor
      
      * fix batch
      
      * style a bit
      
      * Fix some bugs in Flux 2 transformer implementation
      
      * Fix dummy input preparation and fix some test bugs
      
      * fix dtype casting in timestep guidance module.
      
      * resolve conflicts.,
      
      * remove ip adapter stuff.
      
      * Fix Flux 2 transformer consistency test
      
      * Fix bug in Flux2TransformerBlock (double stream block)
      
      * Get remaining Flux 2 transformer tests passing
      
      * make style; make quality; make fix-copies
      
      * remove stuff.
      
      * fix type annotaton.
      
      * remove unneeded stuff from tests
      
      * tests
      
      * up
      
      * up
      
      * add sf support
      
      * Remove unused IP Adapter and ControlNet logic from transformer (#9)
      
      * copied from
      
      * Apply suggestions from code review
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      Co-authored-by: default avatarapolinário <joaopaulo.passos@gmail.com>
      
      * up
      
      * up
      
      * up
      
      * up
      
      * up
      
      * Refactor Flux2Attention into separate classes for double stream and single stream attention
      
      * Add _supports_qkv_fusion to AttentionModuleMixin to allow subclasses to disable QKV fusion
      
      * Have Flux2ParallelSelfAttention inherit from AttentionModuleMixin with _supports_qkv_fusion=False
      
      * Log debug message when calling fuse_projections on a AttentionModuleMixin subclass that does not support QKV fusion
      
      * Address review comments
      
      * Update src/diffusers/pipelines/flux2/pipeline_flux2.py
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      
      * up
      
      * Remove maybe_allow_in_graph decorators for Flux 2 transformer blocks (#12)
      
      * up
      
      * support ostris loras. (#13)
      
      * up
      
      * update schdule
      
      * up
      
      * up (#17)
      
      * add training scripts (#16)
      
      * add training scripts
      Co-authored-by: default avatarLinoy Tsaban <linoytsaban@gmail.com>
      
      * model cpu offload in validation.
      
      * add flux.2 readme
      
      * add img2img and tests
      
      * cpu offload in log validation
      
      * Apply suggestions from code review
      
      * fix
      
      * up
      
      * fixes
      
      * remove i2i training tests for now.
      
      ---------
      Co-authored-by: default avatarLinoy Tsaban <linoytsaban@gmail.com>
      Co-authored-by: default avatarlinoytsaban <linoy@huggingface.co>
      
      * up
      
      ---------
      Co-authored-by: default avataryiyixuxu <yixu310@gmail.com>
      Co-authored-by: default avatarDaniel Gu <dgu8957@gmail.com>
      Co-authored-by: default avataryiyi@huggingface.co <yiyi@ip-10-53-87-203.ec2.internal>
      Co-authored-by: default avatardg845 <58458699+dg845@users.noreply.github.com>
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      Co-authored-by: default avatarapolinário <joaopaulo.passos@gmail.com>
      Co-authored-by: default avataryiyi@huggingface.co <yiyi@ip-26-0-160-103.ec2.internal>
      Co-authored-by: default avatarLinoy Tsaban <linoytsaban@gmail.com>
      Co-authored-by: default avatarlinoytsaban <linoy@huggingface.co>
      5ffb73d4
    • Jerry Wu's avatar
      Add Support for Z-Image Series (#12703) · 4088e8a8
      Jerry Wu authored
      
      
      * Add Support for Z-Image.
      
      * Reformatting with make style, black & isort.
      
      * Remove init, Modify import utils, Merge forward in transformers block, Remove once func in pipeline.
      
      * modified main model forward, freqs_cis left
      
      * refactored to add B dim
      
      * fixed stack issue
      
      * fixed modulation bug
      
      * fixed modulation bug
      
      * fix bug
      
      * remove value_from_time_aware_config
      
      * styling
      
      * Fix neg embed and devide / bug; Reuse pad zero tensor; Turn cat -> repeat; Add hint for attn processor.
      
      * Replace padding with pad_sequence; Add gradient checkpointing.
      
      * Fix flash_attn3 in dispatch attn backend by _flash_attn_forward, replace its origin implement; Add DocString in pipeline for that.
      
      * Fix Docstring and Make Style.
      
      * Revert "Fix flash_attn3 in dispatch attn backend by _flash_attn_forward, replace its origin implement; Add DocString in pipeline for that."
      
      This reverts commit fbf26b7ed11d55146103c97740bad4a5f91744e0.
      
      * update z-image docstring
      
      * Revert attention dispatcher
      
      * update z-image docstring
      
      * styling
      
      * Recover attention_dispatch.py with its origin impl, later would special commit for fa3 compatibility.
      
      * Fix prev bug, and support for prompt_embeds pass in args after prompt pre-encode as List of torch Tensor.
      
      * Remove einop dependency.
      
      * remove redundant imports & make fix-copies
      
      * fix import
      
      ---------
      Co-authored-by: default avatarliudongyang <liudongyang0114@gmail.com>
      4088e8a8
    • Junsong Chen's avatar
      fix typo in docs (#12675) · d33d9f67
      Junsong Chen authored
      
      
      * fix typo in docs
      
      * Update docs/source/en/api/pipelines/sana_video.md
      Co-authored-by: default avatardg845 <58458699+dg845@users.noreply.github.com>
      
      ---------
      Co-authored-by: default avatardg845 <58458699+dg845@users.noreply.github.com>
      d33d9f67
  10. 24 Nov, 2025 3 commits
    • sq's avatar
      Fix variable naming typos in community FluxControlNetFillInpaintPipeline (#12701) · dde8754b
      sq authored
      - Fixed variable naming typos (maskkk -> mask_fill, mask_imagee -> mask_image_fill, masked_imagee -> masked_image_fill, masked_image_latentsss -> masked_latents_fill)
      
      These changes improve code readability without affecting functionality.
      dde8754b
    • cdutr's avatar
      [i8n-pt] Fix grammar and expand Portuguese documentation (#12598) · fbcd3ba6
      cdutr authored
      * Updates Portuguese documentation for Diffusers library
      
      Enhances the Portuguese documentation with:
      - Restructured table of contents for improved navigation
      - Added placeholder page for in-translation content
      - Refined language and improved readability in existing pages
      - Introduced a new page on basic Stable Diffusion performance guidance
      
      Improves overall documentation structure and user experience for Portuguese-speaking users
      
      * Removes untranslated sections from Portuguese documentation
      
      Cleans up the Portuguese documentation table of contents by removing placeholder sections marked as "Em tradução" (In translation)
      
      Removes the in_translation.md file and associated table of contents entries for sections that are not yet translated, improving documentation clarity
      fbcd3ba6
    • Sayak Paul's avatar
      [core] support sage attention + FA2 through `kernels` (#12439) · d176f61f
      Sayak Paul authored
      * up
      
      * support automatic dispatch.
      
      * disable compile support for now./
      
      * up
      
      * flash too.
      
      * document.
      
      * up
      
      * up
      
      * up
      
      * up
      d176f61f