- 23 Oct, 2025 2 commits
-
-
Dhruv Nair authored
* update * update * update
-
kaixuanliu authored
* fix CI bug for kandinsky3_img2img case Signed-off-by:
Liu, Kaixuan <kaixuan.liu@intel.com> * update code Signed-off-by:
Liu, Kaixuan <kaixuan.liu@intel.com> --------- Signed-off-by:
Liu, Kaixuan <kaixuan.liu@intel.com>
-
- 22 Oct, 2025 3 commits
-
-
YiYi Xu authored
add
-
Álvaro Somoza authored
fix
-
David Bertoin authored
* rename photon to prx * rename photon into prx * Revert .gitignore to state before commit b7fb0fe9d63bf766bbe3c42ac154a043796dd370 * rename photon to prx * rename photon into prx * Revert .gitignore to state before commit b7fb0fe9d63bf766bbe3c42ac154a043796dd370 * make fix-copies
-
- 21 Oct, 2025 1 commit
-
-
David Bertoin authored
* Add Photon model and pipeline support This commit adds support for the Photon image generation model: - PhotonTransformer2DModel: Core transformer architecture - PhotonPipeline: Text-to-image generation pipeline - Attention processor updates for Photon-specific attention mechanism - Conversion script for loading Photon checkpoints - Documentation and tests * just store the T5Gemma encoder * enhance_vae_properties if vae is provided only * remove autocast for text encoder forwad * BF16 example * conditioned CFG * remove enhance vae and use vae.config directly when possible * move PhotonAttnProcessor2_0 in transformer_photon * remove einops dependency and now inherits from AttentionMixin * unify the structure of the forward block * update doc * update doc * fix T5Gemma loading from hub * fix timestep shift * remove lora support from doc * Rename EmbedND for PhotoEmbedND * remove modulation dataclass * put _attn_forward and _ffn_forward logic in PhotonBlock's forward * renam LastLayer for FinalLayer * remove lora related code * rename vae_spatial_compression_ratio for vae_scale_factor * support prompt_embeds in call * move xattention conditionning out computation out of the denoising loop * add negative prompts * Use _import_structure for lazy loading * make quality + style * add pipeline test + corresponding fixes * utility function that determines the default resolution given the VAE * Refactor PhotonAttention to match Flux pattern * built-in RMSNorm * Revert accidental .gitignore change * parameter names match the standard diffusers conventions * renaming and remove unecessary attributes setting * Update docs/source/en/api/pipelines/photon.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * quantization example * added doc to toctree * Update docs/source/en/api/pipelines/photon.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/api/pipelines/photon.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/api/pipelines/photon.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * use dispatch_attention_fn for multiple attention backend support * naming changes * make fix copy * Update docs/source/en/api/pipelines/photon.md Co-authored-by:
dg845 <58458699+dg845@users.noreply.github.com> * Add PhotonTransformer2DModel to TYPE_CHECKING imports * make fix-copies * Use Tuple instead of tuple Co-authored-by:
dg845 <58458699+dg845@users.noreply.github.com> * restrict the version of transformers Co-authored-by:
dg845 <58458699+dg845@users.noreply.github.com> * Update tests/pipelines/photon/test_pipeline_photon.py Co-authored-by:
dg845 <58458699+dg845@users.noreply.github.com> * Update tests/pipelines/photon/test_pipeline_photon.py Co-authored-by:
dg845 <58458699+dg845@users.noreply.github.com> * change | for Optional * fix nits. * use typing Dict --------- Co-authored-by:
davidb <davidb@worker-10.soperator-worker-svc.soperator.svc.cluster.local> Co-authored-by:
David Briand <david@photoroom.com> Co-authored-by:
davidb <davidb@worker-8.soperator-worker-svc.soperator.svc.cluster.local> Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> Co-authored-by:
dg845 <58458699+dg845@users.noreply.github.com> Co-authored-by:
sayakpaul <spsayakpaul@gmail.com>
-
- 18 Oct, 2025 1 commit
-
-
Lev Novitskiy authored
* add kandinsky5 transformer pipeline first version --------- Co-authored-by:
Álvaro Somoza <asomoza@users.noreply.github.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com> Co-authored-by:
Charles <charles@huggingface.co>
-
- 17 Oct, 2025 1 commit
-
-
Ali Imran authored
* cleanup of runway model * quality fixes
-
- 15 Oct, 2025 2 commits
-
-
YiYi Xu authored
update Co-authored-by:Aryan <aryan@huggingface.co>
-
Sayak Paul authored
-
- 14 Oct, 2025 1 commit
-
-
Meatfucker authored
Fix missing load_video documentation and load_video import in WanVideoToVideoPipeline example code (#12472) * Update utilities.md Update missing load_video documentation * Update pipeline_wan_video2video.py Fix missing load_video import in example code
-
- 11 Oct, 2025 1 commit
-
-
Steven Liu authored
* fix syntax * fix * style * fix
-
- 07 Oct, 2025 1 commit
-
-
Sayak Paul authored
-
- 05 Oct, 2025 1 commit
-
-
Sayak Paul authored
* up * up * up * up * up * up * remove saves * move things around a bit. * get ready.
-
- 30 Sep, 2025 1 commit
-
-
Steven Liu authored
* change syntax * make style
-
- 29 Sep, 2025 1 commit
-
-
Akshay Babbar authored
* fix: preserve boolean dtype for attention masks in ChromaPipeline - Convert attention masks to bool and prevent dtype corruption - Fix both positive and negative mask handling in _get_t5_prompt_embeds - Remove float conversion in _prepare_attention_mask method Fixes #12116 * test: add ChromaPipeline attention mask dtype tests * test: add slow ChromaPipeline attention mask tests * chore: removed comments * refactor: removing redundant type conversion * Remove dedicated dtype tests as per feedback --------- Co-authored-by:Dhruv Nair <dhruv.nair@gmail.com>
-
- 25 Sep, 2025 1 commit
-
-
Lucain authored
* Support huggingface_hub 0.x and 1.x * httpx
-
- 24 Sep, 2025 1 commit
-
-
Alberto Chimenti authored
Fixed WanVACEPipeline to allow prompt to be None and skip encoding step
-
- 22 Sep, 2025 2 commits
-
-
SahilCarterr authored
* Fixes chroma docs * fix docs fixed docs are now consistent
-
Sayak Paul authored
xfail some kandinsky tests.
-
- 21 Sep, 2025 1 commit
-
-
naykun authored
* feat: add support of qwenimageeditplus * add copies statement * fix copies statement * remove vl_processor reference
-
- 20 Sep, 2025 1 commit
-
-
Dhruv Nair authored
update
-
- 16 Sep, 2025 1 commit
-
-
Sari Hleihil authored
* Added LucyEditPipeline * add import & stype missing copied from * Fix example doc string --------- Co-authored-by:yiyixuxu <yixu310@gmail.com>
-
- 15 Sep, 2025 1 commit
-
-
Linoy Tsaban authored
* support Wan2.2-VACE-Fun-A14B * support Wan2.2-VACE-Fun-A14B * support Wan2.2-VACE-Fun-A14B * Apply style fixes * test --------- Co-authored-by:github-actions[bot] <github-actions[bot]@users.noreply.github.com>
-
- 12 Sep, 2025 1 commit
-
-
Daniel Socek authored
* Use SDP on BF16 in GPU/HPU migration Signed-off-by:
Daniel Socek <daniel.socek@intel.com> * Formatting fix for enabling SDP with BF16 precision on HPU Signed-off-by:
Daniel Socek <daniel.socek@intel.com> --------- Signed-off-by:
Daniel Socek <daniel.socek@intel.com>
-
- 11 Sep, 2025 1 commit
-
-
Sayak Paul authored
* deprecate slicing from flux pipeline. * propagate. * tiling * up * up
-
- 10 Sep, 2025 3 commits
-
-
Justin Ruan authored
Co-authored-by:YiYi Xu <yixu310@gmail.com>
-
ttio2tech authored
fix the device for textencoder
-
Sayak Paul authored
* feat: support group offloading at the pipeline level. * add tests * up * [docs] Pipeline group offloading (#12286) init Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> --------- Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
- 09 Sep, 2025 1 commit
-
-
Frank (Haofan) Wang authored
* add qwen-image-cn-inpaint --------- Co-authored-by:
github-actions[bot] <github-actions[bot]@users.noreply.github.com> Co-authored-by:
yiyixuxu <yixu310@gmail.com>
-
- 08 Sep, 2025 1 commit
-
-
YiYi Xu authored
* add qwen modular
-
- 03 Sep, 2025 1 commit
-
-
co63oc authored
Signed-off-by:co63oc <co63oc@users.noreply.github.com>
-
- 31 Aug, 2025 1 commit
-
-
Nguyễn Trọng Tuấn authored
* add qwenimage-edit inpaint feature * stay up to date with main branch * fix style * fix docs * copies * fix * again * copies --------- Co-authored-by:
“Trgtuan10” <“tuannguyentrong.402@gmail.com”> Co-authored-by:
TuanNT-ZenAI <tuannt.zenai@gmail.com> Co-authored-by:
yiyixuxu <yixu310@gmail.com>
-
- 30 Aug, 2025 1 commit
-
-
Nguyễn Trọng Tuấn authored
add entry Co-authored-by:TuanNT-ZenAI <tuannt.zenai@gmail.com>
-
- 28 Aug, 2025 1 commit
-
-
YiYi Xu authored
* make modular pipeline work with model_index.json * up * style * up * up * style * up more * Fix MultiControlNet import (#12118) fix --------- Co-authored-by:
Álvaro Somoza <asomoza@users.noreply.github.com> Co-authored-by:
Dhruv Nair <dhruv.nair@gmail.com>
-
- 27 Aug, 2025 1 commit
-
-
YiYi Xu authored
* up
-
- 26 Aug, 2025 1 commit
-
-
Sayak Paul authored
* start removing flax stuff. * add deprecation warning. * add warning messages. * more warnings. * remove dockerfiles. * remove more. * Update src/diffusers/models/attention_flax.py Co-authored-by:
Dhruv Nair <dhruv.nair@gmail.com> * up --------- Co-authored-by:
Dhruv Nair <dhruv.nair@gmail.com>
-
- 25 Aug, 2025 1 commit
-
-
sqt authored
-
- 22 Aug, 2025 1 commit
-
-
Frank (Haofan) Wang authored
* support qwen-image-cn-union --------- Co-authored-by:
github-actions[bot] <github-actions[bot]@users.noreply.github.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com>
-
- 20 Aug, 2025 1 commit
-
-
galbria authored
* Add Bria model and pipeline to diffusers - Introduced `BriaTransformer2DModel` and `BriaPipeline` for enhanced image generation capabilities. - Updated import structures across various modules to include the new Bria components. - Added utility functions and output classes specific to the Bria pipeline. - Implemented tests for the Bria pipeline to ensure functionality and output integrity. * with working tests * style and quality pass * adding docs * add to overview * fixes from "make fix-copies" * Refactor transformer_bria.py and pipeline_bria.py: Introduce new EmbedND class for rotary position embedding, and enhance Timestep and TimestepProjEmbeddings classes. Add utility functions for handling negative prompts and generating original sigmas in pipeline_bria.py. * remove redundent and duplicates tests and fix bf16 slow test * style fixes * small doc update * Enhance Bria 3.2 documentation and implementation - Updated the GitHub repository link for Bria 3.2. - Added usage instructions for the gated model access. - Introduced the BriaTransformerBlock and BriaAttention classes to the model architecture. - Refactored existing classes to integrate Bria-specific components, including BriaEmbedND and BriaPipeline. - Updated the pipeline output class to reflect Bria-specific functionality. - Adjusted test cases to align with the new Bria model structure. * Refactor Bria model components and update documentation - Removed outdated inference example from Bria 3.2 documentation. - Introduced the BriaTransformerBlock class to enhance model architecture. - Updated attention handling to use `attention_kwargs` instead of `joint_attention_kwargs`. - Improved import structure in the Bria pipeline to handle optional dependencies. - Adjusted test cases to reflect changes in model dtype assertions. * Update Bria model reference in documentation to reflect new file naming convention * Update docs/source/en/_toctree.yml * Refactor BriaPipeline to inherit from DiffusionPipeline instead of FluxPipeline, updating imports accordingly. * move the __call__ func to the end of file * Update BriaPipeline example to use bfloat16 for precision sensitivity for better result * make style && make quality && make fix-copiessource --------- Co-authored-by:
Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com> Co-authored-by:
Aryan <contact.aryanvs@gmail.com>
-