- 06 May, 2025 4 commits
-
-
Valeriy Selitskiy authored
[lora_conversion] Enhance key handling for OneTrainer components in LORA conversion utility (#11441) (#11487) * [lora_conversion] Enhance key handling for OneTrainer components in LORA conversion utility (#11441) * Update src/diffusers/loaders/lora_conversion_utils.py Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
Aryan authored
* add transformer * add pipeline * fixes * make fix-copies * update * add flux mu shift * update example snippet * debug * cleanup * batch_size=1 optimization * add pipeline test * fix for model cpu offloading' * add last_image support; credits: https://github.com/lllyasviel/FramePack/pull/167 * update example with flf2v * update penguin url * fix test * address review comment: https://github.com/huggingface/diffusers/pull/11428#discussion_r2071032371 * address review comment: https://github.com/huggingface/diffusers/pull/11428#discussion_r2071087689 * Update src/diffusers/pipelines/hunyuan_video/pipeline_hunyuan_video_framepack.py --------- Co-authored-by:
Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com>
-
Sayak Paul authored
* use removeprefix to preserve sanity. * f-string.
-
Sayak Paul authored
* update dep table. * fix
-
- 05 May, 2025 2 commits
-
-
Sayak Paul authored
* feat: enable quantization for hidream lora training. * better handle compute dtype. * finalize. * fix dtype. --------- Co-authored-by:Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com>
-
Connector Switch authored
* implement tiled encode/decode * address review comments
-
- 01 May, 2025 3 commits
-
-
Vladimir Mandic authored
* cache packages_distributions * remove unused exception reference * make style Signed-off-by:
Vladimir Mandic <mandic00@live.com> * change name to _package_map --------- Signed-off-by:
Vladimir Mandic <mandic00@live.com> Co-authored-by:
DN6 <dhruv.nair@gmail.com>
-
co63oc authored
* Fix typos in docs and comments * Apply style fixes --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
github-actions[bot] <github-actions[bot]@users.noreply.github.com>
-
Sayak Paul authored
* [tests] Add torch.compile() test for WanTransformer3DModel * fix wan recompilation issues. * style --------- Co-authored-by:tongyu0924 <winnie920924@gmail.com>
-
- 30 Apr, 2025 4 commits
-
-
Yao Matrix authored
* make autoencoders. controlnet_flux and wan_transformer3d_single_file pass on XPU Signed-off-by:
Yao Matrix <matrix.yao@intel.com> * Apply style fixes --------- Signed-off-by:
Yao Matrix <matrix.yao@intel.com> Co-authored-by:
github-actions[bot] <github-actions[bot]@users.noreply.github.com> Co-authored-by:
Aryan <aryan@huggingface.co>
-
Daniel Socek authored
* Add generic support for Intel Gaudi accelerator (hpu device) Signed-off-by:
Daniel Socek <daniel.socek@intel.com> Co-authored-by:
Libin Tang <libin.tang@intel.com> * Add loggers for generic HPU support Signed-off-by:
Daniel Socek <daniel.socek@intel.com> * Refactor hpu support with is_hpu_available() logic Signed-off-by:
Daniel Socek <daniel.socek@intel.com> * Fix style for hpu support update Signed-off-by:
Daniel Socek <daniel.socek@intel.com> * Decouple soft HPU check from hard device validation to support HPU migration Signed-off-by:
Daniel Socek <daniel.socek@intel.com> --------- Signed-off-by:
Daniel Socek <daniel.socek@intel.com> Co-authored-by:
Libin Tang <libin.tang@intel.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
Aryan authored
udpate
-
Aryan authored
raise warning instead of error
-
- 28 Apr, 2025 3 commits
-
-
urpetkov-amd authored
* Fixing missing provider options argument * Adding if else for provider options * Apply suggestions from code review Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Apply style fixes * Update src/diffusers/pipelines/onnx_utils.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/pipelines/onnx_utils.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> --------- Co-authored-by:
Uros Petkovic <urpektov@amd.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com> Co-authored-by:
github-actions[bot] <github-actions[bot]@users.noreply.github.com>
-
Yao Matrix authored
* enable gguf test cases on XPU Signed-off-by:
YAO Matrix <matrix.yao@intel.com> * make SD35LargeGGUFSingleFileTests::test_pipeline_inference pas Signed-off-by:
root <root@a4bf01945cfe.jf.intel.com> * make FluxControlLoRAGGUFTests::test_lora_loading pass Signed-off-by:
Yao Matrix <matrix.yao@intel.com> * polish code Signed-off-by:
Yao Matrix <matrix.yao@intel.com> * Apply style fixes --------- Signed-off-by:
YAO Matrix <matrix.yao@intel.com> Signed-off-by:
root <root@a4bf01945cfe.jf.intel.com> Signed-off-by:
Yao Matrix <matrix.yao@intel.com> Co-authored-by:
root <root@a4bf01945cfe.jf.intel.com> Co-authored-by:
github-actions[bot] <github-actions[bot]@users.noreply.github.com>
-
Yao Matrix authored
* enable test_layerwise_casting_memory cases on XPU Signed-off-by:
Yao Matrix <matrix.yao@intel.com> * fix style Signed-off-by:
Yao Matrix <matrix.yao@intel.com> --------- Signed-off-by:
Yao Matrix <matrix.yao@intel.com>
-
- 24 Apr, 2025 2 commits
-
-
co63oc authored
-
Ishan Modi authored
update
-
- 23 Apr, 2025 3 commits
-
-
Teriks authored
* Kolors additional pipelines, community contrib --------- Co-authored-by:
Teriks <Teriks@users.noreply.github.com> Co-authored-by:
Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com>
-
Ishan Modi authored
update
-
Aryan authored
* fix * add tests * add message check
-
- 22 Apr, 2025 4 commits
-
-
YiYi Xu authored
up
-
Aryan authored
update
-
Linoy Tsaban authored
* initial commit * initial commit * initial commit * initial commit * initial commit * initial commit * Update examples/dreambooth/train_dreambooth_lora_hidream.py Co-authored-by:
Bagheera <59658056+bghira@users.noreply.github.com> * move prompt embeds, pooled embeds outside * Update examples/dreambooth/train_dreambooth_lora_hidream.py Co-authored-by:
hlky <hlky@hlky.ac> * Update examples/dreambooth/train_dreambooth_lora_hidream.py Co-authored-by:
hlky <hlky@hlky.ac> * fix import * fix import and tokenizer 4, text encoder 4 loading * te * prompt embeds * fix naming * shapes * initial commit to add HiDreamImageLoraLoaderMixin * fix init * add tests * loader * fix model input * add code example to readme * fix default max length of text encoders * prints * nullify training cond in unpatchify for temp fix to incompatible shaping of transformer output during training * smol fix * unpatchify * unpatchify * fix validation * flip pred and loss * fix shift!!! * revert unpatchify changes (for now) * smol fix * Apply style fixes * workaround moe training * workaround moe training * remove prints * to reduce some memory, keep vae in `weight_dtype` same as we have for flux (as it's the same vae) https://github.com/huggingface/diffusers/blob/bbd0c161b55ba2234304f1e6325832dd69c60565/examples/dreambooth/train_dreambooth_lora_flux.py#L1207 * refactor to align with HiDream refactor * refactor to align with HiDream refactor * refactor to align with HiDream refactor * add support for cpu offloading of text encoders * Apply style fixes * adjust lr and rank for train example * fix copies * Apply style fixes * update README * update README * update README * fix license * keep prompt2,3,4 as None in validation * remove reverse ode comment * Update examples/dreambooth/train_dreambooth_lora_hidream.py Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * Update examples/dreambooth/train_dreambooth_lora_hidream.py Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * vae offload change * fix text encoder offloading * Apply style fixes * cleaner to_kwargs * fix module name in copied from * add requirements * fix offloading * fix offloading * fix offloading * update transformers version in reqs * try AutoTokenizer * try AutoTokenizer * Apply style fixes * empty commit * Delete tests/lora/test_lora_layers_hidream.py * change tokenizer_4 to load with AutoTokenizer as well * make text_encoder_four and tokenizer_four configurable * save model card * save model card * revert T5 * fix test * remove non diffusers lumina2 conversion --------- Co-authored-by:
Bagheera <59658056+bghira@users.noreply.github.com> Co-authored-by:
hlky <hlky@hlky.ac> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
github-actions[bot] <github-actions[bot]@users.noreply.github.com>
-
apolinário authored
* Add stochastic sampling to FlowMatchEulerDiscreteScheduler This PR adds stochastic sampling to FlowMatchEulerDiscreteScheduler based on https://github.com/Lightricks/LTX-Video/commit/b1aeddd7ccac85e6d1b0d97762610ddb53c1b408 ltx_video/schedulers/rf.py * Apply style fixes * Use config value directly * Apply style fixes * Swap order * Update src/diffusers/schedulers/scheduling_flow_match_euler_discrete.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/schedulers/scheduling_flow_match_euler_discrete.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> --------- Co-authored-by:
github-actions[bot] <github-actions[bot]@users.noreply.github.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com>
-
- 21 Apr, 2025 6 commits
-
-
Ishan Modi authored
* update * update * addressed PR comments * update --------- Co-authored-by:YiYi Xu <yixu310@gmail.com>
-
josephrocca authored
Co-authored-by:YiYi Xu <yixu310@gmail.com>
-
OleehyO authored
[cogview4][feat] Support attention mechanism with variable-length support and batch packing (#11349) * [cogview4] Enhance attention mechanism with variable-length support and batch packing --------- Co-authored-by:
YiYi Xu <yixu310@gmail.com> Co-authored-by:
github-actions[bot] <github-actions[bot]@users.noreply.github.com>
-
Aamir Nazir authored
* Update outputs.py
-
Ishan Modi authored
* bug fix kadinsky pipeline
-
Aryan authored
update
-
- 19 Apr, 2025 1 commit
-
-
YiYi Xu authored
up
-
- 18 Apr, 2025 2 commits
-
-
YiYi Xu authored
* update transformer --------- Co-authored-by:Aryan <aryan@huggingface.co>
-
Marc Sun authored
fix Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
- 17 Apr, 2025 4 commits
-
-
Frank (Haofan) Wang authored
-
Sayak Paul authored
* improve dtype mismatch handling for bnb + lora. * add a test * fix and updates * update
-
YiYi Xu authored
* add
-
Sayak Paul authored
* propagate hotswap to other load_lora_weights() methods. * simplify documentations. * updates * propagate to load_lora_into_text_encoder. * empty commit
-
- 16 Apr, 2025 2 commits
-
-
nPeppon authored
-
Dhruv Nair authored
* update * update
-