- 26 Apr, 2025 1 commit
-
-
Mert Erbak authored
* Set LANCZOS as default interpolation mode for resizing * [train_dreambooth_lora.py] Set LANCZOS as default interpolation mode for resizing
-
- 24 Apr, 2025 4 commits
-
-
co63oc authored
-
Ishan Modi authored
update
-
Emiliano authored
Fix Flux IP adapter argument in the example IP-Adapter example had a wrong argument. Fix `true_cfg` -> `true_cfg_scale`
-
Linoy Tsaban authored
* 1. add pre-computation of prompt embeddings when custom prompts are used as well 2. save model card even if model is not pushed to hub 3. remove scheduler initialization from code example - not necessary anymore (it's now if the base model's config) 4. add skip_final_inference - to allow to run with validation, but skip the final loading of the pipeline with the lora weights to reduce memory reqs * pre encode validation prompt as well * Update examples/dreambooth/train_dreambooth_lora_hidream.py Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * Update examples/dreambooth/train_dreambooth_lora_hidream.py Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * Update examples/dreambooth/train_dreambooth_lora_hidream.py Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * pre encode validation prompt as well * Apply style fixes * empty commit * change default trained modules * empty commit * address comments + change encoding of validation prompt (before it was only pre-encoded if custom prompts are provided, but should be pre-encoded either way) * Apply style fixes * empty commit * fix validation_embeddings definition * fix final inference condition * fix pipeline deletion in last inference * Apply style fixes * empty commit * layers * remove readme remarks on only pre-computing when instance prompt is provided and change example to 3d icons * smol fix * empty commit --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
github-actions[bot] <github-actions[bot]@users.noreply.github.com>
-
- 23 Apr, 2025 5 commits
-
-
Teriks authored
* Kolors additional pipelines, community contrib --------- Co-authored-by:
Teriks <Teriks@users.noreply.github.com> Co-authored-by:
Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com>
-
Ishan Modi authored
update
-
Ishan Dutta authored
-
Aryan authored
* fix * add tests * add message check
-
Ameer Azam authored
Small change requirements_sana.txt to requirements_hidream.txt
-
- 22 Apr, 2025 4 commits
-
-
YiYi Xu authored
up
-
Aryan authored
update
-
Linoy Tsaban authored
* initial commit * initial commit * initial commit * initial commit * initial commit * initial commit * Update examples/dreambooth/train_dreambooth_lora_hidream.py Co-authored-by:
Bagheera <59658056+bghira@users.noreply.github.com> * move prompt embeds, pooled embeds outside * Update examples/dreambooth/train_dreambooth_lora_hidream.py Co-authored-by:
hlky <hlky@hlky.ac> * Update examples/dreambooth/train_dreambooth_lora_hidream.py Co-authored-by:
hlky <hlky@hlky.ac> * fix import * fix import and tokenizer 4, text encoder 4 loading * te * prompt embeds * fix naming * shapes * initial commit to add HiDreamImageLoraLoaderMixin * fix init * add tests * loader * fix model input * add code example to readme * fix default max length of text encoders * prints * nullify training cond in unpatchify for temp fix to incompatible shaping of transformer output during training * smol fix * unpatchify * unpatchify * fix validation * flip pred and loss * fix shift!!! * revert unpatchify changes (for now) * smol fix * Apply style fixes * workaround moe training * workaround moe training * remove prints * to reduce some memory, keep vae in `weight_dtype` same as we have for flux (as it's the same vae) https://github.com/huggingface/diffusers/blob/bbd0c161b55ba2234304f1e6325832dd69c60565/examples/dreambooth/train_dreambooth_lora_flux.py#L1207 * refactor to align with HiDream refactor * refactor to align with HiDream refactor * refactor to align with HiDream refactor * add support for cpu offloading of text encoders * Apply style fixes * adjust lr and rank for train example * fix copies * Apply style fixes * update README * update README * update README * fix license * keep prompt2,3,4 as None in validation * remove reverse ode comment * Update examples/dreambooth/train_dreambooth_lora_hidream.py Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * Update examples/dreambooth/train_dreambooth_lora_hidream.py Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * vae offload change * fix text encoder offloading * Apply style fixes * cleaner to_kwargs * fix module name in copied from * add requirements * fix offloading * fix offloading * fix offloading * update transformers version in reqs * try AutoTokenizer * try AutoTokenizer * Apply style fixes * empty commit * Delete tests/lora/test_lora_layers_hidream.py * change tokenizer_4 to load with AutoTokenizer as well * make text_encoder_four and tokenizer_four configurable * save model card * save model card * revert T5 * fix test * remove non diffusers lumina2 conversion --------- Co-authored-by:
Bagheera <59658056+bghira@users.noreply.github.com> Co-authored-by:
hlky <hlky@hlky.ac> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
github-actions[bot] <github-actions[bot]@users.noreply.github.com>
-
apolinário authored
* Add stochastic sampling to FlowMatchEulerDiscreteScheduler This PR adds stochastic sampling to FlowMatchEulerDiscreteScheduler based on https://github.com/Lightricks/LTX-Video/commit/b1aeddd7ccac85e6d1b0d97762610ddb53c1b408 ltx_video/schedulers/rf.py * Apply style fixes * Use config value directly * Apply style fixes * Swap order * Update src/diffusers/schedulers/scheduling_flow_match_euler_discrete.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/schedulers/scheduling_flow_match_euler_discrete.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> --------- Co-authored-by:
github-actions[bot] <github-actions[bot]@users.noreply.github.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com>
-
- 21 Apr, 2025 9 commits
-
-
Ishan Modi authored
* update * update * addressed PR comments * update --------- Co-authored-by:YiYi Xu <yixu310@gmail.com>
-
josephrocca authored
Co-authored-by:YiYi Xu <yixu310@gmail.com>
-
OleehyO authored
[cogview4][feat] Support attention mechanism with variable-length support and batch packing (#11349) * [cogview4] Enhance attention mechanism with variable-length support and batch packing --------- Co-authored-by:
YiYi Xu <yixu310@gmail.com> Co-authored-by:
github-actions[bot] <github-actions[bot]@users.noreply.github.com>
-
Aamir Nazir authored
* Update outputs.py
-
Ishan Modi authored
* bug fix kadinsky pipeline
-
Aryan authored
update
-
PromeAI authored
* fix issue that training flux controlnet was unstable and validation results were unstable * del unused code pieces, fix grammar --------- Co-authored-by:
Your Name <you@example.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
Kenneth Gerald Hamilton authored
[train_dreambooth_lora_sdxl.py] Fix the LR Schedulers when num_train_epochs is passed in a distributed training env (#11240) Co-authored-by:Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com>
-
Linoy Tsaban authored
* add fix * add fix * Apply style fixes --------- Co-authored-by:github-actions[bot] <github-actions[bot]@users.noreply.github.com>
-
- 19 Apr, 2025 2 commits
- 18 Apr, 2025 3 commits
-
-
Kazuki Yoda authored
Fix: `StableDiffusionXLControlNetAdapterInpaintPipeline` incorrectly inherited `StableDiffusionLoraLoaderMixin` (#11357) Fix: Inherit `StableDiffusionXLLoraLoaderMixin` `StableDiffusionXLControlNetAdapterInpaintPipeline` used to incorrectly inherit `StableDiffusionLoraLoaderMixin` instead of `StableDiffusionXLLoraLoaderMixin`
-
YiYi Xu authored
* update transformer --------- Co-authored-by:Aryan <aryan@huggingface.co>
-
Marc Sun authored
fix Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
- 17 Apr, 2025 7 commits
-
-
Yao Matrix authored
* enable 2 test cases on XPU Signed-off-by:
YAO Matrix <matrix.yao@intel.com> * Apply style fixes --------- Signed-off-by:
YAO Matrix <matrix.yao@intel.com> Co-authored-by:
github-actions[bot] <github-actions[bot]@users.noreply.github.com> Co-authored-by:
Dhruv Nair <dhruv.nair@gmail.com>
-
Frank (Haofan) Wang authored
-
Sayak Paul authored
* improve dtype mismatch handling for bnb + lora. * add a test * fix and updates * update
-
YiYi Xu authored
* add
-
Sayak Paul authored
* propagate hotswap to other load_lora_weights() methods. * simplify documentations. * updates * propagate to load_lora_into_text_encoder. * empty commit
-
Sayak Paul authored
add note about use_duck_shape in auraflow docs.
-
Sayak Paul authored
fix lora docs utils
-
- 16 Apr, 2025 5 commits
-
-
nPeppon authored
-
Ishan Modi authored
update
-
Dhruv Nair authored
* update * update
-
Sayak Paul authored
* add a snippet for compilation in the auraflow docs. * include speedups.
-
Sayak Paul authored
* enable telemetry for single file loading when using GGUF. * quality
-