- 29 Jul, 2025 1 commit
-
-
Álvaro Somoza authored
* login * more logins * uploads * missed login * another missed login * downloads * examples and more logins * fix * setup * Apply style fixes * fix * Apply style fixes
-
- 26 Jun, 2025 1 commit
-
-
Sayak Paul authored
* post release v0.34.0 * code quality --------- Co-authored-by:YiYi Xu <yixu310@gmail.com>
-
- 19 May, 2025 1 commit
-
-
Quentin Gallouédec authored
* Use HF Papers * Apply style fixes --------- Co-authored-by:github-actions[bot] <github-actions[bot]@users.noreply.github.com>
-
- 05 May, 2025 2 commits
-
-
RogerSinghChugh authored
* Update training script for txt to img sdxl with lora supp with new interpolation. * ran make style and make quality.
-
Parag Ekbote authored
* Add LANCZOS as default interplotation mode. * update script * Update as per code review. * make style.
-
- 28 Apr, 2025 2 commits
- 15 Apr, 2025 1 commit
-
-
Sayak Paul authored
* post release * update * fix deprecations * remaining * update --------- Co-authored-by:YiYi Xu <yixu310@gmail.com>
-
- 09 Apr, 2025 1 commit
-
-
Dhruv Nair authored
* update * update * update * update
-
- 04 Mar, 2025 1 commit
-
-
Alexey Zolotenkov authored
* Fix seed initialization to handle args.seed = 0 correctly * Apply style fixes --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
github-actions[bot] <github-actions[bot]@users.noreply.github.com>
-
- 04 Feb, 2025 1 commit
-
-
Nicolas authored
-
- 23 Jan, 2025 1 commit
-
-
Yaniv Galron authored
We already set the unet to requires grad false at line 506 Co-authored-by:Aryan <aryan@huggingface.co>
-
- 21 Jan, 2025 1 commit
-
-
Sayak Paul authored
change licensing to 2025 from 2024.
-
- 02 Jan, 2025 1 commit
-
-
Doug J authored
Enable VAE hash to be able to change with args change. If not, train_dataset_with_embeddiings may have row number inconsistency with train_dataset_with_vae. Co-authored-by:Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com>
-
- 23 Dec, 2024 1 commit
-
-
Sayak Paul authored
* post release 0.32.0 * stylew
-
- 03 Dec, 2024 1 commit
-
-
aihao authored
* add data_dir parameter to load_dataset --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com> Co-authored-by:
hlky <hlky@hlky.ac>
-
- 22 Oct, 2024 1 commit
-
-
Sayak Paul authored
* post-release * style
-
- 14 Oct, 2024 1 commit
-
-
Leo Jiang authored
* Improve the performance and suitable for NPU * Improve the performance and suitable for NPU computing * Improve the performance and suitable for NPU * Improve the performance and suitable for NPU * Improve the performance and suitable for NPU * Improve the performance and suitable for NPU --------- Co-authored-by:
蒋硕 <jiangshuo9@h-partners.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
- 20 Aug, 2024 1 commit
-
-
Leo Jiang authored
Fix dtype error Co-authored-by:
蒋硕 <jiangshuo9@h-partners.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
- 14 Aug, 2024 1 commit
-
-
Álvaro Somoza authored
* post release * fix quality
-
- 12 Aug, 2024 1 commit
-
-
Dibbla! authored
-
- 08 Aug, 2024 1 commit
-
-
sayantan sadhu authored
* fix for lr scheduler in distributed training * Fixed the recalculation of the total training step section * Fixed lint error --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
- 26 Jul, 2024 1 commit
-
-
Sayak Paul authored
* introduce to promote reusability. * up * add more tests * up * remove comments. * fix fuse_nan test * clarify the scope of fuse_lora and unfuse_lora * remove space * rewrite fuse_lora a bit. * feedback * copy over load_lora_into_text_encoder. * address dhruv's feedback. * fix-copies * fix issubclass. * num_fused_loras * fix * fix * remove mapping * up * fix * style * fix-copies * change to SD3TransformerLoRALoadersMixin * Apply suggestions from code review Co-authored-by:
Dhruv Nair <dhruv.nair@gmail.com> * up * handle wuerstchen * up * move lora to lora_pipeline.py * up * fix-copies * fix documentation. * comment set_adapters(). * fix-copies * fix set_adapters() at the model level. * fix? * fix * loraloadermixin. --------- Co-authored-by:
Dhruv Nair <dhruv.nair@gmail.com>
-
- 25 Jul, 2024 2 commits
-
-
Sayak Paul authored
* introduce to promote reusability. * up * add more tests * up * remove comments. * fix fuse_nan test * clarify the scope of fuse_lora and unfuse_lora * remove space * rewrite fuse_lora a bit. * feedback * copy over load_lora_into_text_encoder. * address dhruv's feedback. * fix-copies * fix issubclass. * num_fused_loras * fix * fix * remove mapping * up * fix * style * fix-copies * change to SD3TransformerLoRALoadersMixin * Apply suggestions from code review Co-authored-by:
Dhruv Nair <dhruv.nair@gmail.com> * up * handle wuerstchen * up * move lora to lora_pipeline.py * up * fix-copies * fix documentation. * comment set_adapters(). * fix-copies * fix set_adapters() at the model level. * fix? * fix --------- Co-authored-by:
Dhruv Nair <dhruv.nair@gmail.com>
- 24 Jun, 2024 2 commits
-
-
Tolga Cangöz authored
* Trim all the trailing white space in the whole repo * Remove unnecessary empty places * make style && make quality * Trim trailing white space * trim --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
drhead authored
Add extra performance features for EMAModel, torch._foreach operations and better support for non-blocking CPU offloading (#7685) * Add support for _foreach operations and non-blocking to EMAModel * default foreach to false * add non-blocking EMA offloading to SD1.5 T2I example script * fix whitespace * move foreach to cli argument * linting * Update README.md re: EMA weight training * correct args.foreach_ema * add tests for foreach ema * code quality * add foreach to from_pretrained * default foreach false * fix linting --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
drhead <a@a.a>
-
- 13 Jun, 2024 1 commit
-
-
Sayak Paul authored
post release
-
- 05 Jun, 2024 1 commit
-
-
Tolga Cangöz authored
* Fix typos * Trim trailing whitespaces * Remove a trailing whitespace * chore: Update MarigoldDepthPipeline checkpoint to prs-eth/marigold-lcm-v1-0 * Revert "chore: Update MarigoldDepthPipeline checkpoint to prs-eth/marigold-lcm-v1-0" This reverts commit fd742b30b4258106008a6af4d0dd4664904f8595. * pokemon -> naruto * `DPMSolverMultistep` -> `DPMSolverMultistepScheduler` * Improve Markdown stylization * Improve style * Improve style * Refactor pipeline variable names for consistency * up style
-
- 30 May, 2024 2 commits
-
-
satani99 authored
* Modularized the train_lora_sdxl file * Modularized the train_lora_sdxl file * Modularized the train_lora_sdxl file --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
Genius Patrick authored
-
- 29 May, 2024 2 commits
-
-
satani99 authored
* Modularized the train_lora file * Modularized the train_lora file * Modularized the train_lora file * Modularized the train_lora file * Modularized the train_lora file --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
Sayak Paul authored
* post release v0.28.0 * style
-
- 14 May, 2024 1 commit
-
-
Kohei authored
Update requirements.txt If the datasets library is old, it will not read the metadata.jsonl and the label will default to an integer of type int. Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
- 12 May, 2024 1 commit
-
-
HelloWorldBeginner authored
Co-authored-by:
mhh001 <mahonghao1@huawei.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
- 11 May, 2024 1 commit
-
-
HelloWorldBeginner authored
fix bugs when using deepspeed Co-authored-by:mhh001 <mahonghao1@huawei.com>
-
- 07 May, 2024 1 commit
-
-
Bagheera authored
* 7879 - adjust documentation to use naruto dataset, since pokemon is now gated * replace references to pokemon in docs * more references to pokemon replaced * Japanese translation update --------- Co-authored-by:bghira <bghira@users.github.com>
-
- 03 May, 2024 1 commit
-
-
HelloWorldBeginner authored
Add Ascend NPU support for SDXL fine-tuning and fix the model saving bug when using DeepSpeed. (#7816) * Add Ascend NPU support for SDXL fine-tuning and fix the model saving bug when using DeepSpeed. * fix check code quality * Decouple the NPU flash attention and make it an independent module. * add doc and unit tests for npu flash attention. --------- Co-authored-by:
mhh001 <mahonghao1@huawei.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
- 27 Apr, 2024 1 commit
-
-
A new function compute_dream_and_update_latents has been added to the training utilities that allows you to do DREAM rectified training in line with the paper https://arxiv.org/abs/2312.00210. The method can be used with an extra argument in the train_text_to_image.py script. Co-authored-by: Jimmy <39@
🇺🇸 .com>
-
- 09 Apr, 2024 1 commit
-
-
Christopher Beckham authored
* Restore unet params back to normal from EMA when validation call is finished * empty commit --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-