- 07 Feb, 2024 1 commit
-
-
Sayak Paul authored
* feat: standarize model card creation for dreambooth training. * correct 'inference * remove comments. * take component out of kwargs * style * add: card template to have a leaner description. * widget support. * propagate changes to train_dreambooth_lora * propagate changes to custom diffusion * make widget properly type-annotated
-
- 16 Jan, 2024 1 commit
-
-
Aryan V S authored
* fix training resume * update * update
-
- 12 Jan, 2024 1 commit
-
-
Vinh H. Pham authored
support compile
-
- 05 Jan, 2024 1 commit
-
-
Sayak Paul authored
* post release * style --------- Co-authored-by:Patrick von Platen <patrick.v.platen@gmail.com>
-
- 28 Dec, 2023 1 commit
-
-
Sayak Paul authored
* add to dreambooth lora. * add: t2i lora. * add: sdxl t2i lora. * style * lcm lora sdxl. * unwrap * fix: enable_adapters().
-
- 24 Dec, 2023 1 commit
-
-
Sayak Paul authored
* initialize alpha too. * add: test * remove config parsing * store rank * debug * remove faulty test
-
- 18 Dec, 2023 1 commit
-
-
Sayak Paul authored
remove depcreated method from lora scripts.
-
- 07 Dec, 2023 1 commit
-
-
Younes Belkada authored
* adapt example scripts to use PEFT * Update examples/text_to_image/train_text_to_image_lora.py * fix * add for SDXL * oops * make sure to install peft * fix * fix * fix dreambooth and lora * more fixes * add peft to requirements.txt * fix * final fix * add peft version in requirements * remove comment * change variable names * add few lines in readme * add to reqs * style * fix issues * fix lora dreambooth xl tests * init_lora_weights to gaussian and add out proj where missing * ammend requirements. * ammend requirements.txt * add correct peft versions --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
- 01 Dec, 2023 1 commit
-
-
Patrick von Platen authored
* Post Release: v0.24.0 * post pone deprecation * post pone deprecation * Add model_index.json
-
- 27 Nov, 2023 1 commit
-
-
Patrick von Platen authored
* add variant * add variant * Apply suggestions from code review * reformat * fix: textual_inversion.py * fix: variant in model_info --------- Co-authored-by:sayakpaul <spsayakpaul@gmail.com>
-
- 21 Nov, 2023 1 commit
-
-
Patrick von Platen authored
* [Lora] Seperate logic * [Lora] Seperate logic * [Lora] Seperate logic * add comments to explain the code better * add comments to explain the code better
-
- 17 Nov, 2023 1 commit
-
-
Lucain authored
* Set usedforsecurity=False in hashlib methods (FIPS compliance) * update version dependency * bump hfh version * bump hfh version
-
- 14 Nov, 2023 1 commit
-
-
Sayak Paul authored
* refactor loaders.py to make it cleaner and leaner. * refactor loaders init * inits. * textual inversion to the init. * inits. * remove certain modules from the main init. * AttnProcsLayers * fix imports * avoid circular import. * fix circular import pt 2. * address PR comments * imports * fix: imports. * remove from main init for avoiding circular deps. * remove spurious deps. * fix-copies. * fix imports. * more debug * more debug * Apply suggestions from code review * Apply suggestions from code review --------- Co-authored-by:Patrick von Platen <patrick.v.platen@gmail.com>
-
- 10 Nov, 2023 1 commit
-
-
Sayak Paul authored
* post release * fix: variant test * up * fix: test
-
- 06 Nov, 2023 1 commit
-
-
Sayak Paul authored
post release
-
- 16 Oct, 2023 1 commit
-
-
Heinz-Alexander Fuetterer authored
* chore: fix typos * Update src/diffusers/pipelines/shap_e/renderer.py Co-authored-by:
psychedelicious <4822129+psychedelicious@users.noreply.github.com> --------- Co-authored-by:
psychedelicious <4822129+psychedelicious@users.noreply.github.com>
-
- 11 Oct, 2023 1 commit
-
-
Sayak Paul authored
* use loralinear instead of depecrecated lora attn procs. * fix parameters() * fix saving * add back support for add kv proj. * fix: param accumul,ation. * propagate the changes.
-
- 08 Oct, 2023 1 commit
-
-
Zeng Xian authored
-
- 02 Oct, 2023 1 commit
-
-
Sayak Paul authored
* fix: how print training resume logs. * propagate changes to text-to-image scripts. * propagate changes to instructpix2pix. * propagate changes to dreambooth * propagate changes to custom diffusion and instructpix2pix * propagate changes to kandinsky * propagate changes to textual inv. * debug * fix: checkpointing. * debug * debug * debug * back to the square * debug * debug * change condition order. * debug * debug * debug * debug * revert to original * clean --------- Co-authored-by:Patrick von Platen <patrick.v.platen@gmail.com>
-
- 14 Sep, 2023 1 commit
-
-
Patrick von Platen authored
* [Release 0.21] Bump version * fix & remove * fix more * fix all, upload
-
- 08 Sep, 2023 1 commit
-
-
Will Berman authored
-
- 25 Aug, 2023 1 commit
-
-
Will Berman authored
-
- 17 Aug, 2023 2 commits
-
-
Sayak Paul authored
-
Patrick von Platen authored
* make safetensors default * set default save method as safetensors * update tests * update to support saving safetensors * update test to account for safetensors default * update example tests to use safetensors * update example to support safetensors * update unet tests for safetensors * fix failing loader tests * fix qc issues * fix pipeline tests * fix example test --------- Co-authored-by:Dhruv Nair <dhruv.nair@gmail.com>
-
- 04 Aug, 2023 1 commit
-
-
Patrick von Platen authored
* correct * correct blocks * finish * finish * finish * Apply suggestions from code review * fix * up * up * up * Update examples/dreambooth/README_sdxl.md Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * Apply suggestions from code review --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
- 01 Aug, 2023 1 commit
-
-
Will Berman authored
-
- 28 Jul, 2023 1 commit
-
-
Sayak Paul authored
* sdxl lora changes. * better name replacement. * better replacement. * debugging * debugging * debugging * debugging * debugging * remove print. * print state dict keys. * print * distingisuih better * debuggable. * fxi: tyests * fix: arg from training script. * access from class. * run style * debug * save intermediate * some simplifications for SDXL LoRA * styling * unet config is not needed in diffusers format. * fix: dynamic SGM block mapping for SDXL kohya loras (#4322) * Use lora compatible layers for linear proj_in/proj_out (#4323) * improve condition for using the sgm_diffusers mapping * informative comment. * load compatible keys and embedding layer maaping. * Get SDXL 1.0 example lora to load * simplify * specif ranks and hidden sizes. * better handling of k rank and hidden * debug * debug * debug * debug * debug * fix: alpha keys * add check for handling LoRAAttnAddedKVProcessor * sanity comment * modifications for text encoder SDXL * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * denugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * up * up * up * up * up * up * unneeded comments. * unneeded comments. * kwargs for the other attention processors. * kwargs for the other attention processors. * debugging * debugging * debugging * debugging * improve * debugging * debugging * more print * Fix alphas * debugging * debugging * debugging * debugging * debugging * debugging * clean up * clean up. * debugging * fix: text --------- Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
Batuhan Taskaya <batuhan@python.org>
-
- 27 Jul, 2023 1 commit
-
-
Xinyang Li authored
-
- 26 Jul, 2023 1 commit
-
-
Patrick von Platen authored
* 0.20.0dev0 * make style
-
- 25 Jul, 2023 1 commit
-
-
Will Berman authored
-
- 18 Jul, 2023 1 commit
-
-
takuoko authored
add rank in dreambooth
-
- 13 Jul, 2023 1 commit
-
-
Ruoxi authored
* Multiply lr scheduler steps by `num_processes`. * Stop multiplying steps by gradient accumulation.
-
- 11 Jul, 2023 1 commit
-
-
Patrick von Platen authored
-
- 09 Jul, 2023 2 commits
-
-
Patrick von Platen authored
-
Will Berman authored
* refactor to support patching LoRA into T5 instantiate the lora linear layer on the same device as the regular linear layer get lora rank from state dict tests fmt can create lora layer in float32 even when rest of model is float16 fix loading model hook remove load_lora_weights_ and T5 dispatching remove Unet#attn_processors_state_dict docstrings * text encoder monkeypatch class method * fix test --------- Co-authored-by:Patrick von Platen <patrick.v.platen@gmail.com>
-
- 07 Jul, 2023 1 commit
-
-
Batuhan Taskaya authored
-
- 15 Jun, 2023 2 commits
-
-
Will Berman authored
* manual check for checkpoints_total_limit instead of using accelerate * remove controlnet_conditioning_embedding_out_channels
-
Patrick von Platen authored
* relax tolerance slightly * correct incorrect naming
-
- 08 Jun, 2023 2 commits
-
-
Patrick von Platen authored
* Post release * Post release
-
Zachary Mueller authored
Apply deprecations
-