- 25 Aug, 2023 1 commit
-
-
Sayak Paul authored
* relax lora loading logic. * cater to the other cases too. * fix: variable name * bring the chaos down. * check * deal with checkpointed files. * Apply suggestions from code review Co-authored-by:
apolinário <joaopaulo.passos@gmail.com> * style --------- Co-authored-by:
apolinário <joaopaulo.passos@gmail.com>
-
- 22 Aug, 2023 2 commits
-
-
Sayak Paul authored
default to None when fc alphas are not available.
-
Sayak Paul authored
* debugging starts * debugging * debugging * debugging * debugging * debugging * debugging ends, but does it? * more robustness.
-
- 18 Aug, 2023 1 commit
-
-
zuojianghua authored
* Update loaders.py add config_file to from_single_file, when the download_from_original_stable_diffusion_ckpt use * Update loaders.py add config_file to from_single_file, when the download_from_original_stable_diffusion_ckpt use * change config_file to original_config_file * make style && make quality --------- Co-authored-by:
jianghua.zuo <jianghua.zuo@weimob.com> Co-authored-by:
Dhruv Nair <dhruv.nair@gmail.com>
-
- 17 Aug, 2023 2 commits
-
-
Patrick von Platen authored
* make safetensors default * set default save method as safetensors * update tests * update to support saving safetensors * update test to account for safetensors default * update example tests to use safetensors * update example to support safetensors * update unet tests for safetensors * fix failing loader tests * fix qc issues * fix pipeline tests * fix example test --------- Co-authored-by:Dhruv Nair <dhruv.nair@gmail.com>
-
Scott Lessans authored
* fixed * add: tests --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
- 11 Aug, 2023 2 commits
-
-
Mystfit authored
* Fixing repo_id regex validation error on windows platforms * Validating correct URL with prefix is provided If we are loading a URL then we don't need to use os.path.join and array slicing to split out a repo_id and file path from an absolute filepath. Checking if the URL prefix is valid first before doing any URL splitting otherwise we raise a ValueError since neither a valid filepath or URL was provided. * Style fixes --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
Abhipsha Das authored
* [WIP] Remove code snippets containing `is_safetensors_available()` * Modifying `import_utils.py` * update pipeline tests for safetensor default * fix test related to cached requests * address import nits --------- Co-authored-by:Dhruv Nair <dhruv.nair@gmail.com>
-
- 02 Aug, 2023 1 commit
-
-
Sayak Paul authored
* temporarily disable text encoder loras. * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debbuging. * modify doc. * rename tests. * print slices. * fix: assertions * Apply suggestions from code review Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> --------- Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
- 28 Jul, 2023 1 commit
-
-
Sayak Paul authored
* sdxl lora changes. * better name replacement. * better replacement. * debugging * debugging * debugging * debugging * debugging * remove print. * print state dict keys. * print * distingisuih better * debuggable. * fxi: tyests * fix: arg from training script. * access from class. * run style * debug * save intermediate * some simplifications for SDXL LoRA * styling * unet config is not needed in diffusers format. * fix: dynamic SGM block mapping for SDXL kohya loras (#4322) * Use lora compatible layers for linear proj_in/proj_out (#4323) * improve condition for using the sgm_diffusers mapping * informative comment. * load compatible keys and embedding layer maaping. * Get SDXL 1.0 example lora to load * simplify * specif ranks and hidden sizes. * better handling of k rank and hidden * debug * debug * debug * debug * debug * fix: alpha keys * add check for handling LoRAAttnAddedKVProcessor * sanity comment * modifications for text encoder SDXL * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * denugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * debugging * up * up * up * up * up * up * unneeded comments. * unneeded comments. * kwargs for the other attention processors. * kwargs for the other attention processors. * debugging * debugging * debugging * debugging * improve * debugging * debugging * more print * Fix alphas * debugging * debugging * debugging * debugging * debugging * debugging * clean up * clean up. * debugging * fix: text --------- Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
Batuhan Taskaya <batuhan@python.org>
-
- 26 Jul, 2023 1 commit
-
-
Patrick von Platen authored
* 0.20.0dev0 * make style
-
- 25 Jul, 2023 5 commits
-
-
Patrick von Platen authored
* fix circular import * fix imports when watermark not specified * fix all pipelines
-
Batuhan Taskaya authored
* Support to load Kohya-ss style LoRA file format (without restrictions) Co-Authored-By:
Takuma Mori <takuma104@gmail.com> Co-Authored-By:
Sayak Paul <spsayakpaul@gmail.com> * tmp: add sdxl to mlp_modules --------- Co-authored-by:
Takuma Mori <takuma104@gmail.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
Patrick von Platen authored
* up * finish * fix final
-
Patrick von Platen authored
* Allow vae to be loaded * up
-
Sayak Paul authored
* Allow low precision sd xl * finish * finish * feat: initial draft for supporting text encoder lora finetuning for SDXL DreamBooth * fix: variable assignments. * add: autocast block. * add debugging * vae dtype hell * fix: vae dtype hell. * fix: vae dtype hell 3. * clean up * lora text encoder loader. * fix: unwrapping models. * add: tests. * docs. * handle unexpected keys. * fix vae dtype in the final inference. * fix scope problem. * fix: save_model_card args. * initialize: prefix to None. * fix: dtype issues. * apply gixes. * debgging. * debugging * debugging * debugging * debugging * debugging * add: fast tests. * pre-tokenize. * address: will's comments. * fix: loader and tests. * fix: dataloader. * simplify dataloader. * length. * simplification. * make style && make quality * simplify state_dict munging * fix: tests. * fix: state_dict packing. * Apply suggestions from code review Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> --------- Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
- 24 Jul, 2023 1 commit
-
-
Jackmin801 authored
fix: missed network_alpha when loading lora from old format
-
- 21 Jul, 2023 2 commits
-
-
Steven Liu authored
* start with stable diffusion * fix * finish stable diffusion pipelines * fix path to pipeline output * fix flax paths * fix copies * add up to score sde ve * finish first pass of pipelines * fix copies * second review * align doc titles * more review fixes * final review
-
Batuhan Taskaya authored
-
- 19 Jul, 2023 1 commit
-
-
Patrick von Platen authored
* Add controlnet from single file * Updates * make style * finish * Apply suggestions from code review Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
- 17 Jul, 2023 1 commit
-
-
Patrick von Platen authored
replace with os path join
-
- 14 Jul, 2023 1 commit
-
-
Sayak Paul authored
* add: test for testing unloading lora. * add :reason to skipif. * initial implementation of lora unload(). * apply styling. * add: doc. * change checkpoints. * reinit generator * finalize slow test. * add fast test for unloading lora.
-
- 11 Jul, 2023 1 commit
-
-
Patrick von Platen authored
* start improving single file load * Fix more * start improving single file load * Fix sd 2.1 * further improve from_single_file
-
- 09 Jul, 2023 1 commit
-
-
Will Berman authored
* refactor to support patching LoRA into T5 instantiate the lora linear layer on the same device as the regular linear layer get lora rank from state dict tests fmt can create lora layer in float32 even when rest of model is float16 fix loading model hook remove load_lora_weights_ and T5 dispatching remove Unet#attn_processors_state_dict docstrings * text encoder monkeypatch class method * fix test --------- Co-authored-by:Patrick von Platen <patrick.v.platen@gmail.com>
-
- 07 Jul, 2023 1 commit
-
-
Yorai Levi authored
* Update pipeline_utils.py typo in safetensors (safetenstors) * Update loaders.py typo in safetensors (safetenstors) * Update modeling_utils.py typo in safetensors (safetenstors)
-
- 06 Jul, 2023 1 commit
-
-
Patrick von Platen authored
* improve sd xl * correct more * finish * make style * fix more
-
- 28 Jun, 2023 1 commit
-
-
Wadim Korablin authored
* Support for manual CLIP loading in StableDiffusionPipeline - txt2img. * Update src/diffusers/pipelines/stable_diffusion/convert_from_ckpt.py * Update variables & according docs to match previous style. * Updated to match style & quality of 'diffusers' --------- Co-authored-by:Patrick von Platen <patrick.v.platen@gmail.com>
-
- 21 Jun, 2023 1 commit
-
-
Steven Liu authored
* clean up loaders * clean up rest of main class apis * apply feedback
-
- 08 Jun, 2023 1 commit
-
-
Patrick von Platen authored
* Post release * Post release
-
- 06 Jun, 2023 5 commits
-
-
Patrick von Platen authored
* Add draft for lora text encoder scale * Improve naming * fix: training dreambooth lora script. * Apply suggestions from code review * Update examples/dreambooth/train_dreambooth_lora.py * Apply suggestions from code review * Apply suggestions from code review * add lora mixin when fit * add lora mixin when fit * add lora mixin when fit * fix more * fix more --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
Jason C.H authored
-
Sayak Paul authored
-
Sayak Paul authored
* feat: add lora attention processor for pt 2.0. * explicit context manager for SDPA. * switch to flash attention * make shapes compatible to work optimally with SDPA. * fix: circular import problem. * explicitly specify the flash attention kernel in sdpa * fall back to efficient attention context manager. * remove explicit dispatch. * fix: removed processor. * fix: remove optional from type annotation. * feat: make changes regarding LoRAAttnProcessor2_0. * remove confusing warning. * formatting. * relax tolerance for PT 2.0 * fix: loading message. * remove unnecessary logging. * add: entry to the docs. * add: network_alpha argument. * relax tolerance.
-
Takuma Mori authored
* merge undoable-monkeypatch * remove TEXT_ENCODER_TARGET_MODULES, refactoring * move create_lora_weight_file
-
- 05 Jun, 2023 1 commit
-
-
Vladislav Lyubimov authored
-
- 02 Jun, 2023 2 commits
-
-
Lachlan Nicholson authored
* iterate over unique tokens to avoid duplicate replacements * added test for multiple references to multi embedding * adhere to black formatting * reorder test post-rebase
-
Takuma Mori authored
* add _convert_kohya_lora_to_diffusers * make style * add scaffold * match result: unet attention only * fix monkey-patch for text_encoder * with CLIPAttention While the terrible images are no longer produced, the results do not match those from the hook ver. This may be due to not setting the network_alpha value. * add to support network_alpha * generate diff image * fix monkey-patch for text_encoder * add test_text_encoder_lora_monkey_patch() * verify that it's okay to release the attn_procs * fix closure version * add comment * Revert "fix monkey-patch for text_encoder" This reverts commit bb9c61e6faecc1935c9c4319c77065837655d616. * Fix to reuse utility functions * make LoRAAttnProcessor targets to self_attn * fix LoRAAttnProcessor target * make style * fix split key * Update src/diffusers/loaders.py * remove TEXT_ENCODER_TARGET_MODULES loop * add print memory usage * remove test_kohya_loras_scaffold.py * add: doc on LoRA civitai * remove print statement and refactor in the doc. * fix state_dict test for kohya-ss style lora * Apply suggestions from code review Co-authored-by:
Takuma Mori <takuma104@gmail.com> --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
- 30 May, 2023 1 commit
-
-
Greg Hunkins authored
* enable state dict for textual inversion loader * Empty-Commit | restart CI * Empty-Commit | restart CI * Empty-Commit | restart CI * Empty-Commit | restart CI * add tests * fix tests * fix tests * fix tests --------- Co-authored-by:Patrick von Platen <patrick.v.platen@gmail.com>
-
- 26 May, 2023 2 commits
-
-
Takuma Mori authored
Fix to apply LoRAXFormersAttnProcessor instead of LoRAAttnProcessor when xFormers is enabled (#3556) * fix to use LoRAXFormersAttnProcessor * add test * using new LoraLoaderMixin.save_lora_weights * add test_lora_save_load_with_xformers
-
Emin Demirci authored
-