1. 01 Dec, 2023 1 commit
  2. 27 Nov, 2023 1 commit
  3. 23 Nov, 2023 2 commits
  4. 21 Nov, 2023 2 commits
    • Patrick von Platen's avatar
      [Lora] Seperate logic (#5809) · 13d73d93
      Patrick von Platen authored
      * [Lora] Seperate logic
      
      * [Lora] Seperate logic
      
      * [Lora] Seperate logic
      
      * add comments to explain the code better
      
      * add comments to explain the code better
      13d73d93
    • Linoy Tsaban's avatar
      Add features to the Dreambooth LoRA SDXL training script (#5508) · 6fac1369
      Linoy Tsaban authored
      
      
      * Additions:
      - support for different lr for text encoder
      - support for Prodigy optimizer
      - support for min snr gamma
      - support for custom captions and dataset loading from the hub
      
      * adjusted --caption_column behaviour (to -not- use the second column of the dataset by default if --caption_column is not provided)
      
      * fixed --output_dir / --model_dir_name confusion
      
      * added --repeats, --adam_weight_decay_text_encoder
      + some fixes
      
      * Update examples/dreambooth/train_dreambooth_lora_sdxl.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update examples/dreambooth/train_dreambooth_lora_sdxl.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Update examples/dreambooth/train_dreambooth_lora_sdxl.py
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * - import compute_snr from diffusers/training_utils.py
      - cluster adamw together
      - when using 'prodigy', if --train_text_encoder == True and --text_encoder_lr != --learning rate, changes the lr of the text encoders optimization params to be --learning_rate (otherwise errors)
      
      * shape fixes when custom captions are used
      
      * formatting and a little cleanup
      
      * code styling
      
      * --repeats default value fixed, changed to 1
      
      * bug fix - removed redundant lines of embedding concatenation when using prior_preservation (that duplicated class_prompt embeddings)
      
      * changed dataset loading logic according to the following usecases (to avoid unnecessary dependency on datasets)-
      1. user provides --dataset_name
      2. user provides local dir --instance_data_dir that contains a metadata .jsonl file
      3. user provides local dir --instance_data_dir that contains only images
      in cases [1,2] we import datasets and use load_dataset method, in case [3] we process the data same as in the original script setting
      
      * styling fix
      
      * arg name fix
      
      * adjusted the --repeats logic
      
      * -removed redundant arg and 'if' when loading local folder with prompts
      -updated readme template
      -some default val fixes
      -custom caption tests
      
      * image path fix for readme
      
      * code style
      
      * bug fix
      
      * --caption_column arg
      
      * readme fix
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarLinoy Tsaban <linoy@huggingface.co>
      6fac1369
  5. 17 Nov, 2023 1 commit
  6. 14 Nov, 2023 1 commit
    • Sayak Paul's avatar
      [Refactor] refactor `loaders.py` to make it cleaner and leaner. (#5771) · ded93f79
      Sayak Paul authored
      
      
      * refactor loaders.py to make it cleaner and leaner.
      
      * refactor loaders init
      
      * inits.
      
      * textual inversion to the init.
      
      * inits.
      
      * remove certain modules from the main init.
      
      * AttnProcsLayers
      
      * fix imports
      
      * avoid circular import.
      
      * fix circular import pt 2.
      
      * address PR comments
      
      * imports
      
      * fix: imports.
      
      * remove from main init for avoiding circular deps.
      
      * remove spurious deps.
      
      * fix-copies.
      
      * fix imports.
      
      * more debug
      
      * more debug
      
      * Apply suggestions from code review
      
      * Apply suggestions from code review
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      ded93f79
  7. 10 Nov, 2023 1 commit
  8. 06 Nov, 2023 1 commit
  9. 27 Oct, 2023 1 commit
  10. 23 Oct, 2023 2 commits
  11. 16 Oct, 2023 1 commit
  12. 11 Oct, 2023 1 commit
  13. 02 Oct, 2023 1 commit
    • Sayak Paul's avatar
      fix: how print training resume logs. (#5117) · d56825e4
      Sayak Paul authored
      
      
      * fix: how print training resume logs.
      
      * propagate changes to text-to-image scripts.
      
      * propagate changes to instructpix2pix.
      
      * propagate changes to dreambooth
      
      * propagate changes to custom diffusion and instructpix2pix
      
      * propagate changes to kandinsky
      
      * propagate changes to textual inv.
      
      * debug
      
      * fix: checkpointing.
      
      * debug
      
      * debug
      
      * debug
      
      * back to the square
      
      * debug
      
      * debug
      
      * change condition order.
      
      * debug
      
      * debug
      
      * debug
      
      * debug
      
      * revert to original
      
      * clean
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      d56825e4
  14. 14 Sep, 2023 1 commit
  15. 08 Sep, 2023 1 commit
  16. 25 Aug, 2023 1 commit
  17. 17 Aug, 2023 1 commit
  18. 07 Aug, 2023 3 commits
  19. 04 Aug, 2023 1 commit
  20. 03 Aug, 2023 1 commit
  21. 28 Jul, 2023 1 commit
    • Sayak Paul's avatar
      [Feat] Support SDXL Kohya-style LoRA (#4287) · 4a4cdd6b
      Sayak Paul authored
      
      
      * sdxl lora changes.
      
      * better name replacement.
      
      * better replacement.
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * remove print.
      
      * print state dict keys.
      
      * print
      
      * distingisuih better
      
      * debuggable.
      
      * fxi: tyests
      
      * fix: arg from training script.
      
      * access from class.
      
      * run style
      
      * debug
      
      * save intermediate
      
      * some simplifications for SDXL LoRA
      
      * styling
      
      * unet config is not needed in diffusers format.
      
      * fix: dynamic SGM block mapping for SDXL kohya loras (#4322)
      
      * Use lora compatible layers for linear proj_in/proj_out (#4323)
      
      * improve condition for using the sgm_diffusers mapping
      
      * informative comment.
      
      * load compatible keys and embedding layer maaping.
      
      * Get SDXL 1.0 example lora to load
      
      * simplify
      
      * specif ranks and hidden sizes.
      
      * better handling of k rank and hidden
      
      * debug
      
      * debug
      
      * debug
      
      * debug
      
      * debug
      
      * fix: alpha keys
      
      * add check for handling LoRAAttnAddedKVProcessor
      
      * sanity comment
      
      * modifications for text encoder SDXL
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * denugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * up
      
      * up
      
      * up
      
      * up
      
      * up
      
      * up
      
      * unneeded comments.
      
      * unneeded comments.
      
      * kwargs for the other attention processors.
      
      * kwargs for the other attention processors.
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * improve
      
      * debugging
      
      * debugging
      
      * more print
      
      * Fix alphas
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * clean up
      
      * clean up.
      
      * debugging
      
      * fix: text
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarBatuhan Taskaya <batuhan@python.org>
      4a4cdd6b
  22. 27 Jul, 2023 1 commit
  23. 26 Jul, 2023 2 commits
  24. 25 Jul, 2023 2 commits
    • Sayak Paul's avatar
      [SDXL DreamBooth LoRA] multiple fixes (#4262) · 161449d5
      Sayak Paul authored
      * add automatic licensing.
      
      * debugging
      
      * debugging
      
      * more debugging
      
      * more debugging.
      
      * run make fix-copies.
      
      * change to default tracker.
      161449d5
    • Sayak Paul's avatar
      [SDXL DreamBooth LoRA] add support for text encoder fine-tuning (#4097) · 365e8461
      Sayak Paul authored
      
      
      * Allow low precision sd xl
      
      * finish
      
      * finish
      
      * feat: initial draft for supporting text encoder lora finetuning for SDXL DreamBooth
      
      * fix: variable assignments.
      
      * add: autocast block.
      
      * add debugging
      
      * vae dtype hell
      
      * fix: vae dtype hell.
      
      * fix: vae dtype hell 3.
      
      * clean up
      
      * lora text encoder loader.
      
      * fix: unwrapping models.
      
      * add: tests.
      
      * docs.
      
      * handle unexpected keys.
      
      * fix vae dtype in the final inference.
      
      * fix scope problem.
      
      * fix: save_model_card args.
      
      * initialize: prefix to None.
      
      * fix: dtype issues.
      
      * apply gixes.
      
      * debgging.
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * debugging
      
      * add: fast tests.
      
      * pre-tokenize.
      
      * address: will's comments.
      
      * fix: loader and tests.
      
      * fix: dataloader.
      
      * simplify dataloader.
      
      * length.
      
      * simplification.
      
      * make style && make quality
      
      * simplify state_dict munging
      
      * fix: tests.
      
      * fix: state_dict packing.
      
      * Apply suggestions from code review
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      365e8461
  25. 11 Jul, 2023 2 commits
  26. 09 Jul, 2023 2 commits
    • Patrick von Platen's avatar
      make style · 4a3e5748
      Patrick von Platen authored
      4a3e5748
    • Will Berman's avatar
      Refactor LoRA (#3778) · c2a28c34
      Will Berman authored
      
      
      * refactor to support patching LoRA into T5
      
      instantiate the lora linear layer on the same device as the regular linear layer
      
      get lora rank from state dict
      
      tests
      
      fmt
      
      can create lora layer in float32 even when rest of model is float16
      
      fix loading model hook
      
      remove load_lora_weights_ and T5 dispatching
      
      remove Unet#attn_processors_state_dict
      
      docstrings
      
      * text encoder monkeypatch class method
      
      * fix test
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      c2a28c34
  27. 07 Jul, 2023 1 commit
  28. 15 Jun, 2023 2 commits
  29. 08 Jun, 2023 2 commits