1. 17 Oct, 2024 2 commits
    • Aryan's avatar
      [refactor] DiffusionPipeline.download (#9557) · 5704376d
      Aryan authored
      
      
      * update
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      5704376d
    • Linoy Tsaban's avatar
      [Flux] Add advanced training script + support textual inversion inference (#9434) · 9a7f8246
      Linoy Tsaban authored
      * add ostris trainer to README & add cache latents of vae
      
      * add ostris trainer to README & add cache latents of vae
      
      * style
      
      * readme
      
      * add test for latent caching
      
      * add ostris noise scheduler
      https://github.com/ostris/ai-toolkit/blob/9ee1ef2a0a2a9a02b92d114a95f21312e5906e54/toolkit/samplers/custom_flowmatch_sampler.py#L95
      
      * style
      
      * fix import
      
      * style
      
      * fix tests
      
      * style
      
      * --change upcasting of transformer?
      
      * update readme according to main
      
      * add pivotal tuning for CLIP
      
      * fix imports, encode_prompt call,add TextualInversionLoaderMixin to FluxPipeline for inference
      
      * TextualInversionLoaderMixin support for FluxPipeline for inference
      
      * move changes to advanced flux script, revert canonical
      
      * add latent caching to canonical script
      
      * revert changes to canonical script to keep it separate from https://github.com/huggingface/diffusers/pull/9160
      
      * revert changes to canonical script to keep it separate from https://github.com/huggingface/diffusers/pull/9160
      
      * style
      
      * remove redundant line and change code block placement to align with logic
      
      * add initializer_token arg
      
      * add transformer frac for range support from pure textual inversion to the orig pivotal tuning
      
      * support pure textual inversion - wip
      
      * adjustments to support pure textual inversion and transformer optimization in only part of the epochs
      
      * fix logic when using initializer token
      
      * fix pure_textual_inversion_condition
      
      * fix ti/pivotal loading of last validation run
      
      * remove embeddings loading for ti in final training run (to avoid adding huggingface hub dependency)
      
      * support pivotal for t5
      
      * adapt pivotal for T5 encoder
      
      * adapt pivotal for T5 encoder and support in flux pipeline
      
      * t5 pivotal support + support fo pivotal for clip only or both
      
      * fix param chaining
      
      * fix param chaining
      
      * README first draft
      
      * readme
      
      * readme
      
      * readme
      
      * style
      
      * fix import
      
      * style
      
      * add fix from https://github.com/huggingface/diffusers/pull/9419
      
      
      
      * add to readme, change function names
      
      * te lr changes
      
      * readme
      
      * change concept tokens logic
      
      * fix indices
      
      * change arg name
      
      * style
      
      * dummy test
      
      * revert dummy test
      
      * reorder pivoting
      
      * add warning in case the token abstraction is not the instance prompt
      
      * experimental - wip - specific block training
      
      * fix documentation and token abstraction processing
      
      * remove transformer block specification feature (for now)
      
      * style
      
      * fix copies
      
      * fix indexing issue when --initializer_concept has different amounts
      
      * add if TextualInversionLoaderMixin to all flux pipelines
      
      * style
      
      * fix import
      
      * fix imports
      
      * address review comments - remove necessary prints & comments, use pin_memory=True, use free_memory utils, unify warning and prints
      
      * style
      
      * logger info fix
      
      * make lora target modules configurable and change the default
      
      * make lora target modules configurable and change the default
      
      * style
      
      * make lora target modules configurable and change the default, add notes to readme
      
      * style
      
      * add tests
      
      * style
      
      * fix repo id
      
      * add updated requirements for advanced flux
      
      * fix indices of t5 pivotal tuning embeddings
      
      * fix path in test
      
      * remove `pin_memory`
      
      * fix filename of embedding
      
      * fix filename of embedding
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      9a7f8246
  2. 16 Oct, 2024 5 commits
  3. 15 Oct, 2024 12 commits
  4. 14 Oct, 2024 2 commits
    • SahilCarterr's avatar
      Added Lora Support to SD3 Img2Img Pipeline (#9659) · 22ed39f5
      SahilCarterr authored
      * add lora
      22ed39f5
    • Yuxuan.Zhang's avatar
      CogView3Plus DiT (#9570) · 8d81564b
      Yuxuan.Zhang authored
      * merge 9588
      
      * max_shard_size="5GB" for colab running
      
      * conversion script updates; modeling test; refactor transformer
      
      * make fix-copies
      
      * Update convert_cogview3_to_diffusers.py
      
      * initial pipeline draft
      
      * make style
      
      * fight bugs 🐛
      
      🪳
      
      * add example
      
      * add tests; refactor
      
      * make style
      
      * make fix-copies
      
      * add co-author
      
      YiYi Xu <yixu310@gmail.com>
      
      * remove files
      
      * add docs
      
      * add co-author
      Co-Authored-By: default avatarYiYi Xu <yixu310@gmail.com>
      
      * fight docs
      
      * address reviews
      
      * make style
      
      * make model work
      
      * remove qkv fusion
      
      * remove qkv fusion tets
      
      * address review comments
      
      * fix make fix-copies error
      
      * remove None and TODO
      
      * for FP16(draft)
      
      * make style
      
      * remove dynamic cfg
      
      * remove pooled_projection_dim as a parameter
      
      * fix tests
      
      ---------
      Co-authored-by: default avatarAryan <aryan@huggingface.co>
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      8d81564b
  5. 11 Oct, 2024 1 commit
  6. 10 Oct, 2024 1 commit
  7. 09 Oct, 2024 4 commits
  8. 08 Oct, 2024 3 commits
  9. 07 Oct, 2024 3 commits
  10. 03 Oct, 2024 1 commit
  11. 02 Oct, 2024 2 commits
  12. 01 Oct, 2024 1 commit
  13. 30 Sep, 2024 2 commits
  14. 28 Sep, 2024 1 commit