• Linoy Tsaban's avatar
    [Flux] Add advanced training script + support textual inversion inference (#9434) · 9a7f8246
    Linoy Tsaban authored
    * add ostris trainer to README & add cache latents of vae
    
    * add ostris trainer to README & add cache latents of vae
    
    * style
    
    * readme
    
    * add test for latent caching
    
    * add ostris noise scheduler
    https://github.com/ostris/ai-toolkit/blob/9ee1ef2a0a2a9a02b92d114a95f21312e5906e54/toolkit/samplers/custom_flowmatch_sampler.py#L95
    
    * style
    
    * fix import
    
    * style
    
    * fix tests
    
    * style
    
    * --change upcasting of transformer?
    
    * update readme according to main
    
    * add pivotal tuning for CLIP
    
    * fix imports, encode_prompt call,add TextualInversionLoaderMixin to FluxPipeline for inference
    
    * TextualInversionLoaderMixin support for FluxPipeline for inference
    
    * move changes to advanced flux script, revert canonical
    
    * add latent caching to canonical script
    
    * revert changes to canonical script to keep it separate from https://github.com/huggingface/diffusers/pull/9160
    
    * revert changes to canonical script to keep it separate from https://github.com/huggingface/diffusers/pull/9160
    
    * style
    
    * remove redundant line and change code block placement to align with logic
    
    * add initializer_token arg
    
    * add transformer frac for range support from pure textual inversion to the orig pivotal tuning
    
    * support pure textual inversion - wip
    
    * adjustments to support pure textual inversion and transformer optimization in only part of the epochs
    
    * fix logic when using initializer token
    
    * fix pure_textual_inversion_condition
    
    * fix ti/pivotal loading of last validation run
    
    * remove embeddings loading for ti in final training run (to avoid adding huggingface hub dependency)
    
    * support pivotal for t5
    
    * adapt pivotal for T5 encoder
    
    * adapt pivotal for T5 encoder and support in flux pipeline
    
    * t5 pivotal support + support fo pivotal for clip only or both
    
    * fix param chaining
    
    * fix param chaining
    
    * README first draft
    
    * readme
    
    * readme
    
    * readme
    
    * style
    
    * fix import
    
    * style
    
    * add fix from https://github.com/huggingface/diffusers/pull/9419
    
    
    
    * add to readme, change function names
    
    * te lr changes
    
    * readme
    
    * change concept tokens logic
    
    * fix indices
    
    * change arg name
    
    * style
    
    * dummy test
    
    * revert dummy test
    
    * reorder pivoting
    
    * add warning in case the token abstraction is not the instance prompt
    
    * experimental - wip - specific block training
    
    * fix documentation and token abstraction processing
    
    * remove transformer block specification feature (for now)
    
    * style
    
    * fix copies
    
    * fix indexing issue when --initializer_concept has different amounts
    
    * add if TextualInversionLoaderMixin to all flux pipelines
    
    * style
    
    * fix import
    
    * fix imports
    
    * address review comments - remove necessary prints & comments, use pin_memory=True, use free_memory utils, unify warning and prints
    
    * style
    
    * logger info fix
    
    * make lora target modules configurable and change the default
    
    * make lora target modules configurable and change the default
    
    * style
    
    * make lora target modules configurable and change the default, add notes to readme
    
    * style
    
    * add tests
    
    * style
    
    * fix repo id
    
    * add updated requirements for advanced flux
    
    * fix indices of t5 pivotal tuning embeddings
    
    * fix path in test
    
    * remove `pin_memory`
    
    * fix filename of embedding
    
    * fix filename of embedding
    
    ---------
    Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
    Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
    9a7f8246
README_flux.md 18.8 KB