• Linoy Tsaban's avatar
    [LoRA] add LoRA support to HiDream and fine-tuning script (#11281) · e30d3bf5
    Linoy Tsaban authored
    
    
    * initial commit
    
    * initial commit
    
    * initial commit
    
    * initial commit
    
    * initial commit
    
    * initial commit
    
    * Update examples/dreambooth/train_dreambooth_lora_hidream.py
    Co-authored-by: default avatarBagheera <59658056+bghira@users.noreply.github.com>
    
    * move prompt embeds, pooled embeds outside
    
    * Update examples/dreambooth/train_dreambooth_lora_hidream.py
    Co-authored-by: default avatarhlky <hlky@hlky.ac>
    
    * Update examples/dreambooth/train_dreambooth_lora_hidream.py
    Co-authored-by: default avatarhlky <hlky@hlky.ac>
    
    * fix import
    
    * fix import and tokenizer 4, text encoder 4 loading
    
    * te
    
    * prompt embeds
    
    * fix naming
    
    * shapes
    
    * initial commit to add HiDreamImageLoraLoaderMixin
    
    * fix init
    
    * add tests
    
    * loader
    
    * fix model input
    
    * add code example to readme
    
    * fix default max length of text encoders
    
    * prints
    
    * nullify training cond in unpatchify for temp fix to incompatible shaping of transformer output during training
    
    * smol fix
    
    * unpatchify
    
    * unpatchify
    
    * fix validation
    
    * flip pred and loss
    
    * fix shift!!!
    
    * revert unpatchify changes (for now)
    
    * smol fix
    
    * Apply style fixes
    
    * workaround moe training
    
    * workaround moe training
    
    * remove prints
    
    * to reduce some memory, keep vae in `weight_dtype` same as we have for flux (as it's the same vae)
    https://github.com/huggingface/diffusers/blob/bbd0c161b55ba2234304f1e6325832dd69c60565/examples/dreambooth/train_dreambooth_lora_flux.py#L1207
    
    
    
    * refactor to align with HiDream refactor
    
    * refactor to align with HiDream refactor
    
    * refactor to align with HiDream refactor
    
    * add support for cpu offloading of text encoders
    
    * Apply style fixes
    
    * adjust lr and rank for train example
    
    * fix copies
    
    * Apply style fixes
    
    * update README
    
    * update README
    
    * update README
    
    * fix license
    
    * keep prompt2,3,4 as None in validation
    
    * remove reverse ode comment
    
    * Update examples/dreambooth/train_dreambooth_lora_hidream.py
    Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
    
    * Update examples/dreambooth/train_dreambooth_lora_hidream.py
    Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
    
    * vae offload change
    
    * fix text encoder offloading
    
    * Apply style fixes
    
    * cleaner to_kwargs
    
    * fix module name in copied from
    
    * add requirements
    
    * fix offloading
    
    * fix offloading
    
    * fix offloading
    
    * update transformers version in reqs
    
    * try AutoTokenizer
    
    * try AutoTokenizer
    
    * Apply style fixes
    
    * empty commit
    
    * Delete tests/lora/test_lora_layers_hidream.py
    
    * change tokenizer_4 to load with AutoTokenizer as well
    
    * make text_encoder_four and tokenizer_four configurable
    
    * save model card
    
    * save model card
    
    * revert T5
    
    * fix test
    
    * remove non diffusers lumina2 conversion
    
    ---------
    Co-authored-by: default avatarBagheera <59658056+bghira@users.noreply.github.com>
    Co-authored-by: default avatarhlky <hlky@hlky.ac>
    Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
    Co-authored-by: default avatargithub-actions[bot] <github-actions[bot]@users.noreply.github.com>
    e30d3bf5
lora_pipeline.py 260 KB