- 12 Dec, 2024 2 commits
-
-
Aryan authored
* transformer * make style & make fix-copies * transformer * add transformer tests * 80% vae * make style * make fix-copies * fix * undo cogvideox changes * update * update * match vae * add docs * t2v pipeline working; scheduler needs to be checked * docs * add pipeline test * update * update * make fix-copies * Apply suggestions from code review Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * update * copy t2v to i2v pipeline * update * apply review suggestions * update * make style * remove framewise encoding/decoding * pack/unpack latents * image2video * update * make fix-copies * update * update * rope scale fix * debug layerwise code * remove debug * Apply suggestions from code review Co-authored-by:
YiYi Xu <yixu310@gmail.com> * propagate precision changes to i2v pipeline * remove downcast * address review comments * fix comment * address review comments * [Single File] LTX support for loading original weights (#10135) * from original file mixin for ltx * undo config mapping fn changes * update * add single file to pipelines * update docs * Update src/diffusers/models/autoencoders/autoencoder_kl_ltx.py * Update src/diffusers/models/autoencoders/autoencoder_kl_ltx.py * rename classes based on ltx review * point to original repository for inference * make style * resolve conflicts correctly --------- Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com>
-
Canva authored
* Add support for XFormers in SD3 * sd3 xformers test * sd3 xformers quality * sd3 xformers update --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com>
-
- 06 Dec, 2024 2 commits
-
-
Juan Acevedo authored
* update ptxla example --------- Co-authored-by:
Juan Acevedo <jfacevedo@google.com> Co-authored-by:
Pei Zhang <zpcore@gmail.com> Co-authored-by:
Pei Zhang <piz@google.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
Pei Zhang <pei@Peis-MacBook-Pro.local> Co-authored-by:
hlky <hlky@hlky.ac>
-
Junsong Chen authored
* first add a script for DC-AE; * DC-AE init * replace triton with custom implementation * 1. rename file and remove un-used codes; * no longer rely on omegaconf and dataclass * replace custom activation with diffuers activation * remove dc_ae attention in attention_processor.py * iinherit from ModelMixin * inherit from ConfigMixin * dc-ae reduce to one file * update downsample and upsample * clean code * support DecoderOutput * remove get_same_padding and val2tuple * remove autocast and some assert * update ResBlock * remove contents within super().__init__ * Update src/diffusers/models/autoencoders/dc_ae.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * remove opsequential * update other blocks to support the removal of build_norm * remove build encoder/decoder project in/out * remove inheritance of RMSNorm2d from LayerNorm * remove reset_parameters for RMSNorm2d Co-authored-by:
YiYi Xu <yixu310@gmail.com> * remove device and dtype in RMSNorm2d __init__ Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/models/autoencoders/dc_ae.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/models/autoencoders/dc_ae.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/models/autoencoders/dc_ae.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * remove op_list & build_block * remove build_stage_main * change file name to autoencoder_dc * move LiteMLA to attention.py * align with other vae decode output; * add DC-AE into init files; * update * make quality && make style; * quick push before dgx disappears again * update * make style * update * update * fix * refactor * refactor * refactor * update * possibly change to nn.Linear * refactor * make fix-copies * replace vae with ae * replace get_block_from_block_type to get_block * replace downsample_block_type from Conv to conv for consistency * add scaling factors * incorporate changes for all checkpoints * make style * move mla to attention processor file; split qkv conv to linears * refactor * add tests * from original file loader * add docs * add standard autoencoder methods * combine attention processor * fix tests * update * minor fix * minor fix * minor fix & in/out shortcut rename * minor fix * make style * fix paper link * update docs * update single file loading * make style * remove single file loading support; todo for DN6 * Apply suggestions from code review Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * add abstract --------- Co-authored-by:
Junyu Chen <chenjydl2003@gmail.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com> Co-authored-by:
chenjy2003 <70215701+chenjy2003@users.noreply.github.com> Co-authored-by:
Aryan <aryan@huggingface.co> Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
- 05 Dec, 2024 1 commit
-
-
YiYi Xu authored
-
- 03 Dec, 2024 1 commit
-
-
Benjamin Paine authored
* fix progress bar updates in SD 1.5 PAG Img2Img pipeline --------- Co-authored-by:
Vinh H. Pham <phamvinh257@gmail.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
- 18 Nov, 2024 1 commit
-
-
ちくわぶ authored
Add all AttnProcessor in `AttentionProcessor` type
-
- 09 Nov, 2024 1 commit
-
-
Eliseu Silva authored
* Feature IP Adapter Xformers Attention Processor: this fix error loading incorrect attention processor when setting Xformers attn after load ip adapter scale, issues: #8863 #8872
-
- 05 Nov, 2024 1 commit
-
-
Aryan authored
* update * udpate * update transformer * make style * fix * add conversion script * update * fix * update * fix * update * fixes * make style * update * update * update * init * update * update * add * up * up * up * update * mochi transformer * remove original implementation * make style * update inits * update conversion script * docs * Update src/diffusers/pipelines/mochi/pipeline_mochi.py Co-authored-by:
Dhruv Nair <dhruv.nair@gmail.com> * Update src/diffusers/pipelines/mochi/pipeline_mochi.py Co-authored-by:
Dhruv Nair <dhruv.nair@gmail.com> * fix docs * pipeline fixes * make style * invert sigmas in scheduler; fix pipeline * fix pipeline num_frames * flip proj and gate in swiglu * make style * fix * make style * fix tests * latent mean and std fix * update * cherry-pick 1069d210e1b9e84a366cdc7a13965626ea258178 * remove additional sigma already handled by flow match scheduler * fix * remove hardcoded value * replace conv1x1 with linear * Update src/diffusers/pipelines/mochi/pipeline_mochi.py Co-authored-by:
Dhruv Nair <dhruv.nair@gmail.com> * framewise decoding and conv_cache * make style * Apply suggestions from code review * mochi vae encoder changes * rebase correctly * Update scripts/convert_mochi_to_diffusers.py * fix tests * fixes * make style * update * make style * update * add framewise and tiled encoding * make style * make original vae implementation behaviour the default; note: framewise encoding does not work * remove framewise encoding implementation due to presence of attn layers * fight test 1 * fight test 2 --------- Co-authored-by:
Dhruv Nair <dhruv.nair@gmail.com> Co-authored-by:
yiyixuxu <yixu310@gmail.com>
-
- 01 Nov, 2024 1 commit
-
-
Leo Jiang authored
* NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX * NPU implementation for FLUX --------- Co-authored-by:蒋硕 <jiangshuo9@h-partners.com>
-
- 29 Oct, 2024 1 commit
-
-
Aryan authored
* update * refactor transformer part 1 * refactor part 2 * refactor part 3 * make style * refactor part 4; modeling tests * make style * refactor part 5 * refactor part 6 * gradient checkpointing * pipeline tests (broken atm) * update * add coauthor Co-Authored-By:
Huan Yang <hyang@fastmail.com> * refactor part 7 * add docs * make style * add coauthor Co-Authored-By:
YiYi Xu <yixu310@gmail.com> * make fix-copies * undo unrelated change * revert changes to embeddings, normalization, transformer * refactor part 8 * make style * refactor part 9 * make style * fix * apply suggestions from review * Apply suggestions from code review Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * update example * remove attention mask for self-attention * update * copied from * update * update --------- Co-authored-by:
Huan Yang <hyang@fastmail.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com> Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
- 21 Oct, 2024 1 commit
-
-
YiYi Xu authored
* update some docs and tests! --------- Co-authored-by:
Aryan <contact.aryanvs@gmail.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
Aryan <aryan@huggingface.co> Co-authored-by:
apolinário <joaopaulo.passos@gmail.com>
-
- 14 Oct, 2024 1 commit
-
-
Yuxuan.Zhang authored
* merge 9588 * max_shard_size="5GB" for colab running * conversion script updates; modeling test; refactor transformer * make fix-copies * Update convert_cogview3_to_diffusers.py * initial pipeline draft * make style * fight bugs
🐛 🪳 * add example * add tests; refactor * make style * make fix-copies * add co-author YiYi Xu <yixu310@gmail.com> * remove files * add docs * add co-author Co-Authored-By:YiYi Xu <yixu310@gmail.com> * fight docs * address reviews * make style * make model work * remove qkv fusion * remove qkv fusion tets * address review comments * fix make fix-copies error * remove None and TODO * for FP16(draft) * make style * remove dynamic cfg * remove pooled_projection_dim as a parameter * fix tests --------- Co-authored-by:
Aryan <aryan@huggingface.co> Co-authored-by:
YiYi Xu <yixu310@gmail.com>
-
- 23 Aug, 2024 2 commits
-
-
Sayak Paul authored
* start fusing flux. * test * finish fusion * fix-copues
-
zR authored
* draft of embedding --------- Co-authored-by:Aryan <aryan@huggingface.co>
-
- 21 Aug, 2024 1 commit
-
-
YiYi Xu authored
* refactor rotary embeds * adding jsmidt as co-author of this PR for https://github.com/huggingface/diffusers/pull/9133 --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
Joseph Smidt <josephsmidt@gmail.com>
-
- 06 Aug, 2024 2 commits
-
-
Ahn Donghoon (안동훈 / suno) authored
add pag sd3 --------- Co-authored-by:
HyoungwonCho <jhw9811@korea.ac.kr> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
crepejung00 <jaewoojung00@naver.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com> Co-authored-by:
Aryan <contact.aryanvs@gmail.com> Co-authored-by:
Aryan <aryan@huggingface.co>
-
Sayak Paul authored
* add fusion support to pixart * add to auraflow. * add tests * apply review feedback. * add back args and kwargs * style
-
- 05 Aug, 2024 1 commit
-
-
Aryan authored
* copy hunyuandit pipeline * pag variant of hunyuan dit * add tests * update docs * make style * make fix-copies * Update src/diffusers/pipelines/pag/pag_utils.py * remove incorrect copied from * remove pag hunyuan attn procs to resolve conflicts * add pag attn procs again * new implementation for pag_utils * revert pag changes * add pag refactor back; update pixart sigma * update pixart pag tests * apply suggestions from review Co-Authored-By: yixu310@gmail.com * make style * update docs, fix tests * fix tests * fix test_components_function since list not accepted as valid __init__ param * apply patch to fix broken tests Co-Authored-By:
Sayak Paul <spsayakpaul@gmail.com> * make style * fix hunyuan tests --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
- 04 Aug, 2024 2 commits
-
-
Aryan authored
* apply qk norm in attention processors * revert attention processor * qk-norm in only attention proc 2.0 and fused variant
-
psychedelicious authored
`None` is valid for `get_attention_scores`, should be typed as such
-
- 01 Aug, 2024 1 commit
-
-
Sayak Paul authored
add flux! Signed-off-by:
Adrien <adrien@huggingface.co> Co-authored-by:
Adrien <adrien.69740@gmail.com> Co-authored-by:
Anatoly Belikov <abelikov@singularitynet.io> Co-authored-by:
Dhruv Nair <dhruv.nair@gmail.com> Co-authored-by:
yiyixuxu <yixu310@gmail.com>
-
- 30 Jul, 2024 1 commit
-
-
Yoach Lacombe authored
* WIP modeling code and pipeline * add custom attention processor + custom activation + add to init * correct ProjectionModel forward * add stable audio to __initèè * add autoencoder and update pipeline and modeling code * add half Rope * add partial rotary v2 * add temporary modfis to scheduler * add EDM DPM Solver * remove TODOs * clean GLU * remove att.group_norm to attn processor * revert back src/diffusers/schedulers/scheduling_dpmsolver_multistep.py * refactor GLU -> SwiGLU * remove redundant args * add channel multiples in autoencoder docstrings * changes in docsrtings and copyright headers * clean pipeline * further cleaning * remove peft and lora and fromoriginalmodel * Delete src/diffusers/pipelines/stable_audio/diffusers.code-workspace * make style * dummy models * fix copied from * add fast oobleck tests * add brownian tree * oobleck autoencoder slow tests * remove TODO * fast stable audio pipeline tests * add slow tests * make style * add first version of docs * wrap is_torchsde_available to the scheduler * fix slow test * test with input waveform * add input waveform * remove some todos * create stableaudio gaussian projection + make style * add pipeline to toctree * fix copied from * make quality * refactor timestep_features->time_proj * refactor joint_attention_kwargs->cross_attention_kwargs * remove forward_chunk * move StableAudioDitModel to transformers folder * correct convert + remove partial rotary embed * apply suggestions from yiyixuxu -> removing attn.kv_heads * remove temb * remove cross_attention_kwargs * further removal of cross_attention_kwargs * remove text encoder autocast to fp16 * continue removing autocast * make style * refactor how text and audio are embedded * add paper * update example code * make style * unify projection model forward + fix device placement * make style * remove fuse qkv * apply suggestions from review * Update src/diffusers/pipelines/stable_audio/pipeline_stable_audio.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * make style * smaller models in fast tests * pass sequential offloading fast tests * add docs for vae and autoencoder * make style and update example * remove useless import * add cosine scheduler * dummy classes * cosine scheduler docs * better description of scheduler --------- Co-authored-by:
YiYi Xu <yixu310@gmail.com>
-
- 26 Jul, 2024 1 commit
-
-
Aryan authored
-
- 24 Jul, 2024 2 commits
-
-
Sayak Paul authored
* remove residual i. * rename to aura_flow in pipeline test
-
Sayak Paul authored
* start debugging the problem, * start * fix * fix * fix imports. * handle hunyuan * remove residuals. * add a check for making sure there's appropriate procs. * add more rigor to the tests. * fix test * remove redundant check * fix-copies * move check_qkv_fusion_matches_attn_procs_length and check_qkv_fusion_processors_exist.
-
- 20 Jul, 2024 1 commit
-
-
shinetzh authored
* fix loop bug in SlicedAttnProcessor --------- Co-authored-by:neoshang <neoshang@tencent.com>
-
- 11 Jul, 2024 1 commit
-
-
Sayak Paul authored
* add lavender flow transformer --------- Co-authored-by:YiYi Xu <yixu310@gmail.com>
-
- 08 Jul, 2024 2 commits
-
-
Tolga Cangöz authored
* Remove unused line --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
PommesPeter authored
--------- Co-authored-by:
zhuole1025 <zhuole1025@gmail.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com>
-
- 03 Jul, 2024 1 commit
-
-
Sayak Paul authored
add dummy lora attention processors to prevent failures in other libs
-
- 25 Jun, 2024 1 commit
-
-
YiYi Xu authored
* first draft --------- Co-authored-by:
yiyixuxu <yixu310@gmail,com> Co-authored-by:
Junhwa Song <ethan9867@gmail.com> Co-authored-by:
Ahn Donghoon (안동훈 / suno) <suno.vivid@gmail.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
- 21 Jun, 2024 1 commit
-
-
Sayak Paul authored
* get rid of the legacy lora remnants and make our codebase lighter * fix depcrecated lora argument * fix * empty commit to trigger ci * remove print * empty
-
- 13 Jun, 2024 1 commit
-
-
ちくわぶ authored
* Fix duplicate variable assignments. * Fix duplicate variable assignments.
-
- 12 Jun, 2024 2 commits
-
-
Dhruv Nair authored
* up * add sd3 * update * update * add tests * fix copies * fix docs * update * add dreambooth lora * add LoRA * update * update * update * update * import fix * update * Update src/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * import fix 2 * update * Update src/diffusers/models/autoencoders/autoencoder_kl.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/models/autoencoders/autoencoder_kl.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/models/autoencoders/autoencoder_kl.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/models/autoencoders/autoencoder_kl.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/models/autoencoders/autoencoder_kl.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/models/autoencoders/autoencoder_kl.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/models/autoencoders/autoencoder_kl.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/models/autoencoders/autoencoder_kl.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/models/autoencoders/autoencoder_kl.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/models/autoencoders/autoencoder_kl.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/models/autoencoders/autoencoder_kl.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * update * update * update * fix ckpt id * fix more ids * update * missing doc * Update src/diffusers/schedulers/scheduling_flow_match_euler_discrete.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update src/diffusers/schedulers/scheduling_flow_match_euler_discrete.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * Update docs/source/en/api/pipelines/stable_diffusion/stable_diffusion_3.md Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * Update docs/source/en/api/pipelines/stable_diffusion/stable_diffusion_3.md Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * update' * fix * update * Update src/diffusers/models/autoencoders/autoencoder_kl.py * Update src/diffusers/models/autoencoders/autoencoder_kl.py * note on gated access. * requirements * licensing --------- Co-authored-by:
sayakpaul <spsayakpaul@gmail.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com>
-
Greg Hunkins authored
* quiet attn parameters * fix lint * make style && make quality --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
- 01 Jun, 2024 1 commit
-
-
XCL authored
* Hunyuan Team: add HunyuanDiT related updates --------- Co-authored-by:
XCLiu <liuxc1996@gmail.com> Co-authored-by:
yiyixuxu <yixu310@gmail.com>
-
- 10 May, 2024 1 commit
-
-
Mark Van Aken authored
* find & replace all FloatTensors to Tensor * apply formatting * Update torch.FloatTensor to torch.Tensor in the remaining files * formatting * Fix the rest of the places where FloatTensor is used as well as in documentation * formatting * Update new file from FloatTensor to Tensor
-
- 03 May, 2024 1 commit
-
-
HelloWorldBeginner authored
Add Ascend NPU support for SDXL fine-tuning and fix the model saving bug when using DeepSpeed. (#7816) * Add Ascend NPU support for SDXL fine-tuning and fix the model saving bug when using DeepSpeed. * fix check code quality * Decouple the NPU flash attention and make it an independent module. * add doc and unit tests for npu flash attention. --------- Co-authored-by:
mhh001 <mahonghao1@huawei.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
- 22 Apr, 2024 1 commit
-
-
Jenyuan-Huang authored
* enable control ip-adapter per-transformer block on-the-fly --------- Co-authored-by:
sayakpaul <spsayakpaul@gmail.com> Co-authored-by:
ResearcherXman <xhs.research@gmail.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com>
-