- 24 Feb, 2025 3 commits
-
-
hlky authored
Co-authored-by:puhuk <wetr235@gmail.com>
-
Sayak Paul authored
* restruct certain keys to be checked for peft config update. * updates * finish./ * finish 2. * updates
-
hlky authored
* Fix `torch_dtype` in Kolors text encoder with `transformers` v4.49 * Default torch_dtype and warning
-
- 22 Feb, 2025 2 commits
-
-
Daniel Regado authored
* More robust from_pretrained init_kwargs type checking * Corrected for Python 3.10 * Type checks subclasses and fixed type warnings * More type corrections and skip tokenizer type checking * make style && make quality * Updated docs and types for Lumina pipelines * Fixed check for empty signature * changed location of helper functions * make style --------- Co-authored-by:hlky <hlky@hlky.ac>
-
Steven Liu authored
* lora * update * update --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
- 21 Feb, 2025 9 commits
-
-
Marc Sun authored
remove check
-
SahilCarterr authored
Fix docs
-
hlky authored
* `device_map` in `load_model_dict_into_meta` * _LOW_CPU_MEM_USAGE_DEFAULT * fix is_peft_version is_bitsandbytes_version
-
Dhruv Nair authored
* update * update --------- Co-authored-by:Sayak Paul <spsayakpaul@gmail.com>
-
Dhruv Nair authored
update
-
Sayak Paul authored
fix
-
Sayak Paul authored
template for remote vae.
-
Sayak Paul authored
* fix: run tests from a pr workflow. * correct * update * checking.
-
Aryan authored
* update * make fix-copies * update * tests * update * update * add co-author Co-Authored-By:
Langdx <82783347+Langdx@users.noreply.github.com> * add co-author Co-Authored-By:
howe <howezhang2018@gmail.com> * update --------- Co-authored-by:
Langdx <82783347+Langdx@users.noreply.github.com> Co-authored-by:
howe <howezhang2018@gmail.com>
-
- 20 Feb, 2025 13 commits
-
-
Aryan authored
* update * update
-
Daniel Regado authored
* Added runtime checkpoint conversion * Updated docs * Fix for quantized model
-
Sayak Paul authored
* run fast gpu tests conditionally on pull requests. * revert unneeded changes. * simplify PR.
-
Sayak Paul authored
install accelerate transformers from .
-
Parag Ekbote authored
Add 5 Notebooks, improve their example scripts and update the missing links for the example README.
-
Haoyun Qin authored
-
Marc Sun authored
* store cls instead of an obj * style
-
Sayak Paul authored
* poc encode_prompt() tests * fix * updates. * fixes * fixes * updates * updates * updates * revert * updates * updates * updates * updates * remove SDXLOptionalComponentsTesterMixin. * remove tests that directly leveraged encode_prompt() in some way or the other. * fix imports. * remove _save_load * fixes * fixes * fixes * fixes
-
AstraliteHeart authored
-
Sayak Paul authored
* add; utility to check if attn_procs,norms,acts are properly documented. * add support listing to the workflows. * change to 2024. * small fixes. * does adding detailed docstrings help? * uncomment image processor check * quality * fix, thanks to @mishig. * Apply suggestions from code review Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * style * JointAttnProcessor2_0 * fixes * fixes * fixes * fixes * fixes * fixes * Update docs/source/en/api/normalization.md Co-authored-by:
hlky <hlky@hlky.ac> --------- Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> Co-authored-by:
hlky <hlky@hlky.ac>
-
Sayak Paul authored
* feat: lora support for Lumina2. * fix-copies. * updates * updates * docs. * fix * add: training script. * tests * updates * updates * major updates. * updates * fixes * docs. * updates * updates
-
Sayak Paul authored
use proper gemma class and config in lumina2 tests.
-
Aryan authored
remove prints
-
- 19 Feb, 2025 4 commits
-
-
Sayak Paul authored
* feat: introduce a style bot. * updates * Apply suggestions from code review Co-authored-by:
Guillaume LEGENDRE <glegendre01@gmail.com> * apply suggestion * fixes * updates --------- Co-authored-by:
Guillaume LEGENDRE <glegendre01@gmail.com>
-
Marc Sun authored
* first draft model loading refactor * revert name change * fix bnb * revert name * fix dduf * fix huanyan * style * Update src/diffusers/models/model_loading_utils.py Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * suggestions from reviews * Update src/diffusers/models/modeling_utils.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * remove safetensors check * fix default value * more fix from suggestions * revert logic for single file * style * typing + fix couple of issues * improve speed * Update src/diffusers/models/modeling_utils.py Co-authored-by:
Aryan <aryan@huggingface.co> * fp8 dtype * add tests * rename resolved_archive_file to resolved_model_file * format * map_location default cpu * add utility function * switch to smaller model + test inference * Apply suggestions from code review Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * rm comment * add log * Apply suggestions from code review Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * add decorator * cosine sim instead * fix use_keep_in_fp32_modules * comm --------- Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> Co-authored-by:
YiYi Xu <yixu310@gmail.com> Co-authored-by:
Aryan <aryan@huggingface.co>
-
Sayak Paul authored
* make set_adapters() robust on silent failures. * fixes to tests * flaky decorator. * fix * flaky to sd3. * remove warning. * sort * quality * skip test_simple_inference_with_text_denoiser_multi_adapter_block_lora * skip testing unsupported features. * raise warning instead of error.
-
hlky authored
DiffusionPipeline mixin `to`+FromOriginalModelMixin/FromSingleFileMixin `from_single_file` type hint (#10811) * DiffusionPipeline mixin `to` type hint * FromOriginalModelMixin from_single_file * FromSingleFileMixin from_single_file
-
- 18 Feb, 2025 2 commits
-
-
Sayak Paul authored
add missing entries to the lora docs.
-
puhuk authored
This PR updates the max_shift value in flux to 1.15 for consistency across the codebase. In addition to modifying max_shift in flux, all related functions that copy and use this logic, such as calculate_shift in `src/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3_img2img.py`, have also been updated to ensure uniform behavior.
-
- 17 Feb, 2025 2 commits
-
-
Sayak Paul authored
update lora support for flux.
-
Sayak Paul authored
fix
-
- 16 Feb, 2025 2 commits
-
-
Parag Ekbote authored
* Add support for callback_on_step_end for AuraFlowPipeline and LuminaText2ImgPipeline. * Apply the suggestions from code review for lumina and auraflow Co-authored-by:
hlky <hlky@hlky.ac> * Update missing inputs and imports. * Add input field. * Apply suggestions from code review-2 Co-authored-by:
hlky <hlky@hlky.ac> * Apply the suggestions from review for unused imports. Co-authored-by:
hlky <hlky@hlky.ac> * make style. * Update pipeline_aura_flow.py * Update pipeline_lumina.py * Update pipeline_lumina.py * Update pipeline_aura_flow.py * Update pipeline_lumina.py --------- Co-authored-by:
hlky <hlky@hlky.ac>
-
Yaniv Galron authored
-
- 15 Feb, 2025 2 commits
-
-
Yuxuan Zhang authored
* init * encode with glm * draft schedule * feat(scheduler): Add CogView scheduler implementation * feat(embeddings): add CogView 2D rotary positional embedding * 1 * Update pipeline_cogview4.py * fix the timestep init and sigma * update latent * draft patch(not work) * fix * [WIP][cogview4]: implement initial CogView4 pipeline Implement the basic CogView4 pipeline structure with the following changes: - Add CogView4 pipeline implementation - Implement DDIM scheduler for CogView4 - Add CogView3Plus transformer architecture - Update embedding models Current limitations: - CFG implementation uses padding for sequence length alignment - Need to verify transformer inference alignment with Megatron TODO: - Consider separate forward passes for condition/uncondition instead of padding approach * [WIP][cogview4][refactor]: Split condition/uncondition forward pass in CogView4 pipeline Split the forward pass for conditional and unconditional predictions in the CogView4 pipeline to match the original implementation. The noise prediction is now done separately for each case before combining them for guidance. However, the results still need improvement. This is a work in progress as the generated images are not yet matching expected quality. * use with -2 hidden state * remove text_projector * 1 * [WIP] Add tensor-reload to align input from transformer block * [WIP] for older glm * use with cogview4 transformers forward twice of u and uc * Update convert_cogview4_to_diffusers.py * remove this * use main example * change back * reset * setback * back * back 4 * Fix qkv conversion logic for CogView4 to Diffusers format * back5 * revert to sat to cogview4 version * update a new convert from megatron * [WIP][cogview4]: implement CogView4 attention processor Add CogView4AttnProcessor class for implementing scaled dot-product attention with rotary embeddings for the CogVideoX model. This processor concatenates encoder and hidden states, applies QKV projections and RoPE, but does not include spatial normalization. TODO: - Fix incorrect QKV projection weights - Resolve ~25% error in RoPE implementation compared to Megatron * [cogview4] implement CogView4 transformer block Implement CogView4 transformer block following the Megatron architecture: - Add multi-modulate and multi-gate mechanisms for adaptive layer normalization - Implement dual-stream attention with encoder-decoder structure - Add feed-forward network with GELU activation - Support rotary position embeddings for image tokens The implementation follows the original CogView4 architecture while adapting it to work within the diffusers framework. * with new attn * [bugfix] fix dimension mismatch in CogView4 attention * [cogview4][WIP]: update final normalization in CogView4 transformer Refactored the final normalization layer in CogView4 transformer to use separate layernorm and AdaLN operations instead of combined AdaLayerNormContinuous. This matches the original implementation but needs validation. Needs verification against reference implementation. * 1 * put back * Update transformer_cogview4.py * change time_shift * Update pipeline_cogview4.py * change timesteps * fix * change text_encoder_id * [cogview4][rope] align RoPE implementation with Megatron - Implement apply_rope method in attention processor to match Megatron's implementation - Update position embeddings to ensure compatibility with Megatron-style rotary embeddings - Ensure consistent rotary position encoding across attention layers This change improves compatibility with Megatron-based models and provides better alignment with the original implementation's positional encoding approach. * [cogview4][bugfix] apply silu activation to time embeddings in CogView4 Applied silu activation to time embeddings before splitting into conditional and unconditional parts in CogView4Transformer2DModel. This matches the original implementation and helps ensure correct time conditioning behavior. * [cogview4][chore] clean up pipeline code - Remove commented out code and debug statements - Remove unused retrieve_timesteps function - Clean up code formatting and documentation This commit focuses on code cleanup in the CogView4 pipeline implementation, removing unnecessary commented code and improving readability without changing functionality. * [cogview4][scheduler] Implement CogView4 scheduler and pipeline * now It work * add timestep * batch * change convert scipt * refactor pt. 1; make style * refactor pt. 2 * refactor pt. 3 * add tests * make fix-copies * update toctree.yml * use flow match scheduler instead of custom * remove scheduling_cogview.py * add tiktoken to test dependencies * Update src/diffusers/models/embeddings.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> * apply suggestions from review * use diffusers apply_rotary_emb * update flow match scheduler to accept timesteps * fix comment * apply review sugestions * Update src/diffusers/schedulers/scheduling_flow_match_euler_discrete.py Co-authored-by:
YiYi Xu <yixu310@gmail.com> --------- Co-authored-by:
三洋三洋 <1258009915@qq.com> Co-authored-by:
OleehyO <leehy0357@gmail.com> Co-authored-by:
Aryan <aryan@huggingface.co> Co-authored-by:
YiYi Xu <yixu310@gmail.com>
-
YiYi Xu authored
* up
-
- 14 Feb, 2025 1 commit
-
-
SahilCarterr authored
-