"magic_pdf/git@developer.sourcefind.cn:wangsen/mineru.git" did not exist on "3b5d3fc83f2e9894270c7aae2bca775da1c1bca6"
  1. 19 May, 2025 1 commit
  2. 13 May, 2025 1 commit
  3. 23 Apr, 2025 1 commit
  4. 21 Apr, 2025 1 commit
  5. 09 Apr, 2025 1 commit
  6. 04 Apr, 2025 1 commit
    • Suprhimp's avatar
      [feat]Add strength in flux_fill pipeline (denoising strength for fluxfill) (#10603) · 94f2c48d
      Suprhimp authored
      * [feat]add strength in flux_fill pipeline
      
      * Update src/diffusers/pipelines/flux/pipeline_flux_fill.py
      
      * Update src/diffusers/pipelines/flux/pipeline_flux_fill.py
      
      * Update src/diffusers/pipelines/flux/pipeline_flux_fill.py
      
      * [refactor] refactor after review
      
      * [fix] change comment
      
      * Apply style fixes
      
      * empty
      
      * fix
      
      * update prepare_latents from flux.img2img pipeline
      
      * style
      
      * Update src/diffusers/pipelines/flux/pipeline_flux_fill.py
      
      ---------
      94f2c48d
  7. 20 Mar, 2025 1 commit
  8. 18 Mar, 2025 1 commit
  9. 06 Mar, 2025 2 commits
  10. 03 Mar, 2025 1 commit
  11. 02 Mar, 2025 1 commit
  12. 25 Feb, 2025 1 commit
  13. 18 Feb, 2025 1 commit
    • puhuk's avatar
      Fix max_shift value in flux and related functions to 1.15 (issue #10675) (#10807) · b75b204a
      puhuk authored
      This PR updates the max_shift value in flux to 1.15 for consistency across the codebase. In addition to modifying max_shift in flux, all related functions that copy and use this logic, such as calculate_shift in `src/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3_img2img.py`, have also been updated to ensure uniform behavior.
      b75b204a
  14. 27 Jan, 2025 1 commit
  15. 23 Jan, 2025 1 commit
  16. 12 Jan, 2025 2 commits
  17. 09 Jan, 2025 1 commit
  18. 07 Jan, 2025 1 commit
  19. 21 Dec, 2024 2 commits
    • Mehmet Yiğit Özgenç's avatar
      flux controlnet inpaint config bug (#10291) · 233dffdc
      Mehmet Yiğit Özgenç authored
      
      
      * flux controlnet inpaint config bug
      
      * Update src/diffusers/pipelines/flux/pipeline_flux_controlnet_inpainting.py
      
      ---------
      Co-authored-by: default avataryigitozgenc <yigit@quantuslabs.ai>
      Co-authored-by: default avatarhlky <hlky@hlky.ac>
      233dffdc
    • hlky's avatar
      Support Flux IP Adapter (#10261) · be207099
      hlky authored
      
      
      * Flux IP-Adapter
      
      * test cfg
      
      * make style
      
      * temp remove copied from
      
      * fix test
      
      * fix test
      
      * v2
      
      * fix
      
      * make style
      
      * temp remove copied from
      
      * Apply suggestions from code review
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      
      * Move encoder_hid_proj to inside FluxTransformer2DModel
      
      * merge
      
      * separate encode_prompt, add copied from, image_encoder offload
      
      * make
      
      * fix test
      
      * fix
      
      * Update src/diffusers/pipelines/flux/pipeline_flux.py
      
      * test_flux_prompt_embeds change not needed
      
      * true_cfg -> true_cfg_scale
      
      * fix merge conflict
      
      * test_flux_ip_adapter_inference
      
      * add fast test
      
      * FluxIPAdapterMixin not test mixin
      
      * Update pipeline_flux.py
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      
      ---------
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      be207099
  20. 18 Dec, 2024 1 commit
  21. 04 Dec, 2024 1 commit
  22. 02 Dec, 2024 1 commit
  23. 23 Nov, 2024 1 commit
  24. 21 Nov, 2024 1 commit
  25. 20 Nov, 2024 1 commit
  26. 14 Nov, 2024 1 commit
    • Sam's avatar
      Update pipeline_flux_img2img.py (#9928) · 5c94937d
      Sam authored
      * Update pipeline_flux_img2img.py
      
      Added FromSingleFileMixin to this pipeline loader like the other FLUX pipelines.
      
      * Update pipeline_flux_img2img.py
      
      typo
      
      * modified:   src/diffusers/pipelines/flux/pipeline_flux_img2img.py
      5c94937d
  27. 08 Nov, 2024 1 commit
  28. 07 Nov, 2024 2 commits
  29. 25 Oct, 2024 2 commits
  30. 23 Oct, 2024 1 commit
  31. 21 Oct, 2024 1 commit
  32. 17 Oct, 2024 1 commit
    • Linoy Tsaban's avatar
      [Flux] Add advanced training script + support textual inversion inference (#9434) · 9a7f8246
      Linoy Tsaban authored
      * add ostris trainer to README & add cache latents of vae
      
      * add ostris trainer to README & add cache latents of vae
      
      * style
      
      * readme
      
      * add test for latent caching
      
      * add ostris noise scheduler
      https://github.com/ostris/ai-toolkit/blob/9ee1ef2a0a2a9a02b92d114a95f21312e5906e54/toolkit/samplers/custom_flowmatch_sampler.py#L95
      
      * style
      
      * fix import
      
      * style
      
      * fix tests
      
      * style
      
      * --change upcasting of transformer?
      
      * update readme according to main
      
      * add pivotal tuning for CLIP
      
      * fix imports, encode_prompt call,add TextualInversionLoaderMixin to FluxPipeline for inference
      
      * TextualInversionLoaderMixin support for FluxPipeline for inference
      
      * move changes to advanced flux script, revert canonical
      
      * add latent caching to canonical script
      
      * revert changes to canonical script to keep it separate from https://github.com/huggingface/diffusers/pull/9160
      
      * revert changes to canonical script to keep it separate from https://github.com/huggingface/diffusers/pull/9160
      
      * style
      
      * remove redundant line and change code block placement to align with logic
      
      * add initializer_token arg
      
      * add transformer frac for range support from pure textual inversion to the orig pivotal tuning
      
      * support pure textual inversion - wip
      
      * adjustments to support pure textual inversion and transformer optimization in only part of the epochs
      
      * fix logic when using initializer token
      
      * fix pure_textual_inversion_condition
      
      * fix ti/pivotal loading of last validation run
      
      * remove embeddings loading for ti in final training run (to avoid adding huggingface hub dependency)
      
      * support pivotal for t5
      
      * adapt pivotal for T5 encoder
      
      * adapt pivotal for T5 encoder and support in flux pipeline
      
      * t5 pivotal support + support fo pivotal for clip only or both
      
      * fix param chaining
      
      * fix param chaining
      
      * README first draft
      
      * readme
      
      * readme
      
      * readme
      
      * style
      
      * fix import
      
      * style
      
      * add fix from https://github.com/huggingface/diffusers/pull/9419
      
      
      
      * add to readme, change function names
      
      * te lr changes
      
      * readme
      
      * change concept tokens logic
      
      * fix indices
      
      * change arg name
      
      * style
      
      * dummy test
      
      * revert dummy test
      
      * reorder pivoting
      
      * add warning in case the token abstraction is not the instance prompt
      
      * experimental - wip - specific block training
      
      * fix documentation and token abstraction processing
      
      * remove transformer block specification feature (for now)
      
      * style
      
      * fix copies
      
      * fix indexing issue when --initializer_concept has different amounts
      
      * add if TextualInversionLoaderMixin to all flux pipelines
      
      * style
      
      * fix import
      
      * fix imports
      
      * address review comments - remove necessary prints & comments, use pin_memory=True, use free_memory utils, unify warning and prints
      
      * style
      
      * logger info fix
      
      * make lora target modules configurable and change the default
      
      * make lora target modules configurable and change the default
      
      * style
      
      * make lora target modules configurable and change the default, add notes to readme
      
      * style
      
      * add tests
      
      * style
      
      * fix repo id
      
      * add updated requirements for advanced flux
      
      * fix indices of t5 pivotal tuning embeddings
      
      * fix path in test
      
      * remove `pin_memory`
      
      * fix filename of embedding
      
      * fix filename of embedding
      
      ---------
      Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
      Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
      9a7f8246
  33. 15 Oct, 2024 1 commit
  34. 11 Oct, 2024 1 commit
  35. 10 Oct, 2024 1 commit