1. 02 Jun, 2025 1 commit
  2. 28 May, 2025 1 commit
  3. 23 May, 2025 1 commit
  4. 19 May, 2025 2 commits
  5. 15 May, 2025 1 commit
  6. 01 May, 2025 1 commit
  7. 08 Apr, 2025 2 commits
  8. 24 Mar, 2025 1 commit
  9. 14 Feb, 2025 1 commit
    • Aryan's avatar
      Module Group Offloading (#10503) · 9a147b82
      Aryan authored
      
      
      * update
      
      * fix
      
      * non_blocking; handle parameters and buffers
      
      * update
      
      * Group offloading with cuda stream prefetching (#10516)
      
      * cuda stream prefetch
      
      * remove breakpoints
      
      * update
      
      * copy model hook implementation from pab
      
      * update; ~very workaround based implementation but it seems to work as expected; needs cleanup and rewrite
      
      * more workarounds to make it actually work
      
      * cleanup
      
      * rewrite
      
      * update
      
      * make sure to sync current stream before overwriting with pinned params
      
      not doing so will lead to erroneous computations on the GPU and cause bad results
      
      * better check
      
      * update
      
      * remove hook implementation to not deal with merge conflict
      
      * re-add hook changes
      
      * why use more memory when less memory do trick
      
      * why still use slightly more memory when less memory do trick
      
      * optimise
      
      * add model tests
      
      * add pipeline tests
      
      * update docs
      
      * add layernorm and groupnorm
      
      * address review comments
      
      * improve tests; add docs
      
      * improve docs
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * apply suggestions from code review
      
      * update tests
      
      * apply suggestions from review
      
      * enable_group_offloading -> enable_group_offload for naming consistency
      
      * raise errors if multiple offloading strategies used; add relevant tests
      
      * handle .to() when group offload applied
      
      * refactor some repeated code
      
      * remove unintentional change from merge conflict
      
      * handle .cuda()
      
      ---------
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      9a147b82
  10. 23 Jan, 2025 1 commit
  11. 22 Jan, 2025 1 commit
    • Aryan's avatar
      [core] Layerwise Upcasting (#10347) · beacaa55
      Aryan authored
      
      
      * update
      
      * update
      
      * make style
      
      * remove dynamo disable
      
      * add coauthor
      Co-Authored-By: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      * update
      
      * update
      
      * update
      
      * update mixin
      
      * add some basic tests
      
      * update
      
      * update
      
      * non_blocking
      
      * improvements
      
      * update
      
      * norm.* -> norm
      
      * apply suggestions from review
      
      * add example
      
      * update hook implementation to the latest changes from pyramid attention broadcast
      
      * deinitialize should raise an error
      
      * update doc page
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * update docs
      
      * update
      
      * refactor
      
      * fix _always_upcast_modules for asym ae and vq_model
      
      * fix lumina embedding forward to not depend on weight dtype
      
      * refactor tests
      
      * add simple lora inference tests
      
      * _always_upcast_modules -> _precision_sensitive_module_patterns
      
      * remove todo comments about review; revert changes to self.dtype in unets because .dtype on ModelMixin should be able to handle fp8 weight case
      
      * check layer dtypes in lora test
      
      * fix UNet1DModelTests::test_layerwise_upcasting_inference
      
      * _precision_sensitive_module_patterns -> _skip_layerwise_casting_patterns based on feedback
      
      * skip test in NCSNppModelTests
      
      * skip tests for AutoencoderTinyTests
      
      * skip tests for AutoencoderOobleckTests
      
      * skip tests for UNet1DModelTests - unsupported pytorch operations
      
      * layerwise_upcasting -> layerwise_casting
      
      * skip tests for UNetRLModelTests; needs next pytorch release for currently unimplemented operation support
      
      * add layerwise fp8 pipeline test
      
      * use xfail
      
      * Apply suggestions from code review
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      * add assertion with fp32 comparison; add tolerance to fp8-fp32 vs fp32-fp32 comparison (required for a few models' test to pass)
      
      * add note about memory consumption on tesla CI runner for failing test
      
      ---------
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      beacaa55
  12. 16 Jan, 2025 1 commit
  13. 25 Oct, 2024 1 commit
  14. 12 Oct, 2024 1 commit
  15. 23 Sep, 2024 1 commit
  16. 16 Sep, 2024 1 commit
  17. 09 Sep, 2024 1 commit
  18. 08 Aug, 2024 1 commit
  19. 05 Jun, 2024 1 commit
    • Tolga Cangöz's avatar
      Errata (#8322) · 98730c5d
      Tolga Cangöz authored
      * Fix typos
      
      * Trim trailing whitespaces
      
      * Remove a trailing whitespace
      
      * chore: Update MarigoldDepthPipeline checkpoint to prs-eth/marigold-lcm-v1-0
      
      * Revert "chore: Update MarigoldDepthPipeline checkpoint to prs-eth/marigold-lcm-v1-0"
      
      This reverts commit fd742b30b4258106008a6af4d0dd4664904f8595.
      
      * pokemon -> naruto
      
      * `DPMSolverMultistep` -> `DPMSolverMultistepScheduler`
      
      * Improve Markdown stylization
      
      * Improve style
      
      * Improve style
      
      * Refactor pipeline variable names for consistency
      
      * up style
      98730c5d
  20. 24 May, 2024 1 commit
  21. 20 May, 2024 1 commit
  22. 10 May, 2024 1 commit
    • Mark Van Aken's avatar
      #7535 Update FloatTensor type hints to Tensor (#7883) · be4afa0b
      Mark Van Aken authored
      * find & replace all FloatTensors to Tensor
      
      * apply formatting
      
      * Update torch.FloatTensor to torch.Tensor in the remaining files
      
      * formatting
      
      * Fix the rest of the places where FloatTensor is used as well as in documentation
      
      * formatting
      
      * Update new file from FloatTensor to Tensor
      be4afa0b
  23. 06 May, 2024 1 commit
  24. 23 Apr, 2024 1 commit
  25. 17 Apr, 2024 2 commits
  26. 25 Feb, 2024 1 commit
  27. 08 Feb, 2024 1 commit
  28. 05 Jan, 2024 1 commit
  29. 16 Nov, 2023 1 commit
  30. 09 Nov, 2023 1 commit
    • M. Tolga Cangöz's avatar
      [`Docs`] Fix typos and update files at Optimization Page (#5674) · 53a8439f
      M. Tolga Cangöz authored
      
      
      * Fix typos, update, trim trailing whitespace
      
      * Trim trailing whitespaces
      
      * Update docs/source/en/optimization/memory.md
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * Update docs/source/en/optimization/memory.md
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * Update _toctree.yml
      
      * Update adapt_a_model.md
      
      * Reverse
      
      * Reverse
      
      * Reverse
      
      * Update dreambooth.md
      
      * Update instructpix2pix.md
      
      * Update lora.md
      
      * Update overview.md
      
      * Update t2i_adapters.md
      
      * Update text2image.md
      
      * Update text_inversion.md
      
      * Update create_dataset.md
      
      * Update create_dataset.md
      
      * Update create_dataset.md
      
      * Update create_dataset.md
      
      * Update coreml.md
      
      * Delete docs/source/en/training/create_dataset.md
      
      * Original create_dataset.md
      
      * Update create_dataset.md
      
      * Delete docs/source/en/training/create_dataset.md
      
      * Add original file
      
      * Delete docs/source/en/training/create_dataset.md
      
      * Add original one
      
      * Delete docs/source/en/training/text2image.md
      
      * Delete docs/source/en/training/instructpix2pix.md
      
      * Delete docs/source/en/training/dreambooth.md
      
      * Add original files
      
      ---------
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      53a8439f
  31. 31 Oct, 2023 1 commit
    • M. Tolga Cangöz's avatar
      [Docs] Fix typos (#5583) · 442017cc
      M. Tolga Cangöz authored
      * Add Copyright info
      
      * Fix typos, improve, update
      
      * Update deepfloyd_if.md
      
      * Update ldm3d_diffusion.md
      
      * Update opt_overview.md
      442017cc
  32. 16 Oct, 2023 1 commit
  33. 27 Sep, 2023 1 commit
  34. 13 Sep, 2023 1 commit
  35. 10 Aug, 2023 2 commits
  36. 02 Aug, 2023 1 commit