1. 13 Feb, 2025 1 commit
    • Aryan's avatar
      Disable PEFT input autocast when using fp8 layerwise casting (#10685) · a0c22997
      Aryan authored
      * disable peft input autocast
      
      * use new peft method name; only disable peft input autocast if submodule layerwise casting active
      
      * add test; reference PeftInputAutocastDisableHook in peft docs
      
      * add load_lora_weights test
      
      * casted -> cast
      
      * Update tests/lora/utils.py
      a0c22997
  2. 22 Jan, 2025 1 commit
    • Aryan's avatar
      [core] Layerwise Upcasting (#10347) · beacaa55
      Aryan authored
      
      
      * update
      
      * update
      
      * make style
      
      * remove dynamo disable
      
      * add coauthor
      Co-Authored-By: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      * update
      
      * update
      
      * update
      
      * update mixin
      
      * add some basic tests
      
      * update
      
      * update
      
      * non_blocking
      
      * improvements
      
      * update
      
      * norm.* -> norm
      
      * apply suggestions from review
      
      * add example
      
      * update hook implementation to the latest changes from pyramid attention broadcast
      
      * deinitialize should raise an error
      
      * update doc page
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * update docs
      
      * update
      
      * refactor
      
      * fix _always_upcast_modules for asym ae and vq_model
      
      * fix lumina embedding forward to not depend on weight dtype
      
      * refactor tests
      
      * add simple lora inference tests
      
      * _always_upcast_modules -> _precision_sensitive_module_patterns
      
      * remove todo comments about review; revert changes to self.dtype in unets because .dtype on ModelMixin should be able to handle fp8 weight case
      
      * check layer dtypes in lora test
      
      * fix UNet1DModelTests::test_layerwise_upcasting_inference
      
      * _precision_sensitive_module_patterns -> _skip_layerwise_casting_patterns based on feedback
      
      * skip test in NCSNppModelTests
      
      * skip tests for AutoencoderTinyTests
      
      * skip tests for AutoencoderOobleckTests
      
      * skip tests for UNet1DModelTests - unsupported pytorch operations
      
      * layerwise_upcasting -> layerwise_casting
      
      * skip tests for UNetRLModelTests; needs next pytorch release for currently unimplemented operation support
      
      * add layerwise fp8 pipeline test
      
      * use xfail
      
      * Apply suggestions from code review
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      
      * add assertion with fp32 comparison; add tolerance to fp8-fp32 vs fp32-fp32 comparison (required for a few models' test to pass)
      
      * add note about memory consumption on tesla CI runner for failing test
      
      ---------
      Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      beacaa55
  3. 10 Jan, 2025 2 commits
  4. 07 Jan, 2025 1 commit
  5. 06 Jan, 2025 2 commits
  6. 02 Jan, 2025 1 commit
  7. 25 Dec, 2024 1 commit
  8. 23 Dec, 2024 5 commits
  9. 20 Dec, 2024 2 commits
  10. 19 Dec, 2024 2 commits
  11. 18 Dec, 2024 2 commits
  12. 17 Dec, 2024 1 commit
  13. 15 Dec, 2024 1 commit
  14. 12 Dec, 2024 1 commit
  15. 10 Dec, 2024 1 commit
  16. 05 Dec, 2024 1 commit
  17. 22 Nov, 2024 1 commit
  18. 20 Nov, 2024 2 commits
  19. 19 Nov, 2024 1 commit
  20. 05 Nov, 2024 1 commit
  21. 02 Nov, 2024 1 commit
  22. 24 Oct, 2024 1 commit
  23. 16 Oct, 2024 1 commit
  24. 14 Oct, 2024 1 commit
  25. 13 Oct, 2024 1 commit
  26. 10 Oct, 2024 1 commit
  27. 09 Oct, 2024 1 commit
  28. 08 Oct, 2024 1 commit
    • Sayak Paul's avatar
      [LoRA] Handle DoRA better (#9547) · 02eeb8e7
      Sayak Paul authored
      * handle dora.
      
      * print test
      
      * debug
      
      * fix
      
      * fix-copies
      
      * update logits
      
      * add warning in the test.
      
      * make is_dora check consistent.
      
      * fix-copies
      02eeb8e7
  29. 07 Oct, 2024 1 commit
  30. 30 Sep, 2024 1 commit