1. 11 Feb, 2021 1 commit
  2. 03 Feb, 2021 1 commit
  3. 02 Feb, 2021 1 commit
  4. 29 Jan, 2021 1 commit
  5. 07 Jan, 2021 1 commit
  6. 05 Jan, 2021 1 commit
  7. 04 Jan, 2021 2 commits
    • Benjamin Lefaudeux's avatar
      [chore] 0.1.2 version bump (#285) · a21f50f9
      Benjamin Lefaudeux authored
      a21f50f9
    • Min Xu's avatar
      [feat] sync adascale from internal repo, support add_param_group (#266) · 3932a1f6
      Min Xu authored
      * [feat] sync adascale from internal repo
      
      - tbd
      
      testing: tbd
      
      * Update argument document of __init__
      
      * update documentation around set_num_gradients_to_accumulate
      
      * added checking code for proper API calling places
      
      * rename internal APIs to make them internal
      
      * updated changelog
      
      * added support for add_param_group and its unit test
      
      * added unit test for set_num_gradients_to_accumulate
      
      * added debias_ewma unit test
      
      * fixed test_set_num_gradients_to_accumulate (need zero_grad() call)
      
      * added missing zero_grad() to test_lr_scheduler
      
      * fixed test_add_param_group with respect to optim.zero_grad()
      
      * added test_gradient_value
      
      * added test_scale_not_equal_default for scale != world_size * grad_accum
      
      * added test_unhook()
      
      * removed print statements
      
      * fixed a typo
      
      * addressed Ben's comment
      3932a1f6
  8. 30 Dec, 2020 1 commit
  9. 24 Dec, 2020 1 commit
    • Min Xu's avatar
      [chore] Update changelog (#268) · 18455bf0
      Min Xu authored
      * Update changelog
      
      missed this item from previous AdaScale commit.
      
      * More change log
      
      * Addressed review comments
      18455bf0
  10. 03 Dec, 2020 1 commit
    • Min Xu's avatar
      [feat] AdaScale: Gradient Accumulation and Add PyTest unit tests (#202) · ce5860ea
      Min Xu authored
      * added AdaScale to README
      
      * [adascale] added gradient accumulation
      
      - added gradient accumulation
      - tested with cifar full trainings with different value of accumulation
      and verified the full accuracy is obtained
      - also removed the patch optimize flag until we need it
      
      * [adascale] adding pytest
      
      - added basic and ddp tests and grad_accum
      - closes #195
      
      * added changelog
      
      * added ddp grad_accum test
      
      * moved ddp and non-ddp tests into separate files
      
      * added checkpoint test
      
      * more doc
      
      * addressed Mike's comments
      ce5860ea
  11. 02 Dec, 2020 1 commit
  12. 01 Dec, 2020 1 commit
  13. 15 Oct, 2020 1 commit
  14. 28 Aug, 2020 1 commit
  15. 31 Jul, 2020 2 commits
  16. 08 Jul, 2020 1 commit