1. 23 May, 2020 1 commit
  2. 14 May, 2020 1 commit
  3. 30 Apr, 2020 1 commit
  4. 28 Apr, 2020 1 commit
  5. 20 Aug, 2019 1 commit
  6. 16 Aug, 2019 2 commits
    • Deyu Fu's avatar
      clean up variance options support by all fused optimizers: · 18062b69
      Deyu Fu authored
      correctly not apply bias correction to epsilon(same as recent upstream change)
      correctly not apply bias correction to weight decay(consistent with upstream AdamW)
      Make adam_w_mode for FusedAdam/LAMB, to do L2 or Weight Decay (Adam vs AdamW)
      Correct document reg_inside_moment differently from adam_w_mode in FusedNovoGrad
      Removed legacy eps_mode from FusedAdam
      Make internal math type float across fused optimizers
      18062b69
    • Deyu Fu's avatar
      add fused lamb, put lamb kernels into one file · c8f9cceb
      Deyu Fu authored
      c8f9cceb
  7. 08 Aug, 2019 1 commit
  8. 31 May, 2019 2 commits
  9. 10 May, 2019 1 commit
  10. 09 Apr, 2019 1 commit
  11. 04 Apr, 2019 1 commit
    • mcarilli's avatar
      WIP: Handle arbitrary combinations of optimizers/models/losses (#232) · 3f87614f
      mcarilli authored
      * Refactor to allow more flexible treatment of multiple optimizers/models/losses
      
      * Adding _process_optimizers.py
      
      * Created L0 tests (now passing).
      
      * fix: minor print typo (#234)
      
      * make L1 results easier to read
      
      * L0 multiple model/optimizer/loss test fleshed out
      
      * Adding test that master params remain synced across distributed processes
      
      * Docstring updates
      
      * Docstring updates
      3f87614f
  12. 19 Mar, 2019 1 commit
  13. 11 Mar, 2019 1 commit
  14. 10 Mar, 2019 1 commit
  15. 08 Mar, 2019 1 commit
  16. 19 Feb, 2019 1 commit