1. 12 May, 2020 1 commit
  2. 11 May, 2020 1 commit
  3. 06 May, 2020 2 commits
  4. 05 May, 2020 1 commit
  5. 04 May, 2020 1 commit
  6. 30 Apr, 2020 2 commits
  7. 29 Apr, 2020 1 commit
    • Samyam Rajbhandari's avatar
      CSR+FP32 fix (#206) · 6cb332f1
      Samyam Rajbhandari authored
      1) CSR parameter names should end with .weight. 
      2) When using basic optimizer directly, DeepSpeed should handle zero_grad. Letting the basic optimizer do the zero_grad resulted in residual gradients in the embedding layer due to unknown reasons.
      6cb332f1
  8. 27 Apr, 2020 1 commit
  9. 25 Apr, 2020 1 commit
  10. 24 Apr, 2020 1 commit
  11. 22 Apr, 2020 2 commits
  12. 21 Apr, 2020 1 commit
  13. 20 Apr, 2020 1 commit
  14. 16 Apr, 2020 1 commit
  15. 12 Apr, 2020 1 commit
  16. 10 Apr, 2020 1 commit
  17. 09 Apr, 2020 1 commit
  18. 07 Apr, 2020 1 commit
  19. 06 Apr, 2020 1 commit
  20. 03 Apr, 2020 1 commit
  21. 28 Mar, 2020 1 commit
  22. 27 Mar, 2020 2 commits
    • Olatunji Ruwase's avatar
      Support multi-output models (#170) · 53c73fe3
      Olatunji Ruwase authored
      * Push to remote
      
      * Correctly handle multi output models by doing loss scaling in backward()
      Unit tests for multi output models
      
      * Fix formatting issues
      
      * Formatting issues fix
      
      * Fix formatting
      
      * Update DeepSpeedExamples submodule
      Enable Megatron model tests
      53c73fe3
    • Calogero Zarbo's avatar
      Add "zero_allow_untested_optimizer" option in conf file (#173) · 43f27332
      Calogero Zarbo authored
      * added zero_allow_untested_optimizer flag helpers
      
      * add zero_allow_untested_optimizer config constants
      
      * zero_allow_untested_optimizer logic with assertion
      
      * Added unit test and CustomOptimizer helper class
      43f27332
  23. 26 Mar, 2020 1 commit
  24. 25 Mar, 2020 1 commit
  25. 23 Mar, 2020 1 commit
  26. 22 Mar, 2020 2 commits
  27. 18 Mar, 2020 4 commits
  28. 17 Mar, 2020 5 commits