1. 13 Aug, 2019 1 commit
    • Deyu Fu's avatar
      Reverse to Fused* naming, clean up accordingly: · 007c5947
      Deyu Fu authored
      FusedSGD now work as before
      FusedAdam now work with o1/o2, no longer fuse scaling and casting
      Removed special backend handling for FusedAdam
      Moved and updated test for FusedAdam into run_optimizers
      Removed legacy tests for optimizers.FP16_optimizer and FusedAdam in run_mixed_adam
      007c5947
  2. 10 Apr, 2019 1 commit
  3. 26 Feb, 2019 1 commit
  4. 05 Feb, 2019 1 commit
    • Jerry Ma's avatar
      Better FP16 support in pytorch fp16 utils. · 713e0fb8
      Jerry Ma authored
      This commit adds an FP16Model class as a successor to network_to_half.
      
      The benefits of this class are:
      
      - Preservation of single-precision for BatchNorm layers. The models
        generated by network_to_half() convert BatchNorm moment tensors to
        half-precision, then back to single-precision, which hurts the
        accuracy of the moment estimators and occasionally results in NaNs.
      - Support for multi-argument nn.Modules (self-explanatory from code).
      713e0fb8
  5. 01 Feb, 2019 1 commit
  6. 30 Oct, 2018 1 commit
    • ngimel's avatar
      Adam tests (#67) · d594826c
      ngimel authored
      * Add unittest for FusedAdam.
      
      * Fix some bugs.
      
      * set seed for adam test
      d594826c
  7. 13 Sep, 2018 1 commit