- 04 Apr, 2019 1 commit
-
-
mcarilli authored
* Refactor to allow more flexible treatment of multiple optimizers/models/losses * Adding _process_optimizers.py * Created L0 tests (now passing). * fix: minor print typo (#234) * make L1 results easier to read * L0 multiple model/optimizer/loss test fleshed out * Adding test that master params remain synced across distributed processes * Docstring updates * Docstring updates
-
- 22 Mar, 2019 1 commit
-
-
mcarilli authored
* Adding Torch + bare-metal nvcc version check and container build tests * Putting a canary in the coalmine * canary proved elusive * Trying direct setup.py install * this should work * Removing canary * hopefully this works
-
- 19 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 13 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 12 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 10 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 08 Mar, 2019 3 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
- 07 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 02 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 01 Mar, 2019 4 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
- 28 Feb, 2019 1 commit
-
-
Michael Carilli authored
-
- 26 Feb, 2019 1 commit
-
-
Michael Carilli authored
-
- 24 Feb, 2019 1 commit
-
-
Michael Carilli authored
-
- 22 Feb, 2019 1 commit
-
-
Michael Carilli authored
Allow multi-tensor unscale to handle FP16 output, so it can also be used for copy-scatter. Rename some options.
-
- 19 Feb, 2019 1 commit
-
-
Michael Carilli authored
-
- 16 Feb, 2019 3 commits
- 13 Feb, 2019 1 commit
-
-
Michael Carilli authored
-
- 08 Feb, 2019 2 commits
-
-
Evgeni Krimer authored
-
Evgeni Krimer authored
-
- 06 Feb, 2019 1 commit
-
-
Michael Carilli authored
-
- 05 Feb, 2019 1 commit
-
-
Jerry Ma authored
This commit adds an FP16Model class as a successor to network_to_half. The benefits of this class are: - Preservation of single-precision for BatchNorm layers. The models generated by network_to_half() convert BatchNorm moment tensors to half-precision, then back to single-precision, which hurts the accuracy of the moment estimators and occasionally results in NaNs. - Support for multi-argument nn.Modules (self-explanatory from code).
-
- 03 Feb, 2019 1 commit
-
-
Michael Carilli authored
-
- 01 Feb, 2019 1 commit
-
-
Michael Carilli authored
-
- 29 Jan, 2019 3 commits
- 28 Jan, 2019 1 commit
-
-
jiej authored
test update to resolve https://github.com/NVIDIA/apex/issues/134#issue-403525480 Using identical learning rate for both DDP with sync BN and single process BN. The previous configure leaves the impression that sync BN requires adjusting lr in the script, which is not true.
-
- 25 Jan, 2019 1 commit
-
-
Michael Carilli authored
-
- 15 Jan, 2019 1 commit
-
-
Jie authored
Added kernel to support sync BN for channel last tensor
-
- 15 Dec, 2018 1 commit
-
-
Deyu Fu authored
-
- 01 Nov, 2018 1 commit
-
-
Michael Carilli authored
-
- 30 Oct, 2018 1 commit
-
-
ngimel authored
* Add unittest for FusedAdam. * Fix some bugs. * set seed for adam test
-
- 29 Oct, 2018 1 commit
-
-
mcarilli authored
* test passes * notes * Using C++-side flatten and unflatten functions * Adding csrc * Persistent synchronization event so it doesn't need to be created and destroyed each time * Interop with parameter flattening in SSD * Added deterministic option to imagenet main.py * Adding options to split gradient averaging and allreduce in pure fp32 * Fixing allreduce_maybe_retain call * Fixing allreduce_fallback * Also sync active_i_buckets from rank 0 * Making retain_allreduce_buffers compatible with/orthogonal to delay_allreduce=True|False * Correcting syntax error, now all seems to work with SSD * Optional cpp extension build * Add mixed precision adam optimizer (#59) * Add FusedAdam Optimizer to Apex that places all the math into a cuda kernel. * Added fixes to fused_adam to get it to work with network. * wip work on python interface for adam with options * fix dispatch for halfs, add python options to handle optional half gradients and params * cleanup, get rid of grid-stride loop
-