- 17 Dec, 2018 3 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
- 14 Dec, 2018 1 commit
-
-
Michael Carilli authored
-
- 12 Dec, 2018 1 commit
-
-
Michael Carilli authored
-
- 11 Dec, 2018 5 commits
- 10 Dec, 2018 2 commits
- 07 Dec, 2018 1 commit
-
-
Deyu Fu authored
-
- 06 Dec, 2018 1 commit
-
-
Deyu Fu authored
-
- 04 Dec, 2018 2 commits
- 03 Dec, 2018 2 commits
-
-
mcarilli authored
adjusted kernel config for better perf. removed divergence in welford warp reduction.
-
jjsjann123 authored
supporting user specified process group
-
- 02 Dec, 2018 1 commit
-
-
ptrblck authored
-
- 30 Nov, 2018 2 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
- 28 Nov, 2018 3 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
- 14 Nov, 2018 1 commit
-
-
mcarilli authored
-
- 10 Nov, 2018 1 commit
-
-
Michael Carilli authored
-
- 06 Nov, 2018 1 commit
-
-
Jie authored
adjusted kernel config for better perf. removed divergence in welford warp reduction.
-
- 01 Nov, 2018 4 commits
-
-
Michael Carilli authored
-
schetlur authored
* Adding some missing fields to adamopt documentation. * Adding some clarification to documentation.
-
Michael Carilli authored
-
Michael Carilli authored
-
- 31 Oct, 2018 1 commit
-
-
Thor Johnsen authored
* Pre-release of fused layer norm apex extension * Remove half and __half2 specializations * Code changes from review
-
- 30 Oct, 2018 7 commits
-
-
ngimel authored
-
mcarilli authored
-
ngimel authored
* Add unittest for FusedAdam. * Fix some bugs. * set seed for adam test
-
ngimel authored
update includes
-
Natalia Gimelshein authored
-
Michael Carilli authored
-
Michael Carilli authored
-
- 29 Oct, 2018 1 commit
-
-
mcarilli authored
* test passes * notes * Using C++-side flatten and unflatten functions * Adding csrc * Persistent synchronization event so it doesn't need to be created and destroyed each time * Interop with parameter flattening in SSD * Added deterministic option to imagenet main.py * Adding options to split gradient averaging and allreduce in pure fp32 * Fixing allreduce_maybe_retain call * Fixing allreduce_fallback * Also sync active_i_buckets from rank 0 * Making retain_allreduce_buffers compatible with/orthogonal to delay_allreduce=True|False * Correcting syntax error, now all seems to work with SSD * Optional cpp extension build * Add mixed precision adam optimizer (#59) * Add FusedAdam Optimizer to Apex that places all the math into a cuda kernel. * Added fixes to fused_adam to get it to work with network. * wip work on python interface for adam with options * fix dispatch for halfs, add python options to handle optional half gradients and params * cleanup, get rid of grid-stride loop
-