- 24 Jun, 2019 1 commit
-
-
mcarilli authored
-
- 09 May, 2019 1 commit
-
-
Tim Zaman authored
-
- 30 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 18 Apr, 2019 1 commit
-
-
Glenn Jocher authored
-
- 12 Mar, 2019 1 commit
-
-
mcarilli authored
-
- 07 Mar, 2019 2 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
- 04 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 01 Mar, 2019 2 commits
- 28 Feb, 2019 1 commit
-
-
vfdev authored
-
- 20 Feb, 2019 5 commits
- 28 Jan, 2019 1 commit
-
-
mcarilli authored
-
- 31 Oct, 2018 1 commit
-
-
Thor Johnsen authored
* Pre-release of fused layer norm apex extension * Remove half and __half2 specializations * Code changes from review
-
- 30 Oct, 2018 1 commit
-
-
Michael Carilli authored
-
- 23 Oct, 2018 1 commit
-
-
jjsjann123 authored
* [syncBN] added syncBN in native pure python apex added fused cuda kernels used for sync BN. Using welford for mean/var optional installation using 'python setup.py install --cuda_ext' added unit test with side to side comparison between apex sync BN with PyTorch BN. Notice that for pytorch BN implementation, because of numerical issue for mean/var, the output will be slightly off. * [syncBN PR] added fp16 support addressing review comments on: 1. updating last pow 2 2. look for import error when importing syncBN kernel * [syncBN PR] added convert function to insert SyncBatchNorm refactored some kernel code * fixing type issue (fp16/fp32/fp64) added Kahan summation editing unit test to use pytorch primitive ops with double, passing reasonable tests now * updating tensor creation calls * fixing the all_reduce contiguous tensor * transposed all reduce results * [syncBN] support fp16 input & fp32 layer for apex fp16 partially fixing launch configs enabling imagenet example to run with --sync_bn * [syncBN PR] Documentation added * adjusting README * adjusting again * added some doc to imagenet example * [syncBN] warp-level reduction bug fix: warp reduction logic updated. check for dummy element to avoid nan. improved launch config for better reduction kernels. Further improvements would be to increase grid size. * [syncBN] fixing undefined behavior in __shfl_down_sync from divergent threads in warp reduction. changing at::native::empty to at::empty (upstream comments)
-
- 08 Oct, 2018 1 commit
-
-
mcarilli authored
-
- 22 Aug, 2018 1 commit
-
-
mcarilli authored
-
- 19 Aug, 2018 1 commit
-
-
Michael Carilli authored
-
- 05 Jul, 2018 1 commit
-
-
mcarilli authored
-
- 15 Jun, 2018 3 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
- 14 Jun, 2018 2 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
- 06 Jun, 2018 2 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
- 25 Apr, 2018 2 commits
-
-
Christian Sarofeen authored
-
Christian Sarofeen authored
-