- 18 Jun, 2020 1 commit
-
-
rohithkrn authored
fix bf16 layernorm bug
-
- 15 Jun, 2020 1 commit
-
-
rohithkrn authored
-
- 08 Jun, 2020 1 commit
-
-
Ashish Farmer authored
IFU_06_05_2020
-
- 03 Jun, 2020 1 commit
-
-
rohithkrn authored
* bfloat16 support for apex DDP * enable mgpu tests for fp16 and bf16 * update Dockerfile
-
- 01 Jun, 2020 3 commits
-
-
mcarilli authored
Co-authored-by:Michael Carilli <mcarilli@nvidia.com>
-
Thor Johnsen authored
Remove distributed lamb from __init__.py
-
Thor Johnsen authored
-
- 31 May, 2020 6 commits
-
-
Thor Johnsen authored
Distributed lamb optimizer
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
- 30 May, 2020 19 commits
-
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
Thor Johnsen authored
-
- 29 May, 2020 6 commits
-
-
Kevin Stephano authored
-
Burc Eryilmaz authored
Fuses dropout and softmax in backward pass, add bias support to CPP MHA, add additive mask support, separate Q/K/V parameters (#854) Co-authored-by:Sukru Eryilmaz <seryilmaz@computelab-dgx1v-32.nvidia.com>
-
Kexin Yu authored
make FusedLAMB async
-
Peng authored
[Documentation] Update the readme for ROCm.
-
Chaitanya Sri Krishna Lolla authored
-
lcskrishna authored
-
- 28 May, 2020 2 commits
-
-
Peng authored
Enable bfloat16 for optimizers
-
Max V. Irgiznov authored
-