- 19 Dec, 2020 16 commits
-
-
Jared Casper authored
-
Jared Casper authored
-
Jared Casper authored
-
Jared Casper authored
-
Jared Casper authored
-
Jared Casper authored
-
Deepak Narayanan authored
-
mohammad authored
-
mohammad authored
-
mohammad authored
-
mohammad authored
-
mohammad authored
-
mohammad authored
-
mohammad authored
-
mohammad authored
Rename --batch-size to --micro-batch-size and drop in-minibatch from --num-micro-batches-in-minibatch
-
Jared Casper authored
-
- 03 Dec, 2020 1 commit
-
-
mohammad authored
-
- 02 Dec, 2020 3 commits
- 30 Nov, 2020 2 commits
- 29 Nov, 2020 2 commits
- 28 Nov, 2020 1 commit
-
-
mohammad authored
-
- 26 Nov, 2020 1 commit
-
-
mohammad authored
-
- 18 Nov, 2020 1 commit
-
-
Jared Casper authored
-
- 17 Nov, 2020 1 commit
-
-
Jared Casper authored
-
- 12 Nov, 2020 12 commits
-
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
Small bugfix in bert_model.py: make sure word_embeddings is initialized before instantiating lm_head
-
Deepak Narayanan authored
Refactor code according to Jared's comments: move pipelining and non-pipelining training loops into separate methods Also, use mpu.get_*_model_parallel_size() instead of args.*_model_parallel_size
-
mshoeybi authored
Allocate tensor in `communicate()` method directly on GPU (instead of allocating on CPU and then moving to GPU)
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-