- 19 Dec, 2020 15 commits
-
-
mohammad authored
-
mshoeybi authored
-
mshoeybi authored
-
mshoeybi authored
-
mshoeybi authored
-
Jared Casper authored
-
Jared Casper authored
-
Jared Casper authored
-
Deepak Narayanan authored
-
mohammad authored
-
mohammad authored
-
mohammad authored
-
mohammad authored
-
mohammad authored
Rename --batch-size to --micro-batch-size and drop in-minibatch from --num-micro-batches-in-minibatch
-
Jared Casper authored
-
- 02 Dec, 2020 1 commit
-
-
mohammad authored
-
- 30 Nov, 2020 1 commit
-
-
mohammad authored
-
- 28 Nov, 2020 1 commit
-
-
mohammad authored
-
- 26 Nov, 2020 1 commit
-
-
mohammad authored
-
- 12 Nov, 2020 21 commits
-
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
Refactor code according to Jared's comments: move pipelining and non-pipelining training loops into separate methods Also, use mpu.get_*_model_parallel_size() instead of args.*_model_parallel_size
-
mshoeybi authored
Allocate tensor in `communicate()` method directly on GPU (instead of allocating on CPU and then moving to GPU)
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
Prevents data_loader from running out of training examples
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
Make sure all forward and backward operations are accounted for
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
Also includes following changes for inter-layer model-parallel implementation: - Refactoring of model implementations - Training loop changes to support inter-layer communication using `ring_exchange` - New groups for inter-layer communication - Checkpoint changes - Command line arguments
-
mohammad authored
-