- 13 Feb, 2021 1 commit
-
-
Deepak Narayanan authored
-
- 09 Feb, 2021 1 commit
-
-
Deepak Narayanan authored
- Split a model's computation into multiple virtual stages as needed, and schedule communication correctly between these virtual stages - Move schedule code into `schedules.py` and communication code into `p2p_communication.py` - Use hyphens instead of spaces in all time logging for consistency - Factor out code in megatron/training.py into helper functions - Refactor evaluate() function: make it use forward_backward_schedule functions
-
- 04 Jan, 2021 1 commit
-
-
Deepak Narayanan authored
-
- 19 Dec, 2020 2 commits
-
-
Jared Casper authored
-
Jared Casper authored
-
- 12 Nov, 2020 2 commits
-
-
Deepak Narayanan authored
-
Deepak Narayanan authored
Also includes following changes for inter-layer model-parallel implementation: - Refactoring of model implementations - Training loop changes to support inter-layer communication using `ring_exchange` - New groups for inter-layer communication - Checkpoint changes - Command line arguments
-
- 06 Aug, 2020 1 commit
-
-
Boris Fomitchev authored
Signed-off-by:Boris Fomitchev <bfomitchev@nvidia.com>
-
- 30 Jul, 2020 1 commit
-
-
Boris Fomitchev authored
Signed-off-by:Boris Fomitchev <bfomitchev@nvidia.com>
-
- 26 Jul, 2020 1 commit
-
-
Neel Kant authored
-
- 05 Jun, 2020 1 commit
-
-
Neel Kant authored
-
- 26 May, 2020 1 commit
-
-
Neel Kant authored
-
- 21 May, 2020 1 commit
-
-
Neel Kant authored
-
- 20 May, 2020 1 commit
-
-
Neel Kant authored
-
- 19 May, 2020 2 commits
- 16 Apr, 2020 1 commit
-
-
Mohammad authored
-
- 10 Feb, 2020 1 commit
-
-
Mohammad Shoeybi authored
-
- 08 Oct, 2019 1 commit
-
-
Mohammad Shoeybi authored
-
- 30 Jul, 2019 1 commit
-
-
Raul Puri authored
Co-authored-by:
shoeybi <shoeybim@gmail.com> Co-authored-by:
raulpuric <raulpuric@berkeley.edu> Co-authored-by:
jaredcasper <jaredcasper@gmail.com> Co-authored-by:
mpatwary <mostofa.patwary@gmail.com> Co-authored-by:
plegresl <plegresl@gmail.com>
-