- 16 May, 2022 2 commits
-
-
Vijay Korthikanti authored
-
Vijay Korthikanti authored
-
- 09 May, 2022 1 commit
-
-
Sangkug Lym authored
change dummy_handler to nullcontext
-
- 25 Mar, 2022 1 commit
-
-
Sangkug Lym authored
-
- 19 Jan, 2022 1 commit
-
-
Vijay Korthikanti authored
-
- 12 Jan, 2022 1 commit
-
-
Vijay Korthikanti authored
-
- 24 Nov, 2021 1 commit
-
-
Szymon Migacz authored
-
- 07 Sep, 2021 1 commit
-
-
Sangkug Lym authored
-
- 02 Sep, 2021 3 commits
-
-
slym authored
-
slym authored
-
Sangkug Lym authored
allreduce overlap with wgrad gemm change custom delay to dummy add
-
- 27 Aug, 2021 2 commits
-
-
Sangkug Lym authored
-
Ryan Prenger authored
-
- 26 Aug, 2021 1 commit
-
-
rprenger authored
-
- 19 Aug, 2021 1 commit
-
-
mshoeybi authored
-
- 16 Aug, 2021 1 commit
-
-
Mohammad Shoeybi authored
-
- 30 Jul, 2021 1 commit
-
-
Deepak Narayanan authored
- Accumulate encoder hidden state gradient to handle skip connection - Correctly compute the number of layers in encoder / decoder for T5 model - Ensure e weights are initialized the same way in embeddings - Synchronize embedding gradients across encoder and decoder for T5 model - Support for checkpoint loading and saving
-
- 02 Jul, 2021 1 commit
-
-
rprenger authored
-
- 18 Mar, 2021 1 commit
-
-
Mohammad Shoeybi authored
-
- 09 Feb, 2021 1 commit
-
-
Deepak Narayanan authored
- Split a model's computation into multiple virtual stages as needed, and schedule communication correctly between these virtual stages - Move schedule code into `schedules.py` and communication code into `p2p_communication.py` - Use hyphens instead of spaces in all time logging for consistency - Factor out code in megatron/training.py into helper functions - Refactor evaluate() function: make it use forward_backward_schedule functions
-
- 26 Jan, 2021 1 commit
-
-
mohammad authored
-
- 30 Dec, 2020 1 commit
-
-
mshoeybi authored
-
- 19 Dec, 2020 1 commit
-
-
mohammad authored
-
- 12 Nov, 2020 3 commits
-
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
Also includes following changes for inter-layer model-parallel implementation: - Refactoring of model implementations - Training loop changes to support inter-layer communication using `ring_exchange` - New groups for inter-layer communication - Checkpoint changes - Command line arguments
-
- 01 Oct, 2020 1 commit
-
-
mohammad authored
-
- 02 Sep, 2020 2 commits
- 27 Aug, 2020 1 commit
-
-
Boris Fomitchev authored
Signed-off-by:Boris Fomitchev <bfomitchev@nvidia.com>
-
- 07 Aug, 2020 2 commits
-
-
Boris Fomitchev authored
Signed-off-by:Boris Fomitchev <bfomitchev@nvidia.com>
-
Boris Fomitchev authored
Signed-off-by:Boris Fomitchev <bfomitchev@nvidia.com>
-
- 06 Aug, 2020 3 commits
-
-
Boris Fomitchev authored
Signed-off-by:Boris Fomitchev <bfomitchev@nvidia.com>
-
Boris Fomitchev authored
Signed-off-by:Boris Fomitchev <bfomitchev@nvidia.com>
-
Boris Fomitchev authored
Signed-off-by:Boris Fomitchev <bfomitchev@nvidia.com>
-
- 30 Jul, 2020 1 commit
-
-
Boris Fomitchev authored
Signed-off-by:Boris Fomitchev <bfomitchev@nvidia.com>
-
- 22 Jul, 2020 1 commit
-
-
Boris Fomitchev authored
Signed-off-by:Boris Fomitchev <bfomitchev@nvidia.com>
-
- 05 Jun, 2020 1 commit
-
-
Neel Kant authored
-
- 21 May, 2020 1 commit
-
-
Neel Kant authored
-
- 19 May, 2020 1 commit
-
-
Neel Kant authored
-