- 04 Jan, 2022 1 commit
-
-
Lawrence McAfee authored
-
- 10 Dec, 2021 1 commit
-
-
Lawrence McAfee authored
-
- 08 Dec, 2021 1 commit
-
-
Lawrence McAfee authored
-
- 07 Dec, 2021 2 commits
-
-
Lawrence McAfee authored
-
Lawrence McAfee authored
-
- 09 Aug, 2021 1 commit
-
-
Lawrence McAfee authored
-
- 30 Jul, 2021 2 commits
-
-
Deepak Narayanan authored
-
Deepak Narayanan authored
- Accumulate encoder hidden state gradient to handle skip connection - Correctly compute the number of layers in encoder / decoder for T5 model - Ensure e weights are initialized the same way in embeddings - Synchronize embedding gradients across encoder and decoder for T5 model - Support for checkpoint loading and saving
-
- 29 Jul, 2021 1 commit
-
-
Deepak Narayanan authored
-
- 21 Jul, 2021 1 commit
-
-
Deepak Narayanan authored
-
- 24 Mar, 2021 3 commits
-
-
Jared Casper authored
-
Jared Casper authored
-
Vijay Korthikanti authored
-
- 03 Mar, 2021 2 commits
-
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
- 26 Feb, 2021 2 commits
-
-
Deepak Narayanan authored
Fix deadlock when get_num_microbatches() < pipeline-parallel size (don't try to measure pipeline stall)
-
Deepak Narayanan authored
-
- 13 Feb, 2021 1 commit
-
-
Deepak Narayanan authored
-
- 10 Feb, 2021 2 commits
-
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
- 09 Feb, 2021 2 commits
-
-
Deepak Narayanan authored
-
Deepak Narayanan authored
- Split a model's computation into multiple virtual stages as needed, and schedule communication correctly between these virtual stages - Move schedule code into `schedules.py` and communication code into `p2p_communication.py` - Use hyphens instead of spaces in all time logging for consistency - Factor out code in megatron/training.py into helper functions - Refactor evaluate() function: make it use forward_backward_schedule functions
-