"vscode:/vscode.git/clone" did not exist on "ea6275dfbc263216c79fcde4abc56363400bde23"
- 09 Dec, 2024 1 commit
-
-
xingjinliang authored
-
- 27 Mar, 2024 1 commit
-
-
liangjing authored
-
- 23 Mar, 2023 1 commit
-
-
Jared Casper authored
-
- 24 Sep, 2022 1 commit
-
-
Jared Casper authored
-
- 23 Sep, 2022 1 commit
-
-
Jared Casper authored
The LICENSE file says everything is 3-clause BSD, which is what we want, but at some point the Apache license was added to the top of some files and that proliferated. This commit removes the Apache license from any files that we own the copyright to. Also updates the copyright year and removes the unnessary coding=utf-8 line.
-
- 10 Aug, 2022 1 commit
-
-
Mohammad Shoeybi authored
-
- 30 Jul, 2021 1 commit
-
-
Deepak Narayanan authored
- Accumulate encoder hidden state gradient to handle skip connection - Correctly compute the number of layers in encoder / decoder for T5 model - Ensure e weights are initialized the same way in embeddings - Synchronize embedding gradients across encoder and decoder for T5 model - Support for checkpoint loading and saving
-
- 16 Apr, 2021 1 commit
-
-
Jared Casper authored
-
- 24 Mar, 2021 1 commit
-
-
Vijay Korthikanti authored
-
- 09 Feb, 2021 1 commit
-
-
Deepak Narayanan authored
- Split a model's computation into multiple virtual stages as needed, and schedule communication correctly between these virtual stages - Move schedule code into `schedules.py` and communication code into `p2p_communication.py` - Use hyphens instead of spaces in all time logging for consistency - Factor out code in megatron/training.py into helper functions - Refactor evaluate() function: make it use forward_backward_schedule functions
-
- 25 Jan, 2021 1 commit
-
-
Mohammad Shoeybi authored
-
- 19 Dec, 2020 1 commit
-
-
mohammad authored
-
- 12 Nov, 2020 5 commits
-
-
Deepak Narayanan authored
-
Deepak Narayanan authored
Refactor code according to Jared's comments: move pipelining and non-pipelining training loops into separate methods Also, use mpu.get_*_model_parallel_size() instead of args.*_model_parallel_size
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
Also includes following changes for inter-layer model-parallel implementation: - Refactoring of model implementations - Training loop changes to support inter-layer communication using `ring_exchange` - New groups for inter-layer communication - Checkpoint changes - Command line arguments
-
- 07 Jul, 2020 1 commit
-
-
Neel Kant authored
-
- 24 Jun, 2020 1 commit
-
-
Neel Kant authored
-
- 05 Jun, 2020 2 commits
- 03 May, 2020 1 commit
-
-
Neel Kant authored
-
- 16 Apr, 2020 1 commit
-
-
Mohammad authored
-
- 13 Apr, 2020 1 commit
-
-
Mohammad authored
-
- 31 Mar, 2020 1 commit
-
-
Mohammad authored
-
- 30 Mar, 2020 2 commits
- 29 Mar, 2020 3 commits
- 28 Mar, 2020 2 commits
- 27 Mar, 2020 1 commit
-
-
Mohammad authored
-
- 17 Mar, 2020 2 commits
-
-
Mohammad Shoeybi authored
-
Mohammad Shoeybi authored
-
- 14 Jan, 2020 1 commit
-
-
Mohammad Shoeybi authored
-
- 21 Nov, 2019 1 commit
-
-
Mohammad Shoeybi authored
-
- 19 Nov, 2019 1 commit
-
-
Jared Casper authored
-
- 31 Oct, 2019 1 commit
-
-
Mohammad Shoeybi authored
-
- 29 Oct, 2019 1 commit
-
-
Mohammad Shoeybi authored
-