"llm/llama.cpp/examples/vscode:/vscode.git/clone" did not exist on "768ab4df541275c05eec5ee5db2f89661302610d"
- 06 Feb, 2021 1 commit
-
-
Deepak Narayanan authored
-
- 12 Jan, 2021 1 commit
-
-
mohammad authored
-
- 05 Jan, 2021 2 commits
- 04 Jan, 2021 1 commit
-
-
Deepak Narayanan authored
-
- 31 Dec, 2020 1 commit
-
-
mohammad authored
-
- 27 Dec, 2020 1 commit
-
-
mohammad authored
-
- 25 Dec, 2020 5 commits
- 23 Dec, 2020 1 commit
-
-
Deepak Narayanan authored
Checkpoint should be saved only after evaluation pass is run to make sure validation losses are identical after loading checkpoint
-
- 22 Dec, 2020 1 commit
-
-
mohammad authored
Add the option for fp32 residual connection (fp32 residual connection machinery still needs to be added)
-
- 19 Dec, 2020 16 commits
-
-
mohammad authored
-
mohammad authored
-
mshoeybi authored
-
mshoeybi authored
-
mshoeybi authored
-
mshoeybi authored
-
Jared Casper authored
-
Jared Casper authored
-
Jared Casper authored
-
Deepak Narayanan authored
-
mohammad authored
-
mohammad authored
-
mohammad authored
-
mohammad authored
-
mohammad authored
Rename --batch-size to --micro-batch-size and drop in-minibatch from --num-micro-batches-in-minibatch
-
Jared Casper authored
-
- 02 Dec, 2020 1 commit
-
-
mohammad authored
-
- 30 Nov, 2020 1 commit
-
-
mohammad authored
-
- 28 Nov, 2020 1 commit
-
-
mohammad authored
-
- 26 Nov, 2020 1 commit
-
-
mohammad authored
-
- 12 Nov, 2020 6 commits
-
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-
Deepak Narayanan authored
Refactor code according to Jared's comments: move pipelining and non-pipelining training loops into separate methods Also, use mpu.get_*_model_parallel_size() instead of args.*_model_parallel_size
-
mshoeybi authored
Allocate tensor in `communicate()` method directly on GPU (instead of allocating on CPU and then moving to GPU)
-
Deepak Narayanan authored
-
Deepak Narayanan authored
-