- 06 Dec, 2022 1 commit
-
-
Hubert Lu authored
* Unskip some unit tests related to issue #82 * Ensure test_state_dict to use capturable=True for torch.optim.Adam * Fix TestFusedAdam tests in test_fused_optimizer.py
-
- 26 Aug, 2022 1 commit
-
-
Hubert Lu authored
* Handle len(cached_x.grad_fn.next_functions) == 1 in cached_cast * Unskip the unit tests related to len(cached_x.grad_fn.next_functions) == 1 Co-authored-by:David Fan <jiafa@microsoft.com>
-
- 08 Aug, 2022 1 commit
-
-
hubertlu-tw authored
-
- 14 Dec, 2021 1 commit
-
-
Hubert Lu authored
* Skip failing unit tests * Modify the test skipping messages
-
- 21 Jan, 2021 1 commit
-
-
Jeff Daily authored
use __launch_bounds__(1024) for multi_tensor_apply, re-enable skipped tests
-
- 15 Jan, 2021 1 commit
-
-
Sarunya Pumma authored
-
- 31 Dec, 2020 1 commit
-
-
lcskrishna authored
-
- 21 May, 2020 1 commit
-
-
lcskrishna authored
-
- 20 May, 2020 2 commits
-
-
lcskrishna authored
-
lcskrishna authored
-
- 19 May, 2020 1 commit
-
-
lcskrishna authored
-
- 15 May, 2020 2 commits
- 13 May, 2020 1 commit
-
-
rohithkrn authored
-
- 22 Apr, 2020 1 commit
-
-
Vinicius Reis authored
The LARC optimizer wraps an underlying optimizer and then needs to be passed to amp.initialize for mixed precision. There were 3 different crashes happening in this situation, fix all of them and add a unit test. I don't know if the 'LARC' in sys.modules check ever worked. In my setup, the entry in sys.modules is 'apex.parallel.LARC'. Checking if the variable is defined seems more reliable though.
-
- 27 Feb, 2020 1 commit
-
-
mcarilli authored
* NHWC support for multi tensor apply * compilation fix for version<=1.4
-
- 03 Oct, 2019 1 commit
-
-
ptrblck authored
* increase atol for Half-Float comparison to 1.5e-4 * disable tests for different opt_levels * reset atol * add bitwise accurate comparison
-
- 27 Aug, 2019 1 commit
-
-
ptrblck authored
* add state_dict, load_state_dict * add test_restoring, test_loss_scale_decrease * disable amp outputs for checkpoint tests * add test for amp.state_dict, cleanup * add state_dict patch, add test * fixed testing, cleanup * add readme for checkpointing * add docs to source/amp * add review changes to doc
-
- 03 Jul, 2019 1 commit
-
-
Michael Carilli authored
-
- 31 May, 2019 1 commit
-
-
mcarilli authored
* Existing tests passing, still need to add per-tensor tests * Test is passing, still need to measure performance * ILP for l2norm functor
-
- 27 May, 2019 2 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
- 16 May, 2019 1 commit
-
-
mcarilli authored
* Support add_param_group * syntax * Test added and passing
-
- 02 May, 2019 1 commit
-
-
Michael Carilli authored
-
- 10 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 04 Apr, 2019 1 commit
-
-
mcarilli authored
* Refactor to allow more flexible treatment of multiple optimizers/models/losses * Adding _process_optimizers.py * Created L0 tests (now passing). * fix: minor print typo (#234) * make L1 results easier to read * L0 multiple model/optimizer/loss test fleshed out * Adding test that master params remain synced across distributed processes * Docstring updates * Docstring updates
-
- 19 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 10 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 26 Feb, 2019 1 commit
-
-
Michael Carilli authored
-