"...git@developer.sourcefind.cn:modelzoo/alphafold2_jax.git" did not exist on "eb93322ba9e65542721fec157cfce6e2b74e0936"
- 28 Nov, 2023 1 commit
-
-
flyingdown authored
-
- 01 Dec, 2021 1 commit
-
-
Hubert Lu authored
-
- 19 Nov, 2021 1 commit
-
-
Hubert Lu authored
-
- 25 Jan, 2021 1 commit
-
-
Jeff Daily authored
- incorrect use of __shfl_down - fix warp size assumptions - update unit tests to exit on failure
-
- 31 Jul, 2020 1 commit
-
-
Chaitanya Sri Krishna Lolla authored
-
- 10 Jul, 2020 1 commit
-
-
Chaitanya Sri Krishna Lolla authored
* Enable sync batchnorm * enable syncbn properly * update the unit tests * update tests * update conditions for welford_merge_element * updated conditions based on comments.
-
- 06 Jul, 2020 1 commit
-
-
jjsjann123 authored
* [sync BN] support non-uniform batch size across process group. TODO: test should be added once cleaned up. * updating unit tests * new unit tests for different inputs * cleaning
-
- 03 Jun, 2020 1 commit
-
-
rohithkrn authored
* bfloat16 support for apex DDP * enable mgpu tests for fp16 and bf16 * update Dockerfile
-
- 06 Nov, 2019 1 commit
-
-
jjsjann123 authored
-
- 26 Jul, 2019 1 commit
-
-
jjsjann123 authored
fixing empty return from python implementation adding proper test to verify functional correctness for python implementation
-
- 12 Jul, 2019 1 commit
-
-
jjsjann123 authored
fixing empty return from python implementation adding proper test to verify functional correctness for python implementation
-
- 01 May, 2019 1 commit
-
-
Michael Carilli authored
-
- 04 Apr, 2019 1 commit
-
-
mcarilli authored
* Refactor to allow more flexible treatment of multiple optimizers/models/losses * Adding _process_optimizers.py * Created L0 tests (now passing). * fix: minor print typo (#234) * make L1 results easier to read * L0 multiple model/optimizer/loss test fleshed out * Adding test that master params remain synced across distributed processes * Docstring updates * Docstring updates
-
- 12 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 26 Feb, 2019 1 commit
-
-
Michael Carilli authored
-
- 03 Feb, 2019 1 commit
-
-
Michael Carilli authored
-
- 01 Nov, 2018 1 commit
-
-
Michael Carilli authored
-
- 29 Oct, 2018 1 commit
-
-
mcarilli authored
* test passes * notes * Using C++-side flatten and unflatten functions * Adding csrc * Persistent synchronization event so it doesn't need to be created and destroyed each time * Interop with parameter flattening in SSD * Added deterministic option to imagenet main.py * Adding options to split gradient averaging and allreduce in pure fp32 * Fixing allreduce_maybe_retain call * Fixing allreduce_fallback * Also sync active_i_buckets from rank 0 * Making retain_allreduce_buffers compatible with/orthogonal to delay_allreduce=True|False * Correcting syntax error, now all seems to work with SSD * Optional cpp extension build * Add mixed precision adam optimizer (#59) * Add FusedAdam Optimizer to Apex that places all the math into a cuda kernel. * Added fixes to fused_adam to get it to work with network. * wip work on python interface for adam with options * fix dispatch for halfs, add python options to handle optional half gradients and params * cleanup, get rid of grid-stride loop
-
- 29 Sep, 2018 2 commits
-
-
Michael Carilli authored
-
mcarilli authored
* beautiful * IT'S WORKING * Hopefully fix race condition for fallback hook * Updating test * shared_param -> delayed_allreduce * Adding a safety check * One more check * syntax...
-
- 14 May, 2018 1 commit
-
-
Michael Carilli authored
-
- 07 May, 2018 1 commit
-
-
Christian Sarofeen authored
-