- 16 Aug, 2019 1 commit
-
-
Deyu Fu authored
-
- 08 Aug, 2019 1 commit
-
-
Deyu Fu authored
-
- 06 Aug, 2019 1 commit
-
-
ngimel authored
* Bug fix for non-affine layer-norm + add backward unit test * clean up tests and add tests for a large batch
-
- 01 Aug, 2019 1 commit
-
-
Natalia Gimelshein authored
-
- 26 Jul, 2019 1 commit
-
-
Edward Z. Yang authored
Signed-off-by:Edward Z. Yang <ezyang@fb.com>
-
- 12 Jul, 2019 1 commit
-
-
Edward Z. Yang authored
Signed-off-by:Edward Z. Yang <ezyang@fb.com>
-
- 03 Jul, 2019 4 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
- 28 Jun, 2019 1 commit
-
-
Thor Johnsen authored
-
- 14 Jun, 2019 1 commit
-
-
Thor Johnsen authored
-
- 11 Jun, 2019 1 commit
-
-
Michael Carilli authored
-
- 31 May, 2019 2 commits
-
-
Thor Johnsen authored
* First draft, for discussion * Fix mistakes in LAMB equations * Add loop over chunk * Bug fix * Bug fix * Bug fix * Undo bug fix * Bug fix * Add multi tensor LAMB optimizer to setup.py * Rename step_size to learning_rate * Fix compilation errors
-
mcarilli authored
* Existing tests passing, still need to add per-tensor tests * Test is passing, still need to measure performance * ILP for l2norm functor
-
- 27 May, 2019 1 commit
-
-
Michael Carilli authored
-
- 10 May, 2019 1 commit
-
-
Michael Carilli authored
-
- 03 May, 2019 1 commit
-
-
Michael Carilli authored
-
- 27 Apr, 2019 1 commit
-
-
jjsjann123 authored
* Persistent group batchnorm added Added persistent grouped batch norm for performance run on strong scaling case: currently only supporting: 1. nhwc layout 2. fp16 3. synchronization only within a node! Environment variable is used to tune LAUNCH_MARGIN that limits the CTAs usage by the persistent kernel. Documentation and examples will follow. * updating type().scalarType() to scalar_type() * moving launch margin to be defined at layer creation, adding a knob cap max ctas per sm * fixing the cta computation * review comment: set device_id through cudaGetDevice() move cudaMemset to cudaMemsetAsync updated __threadfence() to __threadfence_system() inter device write
-
- 26 Apr, 2019 5 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
ptrblck authored
* change .type().ScalarType() to .scalar_type() + at::ScalarType::X to at::kX * revert scalar_type() to type() for AT_DISPATCH_FLOATING_TYPES_AND_HALF * revert scalar_type() to type() in AT_DISPATCH_FLOATING_TYPES * revert scalar_type() to type() for AT_DISPATCH_FLOATING_TYPES_AND_HALF in welford.cu * revert scalar_type() to type() in layer_norm_cuda_kernel.cu * revert at::kType to at::ScalarType::Type * use DISPATCH_FLOAT_AND_HALF to get rid of warnings * add dispatch mechanisms for double+float and double+float+half
-
Michael Carilli authored
-
Michael Carilli authored
-
- 25 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 22 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 18 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 12 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 10 Apr, 2019 2 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
- 09 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 08 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 04 Apr, 2019 1 commit
-
-
mcarilli authored
* Refactor to allow more flexible treatment of multiple optimizers/models/losses * Adding _process_optimizers.py * Created L0 tests (now passing). * fix: minor print typo (#234) * make L1 results easier to read * L0 multiple model/optimizer/loss test fleshed out * Adding test that master params remain synced across distributed processes * Docstring updates * Docstring updates
-
- 21 Mar, 2019 2 commits
-
-
Syed Tousif Ahmed authored
-
Syed Tousif Ahmed authored
-
- 19 Mar, 2019 2 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
- 15 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 12 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 11 Mar, 2019 1 commit
-
-
Simon Layton authored
-