- 23 May, 2019 3 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
- 22 May, 2019 3 commits
-
-
mcarilli authored
-
Michael Carilli authored
-
ptrblck authored
-
- 21 May, 2019 1 commit
-
-
blisc authored
* update larc Signed-off-by:
Jason <jasoli@nvidia.com> * scale_loss fix Signed-off-by:
Jason <jasoli@nvidia.com> * typo Signed-off-by:
Jason <jasoli@nvidia.com> * revert LARC
-
- 17 May, 2019 2 commits
-
-
jjsjann123 authored
update input size check to fix github issue #262 update SyncBatchNorm count check so that size 1 input with cross GPU synchronization runs fine.
-
jjsjann123 authored
resolves issue #254 Added input casting for pure python implementation, this supports mismatched input and layer dtype.
-
- 16 May, 2019 1 commit
-
-
mcarilli authored
* Support add_param_group * syntax * Test added and passing
-
- 15 May, 2019 3 commits
-
-
Michael Glass authored
-
ptrblck authored
-
ptrblck authored
* fix URLs * Update distributed.py
-
- 13 May, 2019 3 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
mcarilli authored
-
- 10 May, 2019 1 commit
-
-
Michael Carilli authored
-
- 09 May, 2019 2 commits
-
-
Tim Zaman authored
-
Wil Kong authored
* Add softmax cross entropy loss with label smoothing support. * Fix deprecation of AT_DISPATCH_XXX and several minor issues. * Fix issues commented by reviewers. * Add FB license. * Remove code generation constraints. * Add a simple unittest for label smoothing.
-
- 08 May, 2019 1 commit
-
-
Michael Carilli authored
-
- 03 May, 2019 1 commit
-
-
Michael Carilli authored
-
- 02 May, 2019 2 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
- 01 May, 2019 3 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
- 30 Apr, 2019 5 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
ptrblck authored
* remove unused tens tensor in example/imagenet/main_amp.py * remove unused tens tensor in deprecated examples and tests/L1
-
mcarilli authored
-
- 29 Apr, 2019 2 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
- 27 Apr, 2019 2 commits
-
-
jjsjann123 authored
* Persistent group batchnorm added Added persistent grouped batch norm for performance run on strong scaling case: currently only supporting: 1. nhwc layout 2. fp16 3. synchronization only within a node! Environment variable is used to tune LAUNCH_MARGIN that limits the CTAs usage by the persistent kernel. Documentation and examples will follow. * updating type().scalarType() to scalar_type() * moving launch margin to be defined at layer creation, adding a knob cap max ctas per sm * fixing the cta computation * review comment: set device_id through cudaGetDevice() move cudaMemset to cudaMemsetAsync updated __threadfence() to __threadfence_system() inter device write
-
Michael Carilli authored
-
- 26 Apr, 2019 5 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
ptrblck authored
* change .type().ScalarType() to .scalar_type() + at::ScalarType::X to at::kX * revert scalar_type() to type() for AT_DISPATCH_FLOATING_TYPES_AND_HALF * revert scalar_type() to type() in AT_DISPATCH_FLOATING_TYPES * revert scalar_type() to type() for AT_DISPATCH_FLOATING_TYPES_AND_HALF in welford.cu * revert scalar_type() to type() in layer_norm_cuda_kernel.cu * revert at::kType to at::ScalarType::Type * use DISPATCH_FLOAT_AND_HALF to get rid of warnings * add dispatch mechanisms for double+float and double+float+half
-