- 15 May, 2019 3 commits
-
-
Michael Glass authored
-
ptrblck authored
-
ptrblck authored
* fix URLs * Update distributed.py
-
- 13 May, 2019 2 commits
-
-
Michael Carilli authored
-
mcarilli authored
-
- 09 May, 2019 1 commit
-
-
Tim Zaman authored
-
- 30 Apr, 2019 5 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
ptrblck authored
* remove unused tens tensor in example/imagenet/main_amp.py * remove unused tens tensor in deprecated examples and tests/L1
-
mcarilli authored
-
- 29 Apr, 2019 2 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
- 26 Apr, 2019 1 commit
-
-
ptrblck authored
* change .type().ScalarType() to .scalar_type() + at::ScalarType::X to at::kX * revert scalar_type() to type() for AT_DISPATCH_FLOATING_TYPES_AND_HALF * revert scalar_type() to type() in AT_DISPATCH_FLOATING_TYPES * revert scalar_type() to type() for AT_DISPATCH_FLOATING_TYPES_AND_HALF in welford.cu * revert scalar_type() to type() in layer_norm_cuda_kernel.cu * revert at::kType to at::ScalarType::Type * use DISPATCH_FLOAT_AND_HALF to get rid of warnings * add dispatch mechanisms for double+float and double+float+half
-
- 23 Apr, 2019 1 commit
-
-
ptrblck authored
-
- 18 Apr, 2019 2 commits
-
-
ptrblck authored
-
Glenn Jocher authored
-
- 16 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 11 Apr, 2019 1 commit
-
-
henrymai authored
The main use of these functions (e.g.: `torch.{conv*, prelu}`) is via their `torch.nn` wrapping layers. The `torch.nn` layers are what contain the weights and call into these lower level functions with the weights as a parameter in their `forward()` method. The `torch.conv*` functions are already in the `FP16_CASTS` list due to amp's philosophy of casting the parameters rather than the model/layer weights. Conceptually `torch.prelu` is the same as the `torch.conv*` case, where its weight parameter is passed in from its wrapper layer `torch.nn.PReLU`.
-
- 10 Apr, 2019 5 commits
-
-
ngimel authored
quick fix: make FusedLayerNorm compatible with cpu
-
Lam Dang authored
-
Lam Dang authored
-
Michael Carilli authored
-
Michael Carilli authored
-
- 09 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 08 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 05 Apr, 2019 3 commits
-
-
-
Michael Carilli authored
-
Michael Carilli authored
-
- 04 Apr, 2019 3 commits
-
-
ngimel authored
Run interpolation in fp32 because it's faster
-
Marek Kolodziej authored
-
mcarilli authored
* Refactor to allow more flexible treatment of multiple optimizers/models/losses * Adding _process_optimizers.py * Created L0 tests (now passing). * fix: minor print typo (#234) * make L1 results easier to read * L0 multiple model/optimizer/loss test fleshed out * Adding test that master params remain synced across distributed processes * Docstring updates * Docstring updates
-
- 03 Apr, 2019 1 commit
-
-
mcarilli authored
-
- 01 Apr, 2019 1 commit
-
-
jjsjann123 authored
Fix a typo in optimized_sync_batchnorm_kernel.py
-
- 31 Mar, 2019 1 commit
-
-
Bingchen Zhao authored
in line 54, running_var should be running_variance..
-
- 27 Mar, 2019 2 commits
- 26 Mar, 2019 2 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
- 23 Mar, 2019 1 commit
-
-
Cubbee authored
-