- 16 Apr, 2019 2 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
- 15 Apr, 2019 3 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
- 12 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 11 Apr, 2019 7 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
Michael Carilli authored
-
henrymai authored
The main use of these functions (e.g.: `torch.{conv*, prelu}`) is via their `torch.nn` wrapping layers. The `torch.nn` layers are what contain the weights and call into these lower level functions with the weights as a parameter in their `forward()` method. The `torch.conv*` functions are already in the `FP16_CASTS` list due to amp's philosophy of casting the parameters rather than the model/layer weights. Conceptually `torch.prelu` is the same as the `torch.conv*` case, where its weight parameter is passed in from its wrapper layer `torch.nn.PReLU`. -
Michael Carilli authored
-
Michael Carilli authored
-
- 10 Apr, 2019 5 commits
-
-
ngimel authored
quick fix: make FusedLayerNorm compatible with cpu
-
Lam Dang authored
-
Lam Dang authored
-
Michael Carilli authored
-
Michael Carilli authored
-
- 09 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 08 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 05 Apr, 2019 3 commits
-
-
-
Michael Carilli authored
-
Michael Carilli authored
-
- 04 Apr, 2019 3 commits
-
-
ngimel authored
Run interpolation in fp32 because it's faster
-
Marek Kolodziej authored
-
mcarilli authored
* Refactor to allow more flexible treatment of multiple optimizers/models/losses * Adding _process_optimizers.py * Created L0 tests (now passing). * fix: minor print typo (#234) * make L1 results easier to read * L0 multiple model/optimizer/loss test fleshed out * Adding test that master params remain synced across distributed processes * Docstring updates * Docstring updates
-
- 03 Apr, 2019 1 commit
-
-
mcarilli authored
-
- 01 Apr, 2019 1 commit
-
-
jjsjann123 authored
Fix a typo in optimized_sync_batchnorm_kernel.py
-
- 31 Mar, 2019 1 commit
-
-
Bingchen Zhao authored
in line 54, running_var should be running_variance..
-
- 27 Mar, 2019 2 commits
- 26 Mar, 2019 2 commits
-
-
Michael Carilli authored
-
Michael Carilli authored
-
- 23 Mar, 2019 1 commit
-
-
Cubbee authored
-
- 22 Mar, 2019 4 commits
-
-
jjsjann123 authored
supporting 2 dimensional input, resolving issue #194 Implementation: for 2d input, switching channel_last flag to true for better memory access pattern in the kernel.
-
henrymai authored
* Add prelu to list of torch overrides This is to fix the following error: File "/opt/conda/lib/python3.6/site-packages/torch/nn/modules/module.py", line 489, in __call__ result = self.forward(*input, **kwargs) File "/opt/conda/lib/python3.6/site-packages/torch/nn/modules/container.py", line 92, in forward input = module(input) File "/opt/conda/lib/python3.6/site-packages/torch/nn/modules/module.py", line 489, in __call__ result = self.forward(*input, **kwargs) File "/opt/conda/lib/python3.6/site-packages/torch/nn/modules/activation.py", line 722, in forward return F.prelu(input, self.weight) File "/opt/conda/lib/python3.6/site-packages/torch/nn/functional.py", line 1040, in prelu return torch.prelu(input, weight) RuntimeError: expected scalar type Half but found Float * Update torch_overrides.py -
enricoschroeder authored
Fix 'local variable 'optimizers_was_list' referenced before assignment' when amp.initialize() is called with optimizers=None (#218)
-
mcarilli authored
* Adding Torch + bare-metal nvcc version check and container build tests * Putting a canary in the coalmine * canary proved elusive * Trying direct setup.py install * this should work * Removing canary * hopefully this works
-
- 21 Mar, 2019 2 commits
-
-
mcarilli authored
Rename IntList to IntArrayRef
-
Syed Tousif Ahmed authored
-