- 06 Feb, 2020 1 commit
-
-
Kevin Stephano authored
* Adding C++ Multihead Attention implementation to contrib. * Add reference test that at least works for forward. * Remove CublasLt support from multihead attention. * Add new Python version of self attention. * Update python model of MHA with backward pass. * Fixed Output Linear connection in MHA. * Clean up compiles and add documentation to PySelfAttention. * Add Encdec Python version of multihead attention. Cleanup files. * Tests for self and encdec multihead attention. * Add reference pytorch implementation of attention with norm and add. * Add cutlass branch definition. * Add cutlass download to compile. * Add norm/add tests. * Add biases to pytorch python versions. * Add tests and fix issues with python version of attention masking. * Create README.md * Update README.md * Update README.md * Update perf test parameters. * Update README.md * Update README.md * Update README.md * Add f...
-
- 21 Jan, 2020 1 commit
-
-
jjsjann123 authored
-
- 08 Jan, 2020 1 commit
-
-
ptrblck authored
* add WAR for pip>=19.3.1 * remove pipmain, use extras_require instead
-
- 04 Oct, 2019 1 commit
-
-
Deyu Fu authored
* move previous fused_adam and fp16_optimizer to contrib * make build contrib.fused_adam optional * change build option name * remove unnecessary try import
-
- 13 Sep, 2019 1 commit
-
-
mcarilli authored
-
- 06 Sep, 2019 1 commit
-
-
mcarilli authored
* Pushing for build tests * Contrib files * Removing deprecated checks
-
- 17 Aug, 2019 1 commit
-
-
Deyu Fu authored
-
- 16 Aug, 2019 1 commit
-
-
Deyu Fu authored
-
- 13 Aug, 2019 1 commit
-
-
Marek Kolodziej authored
Co-authored-by:
Aditya Agrawal <aditya.iitb@gmail.com> Co-authored-by:
Marek Kolodziej <mkolod@gmail.com>
-
- 08 Aug, 2019 1 commit
-
-
Deyu Fu authored
-
- 31 May, 2019 1 commit
-
-
Thor Johnsen authored
* First draft, for discussion * Fix mistakes in LAMB equations * Add loop over chunk * Bug fix * Bug fix * Bug fix * Undo bug fix * Bug fix * Add multi tensor LAMB optimizer to setup.py * Rename step_size to learning_rate * Fix compilation errors
-
- 23 May, 2019 1 commit
-
-
Michael Carilli authored
-
- 22 May, 2019 1 commit
-
-
mcarilli authored
-
- 09 May, 2019 1 commit
-
-
Wil Kong authored
* Add softmax cross entropy loss with label smoothing support. * Fix deprecation of AT_DISPATCH_XXX and several minor issues. * Fix issues commented by reviewers. * Add FB license. * Remove code generation constraints. * Add a simple unittest for label smoothing.
-
- 27 Apr, 2019 1 commit
-
-
jjsjann123 authored
* Persistent group batchnorm added Added persistent grouped batch norm for performance run on strong scaling case: currently only supporting: 1. nhwc layout 2. fp16 3. synchronization only within a node! Environment variable is used to tune LAUNCH_MARGIN that limits the CTAs usage by the persistent kernel. Documentation and examples will follow. * updating type().scalarType() to scalar_type() * moving launch margin to be defined at layer creation, adding a knob cap max ctas per sm * fixing the cta computation * review comment: set device_id through cudaGetDevice() move cudaMemset to cudaMemsetAsync updated __threadfence() to __threadfence_system() inter device write
-
- 18 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 09 Apr, 2019 1 commit
-
-
Michael Carilli authored
-
- 23 Mar, 2019 1 commit
-
-
Cubbee authored
-
- 22 Mar, 2019 1 commit
-
-
mcarilli authored
* Adding Torch + bare-metal nvcc version check and container build tests * Putting a canary in the coalmine * canary proved elusive * Trying direct setup.py install * this should work * Removing canary * hopefully this works
-
- 19 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 13 Mar, 2019 1 commit
-
-
Wil Kong authored
-
- 12 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 10 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 08 Mar, 2019 1 commit
-
-
Simon Layton authored
Initial implementation, all fp32 Tested against torch.optim.sgd
-
- 05 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 04 Mar, 2019 1 commit
-
-
Michael Carilli authored
-
- 19 Feb, 2019 1 commit
-
-
Michael Carilli authored
-
- 11 Feb, 2019 1 commit
-
-
Michael Carilli authored
-
- 04 Feb, 2019 1 commit
-
-
Michael Carilli authored
-
- 12 Dec, 2018 1 commit
-
-
Michael Carilli authored
-
- 31 Oct, 2018 1 commit
-
-
Thor Johnsen authored
* Pre-release of fused layer norm apex extension * Remove half and __half2 specializations * Code changes from review
-
- 30 Oct, 2018 1 commit
-
-
ngimel authored
-
- 29 Oct, 2018 1 commit
-
-
mcarilli authored
* test passes * notes * Using C++-side flatten and unflatten functions * Adding csrc * Persistent synchronization event so it doesn't need to be created and destroyed each time * Interop with parameter flattening in SSD * Added deterministic option to imagenet main.py * Adding options to split gradient averaging and allreduce in pure fp32 * Fixing allreduce_maybe_retain call * Fixing allreduce_fallback * Also sync active_i_buckets from rank 0 * Making retain_allreduce_buffers compatible with/orthogonal to delay_allreduce=True|False * Correcting syntax error, now all seems to work with SSD * Optional cpp extension build * Add mixed precision adam optimizer (#59) * Add FusedAdam Optimizer to Apex that places all the math into a cuda kernel. * Added fixes to fused_adam to get it to work with network. * wip work on python interface for adam with options * fix dispatch for halfs, add python options to handle optional half gradients and params * cleanup, get rid of grid-stride loop
-
- 23 Oct, 2018 1 commit
-
-
jjsjann123 authored
* [syncBN] added syncBN in native pure python apex added fused cuda kernels used for sync BN. Using welford for mean/var optional installation using 'python setup.py install --cuda_ext' added unit test with side to side comparison between apex sync BN with PyTorch BN. Notice that for pytorch BN implementation, because of numerical issue for mean/var, the output will be slightly off. * [syncBN PR] added fp16 support addressing review comments on: 1. updating last pow 2 2. look for import error when importing syncBN kernel * [syncBN PR] added convert function to insert SyncBatchNorm refactored some kernel code * fixing type issue (fp16/fp32/fp64) added Kahan summation editing unit test to use pytorch primitive ops with double, passing reasonable tests now * updating tensor creation calls * fixing the all_reduce contiguous tensor * transposed all reduce results * [syncBN] support fp16 input & fp32 layer for apex fp16 partially fixing launch configs enabling imagenet example to run with --sync_bn * [syncBN PR] Documentation added * adjusting README * adjusting again * added some doc to imagenet example * [syncBN] warp-level reduction bug fix: warp reduction logic updated. check for dummy element to avoid nan. improved launch config for better reduction kernels. Further improvements would be to increase grid size. * [syncBN] fixing undefined behavior in __shfl_down_sync from divergent threads in warp reduction. changing at::native::empty to at::empty (upstream comments)
-
- 23 Jul, 2018 1 commit
-
-
Michael Carilli authored
-
- 05 Jul, 2018 1 commit
-
-
mcarilli authored
-
- 04 Jul, 2018 1 commit
-
-
brett koonce authored
-
- 24 Jun, 2018 1 commit
-
-
Michael Carilli authored
-
- 21 Jun, 2018 1 commit
-
-
cclauss authored
-
- 14 Jun, 2018 1 commit
-
-
Michael Carilli authored
-