[Bug fix] Use shared memory for grad sync when NCCL is not avaliable as...
[Bug fix] Use shared memory for grad sync when NCCL is not avaliable as PyTorch distributed backend. (#3034)
* Use shared memory for grad sync when NCCL is not avaliable as PyTorch distributed backend.
Fix small bugs and update unitests
* Fix bug
* update test
* update test
* Fix unitest
* Fix unitest
* Fix test
* Fix
* simple update
Co-authored-by:
Ubuntu <ubuntu@ip-172-31-24-212.ec2.internal>
Showing
Please register or sign in to comment