README.md 265 Bytes
Newer Older
1
2
3
4
5
6
7
**distributed_data_parallel.py** shows an example using `FP16_Optimizer` with
`torch.nn.parallel.DistributedDataParallel`.
The usage of `FP16_Optimizer` with distributed does not need to change from ordinary 
single-process usage.  Test via
```bash
bash run.sh
```