[Distributed] Pytorch example of distributed GraphSage. (#1495)
* add train_dist.
* Fix sampling example.
* use distributed sampler.
* fix a bug in DistTensor.
* fix distributed training example.
* add graph partition.
* add command
* disable pytorch parallel.
* shutdown correctly.
* load diff graphs.
* add ip_config.txt.
* record timing for each step.
* use ogb
* add profiler.
* fix a bug.
* add train_dist.
* Fix sampling example.
* use distributed sampler.
* fix a bug in DistTensor.
* fix distributed training example.
* add graph partition.
* add command
* disable pytorch parallel.
* shutdown correctly.
* load diff graphs.
* add ip_config.txt.
* record timing for each step.
* use ogb
* add profiler.
* add Ips of the cluster.
* fix exit.
* support multiple clients.
* balance node types and edges.
* move code.
* remove run.sh
* Revert "support multiple clients."
* fix.
* update train_sampling.
* fix.
* fix
* remove run.sh
* update readme.
* update readme.
* use pytorch distributed.
* ensure all trainers run the same number of steps.
* Update README.md
Co-authored-by:
Ubuntu <ubuntu@ip-172-31-16-250.us-west-2.compute.internal>
Showing
Please register or sign in to comment