1. 19 Dec, 2021 2 commits
  2. 17 Dec, 2021 4 commits
  3. 16 Dec, 2021 1 commit
  4. 15 Dec, 2021 2 commits
    • lixiaobai's avatar
      [PinSAGESampler] support PinSAGE sampler on GPU (#3567) · dd762a1e
      lixiaobai authored
      
      
      * Feat: support API "randomwalk_topk" in library
      
      * Feat: use the new API "randomwalk_topk" for PinSAGESampler
      
      * Minor
      
      * Minor
      
      * Refactor: modified codes as checker required
      
      * Minor
      
      * Minor
      
      * Minor
      
      * Minor
      
      * Fix: checking errors in RandomWalkTopk
      
      * Refactor: modified the docstring for randomwalk_topk
      
      * change randomwalk_topk to internal
      
      * fix
      
      * rename
      
      * Minor for pinsage.py
      
      * Feat: support randomwalk and SelectPinSageNeighbors on GPU
      
      Port RandomWalk algorithm on GPU,
      and port SelectPinSageNeighbors on GPU.
      
      * Feat: support GPU on python APIs
      
      * Feat: remove perf print information in FrequenchHashmap
      
      * Fix: modified the code format
      
      Modified the code format as task_lint.sh suggested
      
      * Feat: let test script support PinSAGESampler on GPU
      
      Let test script support PinSAGESampler on GPU,
      minor of "restart_prob".
      
      * Minor
      
      * Minor
      
      * Minor
      
      * Refactor: use the atomic operations from the array module
      
      * Minor: change the long lines
      
      * Refactor: modified the get_node_types for gpu
      
      * Feat: update the contributor date
      
      * Perf: remove unnecessary stream sync
      
      * Feat: support other random walk
      
      But the non-uniform choice is still not supported.
      
      * Fix: add CUDA switch for random walk
      Co-authored-by: default avatarQuan Gan <coin2028@hotmail.com>
      dd762a1e
    • Vasimuddin Md's avatar
      [DistGNN, Graph partitioning] Libra partition (#3376) · 78e0dae6
      Vasimuddin Md authored
      
      
      * added distgnn plus libra codebase
      
      * Dist application codes
      
      * added comments in partition code. changed the interface of partitioning call.
      
      * updated readme
      
      * create libra partitioning branch for the PR
      
      * removed disgnn files for first PR
      
      * updated kernel.cc
      
      * added libra_partition.cc and moved libra code from kernel.cc to libra_partition.cc
      
      * fixed lint error; merged libra2dgl.py and main_Libra.py to libra_partition.py; added graphsage/distgnn folder and partition script.
      
      * removed libra2dgl.py
      
      * fixed the lint error and cleaned the code.
      
      * revisions due to PR comments. added distgnn/tools contains partitions routines
      
      * update 2 PR revision I
      
      * fixed errors; also improved the runtime by 10x.
      
      * fixed minor lint error
      
      * fixed some more lints
      
      * PR revision II changed the interface of libra partition function
      
      * rewrite docstring
      Co-authored-by: default avatarQuan (Andy) Gan <coin2028@hotmail.com>
      78e0dae6
  5. 13 Dec, 2021 3 commits
  6. 09 Dec, 2021 1 commit
  7. 08 Dec, 2021 1 commit
  8. 07 Dec, 2021 1 commit
  9. 06 Dec, 2021 4 commits
  10. 04 Dec, 2021 1 commit
  11. 03 Dec, 2021 2 commits
    • Israt Nisa's avatar
      [Feature] Add Min/max reducer in heterogeneous API for unary message functions (#3514) · cb0e1103
      Israt Nisa authored
      
      
      * min/max support for forward CPU heterograph
      
      * Added etype with each argU values
      
      * scatter_add needs fix
      
      * added scatter_add_hetero. Grads dont match for max reducer
      
      * storing ntype in argX
      
      * fixing scatter_add_hetero
      
      * hetero matches with torch's scatter add
      
      * works copy_e forward+cpu
      
      * added backward for copy_rhs
      
      * Computes gradient for all node types in one kernel
      
      * bug fix
      
      * unnitest for max/min on CPU
      
      * renamed scatter_add_hetero to update_grad_minmax_hetero
      
      * lint check and comment out cuda call for max. Code is for CPU only
      
      * lint check
      
      * replace inf with zero
      
      * minor
      
      * lint check
      
      * removed LIBXSMM code from hetro code
      
      * fixing backward operator of UpdateGradMinMaxHetero
      
      * removed backward from update_grad_minmax_hetero
      
      * docstring
      
      * improved docstring and coding style
      
      * Added pass by pointer for output
      
      * typos and pass by references
      
      * Support for copy_rhs
      
      * Added header <string>
      
      * fix bug in copy_u_max
      
      * Added comments and dimension check of all etypes
      
      * skip mxnet check
      
      * pass by pointer output arrays
      
      * updated docstring
      Co-authored-by: default avatarIsrat Nisa <nisisrat@amazon.com>
      Co-authored-by: default avatarQuan (Andy) Gan <coin2028@hotmail.com>
      cb0e1103
    • Jinjing Zhou's avatar
      769718df
  12. 02 Dec, 2021 3 commits
  13. 30 Nov, 2021 2 commits
  14. 29 Nov, 2021 3 commits
  15. 27 Nov, 2021 1 commit
  16. 24 Nov, 2021 1 commit
  17. 23 Nov, 2021 1 commit
  18. 22 Nov, 2021 2 commits
  19. 21 Nov, 2021 1 commit
  20. 19 Nov, 2021 1 commit
  21. 18 Nov, 2021 1 commit
  22. 17 Nov, 2021 2 commits