"examples/pytorch/git@developer.sourcefind.cn:OpenDAS/dgl.git" did not exist on "20e1bb455b0304d6d39125b9d4f528fe17947c27"
  1. 15 Feb, 2022 1 commit
    • Israt Nisa's avatar
      [Feature] Gather mm (#3641) · b3d3a2c4
      Israt Nisa authored
      
      
      * init
      
      * init
      
      * working cublasGemm
      
      * benchmark high-mem/low-mem, err gather_mm output
      
      * cuda kernel for bmm like kernel
      
      * removed cpu copy for E_per_Rel
      
      * benchmark code from Minjie
      
      * fixed cublas results in gathermm sorted
      
      * use GPU shared mem in unsorted gather mm
      
      * minor
      
      * Added an optimal version of gather_mm_unsorted
      
      * lint
      
      * init gather_mm_scatter
      
      * cublas transpose added
      
      * fixed h_offset for multiple rel
      
      * backward unittest
      
      * cublas support to transpose W
      
      * adding missed file
      
      * forgot to add header file
      
      * lint
      
      * lint
      
      * cleanup
      
      * lint
      
      * docstring
      
      * lint
      
      * added unittest
      
      * lint
      
      * lint
      
      * unittest
      
      * changed err type
      
      * skip cpu test
      
      * skip CPU code
      
      * move in-len loop inside
      
      * lint
      
      * added check different dim length for B
      
      * w_per_len is optional now
      
      * moved gather_mm to pytorch/backend with backward support
      
      * removed a_/b_trans support
      
      * transpose op inside GEMM call
      
      * removed out alloc from API, changed W 2D to 3D
      
      * Added se_gather_mm, Separate API for sortedE
      
      * Fixed gather_mm (unsorted) user interface
      
      * unsorted gmm backward + separate CAPI for un/sorted A
      
      * typecast to float to support atomicAdd
      
      * lint typecast
      
      * lint
      
      * added gather_mm_scatter
      
      * minor
      
      * const
      
      * design changes
      
      * Added idx_a, idx_b support gmm_scatter
      
      * dgl doc
      
      * lint
      
      * adding gather_mm in ops
      
      * lint
      
      * lint
      
      * minor
      
      * removed benchmark files
      
      * minor
      
      * empty commit
      Co-authored-by: default avatarIsrat Nisa <nisisrat@amazon.com>
      b3d3a2c4
  2. 11 Feb, 2022 1 commit
    • ranzhejiang's avatar
      New fused edge_softmax op (#3650) · bc8f8b0b
      ranzhejiang authored
      
      
      * [feature] edge softmax refact.
      
      * delete file
      
      * fix backward and cmake version
      
      * fix backward
      
      * format function
      
      * fix setting
      
      * refix
      
      * refix
      
      * refix
      
      * refix
      
      * refix
      
      * refix
      
      * refix
      
      * refix
      
      * refix
      
      * refix
      
      * refix
      
      * refix
      
      * add cuda kernel for backward and rename some function
      
      * add benchmark for edge_softmax
      
      * fix format
      
      * remove cuda_backwrd
      
      * fix code format and add comment for op on CPU
      
      * fix lint
      Co-authored-by: default avatarJinjing Zhou <VoVAllen@users.noreply.github.com>
      bc8f8b0b
  3. 03 Dec, 2021 1 commit
    • Israt Nisa's avatar
      [Feature] Add Min/max reducer in heterogeneous API for unary message functions (#3514) · cb0e1103
      Israt Nisa authored
      
      
      * min/max support for forward CPU heterograph
      
      * Added etype with each argU values
      
      * scatter_add needs fix
      
      * added scatter_add_hetero. Grads dont match for max reducer
      
      * storing ntype in argX
      
      * fixing scatter_add_hetero
      
      * hetero matches with torch's scatter add
      
      * works copy_e forward+cpu
      
      * added backward for copy_rhs
      
      * Computes gradient for all node types in one kernel
      
      * bug fix
      
      * unnitest for max/min on CPU
      
      * renamed scatter_add_hetero to update_grad_minmax_hetero
      
      * lint check and comment out cuda call for max. Code is for CPU only
      
      * lint check
      
      * replace inf with zero
      
      * minor
      
      * lint check
      
      * removed LIBXSMM code from hetro code
      
      * fixing backward operator of UpdateGradMinMaxHetero
      
      * removed backward from update_grad_minmax_hetero
      
      * docstring
      
      * improved docstring and coding style
      
      * Added pass by pointer for output
      
      * typos and pass by references
      
      * Support for copy_rhs
      
      * Added header <string>
      
      * fix bug in copy_u_max
      
      * Added comments and dimension check of all etypes
      
      * skip mxnet check
      
      * pass by pointer output arrays
      
      * updated docstring
      Co-authored-by: default avatarIsrat Nisa <nisisrat@amazon.com>
      Co-authored-by: default avatarQuan (Andy) Gan <coin2028@hotmail.com>
      cb0e1103
  4. 17 Nov, 2021 1 commit
  5. 22 Jun, 2021 1 commit
    • Israt Nisa's avatar
      [Kernel] Add heterograph support in CUDA kernels (SpMM, SDDMM) (#2925) · 1113f674
      Israt Nisa authored
      
      
      * Added heterograph support SpMM, SDDMM
      
      * bug fix cuda stream
      
      * add cudaStrm destroy and fix whitespace
      
      * Added heterograph support SpMM, SDDMM
      
      * bug fix cuda stream
      
      * add cudaStrm destroy and fix whitespace
      
      * changed max stream = 1
      
      * Fixed ctx
      
      * using default stream
      
      * Added heterograph support SpMM, SDDMM
      
      * bug fix cuda stream
      
      * add cudaStrm destroy and fix whitespace
      
      * changed max stream = 1
      
      * Fixed ctx
      
      * using default stream
      
      * fix bug in copy_rhs
      
      * changed by mistake
      
      * minor datatype change
      
      * added datatype check
      Co-authored-by: default avatarIsrat Nisa <nisisrat@amazon.com>
      1113f674
  6. 03 Jun, 2021 1 commit
  7. 27 Apr, 2021 1 commit
  8. 24 Mar, 2021 1 commit
    • Quan (Andy) Gan's avatar
      [Feature] Sparse-sparse matrix multiplication, addition, and masking (#2753) · 929d8634
      Quan (Andy) Gan authored
      * test
      
      * more stuff
      
      * add test
      
      * fixes
      
      * optimize algo
      
      * replace unordered_map with arrays
      
      * lint
      
      * lint x2
      
      * oops
      
      * disable gpu csrmm tests
      
      * remove gpu invocation
      
      * optimize with openmp
      
      * remove python functions
      
      * add back with docstrings
      
      * lint
      
      * lint
      
      * update python interface
      
      * functionize
      
      * functionize
      
      * lint
      
      * lint
      929d8634
  9. 21 Feb, 2021 1 commit
  10. 28 Jan, 2021 1 commit
  11. 22 Nov, 2020 1 commit
  12. 13 Jul, 2020 1 commit
    • Zihao Ye's avatar
      [Feature] Autograd of gspmm and gsddmm on PyTorch/MXNet/Tensorflow (#1680) · f25b1a06
      Zihao Ye authored
      * init
      
      * reverse(by minjie
      
      * upd
      
      * upd
      
      * upd
      
      * upd
      
      * upd
      
      * upd
      
      * upd
      
      * gpu
      
      * upd
      
      * upd
      
      * upd
      
      * upd
      
      * udp
      
      * upd
      
      * upd
      
      * imidiot
      
      * fix
      
      * upd
      
      * upd
      
      * upd
      
      * udp
      
      * upd
      
      * upd
      
      * fix
      
      * udp
      
      * upd
      
      * upd
      
      * upd
      
      * upd
      
      * upd
      
      * fix
      
      * remove redundency
      
      * upd
      
      * upd
      
      * upd
      
      * cache
      
      * upd
      
      * upd
      
      * upd
      
      * upd
      
      * upd
      
      * udp
      
      * upd
      
      * trigger
      
      * upd
      
      * fix
      
      * upd
      
      * unused import
      
      * upd
      
      * upd
      f25b1a06
  13. 28 Jun, 2020 1 commit
    • Minjie Wang's avatar
      [CUDA][Kernel] More CUDA kernels; Standardize the behavior for sorted COO/CSR (#1704) · 870da747
      Minjie Wang authored
      * add cub; array cumsum
      
      * CSRSliceRows
      
      * fix warning
      
      * operator << for ndarray; CSRSliceRows
      
      * add CSRIsSorted
      
      * add csr_sort
      
      * inplace coosort and outplace csrsort
      
      * WIP: coo is sorted
      
      * mv cuda_utils
      
      * add AllTrue utility
      
      * csr sort
      
      * coo sort
      
      * coo2csr for sorted coo arrays
      
      * CSRToCOO from sorted
      
      * pass tests for the new kernel changes
      
      * cannot use inplace sort
      
      * lint
      
      * try fix msvc error
      
      * Fix g.copy_to and g.asnumbits; ToBlock no longer uses CSC
      
      * stash
      
      * revert some hack
      
      * revert some changes
      
      * address comments
      
      * fix
      
      * fix to_block unittest
      
      * add todo note
      870da747
  14. 22 Jun, 2020 1 commit
    • Zihao Ye's avatar
      [kernel] New SpMM & SDDMM kernel on CPU and CUDA (#1644) · 071cba1f
      Zihao Ye authored
      * udp
      
      * simplify
      
      * sddmm dot cpu
      
      * upd
      
      * format
      
      * upd
      
      * compatible with MJ's PR
      
      * lint
      
      * upd
      
      * upd
      
      * upd
      
      * python end
      
      * upd
      
      * upd
      
      * lint
      
      * lint
      
      * upd
      
      * upd
      
      * upd
      
      * upd
      
      * upd
      
      * lint
      
      * fix mxnet
      
      * upd
      
      * lint
      
      * use minjie's ptr
      
      * macro
      
      * upd
      
      * reorg
      
      * lint
      
      * fix corner cases
      
      * upd
      
      * enrich cpu docs
      
      * upd
      
      * upd
      
      * lint
      
      * lint
      
      * pylint
      
      * sx review
      
      * improve docstring
      
      * python doc
      
      * upd
      
      * restructure
      
      * lint
      
      * upd test
      
      * upd
      
      * pylint
      
      * fix corner cases and test
      071cba1f