- 02 Nov, 2022 1 commit
-
-
Hongzhi (Steve), Chen authored
* [Misc] clang-format auto fix. * manual Co-authored-by:Steve <ubuntu@ip-172-31-34-29.ap-northeast-1.compute.internal>
-
- 29 Oct, 2022 1 commit
-
-
Quan (Andy) Gan authored
* sample neighbors with masks * oops * refactor again * remove * remove debug code * rename macro * address comments * more stuff * remove * fix * try fix unit test * oops * fix test * oops * change name * rename a lot of stuff * oops * ugh * misc fixes * lint * address a lot of comments * lint * lint * fix * that was silly * fix * fix * fix * oops
-
- 28 Oct, 2022 1 commit
-
-
Quan (Andy) Gan authored
* sample neighbors with masks * oops * refactor again * remove * remove debug code * rename macro * address comments * address comment * address comments * rename a lot of stuff * oops
-
- 13 Oct, 2022 1 commit
-
-
Rhett Ying authored
-
- 21 Sep, 2022 1 commit
-
-
Xin Yao authored
* disable warning for tensorpipe * fix warning * enable lint check for cuh files * resolve comments
-
- 19 Sep, 2022 1 commit
-
-
Xin Yao authored
* rename `DLContext` to `DGLContext` * rename `kDLGPU` to `kDLCUDA` * replace DLTensor with DGLArray * fix linting * Unify DGLType and DLDataType to DGLDataType * Fix FFI * rename DLDeviceType to DGLDeviceType * decouple dlpack from the core library * fix bug * fix lint * fix merge * fix build * address comments * rename dl_converter to dlpack_convert * remove redundant comments
-
- 05 Sep, 2022 1 commit
-
-
peizhou001 authored
* enable turn on/off libxsmm at runtime by adding a global config and related API Co-authored-by:Ubuntu <ubuntu@ip-172-31-19-194.ap-northeast-1.compute.internal>
-
- 01 Jul, 2022 2 commits
-
-
Rhett Ying authored
-
Rhett Ying authored
* [Feature] extend sort_csr/csc_by_tag to edge * fix test ffailure in tensorflow * refine sorting by edges * fix docstring * remove unnecessary mem Co-authored-by:Xin Yao <xiny@nvidia.com>
-
- 23 Jun, 2022 1 commit
-
-
Triston authored
* Fix a cub compile error for CUDA 11.5 * Fix comparison of integer expressions of different signedness in coo_sort.cu file * Fix comparison of integer expressions of different signedness in cuda_compact_graph.cu file * Remove never referenced variable in spmm.cu * Fix comparison of integer expressions of different signedness in rowwise_pick.h file * Fix comparison of integer expressions of different signedness in choice.cc file * Remove never referenced variable col_data in spat_op_impl_coo.cc * Remove never referenced variable allowed in global_uniform.cc * Fix comparison of integer expressions of different signedness in graph.cc * Fix comparison of integer expressions of different signedness in graph_apis.cc * Fix the un-used ctx variable in ndarray_partition.cc file for cpu only build * Fix comparison of integer expressions of different signedness in libra_partition.cc * Fix comparison of integer expressions of different signedness in graph_op.cc Co-authored-by:
Triston Cao <tristonc@nvidia.com> Co-authored-by:
Quan (Andy) Gan <coin2028@hotmail.com>
-
- 06 Jun, 2022 1 commit
-
-
Quan (Andy) Gan authored
Co-authored-by:Xin Yao <xiny@nvidia.com>
-
- 28 May, 2022 1 commit
-
-
Quan (Andy) Gan authored
-
- 26 Apr, 2022 1 commit
-
-
ayasar70 authored
* Based on issue #3436. Improving _SegmentCopyKernel s GPU utilization by switching to nonzero based thread assignment * fixing lint issues * Update cub for cuda 11.5 compatibility (#3468) * fixing type mismatch * tx guaranteed to be smaller than nnz. Hence removing last check * minor: updating comment * adding three unit tests for csr slice method to cover some corner cases * timing repeatkernel * clean * clean * clean * updating _SegmentMaskColKernel * Working on requests: removing sorted array check and adding comments to utility functions * fixing lint issue * Optimizing disjoint union kernel * Trying to resolve compilation issue on CI * [EMPTY] Relevant commit message here * applying revision requests on cpu/disjoint_union.cc * removing unnecessary casts * remove extra space Co-authored-by:
Abdurrahman Yasar <ayasar@nvidia.com> Co-authored-by:
nv-dlasalle <63612878+nv-dlasalle@users.noreply.github.com> Co-authored-by:
Jinjing Zhou <VoVAllen@users.noreply.github.com> Co-authored-by:
Quan (Andy) Gan <coin2028@hotmail.com>
-
- 23 Feb, 2022 2 commits
-
-
sanchit-misra authored
-
Minjie Wang authored
* WIP: TypedLinear and new RelGraphConv * wip * further simplify RGCN * a bunch of tweak for performance; add basic cpu support * update on segmm * wip: segment.cu * new backward kernel works * fix a bunch of bugs in kernel; leave idx_a for future * add nn test for typed_linear * rgcn nn test * bugfix in corner case; update RGCN README * doc * fix cpp lint * fix lint * fix ut * wip: hgtconv; presorted flag for rgcn * hgt code and ut; WIP: some fix on reorder graph * better typed linear init * fix ut * fix lint; add docstring
-
- 15 Feb, 2022 1 commit
-
-
Israt Nisa authored
* init * init * working cublasGemm * benchmark high-mem/low-mem, err gather_mm output * cuda kernel for bmm like kernel * removed cpu copy for E_per_Rel * benchmark code from Minjie * fixed cublas results in gathermm sorted * use GPU shared mem in unsorted gather mm * minor * Added an optimal version of gather_mm_unsorted * lint * init gather_mm_scatter * cublas transpose added * fixed h_offset for multiple rel * backward unittest * cublas support to transpose W * adding missed file * forgot to add header file * lint * lint * cleanup * lint * docstring * lint * added unittest * lint * lint * unittest * changed err type * skip cpu test * skip CPU code * move in-len loop inside * lint * added check different dim length for B * w_per_len is optional now * moved gather_mm to pytorch/backend with backward support * removed a_/b_trans support * transpose op inside GEMM call * removed out alloc from API, changed W 2D to 3D * Added se_gather_mm, Separate API for sortedE * Fixed gather_mm (unsorted) user interface * unsorted gmm backward + separate CAPI for un/sorted A * typecast to float to support atomicAdd * lint typecast * lint * added gather_mm_scatter * minor * const * design changes * Added idx_a, idx_b support gmm_scatter * dgl doc * lint * adding gather_mm in ops * lint * lint * minor * removed benchmark files * minor * empty commit Co-authored-by:Israt Nisa <nisisrat@amazon.com>
-
- 11 Feb, 2022 1 commit
-
-
ranzhejiang authored
* [feature] edge softmax refact. * delete file * fix backward and cmake version * fix backward * format function * fix setting * refix * refix * refix * refix * refix * refix * refix * refix * refix * refix * refix * refix * add cuda kernel for backward and rename some function * add benchmark for edge_softmax * fix format * remove cuda_backwrd * fix code format and add comment for op on CPU * fix lint Co-authored-by:Jinjing Zhou <VoVAllen@users.noreply.github.com>
-
- 17 Jan, 2022 1 commit
-
-
Quan (Andy) Gan authored
* oops * test
-
- 11 Jan, 2022 1 commit
-
-
MaoYuan Xian authored
* Pass the std:min argument's type, to avoid the compilation error. * Update parallel_for.h * Update negative_sampling.cc Co-authored-by:Quan (Andy) Gan <coin2028@hotmail.com>
-
- 07 Jan, 2022 1 commit
-
-
Quan (Andy) Gan authored
* first commit * a bunch of fixes * add unique * lint * lint * lint * address comments * Update negative_sampler.py * fix * description * address comments and fix * fix * replace unique with replace * test pylint * Update negative_sampler.py
-
- 16 Dec, 2021 1 commit
-
-
Israt Nisa authored
[Feature] Add CUDA support for `min` and `max` reducer in heterogeneous API for unary message functions (#3566) * CUDA support max/min reducer on forward pass * docstring * concised UpdateGradMinMax_hetero * reorganized UpdateGradMinMax_hetero * CUDA kernels for max/min reducer * variable name * lint check * changed CUDA 2D thread mapping to 1D * removed legacy cusparse for min/max reducer * git CI issue * restarting git CI * adding namespace std Co-authored-by:
Israt Nisa <nisisrat@amazon.com> Co-authored-by:
Quan (Andy) Gan <coin2028@hotmail.com>
-
- 06 Dec, 2021 1 commit
-
-
Quan (Andy) Gan authored
* first commit * second commit * spaghetti unit tests * rewrite test
-
- 03 Dec, 2021 1 commit
-
-
Israt Nisa authored
* min/max support for forward CPU heterograph * Added etype with each argU values * scatter_add needs fix * added scatter_add_hetero. Grads dont match for max reducer * storing ntype in argX * fixing scatter_add_hetero * hetero matches with torch's scatter add * works copy_e forward+cpu * added backward for copy_rhs * Computes gradient for all node types in one kernel * bug fix * unnitest for max/min on CPU * renamed scatter_add_hetero to update_grad_minmax_hetero * lint check and comment out cuda call for max. Code is for CPU only * lint check * replace inf with zero * minor * lint check * removed LIBXSMM code from hetro code * fixing backward operator of UpdateGradMinMaxHetero * removed backward from update_grad_minmax_hetero * docstring * improved docstring and coding style * Added pass by pointer for output * typos and pass by references * Support for copy_rhs * Added header <string> * fix bug in copy_u_max * Added comments and dimension check of all etypes * skip mxnet check * pass by pointer output arrays * updated docstring Co-authored-by:
Israt Nisa <nisisrat@amazon.com> Co-authored-by:
Quan (Andy) Gan <coin2028@hotmail.com>
-
- 17 Nov, 2021 1 commit
-
-
Israt Nisa authored
* Added SDDMMCOO_hetero support * removed redundant CUDA kernels * added benchmark for regression test * fix * fixed bug for single src node type Co-authored-by:
Israt Nisa <nisisrat@amazon.com> Co-authored-by:
Quan (Andy) Gan <coin2028@hotmail.com>
-
- 18 Oct, 2021 1 commit
-
-
David Min authored
* parallelize CSRRowSlice() * use parallel_for for the second loop Co-authored-by:
nv-dlasalle <63612878+nv-dlasalle@users.noreply.github.com> Co-authored-by:
Jinjing Zhou <VoVAllen@users.noreply.github.com>
-
- 17 Sep, 2021 1 commit
-
-
Rhett Ying authored
-
- 14 Sep, 2021 1 commit
-
-
Rhett Ying authored
* [Performance] improve coo2csr space complexity when row is not sorted * [Perf] replace std::vector<> by NDArray * keep both impl of unsorted coo to csr and choose according to graph density dynamically * refine criteria to choose btw Unsorted algos Co-authored-by:Ubuntu <ubuntu@ip-172-31-34-27.us-west-2.compute.internal>
-
- 13 Sep, 2021 2 commits
-
-
sanchit-misra authored
* Fixes bug #3312 * Fixing lint errors Co-authored-by:Mufei Li <mufeili1996@gmail.com>
-
Quan (Andy) Gan authored
-
- 02 Sep, 2021 1 commit
-
-
Tomasz Patejko authored
* [CPU, Parallel] Rewriting omp pragmas with parallel_for * [CPU, Parallel] Decrease number of calls to task function * c[CPU, Parallel] Modify calls to new interface of parallel_for
-
- 01 Sep, 2021 1 commit
-
-
xiang song(charlie.song) authored
[Feature] Add a HINT for the per edge type sampler of heterogeneous DistGraph that highlighting the etypes are sorted already. (#3260) * pass cpp test * distgraph use sorted edge flag. * lint * triger * update test Co-authored-by:Ubuntu <ubuntu@ip-172-31-2-66.ec2.internal>
-
- 31 Aug, 2021 1 commit
-
-
nv-dlasalle authored
* Optimize sampling * Stop initialization of array * Fix includes for linting * Move comment * Fix replace Co-authored-by:Da Zheng <zhengda1936@gmail.com>
-
- 28 Jul, 2021 1 commit
-
-
xiang song(charlie.song) authored
* fix. * fix. * fix. * fix. * Fix test * Deprecate old DistEmbedding impl, use synchronized embedding impl * Basic imple of heterogeneous on homogenenous sampling * make pass * Pass C++ test * Add python test code * lint * lint * Add MultiLayerEtypeNeighborSampler * Add unitest for single machine dataloader * Add dist dataloader test for edge type sampler * Fix lint * fix * support for per etype sample * Fix some bug and enable distributed training with per edge sample * fix * Now distributed training works * turn off some mxnet * turn off mxnet for some dist test * fix * upd * upd according to the comments * Fix * Fix test and now distributed works. * upd * upd * Fix * Fix bug * remove dead code. * upd * Fix * upd * Fix Co-authored-by:
Ubuntu <ubuntu@ip-172-31-71-112.ec2.internal> Co-authored-by:
Ubuntu <ubuntu@ip-172-31-2-66.ec2.internal> Co-authored-by:
Da Zheng <zhengda1936@gmail.com>
-
- 21 Jul, 2021 1 commit
-
-
Jinjing Zhou authored
* remove redundant fill * trigger ci
-
- 13 Jul, 2021 1 commit
-
-
sanchit-misra authored
* optimizations of spmm for CPU * Added names of contributors * Minor code cleanup * Moved the spmm optimization code to a new header file * Moved to DGL's logging method * removed duplicate code between SpMMSumCsr and SpMMCmpCsr * Changes made to follow Google coding style * Fixed lint errors in spmm.h * Fixed some lint errors from spmm_blocking_libxsmm.h * Fixed lint errors from spmm_blocking_libxsmm.h * Added comments to SpMMCreateLibxsmmKernel * to enable building of tests, and other cosmetic changes * disabling libxsmm on windows * Put a condition to avoid opt impl for FP64 as libxsmm does not have FP64 support yet * cosmetic changes and documentation * cosmetic changes * to pass lint tests * replaced multiple allocations for buffers of indices and edges with a single allocation Co-authored-by:Minjie Wang <wmjlyjemaine@gmail.com>
-
- 23 Jun, 2021 1 commit
-
-
Qidong Su authored
* update * update * update * update * lint * lint * update * update * update * update * update * update * update * update * update * update * update * update * update * update * lint * update * clone * update * update * update * update * replace idarray with ndarray * refactor cpp part * refactor python part * debug * refactor interface * test and doc * lint and test * lint * fix * fix * fix * const * doc * fix * fix * fix * fix * fix & doc * fix * fix * update * update * update * merge * doc * doc * lint * fix * more tests * doc * fix * fix * update * update * update * fix * fix Co-authored-by:Minjie Wang <wmjlyjemaine@gmail.com>
-
- 03 Jun, 2021 1 commit
-
-
Israt Nisa authored
* SpMM for heterograph * C APIs SDDMM heterograph * passes initial result * renamed eid with nid * aggregation on same ntype for multiple etypes * fix link check failure * lint check part 2 * lint check part 3 * Fixed SpMMCmpCsr Min op * added mem references * fixed fill(Max/Min), added const * removed newline * brought back docstring Co-authored-by:
Israt Nisa <nisisrat@amazon.com> Co-authored-by:
Da Zheng <zhengda1936@gmail.com>
-
- 01 Jun, 2021 1 commit
-
-
Qidong Su authored
* update * update * update * update * lint * lint * update * update * update * update * update * update * update * update * update * update * update * update * update * update * lint * update * clone * update * update * update * update * replace idarray with ndarray * refactor cpp part * refactor python part * debug * refactor interface * test and doc * lint and test * lint * fix * fix * fix * const * doc * fix * fix * fix * fix * fix & doc * fix * fix * fix * fix * fix * fix * update Co-authored-by:Minjie Wang <wmjlyjemaine@gmail.com>
-
- 20 May, 2021 1 commit
-
-
nv-dlasalle authored
[Feature][Performance] Implement NCCL wrapper for communicating NodeEmbeddings and sparse gradients. (#2825) * Split NCCL wrapper from sparse optimizer and sparse embedding * Add more unit tests for single node nccl * Fix unit test for tf * Switch to device histogram * Fix histgram issues * Finish migration to histogram * Handle cases with zero send/recieve data * Start on partition object * Get compiling * Updates * Add unit tests * Switch to partition object * Fix linting issues * Rename partition file * Add python doc * Fix python assert and finish doxygen comments * Remove stubs for range based partition to satisfy pylint * Wrap unit test in GPU only * Wrap explicit cuda call in ifdef * Merge with partition.py * update docstrings * Cleanup partition_op * Add Workspace object * Switch to using workspace object * Move last remainder based function out of nccl_api * Add error messages * Update docs with examples * Fix linting erros Co-authored-by:xiang song(charlie.song) <classicxsong@gmail.com>
-
- 17 May, 2021 1 commit
-
-
Quan (Andy) Gan authored
* test commit * fixes * oops * add docs * lint * why does it say I have a trailing whitespace * oh ok * fixes * why there's an invalid argument error * address comments * fix * address comments
-