1. 11 Mar, 2022 1 commit
  2. 09 Feb, 2022 1 commit
  3. 03 Feb, 2022 1 commit
    • Amit Aflalo's avatar
      mac arm64(m1) (#200) · d987d295
      Amit Aflalo authored
      * mac arm64(m1)
      
      * linting
      
      * # Compile for mac arm64
      
      * # Compile for mac arm64
      
      * removing .DS_Store
      d987d295
  4. 01 Feb, 2022 1 commit
    • Daniel Falbel's avatar
      Export symbols (#198) · 3bf43eb0
      Daniel Falbel authored
      * Mark exported symbols with `SPARSE_API` so they are available in the DLL on WIndows.
      
      * Export symbols by default in the Python library.
      
      * Sync headers with implementation.
      
      * Include the `SPARSE_API` macro.
      
      * Fix linting issue.
      3bf43eb0
  5. 30 Jan, 2022 1 commit
    • Daniel Falbel's avatar
      Add options to conditionally include Python (#196) · 8cc819d5
      Daniel Falbel authored
      * Add `WITH_PYTHON` to conditionally link to Python.
      
      * Only include `Python.h` when WITH_PYTHON is set.
      
      * Avoid including extensions.h as it includes Python.h.
      
      * Better way to include `getpid()`.
      
      * Define `WITH_PYTHON` when building with setup.py.
      
      * Only include Pyinit when building with Python.
      
      * Only include Pyinit when building with Python.
      8cc819d5
  6. 26 Jan, 2022 1 commit
    • Nick Stathas's avatar
      Skip unnecessary assertions and enable non-blocking data transfers (#195) · fe8c3ce3
      Nick Stathas authored
      * Uses the `trust_data` invariant to skip blocking assertions, when unnecessary, during construction of `SparseStorage` objects.
      * Refactors the dtype and device transfer APIs to align with `torch.Tensor` while maintaining backward compatibility.
      * No longer constructs dummy tensors when changing dtype or device.
      fe8c3ce3
  7. 14 Jan, 2022 1 commit
  8. 13 Nov, 2021 1 commit
  9. 22 Oct, 2021 3 commits
  10. 18 Oct, 2021 5 commits
  11. 17 Oct, 2021 1 commit
    • Feng Shi's avatar
      Update storage.py · c0805819
      Feng Shi authored
      BUG: when sparse_sizes is passed with (None, None), this causes an error
      c0805819
  12. 07 Oct, 2021 1 commit
  13. 20 Sep, 2021 1 commit
  14. 16 Sep, 2021 1 commit
  15. 09 Sep, 2021 8 commits
  16. 08 Sep, 2021 3 commits
  17. 26 Aug, 2021 2 commits
  18. 23 Aug, 2021 3 commits
  19. 18 Aug, 2021 2 commits
  20. 17 Aug, 2021 1 commit
    • Romeo Valentin's avatar
      Let torch determine correct cuda architecture · 407f53e2
      Romeo Valentin authored
      See `pytorch/torch/utils/cpp_extension.cpp:CUDAExtension`:
      >   By default the extension will be compiled to run on all archs of the cards visible during the
      >   building process of the extension, plus PTX. If down the road a new card is installed the
      >   extension may need to be recompiled. If a visible card has a compute capability (CC) that's
      >   newer than the newest version for which your nvcc can build fully-compiled binaries, Pytorch
      >   will make nvcc fall back to building kernels with the newest version of PTX your nvcc does
      >   support (see below for details on PTX).
      
      >   You can override the default behavior using `TORCH_CUDA_ARCH_LIST` to explicitly specify which
      >   CCs you want the extension to support:
      
      >   TORCH_CUDA_ARCH_LIST="6.1 8.6" python build_my_extension.py
      >   TORCH_CUDA_ARCH_LIST="5.2 6.0 6.1 7.0 7.5 8.0 8.6+PTX" python build_my_extension.py
      
      >   The +PTX option causes extension kernel binaries to include PTX instructions for the specified
      >   CC. PTX is an intermediate representation that allows kernels to runtime-compile for any CC >=
      >   the specified CC (for example, 8.6+PTX generates PTX that can runtime-compile for any GPU with
      >   CC >= 8.6). This improves your binary's forward compatibility. However, relying on older PTX to
      >   provide forward compat by runtime-compiling for newer CCs can modestly reduce performance on
      >   those newer CCs. If you know exact CC(s) of the GPUs you want to target, you're always better
      >   off specifying them individually. For example, if you want your extension to run on 8.0 and 8.6,
      >   "8.0+PTX" would work functionally because it includes PTX that can runtime-compile for 8.6, but
      >   "8.0 8.6" would be better.
      
      >   Note that while it's possible to include all supported archs, the more archs get included the
      >   slower the building process will be, as it will build a separate kernel image for each arch.
      407f53e2
  21. 16 Aug, 2021 1 commit