1. 30 Jan, 2022 1 commit
    • Daniel Falbel's avatar
      Add options to conditionally include Python (#196) · 8cc819d5
      Daniel Falbel authored
      * Add `WITH_PYTHON` to conditionally link to Python.
      
      * Only include `Python.h` when WITH_PYTHON is set.
      
      * Avoid including extensions.h as it includes Python.h.
      
      * Better way to include `getpid()`.
      
      * Define `WITH_PYTHON` when building with setup.py.
      
      * Only include Pyinit when building with Python.
      
      * Only include Pyinit when building with Python.
      8cc819d5
  2. 26 Jan, 2022 1 commit
    • Nick Stathas's avatar
      Skip unnecessary assertions and enable non-blocking data transfers (#195) · fe8c3ce3
      Nick Stathas authored
      * Uses the `trust_data` invariant to skip blocking assertions, when unnecessary, during construction of `SparseStorage` objects.
      * Refactors the dtype and device transfer APIs to align with `torch.Tensor` while maintaining backward compatibility.
      * No longer constructs dummy tensors when changing dtype or device.
      fe8c3ce3
  3. 14 Jan, 2022 1 commit
  4. 13 Nov, 2021 1 commit
  5. 22 Oct, 2021 3 commits
  6. 18 Oct, 2021 5 commits
  7. 17 Oct, 2021 1 commit
    • Feng Shi's avatar
      Update storage.py · c0805819
      Feng Shi authored
      BUG: when sparse_sizes is passed with (None, None), this causes an error
      c0805819
  8. 07 Oct, 2021 1 commit
  9. 20 Sep, 2021 1 commit
  10. 16 Sep, 2021 1 commit
  11. 09 Sep, 2021 8 commits
  12. 08 Sep, 2021 3 commits
  13. 26 Aug, 2021 2 commits
  14. 23 Aug, 2021 3 commits
  15. 18 Aug, 2021 2 commits
  16. 17 Aug, 2021 1 commit
    • Romeo Valentin's avatar
      Let torch determine correct cuda architecture · 407f53e2
      Romeo Valentin authored
      See `pytorch/torch/utils/cpp_extension.cpp:CUDAExtension`:
      >   By default the extension will be compiled to run on all archs of the cards visible during the
      >   building process of the extension, plus PTX. If down the road a new card is installed the
      >   extension may need to be recompiled. If a visible card has a compute capability (CC) that's
      >   newer than the newest version for which your nvcc can build fully-compiled binaries, Pytorch
      >   will make nvcc fall back to building kernels with the newest version of PTX your nvcc does
      >   support (see below for details on PTX).
      
      >   You can override the default behavior using `TORCH_CUDA_ARCH_LIST` to explicitly specify which
      >   CCs you want the extension to support:
      
      >   TORCH_CUDA_ARCH_LIST="6.1 8.6" python build_my_extension.py
      >   TORCH_CUDA_ARCH_LIST="5.2 6.0 6.1 7.0 7.5 8.0 8.6+PTX" python build_my_extension.py
      
      >   The +PTX option causes extension kernel binaries to include PTX instructions for the specified
      >   CC. PTX is an intermediate representation that allows kernels to runtime-compile for any CC >=
      >   the specified CC (for example, 8.6+PTX generates PTX that can runtime-compile for any GPU with
      >   CC >= 8.6). This improves your binary's forward compatibility. However, relying on older PTX to
      >   provide forward compat by runtime-compiling for newer CCs can modestly reduce performance on
      >   those newer CCs. If you know exact CC(s) of the GPUs you want to target, you're always better
      >   off specifying them individually. For example, if you want your extension to run on 8.0 and 8.6,
      >   "8.0+PTX" would work functionally because it includes PTX that can runtime-compile for 8.6, but
      >   "8.0 8.6" would be better.
      
      >   Note that while it's possible to include all supported archs, the more archs get included the
      >   slower the building process will be, as it will build a separate kernel image for each arch.
      407f53e2
  17. 16 Aug, 2021 1 commit
  18. 15 Aug, 2021 1 commit
    • rusty1s's avatar
      typo · 3dbd2282
      rusty1s authored
      3dbd2282
  19. 10 Aug, 2021 3 commits