/*! * Copyright (c) 2020-2022 by Contributors * @file tensoradapter.h * @brief Header file for functions exposed by the adapter library. * * Functions in this library must be exported with extern "C" so that DGL can * locate them with dlsym(3) (or GetProcAddress on Windows). */ #ifndef TENSORADAPTER_H_ #define TENSORADAPTER_H_ #ifdef DGL_USE_CUDA #include #endif // DGL_USE_CUDA namespace tensoradapter { extern "C" { /*! * @brief Allocate a piece of CPU memory via * PyTorch's CPUAllocator * * @param nbytes The size to be allocated. * @return Pointer to the allocated memory. */ void* CPURawAlloc(size_t nbytes); /*! * @brief Free the CPU memory. * * @param ptr Pointer to the memory to be freed. */ void CPURawDelete(void* ptr); #ifdef DGL_USE_CUDA /*! * @brief Allocate a piece of GPU memory via * PyTorch's THCCachingAllocator. * * @param nbytes The size to be allocated. * @param stream The stream to be allocated on. * @return Pointer to the allocated memory. */ void* CUDARawAlloc(size_t nbytes, cudaStream_t stream); /*! * @brief Free the GPU memory. * * @param ptr Pointer to the memory to be freed. */ void CUDARawDelete(void* ptr); /*! * @brief Get the current CUDA stream. */ cudaStream_t CUDACurrentStream(); /*! * @brief Let the caching allocator know which streams are using this tensor. * * @param ptr Pointer of the tensor to be recorded. * @param stream The stream that is using this tensor. * @param device_id Device of the tensor. */ void RecordStream(void* ptr, cudaStream_t stream, int device_id); #endif // DGL_USE_CUDA } }; // namespace tensoradapter #endif // TENSORADAPTER_H_