operator.h 2.75 KB
Newer Older
Hang Zhang's avatar
Hang Zhang committed
1
#include <torch/extension.h>
Hang Zhang's avatar
Hang Zhang committed
2
#include <ATen/ATen.h>
Hang Zhang's avatar
Hang Zhang committed
3
4
#include <vector>

Hang Zhang's avatar
Hang Zhang committed
5
at::Tensor ROIAlign_Forward_CUDA(
Hang Zhang's avatar
Hang Zhang committed
6
7
8
9
10
11
12
  const at::Tensor input,
  const at::Tensor rois,
  int64_t pooled_height,
  int64_t pooled_width,
  double spatial_scale,
  int64_t sample_ratio);

Hang Zhang's avatar
Hang Zhang committed
13
at::Tensor ROIAlign_Backward_CUDA(
Hang Zhang's avatar
Hang Zhang committed
14
15
16
17
18
19
20
21
22
23
24
  const at::Tensor rois,
  const at::Tensor grad_output,
  int64_t b_size,
  int64_t channels,
  int64_t height,
  int64_t width,
  int64_t pooled_height,
  int64_t pooled_width,
  double spatial_scale,
  int64_t sampling_ratio);

Hang Zhang's avatar
Hang Zhang committed
25
26
27
28
29
std::vector<at::Tensor> Non_Max_Suppression_CUDA(
  const at::Tensor& input,
  const at::Tensor& scores,
  double thresh);

Hang Zhang's avatar
Hang Zhang committed
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
at::Tensor Aggregate_Forward_CUDA(
  const at::Tensor A_,
  const at::Tensor X_,
  const at::Tensor C_);

std::vector<at::Tensor> Aggregate_Backward_CUDA(
  const at::Tensor GE_,
  const at::Tensor A_,
  const at::Tensor X_,
  const at::Tensor C_);

at::Tensor ScaledL2_Forward_CUDA(
  const at::Tensor X_,
  const at::Tensor C_,
  const at::Tensor S_);

std::vector<at::Tensor> ScaledL2_Backward_CUDA(
  const at::Tensor GSL_,
  const at::Tensor X_,
  const at::Tensor C_,
  const at::Tensor S_,
  const at::Tensor SL_);

at::Tensor BatchNorm_Forward_CUDA(
  const at::Tensor input_, 
  const at::Tensor mean_,
  const at::Tensor std_,
  const at::Tensor gamma_,
Hang Zhang's avatar
Hang Zhang committed
58
59
60
61
62
63
64
65
66
67
  const at::Tensor beta_,
  float eps);

at::Tensor BatchNorm_Forward_Inp_CUDA(
    const at::Tensor input_, 
    const at::Tensor ex_,
    const at::Tensor exs_,
    const at::Tensor gamma_,
    const at::Tensor beta_,
    float eps);
Hang Zhang's avatar
Hang Zhang committed
68
69
70
71

std::vector<at::Tensor> BatchNorm_Backward_CUDA(
  const at::Tensor gradoutput_,
  const at::Tensor input_,
Hang Zhang's avatar
Hang Zhang committed
72
73
74
75
76
77
78
79
80
81
82
  const at::Tensor ex_, 
  const at::Tensor exs_,
  const at::Tensor gamma_,
  const at::Tensor beta_,
  float eps);

std::vector<at::Tensor> BatchNorm_Inp_Backward_CUDA(
  const at::Tensor gradoutput_,
  const at::Tensor output_,
  const at::Tensor ex_, 
  const at::Tensor exs_,
Hang Zhang's avatar
Hang Zhang committed
83
  const at::Tensor gamma_,
Hang Zhang's avatar
Hang Zhang committed
84
85
  const at::Tensor beta_,
  float eps);
Hang Zhang's avatar
Hang Zhang committed
86

Hang Zhang's avatar
Hang Zhang committed
87
std::vector<at::Tensor> Expectation_Forward_CUDA(
Hang Zhang's avatar
Hang Zhang committed
88
89
  const at::Tensor input_);

Hang Zhang's avatar
Hang Zhang committed
90
at::Tensor Expectation_Backward_CUDA(
Hang Zhang's avatar
Hang Zhang committed
91
  const at::Tensor input_,
Hang Zhang's avatar
Hang Zhang committed
92
93
94
95
96
97
98
99
100
101
102
103
104
  const at::Tensor gradEx_,
  const at::Tensor gradExs_);

at::Tensor Expectation_Inp_Backward_CUDA(
  const at::Tensor gradInput_,
  const at::Tensor output_,
  const at::Tensor gradEx_,
  const at::Tensor gradExs_,
  const at::Tensor ex_, 
  const at::Tensor exs_,
  const at::Tensor gamma_,
  const at::Tensor beta_,
  float eps);
Hang Zhang's avatar
Hang Zhang committed
105

Hang Zhang's avatar
Hang Zhang committed
106
107
108
void LeakyRelu_Forward_CUDA(at::Tensor z, float slope);

void LeakyRelu_Backward_CUDA(at::Tensor z, at::Tensor dz, float slope);
Hang Zhang's avatar
Hang Zhang committed
109
110
111
112
113
114
115
116
117

void CONV_RECTIFY_CUDA(
  at::Tensor& output,
  const at::Tensor& input,
  at::IntArrayRef kernel_size,
  at::IntArrayRef stride,
  at::IntArrayRef padding,
  at::IntArrayRef dilation,
  bool avg_mode);