operator.h 3.11 KB
Newer Older
Hang Zhang's avatar
Hang Zhang committed
1
#include <torch/extension.h>
Hang Zhang's avatar
Hang Zhang committed
2
#include <ATen/ATen.h>
Hang Zhang's avatar
Hang Zhang committed
3
4
#include <vector>

Hang Zhang's avatar
Hang Zhang committed
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
std::vector<at::Tensor> box_encoder(
  const int N_img,
  const at::Tensor& bbox_input,
  const at::Tensor& bbox_offsets,
  const at::Tensor& labels_input,
  const at::Tensor& dbox,
  const float criteria = 0.5);

std::vector<at::Tensor> random_horiz_flip(
  at::Tensor& img,
  at::Tensor& bboxes,
  const at::Tensor& bbox_offsets,
  const float p,
  const bool nhwc);

Hang Zhang's avatar
Hang Zhang committed
20
at::Tensor ROIAlign_Forward_CUDA(
Hang Zhang's avatar
Hang Zhang committed
21
22
23
24
25
26
27
  const at::Tensor input,
  const at::Tensor rois,
  int64_t pooled_height,
  int64_t pooled_width,
  double spatial_scale,
  int64_t sample_ratio);

Hang Zhang's avatar
Hang Zhang committed
28
at::Tensor ROIAlign_Backward_CUDA(
Hang Zhang's avatar
Hang Zhang committed
29
30
31
32
33
34
35
36
37
38
39
  const at::Tensor rois,
  const at::Tensor grad_output,
  int64_t b_size,
  int64_t channels,
  int64_t height,
  int64_t width,
  int64_t pooled_height,
  int64_t pooled_width,
  double spatial_scale,
  int64_t sampling_ratio);

Hang Zhang's avatar
Hang Zhang committed
40
41
42
43
44
std::vector<at::Tensor> Non_Max_Suppression_CUDA(
  const at::Tensor& input,
  const at::Tensor& scores,
  double thresh);

Hang Zhang's avatar
Hang Zhang committed
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
at::Tensor Aggregate_Forward_CUDA(
  const at::Tensor A_,
  const at::Tensor X_,
  const at::Tensor C_);

std::vector<at::Tensor> Aggregate_Backward_CUDA(
  const at::Tensor GE_,
  const at::Tensor A_,
  const at::Tensor X_,
  const at::Tensor C_);

at::Tensor ScaledL2_Forward_CUDA(
  const at::Tensor X_,
  const at::Tensor C_,
  const at::Tensor S_);

std::vector<at::Tensor> ScaledL2_Backward_CUDA(
  const at::Tensor GSL_,
  const at::Tensor X_,
  const at::Tensor C_,
  const at::Tensor S_,
  const at::Tensor SL_);

at::Tensor BatchNorm_Forward_CUDA(
  const at::Tensor input_, 
  const at::Tensor mean_,
  const at::Tensor std_,
  const at::Tensor gamma_,
Hang Zhang's avatar
Hang Zhang committed
73
74
75
76
77
78
79
80
81
82
  const at::Tensor beta_,
  float eps);

at::Tensor BatchNorm_Forward_Inp_CUDA(
    const at::Tensor input_, 
    const at::Tensor ex_,
    const at::Tensor exs_,
    const at::Tensor gamma_,
    const at::Tensor beta_,
    float eps);
Hang Zhang's avatar
Hang Zhang committed
83
84
85
86

std::vector<at::Tensor> BatchNorm_Backward_CUDA(
  const at::Tensor gradoutput_,
  const at::Tensor input_,
Hang Zhang's avatar
Hang Zhang committed
87
88
89
90
91
92
93
94
95
96
97
  const at::Tensor ex_, 
  const at::Tensor exs_,
  const at::Tensor gamma_,
  const at::Tensor beta_,
  float eps);

std::vector<at::Tensor> BatchNorm_Inp_Backward_CUDA(
  const at::Tensor gradoutput_,
  const at::Tensor output_,
  const at::Tensor ex_, 
  const at::Tensor exs_,
Hang Zhang's avatar
Hang Zhang committed
98
  const at::Tensor gamma_,
Hang Zhang's avatar
Hang Zhang committed
99
100
  const at::Tensor beta_,
  float eps);
Hang Zhang's avatar
Hang Zhang committed
101

Hang Zhang's avatar
Hang Zhang committed
102
std::vector<at::Tensor> Expectation_Forward_CUDA(
Hang Zhang's avatar
Hang Zhang committed
103
104
  const at::Tensor input_);

Hang Zhang's avatar
Hang Zhang committed
105
at::Tensor Expectation_Backward_CUDA(
Hang Zhang's avatar
Hang Zhang committed
106
  const at::Tensor input_,
Hang Zhang's avatar
Hang Zhang committed
107
108
109
110
111
112
113
114
115
116
117
118
119
  const at::Tensor gradEx_,
  const at::Tensor gradExs_);

at::Tensor Expectation_Inp_Backward_CUDA(
  const at::Tensor gradInput_,
  const at::Tensor output_,
  const at::Tensor gradEx_,
  const at::Tensor gradExs_,
  const at::Tensor ex_, 
  const at::Tensor exs_,
  const at::Tensor gamma_,
  const at::Tensor beta_,
  float eps);
Hang Zhang's avatar
Hang Zhang committed
120

Hang Zhang's avatar
Hang Zhang committed
121
122
123
void LeakyRelu_Forward_CUDA(at::Tensor z, float slope);

void LeakyRelu_Backward_CUDA(at::Tensor z, at::Tensor dz, float slope);
Hang Zhang's avatar
Hang Zhang committed
124
125
126
127
128
129
130
131
132

void CONV_RECTIFY_CUDA(
  at::Tensor& output,
  const at::Tensor& input,
  at::IntArrayRef kernel_size,
  at::IntArrayRef stride,
  at::IntArrayRef padding,
  at::IntArrayRef dilation,
  bool avg_mode);