Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
ColossalAI
Commits
5ab9a712
Commit
5ab9a712
authored
Apr 02, 2022
by
MaxT
Committed by
binmakeswell
Apr 06, 2022
Browse files
[NFC] polish colossalai/kernel/cuda_native/csrc/moe_cuda.cpp code style (#642)
parent
10afec72
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
72 additions
and
91 deletions
+72
-91
colossalai/kernel/cuda_native/csrc/moe_cuda.cpp
colossalai/kernel/cuda_native/csrc/moe_cuda.cpp
+72
-91
No files found.
colossalai/kernel/cuda_native/csrc/moe_cuda.cpp
View file @
5ab9a712
#include <torch/extension.h>
#include <torch/extension.h>
torch
::
Tensor
moe_dispatch_cuda_forward
(
int
s
,
int
ec
,
int
h
,
torch
::
Tensor
moe_dispatch_cuda_forward
(
torch
::
Tensor
batch_tokens
,
int
s
,
int
ec
,
int
h
,
torch
::
Tensor
mask
,
torch
::
Tensor
batch_tokens
,
torch
::
Tensor
dest_idx
);
torch
::
Tensor
mask
,
torch
::
Tensor
dest_idx
);
torch
::
Tensor
moe_dispatch_cuda_backward
(
int
s
,
int
ec
,
int
h
,
torch
::
Tensor
expert_grad
,
torch
::
Tensor
moe_dispatch_cuda_backward
(
torch
::
Tensor
mask
,
int
s
,
int
ec
,
int
h
,
torch
::
Tensor
dest_idx
);
torch
::
Tensor
expert_grad
,
torch
::
Tensor
mask
,
torch
::
Tensor
moe_combine_cuda_forward
(
int
s
,
int
e
,
int
c
,
int
h
,
torch
::
Tensor
dest_idx
);
torch
::
Tensor
expert_tokens
,
torch
::
Tensor
logits
,
torch
::
Tensor
mask
,
torch
::
Tensor
moe_combine_cuda_forward
(
torch
::
Tensor
dest_idx
);
int
s
,
int
e
,
int
c
,
int
h
,
torch
::
Tensor
expert_tokens
,
std
::
vector
<
torch
::
Tensor
>
torch
::
Tensor
logits
,
moe_combine_cuda_backward
(
int
s
,
int
e
,
int
c
,
int
h
,
torch
::
Tensor
tokens_grad
,
torch
::
Tensor
mask
,
torch
::
Tensor
expert_tokens
,
torch
::
Tensor
logits
,
torch
::
Tensor
dest_idx
);
torch
::
Tensor
mask
,
torch
::
Tensor
dest_idx
);
std
::
vector
<
torch
::
Tensor
>
moe_combine_cuda_backward
(
int
s
,
int
e
,
int
c
,
int
h
,
torch
::
Tensor
tokens_grad
,
torch
::
Tensor
expert_tokens
,
torch
::
Tensor
logits
,
torch
::
Tensor
mask
,
torch
::
Tensor
dest_idx
);
torch
::
Tensor
cumsum_sub_one_in_dim0
(
torch
::
Tensor
mask
);
torch
::
Tensor
cumsum_sub_one_in_dim0
(
torch
::
Tensor
mask
);
#define CHECK_CUDA(x) TORCH_CHECK(x.device().is_cuda(), #x " must be a CUDA tensor")
#define CHECK_CUDA(x) \
#define CHECK_CONTIGUOUS(x) TORCH_CHECK(x.is_contiguous(), #x " must be contiguous")
TORCH_CHECK
(
x
.
device
().
is_cuda
(),
#
x
" must be a CUDA tensor"
)
#define CHECK_INPUT(x) CHECK_CUDA(x); CHECK_CONTIGUOUS(x)
#define CHECK_CONTIGUOUS(x) \
TORCH_CHECK
(
x
.
is_contiguous
(),
#
x
" must be contiguous"
)
#define CHECK_INPUT(x) \
CHECK_CUDA
(
x
);
\
CHECK_CONTIGUOUS
(
x
)
torch
::
Tensor
moe_dispatch_forward
(
torch
::
Tensor
moe_dispatch_forward
(
int
s
,
int
ec
,
int
h
,
int
s
,
int
ec
,
int
h
,
torch
::
Tensor
batch_tokens
,
torch
::
Tensor
batch_tokens
,
torch
::
Tensor
mask
,
torch
::
Tensor
dest_idx
)
{
torch
::
Tensor
mask
,
torch
::
Tensor
dest_idx
)
{
CHECK_INPUT
(
batch_tokens
);
CHECK_INPUT
(
batch_tokens
);
CHECK_CUDA
(
mask
);
CHECK_CUDA
(
mask
);
CHECK_CUDA
(
dest_idx
);
CHECK_CUDA
(
dest_idx
);
return
moe_dispatch_cuda_forward
(
return
moe_dispatch_cuda_forward
(
s
,
ec
,
h
,
batch_tokens
,
mask
,
dest_idx
);
s
,
ec
,
h
,
batch_tokens
,
mask
,
dest_idx
);
}
}
torch
::
Tensor
moe_dispatch_backward
(
torch
::
Tensor
moe_dispatch_backward
(
int
s
,
int
ec
,
int
h
,
int
s
,
int
ec
,
int
h
,
torch
::
Tensor
expert_grad
,
torch
::
Tensor
expert_grad
,
torch
::
Tensor
mask
,
torch
::
Tensor
mask
,
torch
::
Tensor
dest_idx
)
{
torch
::
Tensor
dest_idx
)
{
CHECK_INPUT
(
expert_grad
);
CHECK_INPUT
(
expert_grad
);
CHECK_CUDA
(
mask
);
CHECK_CUDA
(
mask
);
CHECK_CUDA
(
dest_idx
);
CHECK_CUDA
(
dest_idx
);
return
moe_dispatch_cuda_backward
(
return
moe_dispatch_cuda_backward
(
s
,
ec
,
h
,
expert_grad
,
mask
,
dest_idx
);
s
,
ec
,
h
,
expert_grad
,
mask
,
dest_idx
);
}
}
torch
::
Tensor
moe_combine_forward
(
torch
::
Tensor
moe_combine_forward
(
int
s
,
int
e
,
int
c
,
int
h
,
int
s
,
int
e
,
int
c
,
int
h
,
torch
::
Tensor
expert_tokens
,
torch
::
Tensor
expert_tokens
,
torch
::
Tensor
logits
,
torch
::
Tensor
mask
,
torch
::
Tensor
logits
,
torch
::
Tensor
dest_idx
)
{
torch
::
Tensor
mask
,
torch
::
Tensor
dest_idx
)
{
CHECK_INPUT
(
expert_tokens
);
CHECK_INPUT
(
logits
);
CHECK_INPUT
(
expert_tokens
);
CHECK_CUDA
(
mask
);
CHECK_INPUT
(
logits
);
CHECK_CUDA
(
dest_idx
);
CHECK_CUDA
(
mask
);
CHECK_CUDA
(
dest_idx
);
return
moe_combine_cuda_forward
(
s
,
e
,
c
,
h
,
expert_tokens
,
logits
,
mask
,
dest_idx
);
return
moe_combine_cuda_forward
(
s
,
e
,
c
,
h
,
expert_tokens
,
logits
,
mask
,
dest_idx
);
}
}
std
::
vector
<
torch
::
Tensor
>
moe_combine_backward
(
std
::
vector
<
torch
::
Tensor
>
int
s
,
int
e
,
int
c
,
int
h
,
moe_combine_backward
(
int
s
,
int
e
,
int
c
,
int
h
,
torch
::
Tensor
tokens_grad
,
torch
::
Tensor
tokens_grad
,
torch
::
Tensor
expert_tokens
,
torch
::
Tensor
logits
,
torch
::
Tensor
expert_tokens
,
torch
::
Tensor
mask
,
torch
::
Tensor
dest_idx
)
{
torch
::
Tensor
logits
,
torch
::
Tensor
mask
,
CHECK_INPUT
(
tokens_grad
);
torch
::
Tensor
dest_idx
)
{
CHECK_INPUT
(
logits
);
CHECK_CUDA
(
mask
);
CHECK_INPUT
(
tokens_grad
);
CHECK_CUDA
(
dest_idx
);
CHECK_INPUT
(
logits
);
CHECK_CUDA
(
mask
);
return
moe_combine_cuda_backward
(
s
,
e
,
c
,
h
,
tokens_grad
,
expert_tokens
,
CHECK_CUDA
(
dest_idx
);
logits
,
mask
,
dest_idx
);
return
moe_combine_cuda_backward
(
s
,
e
,
c
,
h
,
tokens_grad
,
expert_tokens
,
logits
,
mask
,
dest_idx
);
}
}
torch
::
Tensor
moe_cumsum
(
torch
::
Tensor
mask
)
{
torch
::
Tensor
moe_cumsum
(
torch
::
Tensor
mask
)
{
CHECK_INPUT
(
mask
);
CHECK_INPUT
(
mask
);
return
cumsum_sub_one_in_dim0
(
mask
);
return
cumsum_sub_one_in_dim0
(
mask
);
}
}
PYBIND11_MODULE
(
TORCH_EXTENSION_NAME
,
m
)
{
PYBIND11_MODULE
(
TORCH_EXTENSION_NAME
,
m
)
{
m
.
def
(
"cumsum_sub_one"
,
&
moe_cumsum
,
m
.
def
(
"cumsum_sub_one"
,
&
moe_cumsum
,
"Fast cumsum operation in dim0"
);
"Fast cumsum operation in dim0"
);
m
.
def
(
"dispatch_forward"
,
&
moe_dispatch_forward
,
m
.
def
(
"dispatch_forward"
,
&
moe_dispatch_forward
,
"Forward operation in MoE dispatch function"
);
"Forward operation in MoE dispatch function"
);
m
.
def
(
"dispatch_backward"
,
&
moe_dispatch_backward
,
m
.
def
(
"dispatch_backward"
,
&
moe_dispatch_backward
,
"Backward operation in MoE dispatch function"
);
"Backward operation in MoE dispatch function"
);
m
.
def
(
"combine_forward"
,
&
moe_combine_forward
,
m
.
def
(
"combine_forward"
,
&
moe_combine_forward
,
"Combine operation in MoE combine function"
);
"Combine operation in MoE combine function"
);
m
.
def
(
"combine_backward"
,
&
moe_combine_backward
,
m
.
def
(
"combine_backward"
,
&
moe_combine_backward
,
"Combine operation in MoE combine function"
);
"Combine operation in MoE combine function"
);
}
}
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment