Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
composable_kernel
Commits
b19fe02b
Commit
b19fe02b
authored
Sep 26, 2023
by
letaoqin
Browse files
change 'fwd' inc file name to 'infer'
parent
6dbced07
Changes
6
Hide whitespace changes
Inline
Side-by-side
Showing
6 changed files
with
6 additions
and
6 deletions
+6
-6
example/52_flash_atten_bias/batched_gemm_multihead_attention_bias_infer.cpp
...tten_bias/batched_gemm_multihead_attention_bias_infer.cpp
+1
-1
example/52_flash_atten_bias/batched_gemm_multihead_attention_infer.cpp
...ash_atten_bias/batched_gemm_multihead_attention_infer.cpp
+1
-1
example/52_flash_atten_bias/grouped_mutihead_attention_bias_infer.cpp
...lash_atten_bias/grouped_mutihead_attention_bias_infer.cpp
+1
-1
example/52_flash_atten_bias/run_batched_multihead_attention_bias_infer.inc
...atten_bias/run_batched_multihead_attention_bias_infer.inc
+1
-1
example/52_flash_atten_bias/run_batched_multihead_attention_infer.inc
...lash_atten_bias/run_batched_multihead_attention_infer.inc
+1
-1
example/52_flash_atten_bias/run_grouped_multihead_attention_bias_infer.inc
...atten_bias/run_grouped_multihead_attention_bias_infer.inc
+1
-1
No files found.
example/52_flash_atten_bias/batched_gemm_multihead_attention_bias_infer.cpp
View file @
b19fe02b
...
...
@@ -157,6 +157,6 @@ using ReferenceGemm1Instance = ck::tensor_operation::host::ReferenceBatchedGemm<
B1ElementOp
,
CElementOp
>
;
#include "run_batched_multihead_attention_bias_
forward
.inc"
#include "run_batched_multihead_attention_bias_
infer
.inc"
int
main
(
int
argc
,
char
*
argv
[])
{
return
run
(
argc
,
argv
);
}
example/52_flash_atten_bias/batched_gemm_multihead_attention_infer.cpp
View file @
b19fe02b
...
...
@@ -157,6 +157,6 @@ using ReferenceGemm1Instance = ck::tensor_operation::host::ReferenceBatchedGemm<
B1ElementOp
,
CElementOp
>
;
#include "run_batched_multihead_attention_
forward
.inc"
#include "run_batched_multihead_attention_
infer
.inc"
int
main
(
int
argc
,
char
*
argv
[])
{
return
run
(
argc
,
argv
);
}
example/52_flash_atten_bias/grouped_mutihead_attention_bias_infer.cpp
View file @
b19fe02b
...
...
@@ -156,6 +156,6 @@ using ReferenceGemm1Instance = ck::tensor_operation::host::ReferenceBatchedGemm<
B1ElementOp
,
CElementOp
>
;
#include "run_grouped_multihead_attention_bias_
forward
.inc"
#include "run_grouped_multihead_attention_bias_
infer
.inc"
int
main
(
int
argc
,
char
*
argv
[])
{
return
run
(
argc
,
argv
);
}
example/52_flash_atten_bias/run_batched_multihead_attention_bias_
forward
.inc
→
example/52_flash_atten_bias/run_batched_multihead_attention_bias_
infer
.inc
View file @
b19fe02b
...
...
@@ -5,7 +5,7 @@ int run(int argc, char* argv[])
{
bool
do_verification
=
true
;
int
init_method
=
1
;
bool
time_kernel
=
tru
e
;
bool
time_kernel
=
fals
e
;
// GEMM shape for A/B0/B1/C
// C_g_m_o = A_g_m_k * B0_g_k_n * B1_g_n_o
...
...
example/52_flash_atten_bias/run_batched_multihead_attention_
forward
.inc
→
example/52_flash_atten_bias/run_batched_multihead_attention_
infer
.inc
View file @
b19fe02b
...
...
@@ -5,7 +5,7 @@ int run(int argc, char* argv[])
{
bool
do_verification
=
true
;
int
init_method
=
1
;
bool
time_kernel
=
tru
e
;
bool
time_kernel
=
fals
e
;
// GEMM shape for A/B0/B1/C
// C_g_m_o = A_g_m_k * B0_g_k_n * B1_g_n_o
...
...
example/52_flash_atten_bias/run_grouped_multihead_attention_bias_
forward
.inc
→
example/52_flash_atten_bias/run_grouped_multihead_attention_bias_
infer
.inc
View file @
b19fe02b
...
...
@@ -5,7 +5,7 @@ int run(int argc, char* argv[])
{
bool
do_verification
=
true
;
int
init_method
=
1
;
bool
time_kernel
=
tru
e
;
bool
time_kernel
=
fals
e
;
bool
input_permute
=
false
;
bool
output_permute
=
true
;
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment