Commit 24b257f1 authored by sunzhq2's avatar sunzhq2
Browse files

init

parent 920b3c0f
{
"operator": "gelu",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072],
[8192]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
]
}
\ No newline at end of file
{
"operator": "gemm",
"iterations": 100,
"input_shape_groups": {
"M": [4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072],
"KN": [
[1024, 1024],
[4096, 4096],
[8192, 8192],
[12288, 12288],
[16384, 32],
[16384, 128],
[16384, 1024],
[32, 16384],
[128, 16384],
[1024, 16384]
]
},
"dtype": [
"float32",
"bfloat16",
"float16",
"int8"
]
}
\ No newline at end of file
{
"operator": "gemv",
"iterations": 100,
"input_shape_groups": [
{
"M": [1],
"K": [16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192],
"N": [4096, 8192]
},
{
"M": [1],
"K": [4096, 8192],
"N": [16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192]
}
],
"dtype": [
"float32",
"bfloat16",
"float16",
"int8"
]
}
\ No newline at end of file
{
"operator": "group_gemm",
"iterations": 100,
"input_shape_groups": {
"gemm_group": [1, 2, 3, 4, 5, 6, 7, 8],
"batch": [1, 2, 4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192],
"KN": [[32, 16384], [16384, 32], [16384, 16384]]
},
"dtype": [
"float32",
"bfloat16",
"float16",
"int8"
]
}
\ No newline at end of file
{
"operator": "host2device",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072],
[1024]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
]
}
\ No newline at end of file
{
"operator": "index_add",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[1024],
[1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072, 262144, 524288]
],
[
[1024],
[1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072, 262144, 524288]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
]
}
\ No newline at end of file
{
"operator": "layernorm",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[1024],
[1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072, 262144, 524288]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
]
}
\ No newline at end of file
{
"operator": "log",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072],
[8192]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
]
}
\ No newline at end of file
{
"operator": "mul",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072],
[8192]
],
[
[4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072],
[8192]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
]
}
\ No newline at end of file
{
"operator": "p2p",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072, 262144, 524288, 1048576, 2097152],
[1024]
],
[
[8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072, 262144, 524288, 1048576, 2097152],
[1024]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
],
"group": [
2,
4,
8
]
}
\ No newline at end of file
{
"operator": "reduce_max",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[1024],
[1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072, 262144, 524288]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
]
}
\ No newline at end of file
{
"operator": "reduce_min",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[1024],
[1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072, 262144, 524288]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
]
}
\ No newline at end of file
{
"operator": "reduce_sum",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[1024],
[1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072, 262144, 524288]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
]
}
\ No newline at end of file
{
"operator": "reducescatter",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072, 262144, 524288, 1048576, 2097152],
[1024]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
],
"group": [
2,
4,
8
]
}
\ No newline at end of file
{
"operator": "scatter",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[1024],
[1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072, 262144, 524288]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
]
}
\ No newline at end of file
{
"operator": "silu",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072],
[8192]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
]
}
\ No newline at end of file
{
"operator": "sin",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072],
[8192]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
]
}
\ No newline at end of file
{
"operator": "softmax",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[1024],
[1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072, 262144, 524288]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
]
}
\ No newline at end of file
{
"operator": "sort",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[1024],
[1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072, 262144, 524288]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
]
}
\ No newline at end of file
{
"operator": "sqrt",
"iterations": 100,
"input_shape_groups": {
"inputs": [
[
[4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072],
[8192]
]
]
},
"dtype": [
"float32",
"bfloat16",
"float16"
]
}
\ No newline at end of file
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment