- 02 Dec, 2021 15 commits
-
-
Chao Liu authored
-
Jing Zhang authored
-
Chao Liu authored
-
Chao Liu authored
-
Chao Liu authored
-
Chao Liu authored
update readme for example/3_conv_xdl
-
Chao Liu authored
update readme for example/2_gemm_xdl_bias_relu_add
-
Chao Liu authored
-
Chao Liu authored
-
Chao Liu authored
-
Chao Liu authored
-
Jing Zhang authored
-
Jing Zhang authored
-
Chao Liu authored
-
Chao Liu authored
-
- 01 Dec, 2021 2 commits
- 30 Nov, 2021 2 commits
- 24 Nov, 2021 1 commit
-
-
zjing14 authored
-
- 23 Nov, 2021 6 commits
- 22 Nov, 2021 3 commits
- 20 Nov, 2021 4 commits
- 18 Nov, 2021 3 commits
-
-
Chao Liu authored
* reworking vector_type * use __builtin_memcpy for bit_cast and vector access of scalar pointer * clean up
-
zjing14 authored
* init * refactor for 1x1 * rename e0_e1 * add e1 with bugs * debug * fixed * fixed e1 * add timer * imprve threadwise gemm with dot2 * add e2 * tuning * seperate c2 * add nhwc * restore nchwc * clean * opt * fixed; tuning * add BGlobalMoveSliceWindowStepHacks{} * tuning * repeat running * adjust * merge v5r1 nchwc * add adaptors * split k0 k1 in c_thread_grid * split h and w * remove v5r1 nhwc * clean for pr * remove host_conv_add * clean code * clean * add dynamic support * static mode * test static * add conv+add fusion * fixed validation * naming fix * use activ_enum * make static * refactor conv_add for InMem::add * add bias * add conv_out * add configurable makeddesc * add maxpool fusion * add maxpool host for validation * enable static desc * conv-only use v5r1_add * test * test * for binary dumps * fixed incorrect results due to typo * clean * debugging maxpool * workaround with offset trick * clean code * modularize ops of fusion * add gridwise_gemm_v3 * create seperate fusion fun * enable dynamic mode of conv and conv+resize_add * add dynamic mode of maxpool * add pass by point * add activ_type as arguments * merge develop * clean * reset config to old default Co-authored-by:Chao Liu <chao.liu2@amd.com>
-
zjing14 authored
* fixed bfloat16 issues * refactor type_convert * fixed host_convolution_forward for ushort Co-authored-by:Chao Liu <chao.liu2@amd.com>
-
- 16 Nov, 2021 2 commits
-
-
zjing14 authored
* fixed bfloat16 issues * refactor type_convert Co-authored-by:Chao Liu <chao.liu2@amd.com>
-
Jing Zhang authored
-
- 15 Nov, 2021 2 commits
-
-
zjing14 authored
* init StaticBufferV2 * clean * adopt old output stage for staticBufferV2 * clean * remove hack * clean * clean * add parameters * clean code * move c_buffer alloc into blockwise gemm * add adaptors for m/n_thread_data_on_grid * tweak gemm * adjust blockwise_gemm_xdlops * tweak * update conv * update script * adding bwd 1x1 * update script * adding 1x1 bwd * debugging bwd 1x1 failure * update script * update script * test * test v100 * add bf16_1k * clang-format * clean * add bfp16 for gfx908 * add verification * clean up * clean code * restore bfl16 * clean * add bfp16 support into gemm_driver * apply new generator to other drivers * add int8 support * cleanb * clean * clean * clean Co-authored-by:
Chao Liu <chao.liu2@amd.com> Co-authored-by:
Chao Liu <lc.roy86@gmail.com> Co-authored-by:
root <root@hayabusa6111.amd.com>
-
Chao Liu authored
* start fixing 16bit data packing * adding StaticTensor * adding StaticTensor * adding StaticTensor * add missing constexpr * adding static tensor * adding static tensor * adding transpose * add inline asm for transpose 2x2 of half_t * add general transpose_vectors(), but have unnecessary register initialization using v_mov * fix unnecessary register initialization in transpose_vector by using more pass-by-reference * add hardcoded logic for NHWC wrw * improve asm for v_pack * make ThreadwiseTensorSliceTransfer_v3r2 support any tensor * tweak * reorganize file
-