- 28 Feb, 2023 3 commits
- 27 Feb, 2023 2 commits
- 23 Feb, 2023 5 commits
- 22 Feb, 2023 1 commit
-
-
danyao12 authored
-
- 21 Feb, 2023 4 commits
-
-
guangzlu authored
* saved dropout random number in gridwise_batched_gemm_softmax_gemm_xdl_cshuffle_v2.hpp * modified device_grouped_gemm_softmax_gemm_permute_train_xdl_cshuffle.hpp * added z tensor for dropout storing * added z in example * can compile now * change fp16 xdl to bf16 * fixed some bugs in example * changed fwd file names * fixed some bugs in fwd drop verify * Delete device_grouped_multihead_attention_forward_xdl_cshuffle * Fwd drop verify2 (#585) * fixed some bugs in fwd drop verify * Delete device_grouped_multihead_attention_forward_xdl_cshuffle * added group fwd mha dropout verify * added dropout verify for grouped mha fp16 fwd * added bf16 fwd attn dropout verify * added dropout verify to batched mha fwd * added batched fla fwd bf16 dropout verify * changed some format * added switch for lse storing in attn fwd * added switch for lse storing in attn fwd * resolved conflicts in reference_dropout.hpp --------- Co-authored-by:ltqin <letao.qin@amd.com>
-
danyao12 authored
-
danyao12 authored
-
ltqin authored
-
- 17 Feb, 2023 1 commit
-
-
danyao12 authored
-
- 16 Feb, 2023 2 commits
- 15 Feb, 2023 6 commits
- 13 Feb, 2023 5 commits
- 12 Feb, 2023 4 commits
- 10 Feb, 2023 6 commits
- 09 Feb, 2023 1 commit
-
-
guangzlu authored
-