1. 28 Feb, 2023 3 commits
  2. 27 Feb, 2023 2 commits
  3. 23 Feb, 2023 5 commits
  4. 22 Feb, 2023 1 commit
  5. 21 Feb, 2023 4 commits
    • guangzlu's avatar
      Added dropout verification for flash attention forward (#593) · 042e4b8c
      guangzlu authored
      
      
      * saved dropout random number in gridwise_batched_gemm_softmax_gemm_xdl_cshuffle_v2.hpp
      
      * modified device_grouped_gemm_softmax_gemm_permute_train_xdl_cshuffle.hpp
      
      * added z tensor for dropout storing
      
      * added z in example
      
      * can compile now
      
      * change fp16 xdl to bf16
      
      * fixed some bugs in example
      
      * changed fwd file names
      
      * fixed some bugs in fwd drop verify
      
      * Delete device_grouped_multihead_attention_forward_xdl_cshuffle
      
      * Fwd drop verify2 (#585)
      
      * fixed some bugs in fwd drop verify
      
      * Delete device_grouped_multihead_attention_forward_xdl_cshuffle
      
      * added group fwd mha dropout verify
      
      * added dropout verify for grouped mha fp16 fwd
      
      * added bf16 fwd attn dropout verify
      
      * added dropout verify to batched mha fwd
      
      * added batched fla fwd bf16 dropout verify
      
      * changed some format
      
      * added switch for lse storing in attn fwd
      
      * added switch for lse storing in attn fwd
      
      * resolved conflicts in reference_dropout.hpp
      
      ---------
      Co-authored-by: default avatarltqin <letao.qin@amd.com>
      042e4b8c
    • danyao12's avatar
      fix drop==0 compiler issue in prototype1 · 272b7574
      danyao12 authored
      272b7574
    • danyao12's avatar
      63c2d069
    • ltqin's avatar
      fix drop==0, compiler issue · 82ce7f4e
      ltqin authored
      82ce7f4e
  6. 17 Feb, 2023 1 commit
  7. 16 Feb, 2023 2 commits
  8. 15 Feb, 2023 6 commits
  9. 13 Feb, 2023 5 commits
  10. 12 Feb, 2023 4 commits
  11. 10 Feb, 2023 6 commits
  12. 09 Feb, 2023 1 commit