Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
flash-attention
Commits
215930bce365dc31ed890edeee8517102ca2ee35
Switch branch/tag
flash-attention
flash_attn
flash_attn_triton.py
31 Oct, 2022
9 commits
Fix EVEN_M & EVEN_HEADDIM for headdim=40 in Triton bwd
· 215930bc
Tri Dao
authored
Oct 31, 2022
215930bc
Add debug_barrier for all headdims in Triton bwd
· 4f81aff4
Tri Dao
authored
Oct 31, 2022
4f81aff4
Disable some autotune configs that give wrong results in Triton bwd
· bedcbd6a
Tri Dao
authored
Oct 31, 2022
bedcbd6a
[WIP] Support all head dimensions up to 128 in the Triton bwd
· e78d509c
Tri Dao
authored
Oct 31, 2022
WIP because there seems to be some race conditions for head dimensions other than 16, 32, 64, 128.
e78d509c
Support all head dimensions up to 128 in the Triton fwd
· 008951f1
Tri Dao
authored
Oct 30, 2022
008951f1
Support arbitrary seqlens (both q & k) in Triton bwd
· b910bf14
Tri Dao
authored
Oct 30, 2022
b910bf14
Support arbitrary seqlen_k in Triton bwd
· dc554693
Tri Dao
authored
Oct 30, 2022
dc554693
Fix Triton fwd to support seqlen not multiples of 128
· d11341fd
Tri Dao
authored
Oct 30, 2022
d11341fd
Implement FlashAttention in Triton
· b0c0db81
Tri Dao
authored
Oct 30, 2022
b0c0db81