Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
flash-attention
Commits
bedcbd6a71cbf8470686a20934ac1ef2e70123cd
Switch branch/tag
flash-attention
flash_attn
flash_attn_triton.py
31 Oct, 2022
7 commits
Disable some autotune configs that give wrong results in Triton bwd
· bedcbd6a
Tri Dao
authored
Oct 31, 2022
bedcbd6a
[WIP] Support all head dimensions up to 128 in the Triton bwd
· e78d509c
Tri Dao
authored
Oct 31, 2022
WIP because there seems to be some race conditions for head dimensions other than 16, 32, 64, 128.
e78d509c
Support all head dimensions up to 128 in the Triton fwd
· 008951f1
Tri Dao
authored
Oct 30, 2022
008951f1
Support arbitrary seqlens (both q & k) in Triton bwd
· b910bf14
Tri Dao
authored
Oct 30, 2022
b910bf14
Support arbitrary seqlen_k in Triton bwd
· dc554693
Tri Dao
authored
Oct 30, 2022
dc554693
Fix Triton fwd to support seqlen not multiples of 128
· d11341fd
Tri Dao
authored
Oct 30, 2022
d11341fd
Implement FlashAttention in Triton
· b0c0db81
Tri Dao
authored
Oct 30, 2022
b0c0db81