Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
flash-attention
Commits
d4da6bfc8f26b1403a453bd9743cb19b96df6fbd
Switch branch/tag
flash-attention
csrc
flash_attn
src
utils.h
21 Jan, 2024
3 commits
Move rotary device functions to a separate file
· 395e5a0d
Tri Dao
authored
Jan 20, 2024
395e5a0d
Refactor masking in fwd pass into 1 object
· 66a127ae
Tri Dao
authored
Jan 20, 2024
66a127ae
Change inline to __forceinline__, use __grid_constant__ param
· ed4959b2
Tri Dao
authored
Jan 20, 2024
ed4959b2
14 Jan, 2024
1 commit
apply_dropout now takes tensor of rowcol layout
· 10dad612
Tri Dao
authored
Jan 14, 2024
10dad612
12 Jan, 2024
1 commit
Simplify SmemLayoutVtransposed in kernel_traits.h
· 8d1b169e
Tri Dao
authored
Jan 12, 2024
8d1b169e
16 Sep, 2023
1 commit
Implement rotary embedding in flash_attn_with_kvcache
· ccbb14f3
Tri Dao
authored
Sep 16, 2023
ccbb14f3
04 Sep, 2023
1 commit
Implement flash_attn_with_kvcache
· 37c6e054
Tri Dao
authored
Sep 04, 2023
37c6e054
01 Sep, 2023
1 commit
Remove old code in utils.h (#511)
· dd8a7549
Sophia Wisdom
authored
Sep 01, 2023
dd8a7549
13 Aug, 2023
2 commits
Prepare for Cutlass 3.2
· dbd79237
Tri Dao
authored
Aug 13, 2023
dbd79237
Update to Cutlass 3.1
· 3524e13c
Tri Dao
authored
Aug 13, 2023
3524e13c
17 Jul, 2023
1 commit
FlashAttention-2 release
· 4f285b35
Tri Dao
authored
Jul 17, 2023
4f285b35