Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
flash-attention
Commits
dc08ea1c33afca500a3d4ada907608f7815a11d9
Switch branch/tag
flash-attention
csrc
rotary
15 Mar, 2023
1 commit
Support H100 for other CUDA extensions
· dc08ea1c
Tri Dao
authored
Mar 15, 2023
dc08ea1c
14 Mar, 2023
1 commit
[Rotary] Implement GPT-J style (interleaved) rotary
· e45a46a5
Tri Dao
authored
Mar 14, 2023
e45a46a5
25 Dec, 2022
1 commit
Implement TensorParallel for MHA
· 1e712ea8
Tri Dao
authored
Dec 24, 2022
1e712ea8
05 Nov, 2022
1 commit
Implement rotary embedding in CUDA
· ca81f32e
Tri Dao
authored
Nov 04, 2022
ca81f32e