Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
flash-attention
Commits
6ac8e63a85acdf717783873c67eb4cb648c621a4
Switch branch/tag
flash-attention
vllm_flash_attn
ops
triton
rotary.py
28 Mar, 2024
1 commit
flash_attn -> vllm_flash_attn
· 6ac8e63a
Woosuk Kwon
authored
Mar 28, 2024
6ac8e63a
11 Sep, 2023
1 commit
[Gen] Fix calling update_graph_cache in tests
· 8a733cbd
Tri Dao
authored
Sep 10, 2023
8a733cbd
06 Sep, 2023
1 commit
[Rotary] Set device before launching Triton kernel to avoid error
· 97951590
Tri Dao
authored
Sep 05, 2023
97951590
04 Sep, 2023
1 commit
[Rotary] Implement varlen rotary
· b28ec236
Tri Dao
authored
Sep 03, 2023
b28ec236
03 Sep, 2023
3 commits
[Rotary] Clean up rotary Triton implementation a bit
· 861c8257
Tri Dao
authored
Sep 03, 2023
861c8257
[Rotary] Speed up rotary kernel when interleaved=True
· 1c523c1c
Tri Dao
authored
Sep 03, 2023
1c523c1c
[Rotary] Implement rotary in Triton
· 942fcbf0
Tri Dao
authored
Sep 03, 2023
942fcbf0