Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
flash-attention
Commits
419ea45b64fc5e34c61b48387a3e6af6ee1ff4e5
Switch branch/tag
flash-attention
training
configs
20 Jan, 2023
1 commit
fix typo in default.yaml
· 419ea45b
Ikko Eltociear Ashimine
authored
Jan 21, 2023
additionaly -> additionally
419ea45b
18 Jan, 2023
1 commit
[FusedDense] Support relu, rename FusedDenseGeluDense -> FusedMLP
· 88173a1a
Tri Dao
authored
Jan 17, 2023
88173a1a
01 Jan, 2023
1 commit
[Loss] Use flash_attn.losses.cross_entropy.CrossEntropyLoss
· 71befc19
Tri Dao
authored
Dec 31, 2022
71befc19
23 Dec, 2022
1 commit
Add smoothing for CrossEntropyParallel, rename to CrossEntropyLoss
· dff68c2b
Tri Dao
authored
Dec 23, 2022
dff68c2b
21 Dec, 2022
1 commit
Fix typo in config: train.gpu -> train.gpu_mem
· c2407dec
Tri Dao
authored
Dec 21, 2022
c2407dec
29 Nov, 2022
2 commits
Update configs, add results
· 4a6eaa9f
Tri Dao
authored
Nov 29, 2022
4a6eaa9f
Release training code
· 0bf5e500
Tri Dao
authored
Nov 28, 2022
0bf5e500