- 23 Jan, 2023 1 commit
-
-
Tri Dao authored
-
- 19 Jan, 2023 1 commit
-
-
Tri Dao authored
-
- 18 Jan, 2023 2 commits
- 16 Jan, 2023 3 commits
- 15 Jan, 2023 1 commit
-
-
Tri Dao authored
-
- 13 Jan, 2023 1 commit
-
-
Kiarash Jamali authored
Documentation says default is 0.1, but the code has attention_dropout default at 0.0
-
- 08 Jan, 2023 4 commits
- 07 Jan, 2023 4 commits
-
-
Tri Dao authored
-
Tri Dao authored
-
Darius Lam authored
-
Tri Dao authored
-
- 04 Jan, 2023 1 commit
-
-
Tri Dao authored
-
- 02 Jan, 2023 2 commits
- 01 Jan, 2023 3 commits
- 30 Dec, 2022 1 commit
-
-
Tri Dao authored
-
- 28 Dec, 2022 2 commits
- 27 Dec, 2022 3 commits
- 25 Dec, 2022 3 commits
- 24 Dec, 2022 1 commit
-
-
Tri Dao authored
-
- 23 Dec, 2022 2 commits
- 21 Dec, 2022 1 commit
-
-
Tri Dao authored
-
- 20 Dec, 2022 1 commit
-
-
Tri Dao authored
-
- 19 Dec, 2022 1 commit
-
-
Tri Dao authored
-
- 17 Dec, 2022 1 commit
-
-
Alexander Ploshkin authored
-
- 16 Dec, 2022 1 commit
-
-
Alexander Ploshkin authored
-