- 14 Nov, 2022 4 commits
- 13 Nov, 2022 1 commit
-
-
Tri Dao authored
-
- 10 Nov, 2022 1 commit
-
-
Tri Dao authored
To avoid import error if one doesn't have rotary_emb installed
-
- 07 Nov, 2022 1 commit
-
-
Tri Dao authored
-
- 06 Nov, 2022 1 commit
-
-
Tri Dao authored
-
- 05 Nov, 2022 2 commits
- 04 Nov, 2022 3 commits
- 03 Nov, 2022 1 commit
-
-
Tri Dao authored
-
- 02 Nov, 2022 1 commit
-
-
Tri Dao authored
-
- 01 Nov, 2022 2 commits
- 31 Oct, 2022 10 commits
- 24 Oct, 2022 3 commits
- 23 Oct, 2022 2 commits
- 21 Oct, 2022 2 commits
- 14 Oct, 2022 2 commits
- 06 Oct, 2022 1 commit
-
-
Antoine Adam authored
According to the `setup.py` file, only dependencies are torch and einops. But the `bert_padding.py` file requires `numpy` only to multiply the elements of a `torch.Size` object. This change aims at allowing the use of FlashAttention without numpy.
-
- 11 Sep, 2022 1 commit
-
-
Tri Dao authored
-
- 06 Sep, 2022 1 commit
-
-
eric-tc-wong authored
Recasting query and key after rotary_emb()
-
- 09 Aug, 2022 1 commit
-
-
Tri Dao authored
-