- 31 Oct, 2022 8 commits
- 25 Oct, 2022 1 commit
-
-
Tri Dao authored
-
- 24 Oct, 2022 4 commits
- 23 Oct, 2022 4 commits
- 22 Oct, 2022 1 commit
-
-
Tri Dao authored
-
- 21 Oct, 2022 3 commits
- 17 Oct, 2022 4 commits
-
-
Tri Dao authored
fix typo in function mha_fwd
-
robotcator authored
-
robotcator authored
-
YangShu authored
as title.
-
- 16 Oct, 2022 1 commit
-
-
Tri Dao authored
-
- 14 Oct, 2022 2 commits
- 10 Oct, 2022 1 commit
-
-
Tri Dao authored
build wheel workflow
-
- 06 Oct, 2022 2 commits
-
-
Tri Dao authored
remove numpy dependency
-
Antoine Adam authored
According to the `setup.py` file, only dependencies are torch and einops. But the `bert_padding.py` file requires `numpy` only to multiply the elements of a `torch.Size` object. This change aims at allowing the use of FlashAttention without numpy.
-
- 05 Oct, 2022 5 commits
-
-
Tri Dao authored
Make flash attention compile on Windows.
-
Eric Engelhart authored
-
Eric Engelhart authored
-
Eric Engelhart authored
-
Tri Dao authored
-
- 26 Sep, 2022 2 commits
-
-
robotcator authored
-
robotcator authored
-
- 12 Sep, 2022 1 commit
-
-
Tri Dao authored
-
- 11 Sep, 2022 1 commit
-
-
Tri Dao authored
-