Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
flash-attention
Commits
1bbebccc
Commit
1bbebccc
authored
Jul 09, 2022
by
Tri Dao
Browse files
Edit README to mention bf16 support
parent
de19de7a
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
2 additions
and
2 deletions
+2
-2
README.md
README.md
+2
-2
No files found.
README.md
View file @
1bbebccc
...
...
@@ -23,8 +23,8 @@ PYTHONPATH=$PWD python benchmarks/benchmark_flash_attention.py
FlashAttention currently supports:
1.
Turing or Ampere GPUs (e.g., A100, RTX 3090, T4, RTX 2080).
2.
fp16.
3.
Head dimensions 16, 32, 64, 128 (
bw
d requires A100).
2.
fp16
and bf16 (bf16 requires Ampere GPUs)
.
3.
Head dimensions 16, 32, 64, 128 (
head dim 128 backwar
d requires A100).
Our tentative roadmap:
1.
[Jun 2022] Make package pip-installable.
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment