Unverified Commit 6ababeb7 authored by Philip Turner's avatar Philip Turner Committed by GitHub
Browse files

Update usage.md

parent 72ad03ea
......@@ -123,3 +123,5 @@ yields the fastest BERT training on cloud instances in MLPerf training 2.0 (June
- [Jax](https://github.com/google/jax): an [implementation](https://github.com/lucidrains/flash-attention-jax)
in Jax by [lucidrains](https://github.com/lucidrains/).
- [Metal](https://developer.apple.com/metal/): an [implementation](https://github.com/philipturner/metal-flash-attention) by Philip Turner. This ports FlashAttention to mobile GPU architectures such as Apple silicon.
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment