Commit b0ed0a73 authored by Tri Dao's avatar Tri Dao
Browse files

Mention DeepSpeed inference in usage.md

parent 25387b24
...@@ -15,6 +15,9 @@ PR or email us. We'd very much like to hear from you! ...@@ -15,6 +15,9 @@ PR or email us. We'd very much like to hear from you!
[On-going](https://github.com/huggingface/transformers/pull/18439), blogpost [On-going](https://github.com/huggingface/transformers/pull/18439), blogpost
coming soon. coming soon.
- Microsoft's [DeepSpeed](https://github.com/microsoft/DeepSpeed):
FlashAttention is [integrated](https://github.com/microsoft/DeepSpeed/blob/ec13da6ba7cabc44bb4745a64a208b8580792954/deepspeed/ops/transformer/inference/triton_ops.py) into DeepSpeed's inference engine.
- MosaicML [Composer](https://github.com/mosaicml/composer) - MosaicML [Composer](https://github.com/mosaicml/composer)
[library](https://www.mosaicml.com/blog/gpt-3-quality-for-500k). Composer is a [library](https://www.mosaicml.com/blog/gpt-3-quality-for-500k). Composer is a
library for efficient neural network training. library for efficient neural network training.
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment