Unverified Commit 734d8991 authored by Jeff Rasley's avatar Jeff Rasley Committed by GitHub
Browse files
parent b652395e
Transformer Kernels
===================
The transformer kernel API in DeepSpeed can be used to create BERT transformer layer for
more efficient pre-training and fine-tuning, it includes the transformer layer configurations and
transformer layer module initialization.
Here we present the transformer kernel API.
Please see the `BERT pre-training tutorial <https://www.deepspeed.ai/tutorials/bert-pretraining/>`_ for usage details.
DeepSpeed Transformer Config
----------------------------
.. autoclass:: deepspeed.DeepSpeedTransformerConfig
DeepSpeed Transformer Layer
----------------------------
.. autoclass:: deepspeed.DeepSpeedTransformerLayer
...@@ -8,3 +8,4 @@ tensorflow-gpu==1.15.2 ...@@ -8,3 +8,4 @@ tensorflow-gpu==1.15.2
pytest pytest
pytest-forked pytest-forked
pre-commit pre-commit
clang-format
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment