"vscode:/vscode.git/clone" did not exist on "1f2c3c31008f2f8221cbdd4e1749a7faf5b2e548"
Unverified Commit 734d8991 authored by Jeff Rasley's avatar Jeff Rasley Committed by GitHub
Browse files
parent b652395e
DeepSpeed Activation Checkpointing
======================
The activation checkpointing API's in DeepSpeed can be used to enable a range of memory optimizations relating
to activation checkpointing. These include activation partitioning across
GPUs when using model parallelism, CPU Checkpointing, contiguous memory optimizations, etc.
Please see the `DeepSpeed JSON config <https://www.deepspeed.ai/docs/config-json/>`_ for the full set.
Here we present the activation checkpointing API's.
Please see the enabling DeepSpeed for Megatron-LM tutorial for usage details.
.. autofunction:: deepspeed.checkpointing.configure
.. autofunction:: deepspeed.checkpointing.is_configured
.. autofunction:: deepspeed.checkpointing.checkpoint
.. autofunction:: deepspeed.checkpointing.reset
.. autofunction:: deepspeed.checkpointing.get_cuda_rng_tracker
.. autofunction:: deepspeed.checkpointing.model_parallel_cuda_manual_seed
.. autoclass:: deepspeed.checkpointing.CudaRNGStatesTracker
.. autoclass:: deepspeed.checkpointing.CheckpointFunction
......@@ -76,4 +76,11 @@ add_module_names = True
autoclass_content = 'both'
autodoc_mock_imports = ["torch", "apex", "mpi4py", "tensorboardX"]
autodoc_mock_imports = [
"torch",
"apex",
"mpi4py",
"tensorboardX",
"deepspeed_transformer_cuda",
"deepspeed_stochastic_transformer_cuda",
]
......@@ -8,7 +8,6 @@ Model Setup
:maxdepth: 2
initialize
checkpointing
Training API
------------
......@@ -28,6 +27,14 @@ Checkpointing API
activation-checkpointing
Transformer Kernel API
----------------------
.. toctree::
:maxdepth: 2
kernel
Indices and tables
------------------
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment