Commit 00129014 authored by Vijay Korthikanti's avatar Vijay Korthikanti
Browse files

avoid 'reverse checkpointed' term in argument description

parent 5d2e13a6
...@@ -492,9 +492,9 @@ def _add_training_args(parser): ...@@ -492,9 +492,9 @@ def _add_training_args(parser):
help='Checkpoint activatins to allow for training ' help='Checkpoint activatins to allow for training '
'with larger models, sequences, and batch sizes. ' 'with larger models, sequences, and batch sizes. '
'It is supported at two granularities 1) full: ' 'It is supported at two granularities 1) full: '
'whole transformer layer is reverse checkpointed, ' 'whole transformer layer is checkpointed, '
'2) selective: core attention part of the transformer ' '2) selective: core attention part of the transformer '
'layer is reverse checkpointed.') 'layer is checkpointed.')
group.add_argument('--distribute-checkpointed-activations', group.add_argument('--distribute-checkpointed-activations',
action='store_true', action='store_true',
help='If set, distribute checkpointed activations ' help='If set, distribute checkpointed activations '
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment