Commit 00129014 authored by Vijay Korthikanti's avatar Vijay Korthikanti
Browse files

avoid 'reverse checkpointed' term in argument description

parent 5d2e13a6
......@@ -492,9 +492,9 @@ def _add_training_args(parser):
help='Checkpoint activatins to allow for training '
'with larger models, sequences, and batch sizes. '
'It is supported at two granularities 1) full: '
'whole transformer layer is reverse checkpointed, '
'whole transformer layer is checkpointed, '
'2) selective: core attention part of the transformer '
'layer is reverse checkpointed.')
'layer is checkpointed.')
group.add_argument('--distribute-checkpointed-activations',
action='store_true',
help='If set, distribute checkpointed activations '
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment