Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
0311ba21
Unverified
Commit
0311ba21
authored
Apr 08, 2021
by
Stas Bekman
Committed by
GitHub
Apr 08, 2021
Browse files
typo (#11152)
* typo * style
parent
269c9638
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
3 additions
and
3 deletions
+3
-3
docs/source/main_classes/trainer.rst
docs/source/main_classes/trainer.rst
+3
-3
No files found.
docs/source/main_classes/trainer.rst
View file @
0311ba21
...
...
@@ -355,9 +355,9 @@ Notes:
able
to
use
significantly
larger
batch
sizes
using
the
same
hardware
(
e
.
g
.
3
x
and
even
bigger
)
which
should
lead
to
significantly
shorter
training
time
.
3.
To
use
the
second
version
of
Sharded
data
-
parallelism
,
add
``--
sharded_ddp
zero_dp_2
``
or
``--
sharded_ddp
zero_dp_3
`
to
the
command
line
arguments
,
and
make
sure
you
have
added
the
distributed
launcher
``-
m
torch
.
distributed
.
launch
--
nproc_per_node
=
NUMBER_OF_GPUS_YOU_HAVE
``
if
you
haven
't been using it already.
3.
To
use
the
second
version
of
Sharded
data
-
parallelism
,
add
``--
sharded_ddp
zero_dp_2
``
or
``--
sharded_ddp
zero_dp_3
``
to
the
command
line
arguments
,
and
make
sure
you
have
added
the
distributed
launcher
``-
m
torch
.
distributed
.
launch
--
nproc_per_node
=
NUMBER_OF_GPUS_YOU_HAVE
``
if
you
haven
't been using it already.
For example here is how you could use it for ``run_translation.py`` with 2 GPUs:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment