Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
Megatron-LM
Commits
9dec5374
Commit
9dec5374
authored
Aug 16, 2021
by
Lawrence McAfee
Browse files
added evaluation logic; finalized flag levels
parent
3bd2e973
Changes
2
Show whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
6 additions
and
7 deletions
+6
-7
megatron/arguments.py
megatron/arguments.py
+0
-4
megatron/training.py
megatron/training.py
+6
-3
No files found.
megatron/arguments.py
View file @
9dec5374
...
@@ -601,10 +601,6 @@ def _add_distributed_args(parser):
...
@@ -601,10 +601,6 @@ def _add_distributed_args(parser):
group
.
add_argument
(
'--use-cpu-initialization'
,
action
=
'store_true'
,
group
.
add_argument
(
'--use-cpu-initialization'
,
action
=
'store_true'
,
default
=
None
,
help
=
'If set, affine parallel weights '
default
=
None
,
help
=
'If set, affine parallel weights '
'initialization uses CPU'
)
'initialization uses CPU'
)
# group.add_argument('--empty-unused-memory-each-iter', action='store_true',
# help='Call torch.cuda.empty_cache() each iteration '
# '(training and eval), to reduce fragmentation',
# default=False)
group
.
add_argument
(
'--empty-unused-memory-each-iter'
,
default
=
0
,
type
=
int
,
group
.
add_argument
(
'--empty-unused-memory-each-iter'
,
default
=
0
,
type
=
int
,
choices
=
[
0
,
1
,
2
],
choices
=
[
0
,
1
,
2
],
help
=
'Call torch.cuda.empty_cache() each iteration '
help
=
'Call torch.cuda.empty_cache() each iteration '
...
...
megatron/training.py
View file @
9dec5374
...
@@ -363,8 +363,7 @@ def train_step(forward_step_func, data_iterator,
...
@@ -363,8 +363,7 @@ def train_step(forward_step_func, data_iterator,
optimizer
,
timers
,
forward_only
=
False
)
optimizer
,
timers
,
forward_only
=
False
)
# Empty unused memory
# Empty unused memory
if
args
.
empty_unused_memory_each_iter
>=
2
:
if
args
.
empty_unused_memory_each_iter
>=
1
:
raise
Exception
(
"hi."
)
torch
.
cuda
.
empty_cache
()
torch
.
cuda
.
empty_cache
()
# All-reduce if needed.
# All-reduce if needed.
...
@@ -414,7 +413,7 @@ def train_step(forward_step_func, data_iterator,
...
@@ -414,7 +413,7 @@ def train_step(forward_step_func, data_iterator,
skipped_iter
=
1
skipped_iter
=
1
# Empty unused memory
# Empty unused memory
if
args
.
empty_unused_memory_each_iter
>=
1
:
if
args
.
empty_unused_memory_each_iter
>=
2
:
torch
.
cuda
.
empty_cache
()
torch
.
cuda
.
empty_cache
()
if
mpu
.
is_pipeline_last_stage
(
ignore_virtual
=
True
):
if
mpu
.
is_pipeline_last_stage
(
ignore_virtual
=
True
):
...
@@ -725,6 +724,10 @@ def evaluate(forward_step_func, data_iterator, model, verbose=False):
...
@@ -725,6 +724,10 @@ def evaluate(forward_step_func, data_iterator, model, verbose=False):
forward_step_func
,
data_iterator
,
model
,
optimizer
=
None
,
forward_step_func
,
data_iterator
,
model
,
optimizer
=
None
,
timers
=
None
,
forward_only
=
True
)
timers
=
None
,
forward_only
=
True
)
# Empty unused memory
if
args
.
empty_unused_memory_each_iter
>=
1
:
torch
.
cuda
.
empty_cache
()
if
mpu
.
is_pipeline_last_stage
(
ignore_virtual
=
True
):
if
mpu
.
is_pipeline_last_stage
(
ignore_virtual
=
True
):
# Reduce across processes.
# Reduce across processes.
for
loss_dict
in
loss_dicts
:
for
loss_dict
in
loss_dicts
:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment