Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
Megatron-LM
Commits
6e9d5cb0
"model/git@developer.sourcefind.cn:OpenDAS/ollama.git" did not exist on "3b96a93672377129f2a2aafc447e79ef1ca48c5f"
Commit
6e9d5cb0
authored
Dec 12, 2020
by
mohammad
Committed by
Deepak Narayanan
Dec 19, 2020
Browse files
Move args writer to the beginning of training
parent
8a6e56b8
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
7 additions
and
4 deletions
+7
-4
megatron/initialize.py
megatron/initialize.py
+3
-4
megatron/training.py
megatron/training.py
+4
-0
No files found.
megatron/initialize.py
View file @
6e9d5cb0
...
@@ -79,8 +79,6 @@ def initialize_megatron(extra_args_provider=None, args_defaults={},
...
@@ -79,8 +79,6 @@ def initialize_megatron(extra_args_provider=None, args_defaults={},
# Autoresume.
# Autoresume.
_init_autoresume
()
_init_autoresume
()
# Write arguments to tensorboard.
_write_args_to_tensorboard
()
# No continuation function
# No continuation function
return
None
return
None
...
@@ -154,13 +152,14 @@ def _set_random_seed(seed_):
...
@@ -154,13 +152,14 @@ def _set_random_seed(seed_):
raise
ValueError
(
'Seed ({}) should be a positive integer.'
.
format
(
seed
))
raise
ValueError
(
'Seed ({}) should be a positive integer.'
.
format
(
seed
))
def
_
write_args_to_tensorboard
():
def
write_args_to_tensorboard
():
"""Write arguments to tensorboard."""
"""Write arguments to tensorboard."""
args
=
get_args
()
args
=
get_args
()
writer
=
get_tensorboard_writer
()
writer
=
get_tensorboard_writer
()
if
writer
:
if
writer
:
for
arg
in
vars
(
args
):
for
arg
in
vars
(
args
):
writer
.
add_text
(
arg
,
str
(
getattr
(
args
,
arg
)))
writer
.
add_text
(
arg
,
str
(
getattr
(
args
,
arg
)),
global_step
=
args
.
iteration
)
def
_initialize_mem_buffs
():
def
_initialize_mem_buffs
():
...
...
megatron/training.py
View file @
6e9d5cb0
...
@@ -41,6 +41,7 @@ from megatron.checkpointing import save_checkpoint
...
@@ -41,6 +41,7 @@ from megatron.checkpointing import save_checkpoint
from
megatron.fp16
import
FP16_Module
from
megatron.fp16
import
FP16_Module
from
megatron.fp16
import
FP16_Optimizer
from
megatron.fp16
import
FP16_Optimizer
from
megatron.initialize
import
initialize_megatron
from
megatron.initialize
import
initialize_megatron
from
megatron.initialize
import
write_args_to_tensorboard
from
megatron.learning_rates
import
AnnealingLR
from
megatron.learning_rates
import
AnnealingLR
from
megatron.model
import
DistributedDataParallel
as
LocalDDP
from
megatron.model
import
DistributedDataParallel
as
LocalDDP
from
megatron.model
import
get_params_for_weight_decay_optimization
from
megatron.model
import
get_params_for_weight_decay_optimization
...
@@ -811,6 +812,9 @@ def train(forward_step_func, model, optimizer, lr_scheduler,
...
@@ -811,6 +812,9 @@ def train(forward_step_func, model, optimizer, lr_scheduler,
args
=
get_args
()
args
=
get_args
()
timers
=
get_timers
()
timers
=
get_timers
()
# Write args to tensorboard
write_args_to_tensorboard
()
# Turn on training mode which enables dropout.
# Turn on training mode which enables dropout.
model
.
train
()
model
.
train
()
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment