Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
Megatron-LM
Commits
f86bb671
Commit
f86bb671
authored
Dec 26, 2019
by
Mohammad Shoeybi
Browse files
checked and bert, gpt, and albert albert run
parent
d6485684
Changes
5
Hide whitespace changes
Inline
Side-by-side
Showing
5 changed files
with
13 additions
and
4 deletions
+13
-4
arguments.py
arguments.py
+1
-1
configure_data.py
configure_data.py
+2
-0
gpt2_data_loader.py
gpt2_data_loader.py
+1
-1
pretrain_albert.py
pretrain_albert.py
+2
-1
pretrain_gpt2.py
pretrain_gpt2.py
+7
-1
No files found.
arguments.py
View file @
f86bb671
...
@@ -278,7 +278,7 @@ def add_data_args(parser):
...
@@ -278,7 +278,7 @@ def add_data_args(parser):
help
=
'path(s) to the validation data.'
)
help
=
'path(s) to the validation data.'
)
group
.
add_argument
(
'--test-data'
,
nargs
=
'*'
,
default
=
None
,
group
.
add_argument
(
'--test-data'
,
nargs
=
'*'
,
default
=
None
,
help
=
'path(s) to the testing data.'
)
help
=
'path(s) to the testing data.'
)
group
.
add_argument
(
'--data-path'
,
type
=
str
,
default
=
None
,
group
.
add_argument
(
'--data-path'
,
nargs
=
'+'
,
default
=
None
,
help
=
'path to combined dataset to split'
)
help
=
'path to combined dataset to split'
)
group
.
add_argument
(
'--split'
,
default
=
'1000,1,1'
,
group
.
add_argument
(
'--split'
,
default
=
'1000,1,1'
,
help
=
'comma-separated list of proportions for training,'
help
=
'comma-separated list of proportions for training,'
...
...
configure_data.py
View file @
f86bb671
...
@@ -131,6 +131,8 @@ def make_loaders(args):
...
@@ -131,6 +131,8 @@ def make_loaders(args):
if
eval_seq_length
is
not
None
and
eval_seq_length
<
0
:
if
eval_seq_length
is
not
None
and
eval_seq_length
<
0
:
eval_seq_length
=
eval_seq_length
*
world_size
eval_seq_length
=
eval_seq_length
*
world_size
split
=
get_split
(
args
)
split
=
get_split
(
args
)
if
args
.
data_path
is
not
None
:
args
.
train_data
=
args
.
data_path
data_set_args
=
{
data_set_args
=
{
'path'
:
args
.
train_data
,
'path'
:
args
.
train_data
,
'seq_length'
:
seq_length
,
'seq_length'
:
seq_length
,
...
...
gpt2_data_loader.py
View file @
f86bb671
...
@@ -57,7 +57,7 @@ def make_gpt2_dataloaders(args):
...
@@ -57,7 +57,7 @@ def make_gpt2_dataloaders(args):
pin_memory
=
True
)
pin_memory
=
True
)
train
=
make_data_loader_
(
args
.
train_data
)
train
=
make_data_loader_
(
args
.
train_data
)
valid
=
make_data_loader_
(
args
.
val_data
)
valid
=
make_data_loader_
(
args
.
val
id
_data
)
test
=
make_data_loader_
(
args
.
test_data
)
test
=
make_data_loader_
(
args
.
test_data
)
args
.
do_train
=
False
args
.
do_train
=
False
...
...
pretrain_albert.py
View file @
f86bb671
...
@@ -143,9 +143,10 @@ def get_train_val_test_data(args):
...
@@ -143,9 +143,10 @@ def get_train_val_test_data(args):
print_rank_0
(
' validation: {}'
.
format
(
train_val_test_num_samples
[
1
]))
print_rank_0
(
' validation: {}'
.
format
(
train_val_test_num_samples
[
1
]))
print_rank_0
(
' test: {}'
.
format
(
train_val_test_num_samples
[
2
]))
print_rank_0
(
' test: {}'
.
format
(
train_val_test_num_samples
[
2
]))
assert
len
(
args
.
data_path
)
==
1
train_ds
,
valid_ds
,
test_ds
=
build_train_valid_test_datasets
(
train_ds
,
valid_ds
,
test_ds
=
build_train_valid_test_datasets
(
vocab_file
=
args
.
vocab
,
vocab_file
=
args
.
vocab
,
data_prefix
=
args
.
data_path
,
data_prefix
=
args
.
data_path
[
0
]
,
data_impl
=
args
.
data_impl
,
data_impl
=
args
.
data_impl
,
splits_string
=
args
.
split
,
splits_string
=
args
.
split
,
train_valid_test_num_samples
=
train_val_test_num_samples
,
train_valid_test_num_samples
=
train_val_test_num_samples
,
...
...
pretrain_gpt2.py
View file @
f86bb671
...
@@ -113,9 +113,15 @@ def get_train_val_test_data(args):
...
@@ -113,9 +113,15 @@ def get_train_val_test_data(args):
# Data loader only on rank 0 of each model parallel group.
# Data loader only on rank 0 of each model parallel group.
if
mpu
.
get_model_parallel_rank
()
==
0
:
if
mpu
.
get_model_parallel_rank
()
==
0
:
if
args
.
data_loader
==
'numpy'
:
if
args
.
data_loader
==
'numpy'
:
assert
len
(
args
.
train_data
)
==
1
args
.
train_data
=
args
.
train_data
[
0
]
assert
len
(
args
.
valid_data
)
==
1
args
.
valid_data
=
args
.
valid_data
[
0
]
assert
len
(
args
.
test_data
)
==
1
args
.
test_data
=
args
.
test_data
[
0
]
(
train_data
,
val_data
,
test_data
),
num_tokens
,
\
(
train_data
,
val_data
,
test_data
),
num_tokens
,
\
eod_token
=
make_gpt2_dataloaders
(
args
)
eod_token
=
make_gpt2_dataloaders
(
args
)
elif
args
.
data_loader
==
'raw'
or
args
.
data_loader
==
'lazy'
elif
args
.
data_loader
==
'raw'
or
args
.
data_loader
==
'lazy'
:
data_config
=
configure_data
()
data_config
=
configure_data
()
data_config
.
set_defaults
(
data_set_type
=
'GPT2'
,
transpose
=
False
)
data_config
.
set_defaults
(
data_set_type
=
'GPT2'
,
transpose
=
False
)
(
train_data
,
val_data
,
test_data
),
tokenizer
=
data_config
.
apply
(
(
train_data
,
val_data
,
test_data
),
tokenizer
=
data_config
.
apply
(
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment