Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
Megatron-LM
Commits
b886b7bb
Commit
b886b7bb
authored
Oct 08, 2019
by
Mohammad Shoeybi
Browse files
created megatron package
parent
4947002d
Changes
45
Hide whitespace changes
Inline
Side-by-side
Showing
5 changed files
with
37 additions
and
37 deletions
+37
-37
megatron/mpu/utils.py
megatron/mpu/utils.py
+0
-0
megatron/utils.py
megatron/utils.py
+3
-3
openwebtext/tokenizer.py
openwebtext/tokenizer.py
+1
-1
pretrain_bert.py
pretrain_bert.py
+17
-17
pretrain_gpt2.py
pretrain_gpt2.py
+16
-16
No files found.
mpu/utils.py
→
megatron/
mpu/utils.py
View file @
b886b7bb
File moved
utils.py
→
megatron/
utils.py
View file @
b886b7bb
...
...
@@ -22,9 +22,9 @@ import numpy as np
import
torch
from
torch.nn.parallel.distributed
import
DistributedDataParallel
as
torchDDP
from
fp16
import
FP16_Optimizer
import
mpu
import
model
from
megatron.
fp16
import
FP16_Optimizer
from
megatron
import
mpu
from
megatron
import
model
def
print_rank_0
(
message
):
...
...
openwebtext/tokenizer.py
View file @
b886b7bb
...
...
@@ -16,7 +16,7 @@
import
sys
sys
.
path
.
append
(
'..'
)
from
data_utils.tokenization_gpt2
import
GPT2Tokenizer
from
megatron.
data_utils.tokenization_gpt2
import
GPT2Tokenizer
class
Tokenizer
:
...
...
pretrain_bert.py
View file @
b886b7bb
...
...
@@ -24,24 +24,24 @@ import torch.nn.functional as F
from
arguments
import
get_args
from
configure_data
import
configure_data
from
fp16
import
FP16_Module
from
fp16
import
FP16_Optimizer
from
learning_rates
import
AnnealingLR
from
model
import
BertModel
from
model
import
get_params_for_weight_decay_optimization
from
model
import
gpt2_get_params_for_weight_decay_optimization
from
model
import
DistributedDataParallel
as
LocalDDP
import
mpu
from
megatron.
fp16
import
FP16_Module
from
megatron.
fp16
import
FP16_Optimizer
from
megatron.
learning_rates
import
AnnealingLR
from
megatron.
model
import
BertModel
from
megatron.
model
import
get_params_for_weight_decay_optimization
from
megatron.
model
import
gpt2_get_params_for_weight_decay_optimization
from
megatron.
model
import
DistributedDataParallel
as
LocalDDP
from
megatron
import
mpu
from
apex.optimizers
import
FusedAdam
as
Adam
from
utils
import
Timers
from
utils
import
save_checkpoint
from
utils
import
load_checkpoint
from
utils
import
report_memory
from
utils
import
print_args
from
utils
import
print_params_min_max_norm
from
utils
import
print_rank_0
from
utils
import
enable_adlr_autoresume
from
utils
import
check_adlr_autoresume_termination
from
megatron.
utils
import
Timers
from
megatron.
utils
import
save_checkpoint
from
megatron.
utils
import
load_checkpoint
from
megatron.
utils
import
report_memory
from
megatron.
utils
import
print_args
from
megatron.
utils
import
print_params_min_max_norm
from
megatron.
utils
import
print_rank_0
from
megatron.
utils
import
enable_adlr_autoresume
from
megatron.
utils
import
check_adlr_autoresume_termination
def
get_model
(
args
):
"""Build the model."""
...
...
pretrain_gpt2.py
View file @
b886b7bb
...
...
@@ -24,23 +24,23 @@ import torch
from
arguments
import
get_args
from
configure_data
import
configure_data
from
fp16
import
FP16_Module
from
fp16
import
FP16_Optimizer
from
learning_rates
import
AnnealingLR
from
model
import
GPT2Model
from
model
import
gpt2_get_params_for_weight_decay_optimization
from
model
import
DistributedDataParallel
as
LocalDDP
import
mpu
from
megatron.
fp16
import
FP16_Module
from
megatron.
fp16
import
FP16_Optimizer
from
megatron.
learning_rates
import
AnnealingLR
from
megatron.
model
import
GPT2Model
from
megatron.
model
import
gpt2_get_params_for_weight_decay_optimization
from
megatron.
model
import
DistributedDataParallel
as
LocalDDP
from
megatron
import
mpu
from
apex.optimizers
import
FusedAdam
as
Adam
from
utils
import
Timers
from
utils
import
save_checkpoint
from
utils
import
load_checkpoint
from
utils
import
report_memory
from
utils
import
print_args
from
utils
import
print_params_min_max_norm
from
utils
import
print_rank_0
from
utils
import
enable_adlr_autoresume
from
utils
import
check_adlr_autoresume_termination
from
megatron.
utils
import
Timers
from
megatron.
utils
import
save_checkpoint
from
megatron.
utils
import
load_checkpoint
from
megatron.
utils
import
report_memory
from
megatron.
utils
import
print_args
from
megatron.
utils
import
print_params_min_max_norm
from
megatron.
utils
import
print_rank_0
from
megatron.
utils
import
enable_adlr_autoresume
from
megatron.
utils
import
check_adlr_autoresume_termination
from
gpt2_data_loader
import
make_gpt2_dataloaders
...
...
Prev
1
2
3
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment