Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
fa2ccbc0
"...git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "6015f91a5a28548a597f8d24341d089fe04994e8"
Commit
fa2ccbc0
authored
Dec 21, 2019
by
Aymeric Augustin
Browse files
Fix E266 flake8 warning (x90).
parent
2ab78325
Changes
30
Show whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
45 additions
and
45 deletions
+45
-45
examples/contrib/run_swag.py
examples/contrib/run_swag.py
+2
-2
examples/distillation/run_squad_w_distillation.py
examples/distillation/run_squad_w_distillation.py
+2
-2
examples/distillation/scripts/extract.py
examples/distillation/scripts/extract.py
+3
-3
examples/distillation/train.py
examples/distillation/train.py
+9
-9
examples/mm-imdb/run_mmimdb.py
examples/mm-imdb/run_mmimdb.py
+2
-2
examples/run_bertology.py
examples/run_bertology.py
+2
-2
examples/run_glue.py
examples/run_glue.py
+2
-2
examples/run_lm_finetuning.py
examples/run_lm_finetuning.py
+2
-2
examples/run_multiple_choice.py
examples/run_multiple_choice.py
+2
-2
examples/run_ner.py
examples/run_ner.py
+2
-2
examples/run_xnli.py
examples/run_xnli.py
+2
-2
templates/adding_a_new_example_script/run_xxx.py
templates/adding_a_new_example_script/run_xxx.py
+2
-2
templates/adding_a_new_model/convert_xxx_original_tf_checkpoint_to_pytorch.py
...ew_model/convert_xxx_original_tf_checkpoint_to_pytorch.py
+1
-1
transformers/convert_albert_original_tf_checkpoint_to_pytorch.py
...rmers/convert_albert_original_tf_checkpoint_to_pytorch.py
+1
-1
transformers/convert_bert_original_tf_checkpoint_to_pytorch.py
...formers/convert_bert_original_tf_checkpoint_to_pytorch.py
+1
-1
transformers/convert_gpt2_original_tf_checkpoint_to_pytorch.py
...formers/convert_gpt2_original_tf_checkpoint_to_pytorch.py
+1
-1
transformers/convert_openai_original_tf_checkpoint_to_pytorch.py
...rmers/convert_openai_original_tf_checkpoint_to_pytorch.py
+1
-1
transformers/convert_pytorch_checkpoint_to_tf2.py
transformers/convert_pytorch_checkpoint_to_tf2.py
+1
-1
transformers/convert_roberta_original_pytorch_checkpoint_to_pytorch.py
...convert_roberta_original_pytorch_checkpoint_to_pytorch.py
+6
-6
transformers/convert_t5_original_tf_checkpoint_to_pytorch.py
transformers/convert_t5_original_tf_checkpoint_to_pytorch.py
+1
-1
No files found.
examples/contrib/run_swag.py
View file @
fa2ccbc0
...
@@ -487,7 +487,7 @@ def evaluate(args, model, tokenizer, prefix=""):
...
@@ -487,7 +487,7 @@ def evaluate(args, model, tokenizer, prefix=""):
def
main
():
def
main
():
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--train_file"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"SWAG csv for training. E.g., train.csv"
"--train_file"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"SWAG csv for training. E.g., train.csv"
)
)
...
@@ -520,7 +520,7 @@ def main():
...
@@ -520,7 +520,7 @@ def main():
help
=
"The output directory where the model checkpoints and predictions will be written."
,
help
=
"The output directory where the model checkpoints and predictions will be written."
,
)
)
#
#
Other parameters
# Other parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--config_name"
,
default
=
""
,
type
=
str
,
help
=
"Pretrained config name or path if not the same as model_name"
"--config_name"
,
default
=
""
,
type
=
str
,
help
=
"Pretrained config name or path if not the same as model_name"
)
)
...
...
examples/distillation/run_squad_w_distillation.py
View file @
fa2ccbc0
...
@@ -430,7 +430,7 @@ def load_and_cache_examples(args, tokenizer, evaluate=False, output_examples=Fal
...
@@ -430,7 +430,7 @@ def load_and_cache_examples(args, tokenizer, evaluate=False, output_examples=Fal
def
main
():
def
main
():
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--train_file"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"SQuAD json for training. E.g., train-v1.1.json"
"--train_file"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"SQuAD json for training. E.g., train-v1.1.json"
)
)
...
@@ -486,7 +486,7 @@ def main():
...
@@ -486,7 +486,7 @@ def main():
"--temperature"
,
default
=
2.0
,
type
=
float
,
help
=
"Distillation temperature. Only for distillation."
"--temperature"
,
default
=
2.0
,
type
=
float
,
help
=
"Distillation temperature. Only for distillation."
)
)
#
#
Other parameters
# Other parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--config_name"
,
default
=
""
,
type
=
str
,
help
=
"Pretrained config name or path if not the same as model_name"
"--config_name"
,
default
=
""
,
type
=
str
,
help
=
"Pretrained config name or path if not the same as model_name"
)
)
...
...
examples/distillation/scripts/extract.py
View file @
fa2ccbc0
...
@@ -43,7 +43,7 @@ if __name__ == "__main__":
...
@@ -43,7 +43,7 @@ if __name__ == "__main__":
state_dict
=
model
.
state_dict
()
state_dict
=
model
.
state_dict
()
compressed_sd
=
{}
compressed_sd
=
{}
#
##
Embeddings #
##
# Embeddings #
if
args
.
model_type
==
"gpt2"
:
if
args
.
model_type
==
"gpt2"
:
for
param_name
in
[
"wte.weight"
,
"wpe.weight"
]:
for
param_name
in
[
"wte.weight"
,
"wpe.weight"
]:
compressed_sd
[
f
"
{
prefix
}
.
{
param_name
}
"
]
=
state_dict
[
f
"
{
prefix
}
.
{
param_name
}
"
]
compressed_sd
[
f
"
{
prefix
}
.
{
param_name
}
"
]
=
state_dict
[
f
"
{
prefix
}
.
{
param_name
}
"
]
...
@@ -55,7 +55,7 @@ if __name__ == "__main__":
...
@@ -55,7 +55,7 @@ if __name__ == "__main__":
param_name
=
f
"
{
prefix
}
.embeddings.LayerNorm.
{
w
}
"
param_name
=
f
"
{
prefix
}
.embeddings.LayerNorm.
{
w
}
"
compressed_sd
[
param_name
]
=
state_dict
[
param_name
]
compressed_sd
[
param_name
]
=
state_dict
[
param_name
]
#
##
Transformer Blocks #
##
# Transformer Blocks #
std_idx
=
0
std_idx
=
0
for
teacher_idx
in
[
0
,
2
,
4
,
7
,
9
,
11
]:
for
teacher_idx
in
[
0
,
2
,
4
,
7
,
9
,
11
]:
if
args
.
model_type
==
"gpt2"
:
if
args
.
model_type
==
"gpt2"
:
...
@@ -82,7 +82,7 @@ if __name__ == "__main__":
...
@@ -82,7 +82,7 @@ if __name__ == "__main__":
]
]
std_idx
+=
1
std_idx
+=
1
#
##
Language Modeling Head ###s
# Language Modeling Head ###s
if
args
.
model_type
==
"roberta"
:
if
args
.
model_type
==
"roberta"
:
for
layer
in
[
"lm_head.decoder.weight"
,
"lm_head.bias"
]:
for
layer
in
[
"lm_head.decoder.weight"
,
"lm_head.bias"
]:
compressed_sd
[
f
"
{
layer
}
"
]
=
state_dict
[
f
"
{
layer
}
"
]
compressed_sd
[
f
"
{
layer
}
"
]
=
state_dict
[
f
"
{
layer
}
"
]
...
...
examples/distillation/train.py
View file @
fa2ccbc0
...
@@ -219,7 +219,7 @@ def main():
...
@@ -219,7 +219,7 @@ def main():
args
=
parser
.
parse_args
()
args
=
parser
.
parse_args
()
sanity_checks
(
args
)
sanity_checks
(
args
)
#
#
ARGS #
#
# ARGS #
init_gpu_params
(
args
)
init_gpu_params
(
args
)
set_seed
(
args
)
set_seed
(
args
)
if
args
.
is_master
:
if
args
.
is_master
:
...
@@ -236,7 +236,7 @@ def main():
...
@@ -236,7 +236,7 @@ def main():
os
.
makedirs
(
args
.
dump_path
)
os
.
makedirs
(
args
.
dump_path
)
logger
.
info
(
f
"Experiment will be dumped and logged in
{
args
.
dump_path
}
"
)
logger
.
info
(
f
"Experiment will be dumped and logged in
{
args
.
dump_path
}
"
)
#
##
SAVE PARAMS #
##
# SAVE PARAMS #
logger
.
info
(
f
"Param:
{
args
}
"
)
logger
.
info
(
f
"Param:
{
args
}
"
)
with
open
(
os
.
path
.
join
(
args
.
dump_path
,
"parameters.json"
),
"w"
)
as
f
:
with
open
(
os
.
path
.
join
(
args
.
dump_path
,
"parameters.json"
),
"w"
)
as
f
:
json
.
dump
(
vars
(
args
),
f
,
indent
=
4
)
json
.
dump
(
vars
(
args
),
f
,
indent
=
4
)
...
@@ -245,7 +245,7 @@ def main():
...
@@ -245,7 +245,7 @@ def main():
student_config_class
,
student_model_class
,
_
=
MODEL_CLASSES
[
args
.
student_type
]
student_config_class
,
student_model_class
,
_
=
MODEL_CLASSES
[
args
.
student_type
]
teacher_config_class
,
teacher_model_class
,
teacher_tokenizer_class
=
MODEL_CLASSES
[
args
.
teacher_type
]
teacher_config_class
,
teacher_model_class
,
teacher_tokenizer_class
=
MODEL_CLASSES
[
args
.
teacher_type
]
#
##
TOKENIZER #
##
# TOKENIZER #
tokenizer
=
teacher_tokenizer_class
.
from_pretrained
(
args
.
teacher_name
)
tokenizer
=
teacher_tokenizer_class
.
from_pretrained
(
args
.
teacher_name
)
special_tok_ids
=
{}
special_tok_ids
=
{}
for
tok_name
,
tok_symbol
in
tokenizer
.
special_tokens_map
.
items
():
for
tok_name
,
tok_symbol
in
tokenizer
.
special_tokens_map
.
items
():
...
@@ -255,7 +255,7 @@ def main():
...
@@ -255,7 +255,7 @@ def main():
args
.
special_tok_ids
=
special_tok_ids
args
.
special_tok_ids
=
special_tok_ids
args
.
max_model_input_size
=
tokenizer
.
max_model_input_sizes
[
args
.
teacher_name
]
args
.
max_model_input_size
=
tokenizer
.
max_model_input_sizes
[
args
.
teacher_name
]
#
#
DATA LOADER #
#
# DATA LOADER #
logger
.
info
(
f
"Loading data from
{
args
.
data_file
}
"
)
logger
.
info
(
f
"Loading data from
{
args
.
data_file
}
"
)
with
open
(
args
.
data_file
,
"rb"
)
as
fp
:
with
open
(
args
.
data_file
,
"rb"
)
as
fp
:
data
=
pickle
.
load
(
fp
)
data
=
pickle
.
load
(
fp
)
...
@@ -275,7 +275,7 @@ def main():
...
@@ -275,7 +275,7 @@ def main():
train_lm_seq_dataset
=
LmSeqsDataset
(
params
=
args
,
data
=
data
)
train_lm_seq_dataset
=
LmSeqsDataset
(
params
=
args
,
data
=
data
)
logger
.
info
(
f
"Data loader created."
)
logger
.
info
(
f
"Data loader created."
)
#
#
STUDENT #
#
# STUDENT #
logger
.
info
(
f
"Loading student config from
{
args
.
student_config
}
"
)
logger
.
info
(
f
"Loading student config from
{
args
.
student_config
}
"
)
stu_architecture_config
=
student_config_class
.
from_pretrained
(
args
.
student_config
)
stu_architecture_config
=
student_config_class
.
from_pretrained
(
args
.
student_config
)
stu_architecture_config
.
output_hidden_states
=
True
stu_architecture_config
.
output_hidden_states
=
True
...
@@ -290,26 +290,26 @@ def main():
...
@@ -290,26 +290,26 @@ def main():
student
.
to
(
f
"cuda:
{
args
.
local_rank
}
"
)
student
.
to
(
f
"cuda:
{
args
.
local_rank
}
"
)
logger
.
info
(
f
"Student loaded."
)
logger
.
info
(
f
"Student loaded."
)
#
#
TEACHER #
#
# TEACHER #
teacher
=
teacher_model_class
.
from_pretrained
(
args
.
teacher_name
,
output_hidden_states
=
True
)
teacher
=
teacher_model_class
.
from_pretrained
(
args
.
teacher_name
,
output_hidden_states
=
True
)
if
args
.
n_gpu
>
0
:
if
args
.
n_gpu
>
0
:
teacher
.
to
(
f
"cuda:
{
args
.
local_rank
}
"
)
teacher
.
to
(
f
"cuda:
{
args
.
local_rank
}
"
)
logger
.
info
(
f
"Teacher loaded from
{
args
.
teacher_name
}
."
)
logger
.
info
(
f
"Teacher loaded from
{
args
.
teacher_name
}
."
)
#
#
FREEZING #
#
# FREEZING #
if
args
.
freeze_pos_embs
:
if
args
.
freeze_pos_embs
:
freeze_pos_embeddings
(
student
,
args
)
freeze_pos_embeddings
(
student
,
args
)
if
args
.
freeze_token_type_embds
:
if
args
.
freeze_token_type_embds
:
freeze_token_type_embeddings
(
student
,
args
)
freeze_token_type_embeddings
(
student
,
args
)
#
#
SANITY CHECKS #
#
# SANITY CHECKS #
assert
student
.
config
.
vocab_size
==
teacher
.
config
.
vocab_size
assert
student
.
config
.
vocab_size
==
teacher
.
config
.
vocab_size
assert
student
.
config
.
hidden_size
==
teacher
.
config
.
hidden_size
assert
student
.
config
.
hidden_size
==
teacher
.
config
.
hidden_size
assert
student
.
config
.
max_position_embeddings
==
teacher
.
config
.
max_position_embeddings
assert
student
.
config
.
max_position_embeddings
==
teacher
.
config
.
max_position_embeddings
if
args
.
mlm
:
if
args
.
mlm
:
assert
token_probs
.
size
(
0
)
==
stu_architecture_config
.
vocab_size
assert
token_probs
.
size
(
0
)
==
stu_architecture_config
.
vocab_size
#
#
DISTILLER #
#
# DISTILLER #
torch
.
cuda
.
empty_cache
()
torch
.
cuda
.
empty_cache
()
distiller
=
Distiller
(
distiller
=
Distiller
(
params
=
args
,
dataset
=
train_lm_seq_dataset
,
token_probs
=
token_probs
,
student
=
student
,
teacher
=
teacher
params
=
args
,
dataset
=
train_lm_seq_dataset
,
token_probs
=
token_probs
,
student
=
student
,
teacher
=
teacher
...
...
examples/mm-imdb/run_mmimdb.py
View file @
fa2ccbc0
...
@@ -344,7 +344,7 @@ def load_examples(args, tokenizer, evaluate=False):
...
@@ -344,7 +344,7 @@ def load_examples(args, tokenizer, evaluate=False):
def
main
():
def
main
():
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--data_dir"
,
"--data_dir"
,
default
=
None
,
default
=
None
,
...
@@ -374,7 +374,7 @@ def main():
...
@@ -374,7 +374,7 @@ def main():
help
=
"The output directory where the model predictions and checkpoints will be written."
,
help
=
"The output directory where the model predictions and checkpoints will be written."
,
)
)
#
#
Other parameters
# Other parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--config_name"
,
default
=
""
,
type
=
str
,
help
=
"Pretrained config name or path if not the same as model_name"
"--config_name"
,
default
=
""
,
type
=
str
,
help
=
"Pretrained config name or path if not the same as model_name"
)
)
...
...
examples/run_bertology.py
View file @
fa2ccbc0
...
@@ -242,7 +242,7 @@ def prune_heads(args, model, eval_dataloader, head_mask):
...
@@ -242,7 +242,7 @@ def prune_heads(args, model, eval_dataloader, head_mask):
def
main
():
def
main
():
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--data_dir"
,
"--data_dir"
,
default
=
None
,
default
=
None
,
...
@@ -272,7 +272,7 @@ def main():
...
@@ -272,7 +272,7 @@ def main():
help
=
"The output directory where the model predictions and checkpoints will be written."
,
help
=
"The output directory where the model predictions and checkpoints will be written."
,
)
)
#
#
Other parameters
# Other parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--config_name"
,
"--config_name"
,
default
=
""
,
default
=
""
,
...
...
examples/run_glue.py
View file @
fa2ccbc0
...
@@ -410,7 +410,7 @@ def load_and_cache_examples(args, task, tokenizer, evaluate=False):
...
@@ -410,7 +410,7 @@ def load_and_cache_examples(args, task, tokenizer, evaluate=False):
def
main
():
def
main
():
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--data_dir"
,
"--data_dir"
,
default
=
None
,
default
=
None
,
...
@@ -447,7 +447,7 @@ def main():
...
@@ -447,7 +447,7 @@ def main():
help
=
"The output directory where the model predictions and checkpoints will be written."
,
help
=
"The output directory where the model predictions and checkpoints will be written."
,
)
)
#
#
Other parameters
# Other parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--config_name"
,
default
=
""
,
type
=
str
,
help
=
"Pretrained config name or path if not the same as model_name"
"--config_name"
,
default
=
""
,
type
=
str
,
help
=
"Pretrained config name or path if not the same as model_name"
)
)
...
...
examples/run_lm_finetuning.py
View file @
fa2ccbc0
...
@@ -422,7 +422,7 @@ def evaluate(args, model, tokenizer, prefix=""):
...
@@ -422,7 +422,7 @@ def evaluate(args, model, tokenizer, prefix=""):
def
main
():
def
main
():
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--train_data_file"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"The input training data file (a text file)."
"--train_data_file"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"The input training data file (a text file)."
)
)
...
@@ -434,7 +434,7 @@ def main():
...
@@ -434,7 +434,7 @@ def main():
help
=
"The output directory where the model predictions and checkpoints will be written."
,
help
=
"The output directory where the model predictions and checkpoints will be written."
,
)
)
#
#
Other parameters
# Other parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--eval_data_file"
,
"--eval_data_file"
,
default
=
None
,
default
=
None
,
...
...
examples/run_multiple_choice.py
View file @
fa2ccbc0
...
@@ -385,7 +385,7 @@ def load_and_cache_examples(args, task, tokenizer, evaluate=False, test=False):
...
@@ -385,7 +385,7 @@ def load_and_cache_examples(args, task, tokenizer, evaluate=False, test=False):
def
main
():
def
main
():
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--data_dir"
,
"--data_dir"
,
default
=
None
,
default
=
None
,
...
@@ -422,7 +422,7 @@ def main():
...
@@ -422,7 +422,7 @@ def main():
help
=
"The output directory where the model predictions and checkpoints will be written."
,
help
=
"The output directory where the model predictions and checkpoints will be written."
,
)
)
#
#
Other parameters
# Other parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--config_name"
,
default
=
""
,
type
=
str
,
help
=
"Pretrained config name or path if not the same as model_name"
"--config_name"
,
default
=
""
,
type
=
str
,
help
=
"Pretrained config name or path if not the same as model_name"
)
)
...
...
examples/run_ner.py
View file @
fa2ccbc0
...
@@ -385,7 +385,7 @@ def load_and_cache_examples(args, tokenizer, labels, pad_token_label_id, mode):
...
@@ -385,7 +385,7 @@ def load_and_cache_examples(args, tokenizer, labels, pad_token_label_id, mode):
def
main
():
def
main
():
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--data_dir"
,
"--data_dir"
,
default
=
None
,
default
=
None
,
...
@@ -415,7 +415,7 @@ def main():
...
@@ -415,7 +415,7 @@ def main():
help
=
"The output directory where the model predictions and checkpoints will be written."
,
help
=
"The output directory where the model predictions and checkpoints will be written."
,
)
)
#
#
Other parameters
# Other parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--labels"
,
"--labels"
,
default
=
""
,
default
=
""
,
...
...
examples/run_xnli.py
View file @
fa2ccbc0
...
@@ -377,7 +377,7 @@ def load_and_cache_examples(args, task, tokenizer, evaluate=False):
...
@@ -377,7 +377,7 @@ def load_and_cache_examples(args, task, tokenizer, evaluate=False):
def
main
():
def
main
():
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--data_dir"
,
"--data_dir"
,
default
=
None
,
default
=
None
,
...
@@ -417,7 +417,7 @@ def main():
...
@@ -417,7 +417,7 @@ def main():
help
=
"The output directory where the model predictions and checkpoints will be written."
,
help
=
"The output directory where the model predictions and checkpoints will be written."
,
)
)
#
#
Other parameters
# Other parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--config_name"
,
default
=
""
,
type
=
str
,
help
=
"Pretrained config name or path if not the same as model_name"
"--config_name"
,
default
=
""
,
type
=
str
,
help
=
"Pretrained config name or path if not the same as model_name"
)
)
...
...
templates/adding_a_new_example_script/run_xxx.py
View file @
fa2ccbc0
...
@@ -401,7 +401,7 @@ def load_and_cache_examples(args, tokenizer, evaluate=False, output_examples=Fal
...
@@ -401,7 +401,7 @@ def load_and_cache_examples(args, tokenizer, evaluate=False, output_examples=Fal
def
main
():
def
main
():
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--train_file"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"SQuAD json for training. E.g., train-v1.1.json"
"--train_file"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"SQuAD json for training. E.g., train-v1.1.json"
)
)
...
@@ -434,7 +434,7 @@ def main():
...
@@ -434,7 +434,7 @@ def main():
help
=
"The output directory where the model checkpoints and predictions will be written."
,
help
=
"The output directory where the model checkpoints and predictions will be written."
,
)
)
#
#
Other parameters
# Other parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--config_name"
,
default
=
""
,
type
=
str
,
help
=
"Pretrained config name or path if not the same as model_name"
"--config_name"
,
default
=
""
,
type
=
str
,
help
=
"Pretrained config name or path if not the same as model_name"
)
)
...
...
templates/adding_a_new_model/convert_xxx_original_tf_checkpoint_to_pytorch.py
View file @
fa2ccbc0
...
@@ -43,7 +43,7 @@ def convert_tf_checkpoint_to_pytorch(tf_checkpoint_path, config_file, pytorch_du
...
@@ -43,7 +43,7 @@ def convert_tf_checkpoint_to_pytorch(tf_checkpoint_path, config_file, pytorch_du
if
__name__
==
"__main__"
:
if
__name__
==
"__main__"
:
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--tf_checkpoint_path"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"Path to the TensorFlow checkpoint path."
"--tf_checkpoint_path"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"Path to the TensorFlow checkpoint path."
)
)
...
...
transformers/convert_albert_original_tf_checkpoint_to_pytorch.py
View file @
fa2ccbc0
...
@@ -43,7 +43,7 @@ def convert_tf_checkpoint_to_pytorch(tf_checkpoint_path, albert_config_file, pyt
...
@@ -43,7 +43,7 @@ def convert_tf_checkpoint_to_pytorch(tf_checkpoint_path, albert_config_file, pyt
if
__name__
==
"__main__"
:
if
__name__
==
"__main__"
:
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--tf_checkpoint_path"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"Path to the TensorFlow checkpoint path."
"--tf_checkpoint_path"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"Path to the TensorFlow checkpoint path."
)
)
...
...
transformers/convert_bert_original_tf_checkpoint_to_pytorch.py
View file @
fa2ccbc0
...
@@ -43,7 +43,7 @@ def convert_tf_checkpoint_to_pytorch(tf_checkpoint_path, bert_config_file, pytor
...
@@ -43,7 +43,7 @@ def convert_tf_checkpoint_to_pytorch(tf_checkpoint_path, bert_config_file, pytor
if
__name__
==
"__main__"
:
if
__name__
==
"__main__"
:
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--tf_checkpoint_path"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"Path to the TensorFlow checkpoint path."
"--tf_checkpoint_path"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"Path to the TensorFlow checkpoint path."
)
)
...
...
transformers/convert_gpt2_original_tf_checkpoint_to_pytorch.py
View file @
fa2ccbc0
...
@@ -51,7 +51,7 @@ def convert_gpt2_checkpoint_to_pytorch(gpt2_checkpoint_path, gpt2_config_file, p
...
@@ -51,7 +51,7 @@ def convert_gpt2_checkpoint_to_pytorch(gpt2_checkpoint_path, gpt2_config_file, p
if
__name__
==
"__main__"
:
if
__name__
==
"__main__"
:
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--gpt2_checkpoint_path"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"Path to the TensorFlow checkpoint path."
"--gpt2_checkpoint_path"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"Path to the TensorFlow checkpoint path."
)
)
...
...
transformers/convert_openai_original_tf_checkpoint_to_pytorch.py
View file @
fa2ccbc0
...
@@ -51,7 +51,7 @@ def convert_openai_checkpoint_to_pytorch(openai_checkpoint_folder_path, openai_c
...
@@ -51,7 +51,7 @@ def convert_openai_checkpoint_to_pytorch(openai_checkpoint_folder_path, openai_c
if
__name__
==
"__main__"
:
if
__name__
==
"__main__"
:
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--openai_checkpoint_folder_path"
,
"--openai_checkpoint_folder_path"
,
default
=
None
,
default
=
None
,
...
...
transformers/convert_pytorch_checkpoint_to_tf2.py
View file @
fa2ccbc0
...
@@ -410,7 +410,7 @@ def convert_all_pt_checkpoints_to_tf(
...
@@ -410,7 +410,7 @@ def convert_all_pt_checkpoints_to_tf(
if
__name__
==
"__main__"
:
if
__name__
==
"__main__"
:
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--tf_dump_path"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"Path to the output Tensorflow dump file."
"--tf_dump_path"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"Path to the output Tensorflow dump file."
)
)
...
...
transformers/convert_roberta_original_pytorch_checkpoint_to_pytorch.py
View file @
fa2ccbc0
...
@@ -94,7 +94,7 @@ def convert_roberta_checkpoint_to_pytorch(roberta_checkpoint_path, pytorch_dump_
...
@@ -94,7 +94,7 @@ def convert_roberta_checkpoint_to_pytorch(roberta_checkpoint_path, pytorch_dump_
layer
:
BertLayer
=
model
.
roberta
.
encoder
.
layer
[
i
]
layer
:
BertLayer
=
model
.
roberta
.
encoder
.
layer
[
i
]
roberta_layer
:
TransformerSentenceEncoderLayer
=
roberta_sent_encoder
.
layers
[
i
]
roberta_layer
:
TransformerSentenceEncoderLayer
=
roberta_sent_encoder
.
layers
[
i
]
#
##
self attention
# self attention
self_attn
:
BertSelfAttention
=
layer
.
attention
.
self
self_attn
:
BertSelfAttention
=
layer
.
attention
.
self
assert
(
assert
(
roberta_layer
.
self_attn
.
k_proj
.
weight
.
data
.
shape
roberta_layer
.
self_attn
.
k_proj
.
weight
.
data
.
shape
...
@@ -110,7 +110,7 @@ def convert_roberta_checkpoint_to_pytorch(roberta_checkpoint_path, pytorch_dump_
...
@@ -110,7 +110,7 @@ def convert_roberta_checkpoint_to_pytorch(roberta_checkpoint_path, pytorch_dump_
self_attn
.
value
.
weight
.
data
=
roberta_layer
.
self_attn
.
v_proj
.
weight
self_attn
.
value
.
weight
.
data
=
roberta_layer
.
self_attn
.
v_proj
.
weight
self_attn
.
value
.
bias
.
data
=
roberta_layer
.
self_attn
.
v_proj
.
bias
self_attn
.
value
.
bias
.
data
=
roberta_layer
.
self_attn
.
v_proj
.
bias
#
##
self-attention output
# self-attention output
self_output
:
BertSelfOutput
=
layer
.
attention
.
output
self_output
:
BertSelfOutput
=
layer
.
attention
.
output
assert
self_output
.
dense
.
weight
.
shape
==
roberta_layer
.
self_attn
.
out_proj
.
weight
.
shape
assert
self_output
.
dense
.
weight
.
shape
==
roberta_layer
.
self_attn
.
out_proj
.
weight
.
shape
self_output
.
dense
.
weight
=
roberta_layer
.
self_attn
.
out_proj
.
weight
self_output
.
dense
.
weight
=
roberta_layer
.
self_attn
.
out_proj
.
weight
...
@@ -118,20 +118,20 @@ def convert_roberta_checkpoint_to_pytorch(roberta_checkpoint_path, pytorch_dump_
...
@@ -118,20 +118,20 @@ def convert_roberta_checkpoint_to_pytorch(roberta_checkpoint_path, pytorch_dump_
self_output
.
LayerNorm
.
weight
=
roberta_layer
.
self_attn_layer_norm
.
weight
self_output
.
LayerNorm
.
weight
=
roberta_layer
.
self_attn_layer_norm
.
weight
self_output
.
LayerNorm
.
bias
=
roberta_layer
.
self_attn_layer_norm
.
bias
self_output
.
LayerNorm
.
bias
=
roberta_layer
.
self_attn_layer_norm
.
bias
#
##
intermediate
# intermediate
intermediate
:
BertIntermediate
=
layer
.
intermediate
intermediate
:
BertIntermediate
=
layer
.
intermediate
assert
intermediate
.
dense
.
weight
.
shape
==
roberta_layer
.
fc1
.
weight
.
shape
assert
intermediate
.
dense
.
weight
.
shape
==
roberta_layer
.
fc1
.
weight
.
shape
intermediate
.
dense
.
weight
=
roberta_layer
.
fc1
.
weight
intermediate
.
dense
.
weight
=
roberta_layer
.
fc1
.
weight
intermediate
.
dense
.
bias
=
roberta_layer
.
fc1
.
bias
intermediate
.
dense
.
bias
=
roberta_layer
.
fc1
.
bias
#
##
output
# output
bert_output
:
BertOutput
=
layer
.
output
bert_output
:
BertOutput
=
layer
.
output
assert
bert_output
.
dense
.
weight
.
shape
==
roberta_layer
.
fc2
.
weight
.
shape
assert
bert_output
.
dense
.
weight
.
shape
==
roberta_layer
.
fc2
.
weight
.
shape
bert_output
.
dense
.
weight
=
roberta_layer
.
fc2
.
weight
bert_output
.
dense
.
weight
=
roberta_layer
.
fc2
.
weight
bert_output
.
dense
.
bias
=
roberta_layer
.
fc2
.
bias
bert_output
.
dense
.
bias
=
roberta_layer
.
fc2
.
bias
bert_output
.
LayerNorm
.
weight
=
roberta_layer
.
final_layer_norm
.
weight
bert_output
.
LayerNorm
.
weight
=
roberta_layer
.
final_layer_norm
.
weight
bert_output
.
LayerNorm
.
bias
=
roberta_layer
.
final_layer_norm
.
bias
bert_output
.
LayerNorm
.
bias
=
roberta_layer
.
final_layer_norm
.
bias
#
###
end of layer
# end of layer
if
classification_head
:
if
classification_head
:
model
.
classifier
.
dense
.
weight
=
roberta
.
model
.
classification_heads
[
"mnli"
].
dense
.
weight
model
.
classifier
.
dense
.
weight
=
roberta
.
model
.
classification_heads
[
"mnli"
].
dense
.
weight
...
@@ -170,7 +170,7 @@ def convert_roberta_checkpoint_to_pytorch(roberta_checkpoint_path, pytorch_dump_
...
@@ -170,7 +170,7 @@ def convert_roberta_checkpoint_to_pytorch(roberta_checkpoint_path, pytorch_dump_
if
__name__
==
"__main__"
:
if
__name__
==
"__main__"
:
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--roberta_checkpoint_path"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"Path the official PyTorch dump."
"--roberta_checkpoint_path"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"Path the official PyTorch dump."
)
)
...
...
transformers/convert_t5_original_tf_checkpoint_to_pytorch.py
View file @
fa2ccbc0
...
@@ -43,7 +43,7 @@ def convert_tf_checkpoint_to_pytorch(tf_checkpoint_path, config_file, pytorch_du
...
@@ -43,7 +43,7 @@ def convert_tf_checkpoint_to_pytorch(tf_checkpoint_path, config_file, pytorch_du
if
__name__
==
"__main__"
:
if
__name__
==
"__main__"
:
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
#
#
Required parameters
# Required parameters
parser
.
add_argument
(
parser
.
add_argument
(
"--tf_checkpoint_path"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"Path to the TensorFlow checkpoint path."
"--tf_checkpoint_path"
,
default
=
None
,
type
=
str
,
required
=
True
,
help
=
"Path to the TensorFlow checkpoint path."
)
)
...
...
Prev
1
2
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment