Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
88368c2a
Commit
88368c2a
authored
Sep 06, 2019
by
LysandreJik
Browse files
Added DistilBERT to `run_lm_finetuning`
parent
2d8ec5a6
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
5 additions
and
3 deletions
+5
-3
examples/run_lm_finetuning.py
examples/run_lm_finetuning.py
+5
-3
No files found.
examples/run_lm_finetuning.py
View file @
88368c2a
...
@@ -39,7 +39,8 @@ from pytorch_transformers import (WEIGHTS_NAME, AdamW, WarmupLinearSchedule,
...
@@ -39,7 +39,8 @@ from pytorch_transformers import (WEIGHTS_NAME, AdamW, WarmupLinearSchedule,
BertConfig
,
BertForMaskedLM
,
BertTokenizer
,
BertConfig
,
BertForMaskedLM
,
BertTokenizer
,
GPT2Config
,
GPT2LMHeadModel
,
GPT2Tokenizer
,
GPT2Config
,
GPT2LMHeadModel
,
GPT2Tokenizer
,
OpenAIGPTConfig
,
OpenAIGPTLMHeadModel
,
OpenAIGPTTokenizer
,
OpenAIGPTConfig
,
OpenAIGPTLMHeadModel
,
OpenAIGPTTokenizer
,
RobertaConfig
,
RobertaForMaskedLM
,
RobertaTokenizer
)
RobertaConfig
,
RobertaForMaskedLM
,
RobertaTokenizer
,
DistilBertConfig
,
DistilBertForMaskedLM
,
DistilBertTokenizer
)
logger
=
logging
.
getLogger
(
__name__
)
logger
=
logging
.
getLogger
(
__name__
)
...
@@ -49,7 +50,8 @@ MODEL_CLASSES = {
...
@@ -49,7 +50,8 @@ MODEL_CLASSES = {
'gpt2'
:
(
GPT2Config
,
GPT2LMHeadModel
,
GPT2Tokenizer
),
'gpt2'
:
(
GPT2Config
,
GPT2LMHeadModel
,
GPT2Tokenizer
),
'openai-gpt'
:
(
OpenAIGPTConfig
,
OpenAIGPTLMHeadModel
,
OpenAIGPTTokenizer
),
'openai-gpt'
:
(
OpenAIGPTConfig
,
OpenAIGPTLMHeadModel
,
OpenAIGPTTokenizer
),
'bert'
:
(
BertConfig
,
BertForMaskedLM
,
BertTokenizer
),
'bert'
:
(
BertConfig
,
BertForMaskedLM
,
BertTokenizer
),
'roberta'
:
(
RobertaConfig
,
RobertaForMaskedLM
,
RobertaTokenizer
)
'roberta'
:
(
RobertaConfig
,
RobertaForMaskedLM
,
RobertaTokenizer
),
'distilbert'
:
(
DistilBertConfig
,
DistilBertForMaskedLM
,
DistilBertTokenizer
)
}
}
...
@@ -380,7 +382,7 @@ def main():
...
@@ -380,7 +382,7 @@ def main():
parser
.
add_argument
(
'--server_port'
,
type
=
str
,
default
=
''
,
help
=
"For distant debugging."
)
parser
.
add_argument
(
'--server_port'
,
type
=
str
,
default
=
''
,
help
=
"For distant debugging."
)
args
=
parser
.
parse_args
()
args
=
parser
.
parse_args
()
if
args
.
model_type
in
[
"bert"
,
"roberta"
]
and
not
args
.
mlm
:
if
args
.
model_type
in
[
"bert"
,
"roberta"
,
"distilbert"
]
and
not
args
.
mlm
:
raise
ValueError
(
"BERT and RoBERTa do not have LM heads but masked LM heads. They must be run using the --mlm "
raise
ValueError
(
"BERT and RoBERTa do not have LM heads but masked LM heads. They must be run using the --mlm "
"flag (masked language modeling)."
)
"flag (masked language modeling)."
)
if
args
.
eval_data_file
is
None
and
args
.
do_eval
:
if
args
.
eval_data_file
is
None
and
args
.
do_eval
:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment