Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
053efc5d
Unverified
Commit
053efc5d
authored
Jan 19, 2021
by
Sylvain Gugger
Committed by
GitHub
Jan 19, 2021
Browse files
Fix imports in conversion scripts (#9674)
parent
2390c16f
Changes
28
Show whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
47 additions
and
41 deletions
+47
-41
src/transformers/models/albert/convert_albert_original_tf_checkpoint_to_pytorch.py
...lbert/convert_albert_original_tf_checkpoint_to_pytorch.py
+2
-2
src/transformers/models/bart/convert_bart_original_pytorch_checkpoint_to_pytorch.py
...rt/convert_bart_original_pytorch_checkpoint_to_pytorch.py
+9
-3
src/transformers/models/bert/convert_bert_original_tf2_checkpoint_to_pytorch.py
...s/bert/convert_bert_original_tf2_checkpoint_to_pytorch.py
+2
-2
src/transformers/models/bert/convert_bert_original_tf_checkpoint_to_pytorch.py
...ls/bert/convert_bert_original_tf_checkpoint_to_pytorch.py
+2
-2
src/transformers/models/bert/convert_bert_pytorch_checkpoint_to_original_tf.py
...ls/bert/convert_bert_pytorch_checkpoint_to_original_tf.py
+1
-1
src/transformers/models/blenderbot/convert_blenderbot_original_pytorch_checkpoint_to_pytorch.py
...vert_blenderbot_original_pytorch_checkpoint_to_pytorch.py
+2
-2
src/transformers/models/dialogpt/convert_dialogpt_original_pytorch_checkpoint_to_pytorch.py
...onvert_dialogpt_original_pytorch_checkpoint_to_pytorch.py
+1
-1
src/transformers/models/dpr/convert_dpr_original_checkpoint_to_pytorch.py
.../models/dpr/convert_dpr_original_checkpoint_to_pytorch.py
+1
-2
src/transformers/models/electra/convert_electra_original_tf_checkpoint_to_pytorch.py
...ctra/convert_electra_original_tf_checkpoint_to_pytorch.py
+2
-2
src/transformers/models/fsmt/convert_fsmt_original_pytorch_checkpoint_to_pytorch.py
...mt/convert_fsmt_original_pytorch_checkpoint_to_pytorch.py
+5
-4
src/transformers/models/funnel/__init__.py
src/transformers/models/funnel/__init__.py
+1
-0
src/transformers/models/funnel/convert_funnel_original_tf_checkpoint_to_pytorch.py
...unnel/convert_funnel_original_tf_checkpoint_to_pytorch.py
+3
-3
src/transformers/models/gpt2/convert_gpt2_original_tf_checkpoint_to_pytorch.py
...ls/gpt2/convert_gpt2_original_tf_checkpoint_to_pytorch.py
+3
-3
src/transformers/models/longformer/convert_longformer_original_pytorch_lightning_to_pytorch.py
...nvert_longformer_original_pytorch_lightning_to_pytorch.py
+1
-1
src/transformers/models/lxmert/convert_lxmert_original_tf_checkpoint_to_pytorch.py
...xmert/convert_lxmert_original_tf_checkpoint_to_pytorch.py
+3
-3
src/transformers/models/marian/convert_marian_tatoeba_to_pytorch.py
...ormers/models/marian/convert_marian_tatoeba_to_pytorch.py
+1
-1
src/transformers/models/marian/convert_marian_to_pytorch.py
src/transformers/models/marian/convert_marian_to_pytorch.py
+2
-2
src/transformers/models/mbart/convert_mbart_original_checkpoint_to_pytorch.py
...els/mbart/convert_mbart_original_checkpoint_to_pytorch.py
+2
-3
src/transformers/models/mobilebert/convert_mobilebert_original_tf_checkpoint_to_pytorch.py
...t/convert_mobilebert_original_tf_checkpoint_to_pytorch.py
+2
-2
src/transformers/models/pegasus/convert_pegasus_tf_to_pytorch.py
...nsformers/models/pegasus/convert_pegasus_tf_to_pytorch.py
+2
-2
No files found.
src/transformers/models/albert/convert_albert_original_tf_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -19,8 +19,8 @@ import argparse
...
@@ -19,8 +19,8 @@ import argparse
import
torch
import
torch
from
...utils
import
logging
from
transformers
import
AlbertConfig
,
AlbertForPreTraining
,
load_tf_weights_in_albert
from
.
import
AlbertConfig
,
AlbertForPreTraining
,
load_tf_weights_in_albert
from
transformers.utils
import
logging
logging
.
set_verbosity_info
()
logging
.
set_verbosity_info
()
...
...
src/transformers/models/bart/convert_bart_original_pytorch_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -23,9 +23,15 @@ import fairseq
...
@@ -23,9 +23,15 @@ import fairseq
import
torch
import
torch
from
packaging
import
version
from
packaging
import
version
from
...utils
import
logging
from
transformers
import
(
from
.
import
BartConfig
,
BartForConditionalGeneration
,
BartForSequenceClassification
,
BartModel
,
BartTokenizer
BartConfig
,
from
.modeling_bart
import
_make_linear_from_emb
BartForConditionalGeneration
,
BartForSequenceClassification
,
BartModel
,
BartTokenizer
,
)
from
transformers.models.bart.modeling_bart
import
_make_linear_from_emb
from
transformers.utils
import
logging
FAIRSEQ_MODELS
=
[
"bart.large"
,
"bart.large.mnli"
,
"bart.large.cnn"
,
"bart_xsum/model.pt"
]
FAIRSEQ_MODELS
=
[
"bart.large"
,
"bart.large.mnli"
,
"bart.large.cnn"
,
"bart_xsum/model.pt"
]
...
...
src/transformers/models/bert/convert_bert_original_tf2_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -28,8 +28,8 @@ import re
...
@@ -28,8 +28,8 @@ import re
import
tensorflow
as
tf
import
tensorflow
as
tf
import
torch
import
torch
from
...utils
import
logging
from
transformers
import
BertConfig
,
BertModel
from
.
import
BertConfig
,
BertModel
from
transformers.utils
import
logging
logging
.
set_verbosity_info
()
logging
.
set_verbosity_info
()
...
...
src/transformers/models/bert/convert_bert_original_tf_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -19,8 +19,8 @@ import argparse
...
@@ -19,8 +19,8 @@ import argparse
import
torch
import
torch
from
...utils
import
logging
from
transformers
import
BertConfig
,
BertForPreTraining
,
load_tf_weights_in_bert
from
.
import
BertConfig
,
BertForPreTraining
,
load_tf_weights_in_bert
from
transformers.utils
import
logging
logging
.
set_verbosity_info
()
logging
.
set_verbosity_info
()
...
...
src/transformers/models/bert/convert_bert_pytorch_checkpoint_to_original_tf.py
View file @
053efc5d
...
@@ -22,7 +22,7 @@ import numpy as np
...
@@ -22,7 +22,7 @@ import numpy as np
import
tensorflow
as
tf
import
tensorflow
as
tf
import
torch
import
torch
from
.
import
BertModel
from
transformers
import
BertModel
def
convert_pytorch_checkpoint_to_tf
(
model
:
BertModel
,
ckpt_dir
:
str
,
model_name
:
str
):
def
convert_pytorch_checkpoint_to_tf
(
model
:
BertModel
,
ckpt_dir
:
str
,
model_name
:
str
):
...
...
src/transformers/models/blenderbot/convert_blenderbot_original_pytorch_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -18,8 +18,8 @@ import argparse
...
@@ -18,8 +18,8 @@ import argparse
import
torch
import
torch
from
...models.bart
import
BartConfig
,
BartForConditionalGeneration
from
transformers
import
BartConfig
,
BartForConditionalGeneration
from
..
.utils
import
logging
from
transformers
.utils
import
logging
logging
.
set_verbosity_info
()
logging
.
set_verbosity_info
()
...
...
src/transformers/models/dialogpt/convert_dialogpt_original_pytorch_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -17,7 +17,7 @@ import os
...
@@ -17,7 +17,7 @@ import os
import
torch
import
torch
from
..
.file_utils
import
WEIGHTS_NAME
from
transformers
.file_utils
import
WEIGHTS_NAME
DIALOGPT_MODELS
=
[
"small"
,
"medium"
,
"large"
]
DIALOGPT_MODELS
=
[
"small"
,
"medium"
,
"large"
]
...
...
src/transformers/models/dpr/convert_dpr_original_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -19,8 +19,7 @@ from pathlib import Path
...
@@ -19,8 +19,7 @@ from pathlib import Path
import
torch
import
torch
from
torch.serialization
import
default_restore_location
from
torch.serialization
import
default_restore_location
from
...models.bert
import
BertConfig
from
.transformers
import
BertConfig
,
DPRConfig
,
DPRContextEncoder
,
DPRQuestionEncoder
,
DPRReader
from
.
import
DPRConfig
,
DPRContextEncoder
,
DPRQuestionEncoder
,
DPRReader
CheckpointState
=
collections
.
namedtuple
(
CheckpointState
=
collections
.
namedtuple
(
...
...
src/transformers/models/electra/convert_electra_original_tf_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -19,8 +19,8 @@ import argparse
...
@@ -19,8 +19,8 @@ import argparse
import
torch
import
torch
from
...utils
import
logging
from
transformers
import
ElectraConfig
,
ElectraForMaskedLM
,
ElectraForPreTraining
,
load_tf_weights_in_electra
from
.
import
ElectraConfig
,
ElectraForMaskedLM
,
ElectraForPreTraining
,
load_tf_weights_in_electra
from
transformers.utils
import
logging
logging
.
set_verbosity_info
()
logging
.
set_verbosity_info
()
...
...
src/transformers/models/fsmt/convert_fsmt_original_pytorch_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -31,10 +31,11 @@ import torch
...
@@ -31,10 +31,11 @@ import torch
from
fairseq
import
hub_utils
from
fairseq
import
hub_utils
from
fairseq.data.dictionary
import
Dictionary
from
fairseq.data.dictionary
import
Dictionary
from
...file_utils
import
WEIGHTS_NAME
from
transfomers.models.fsmt.tokenization_fsmt
import
VOCAB_FILES_NAMES
from
...tokenization_utils_base
import
TOKENIZER_CONFIG_FILE
from
transformers
import
FSMTConfig
,
FSMTForConditionalGeneration
from
...utils
import
logging
from
transformers.file_utils
import
WEIGHTS_NAME
from
.
import
VOCAB_FILES_NAMES
,
FSMTConfig
,
FSMTForConditionalGeneration
from
transformers.tokenization_utils_base
import
TOKENIZER_CONFIG_FILE
from
transformers.utils
import
logging
logging
.
set_verbosity_warning
()
logging
.
set_verbosity_warning
()
...
...
src/transformers/models/funnel/__init__.py
View file @
053efc5d
...
@@ -23,6 +23,7 @@ from ...file_utils import _BaseLazyModule, is_tf_available, is_tokenizers_availa
...
@@ -23,6 +23,7 @@ from ...file_utils import _BaseLazyModule, is_tf_available, is_tokenizers_availa
_import_structure
=
{
_import_structure
=
{
"configuration_funnel"
:
[
"FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP"
,
"FunnelConfig"
],
"configuration_funnel"
:
[
"FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP"
,
"FunnelConfig"
],
"convert_funnel_original_tf_checkpoint_to_pytorch"
:
[],
"tokenization_funnel"
:
[
"FunnelTokenizer"
],
"tokenization_funnel"
:
[
"FunnelTokenizer"
],
}
}
...
...
src/transformers/models/funnel/convert_funnel_original_tf_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -16,14 +16,14 @@
...
@@ -16,14 +16,14 @@
import
argparse
import
argparse
import
logging
import
torch
import
torch
from
.
import
FunnelConfig
,
FunnelForPreTraining
,
load_tf_weights_in_funnel
from
transformers
import
FunnelConfig
,
FunnelForPreTraining
,
load_tf_weights_in_funnel
from
transformers.utils
import
logging
logging
.
basicConfig
(
level
=
logging
.
INFO
)
logging
.
set_verbosity_info
(
)
def
convert_tf_checkpoint_to_pytorch
(
tf_checkpoint_path
,
config_file
,
pytorch_dump_path
):
def
convert_tf_checkpoint_to_pytorch
(
tf_checkpoint_path
,
config_file
,
pytorch_dump_path
):
...
...
src/transformers/models/gpt2/convert_gpt2_original_tf_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -19,9 +19,9 @@ import argparse
...
@@ -19,9 +19,9 @@ import argparse
import
torch
import
torch
from
...file_utils
import
CONFIG_NAME
,
WEIGHTS_NAME
from
transformers
import
GPT2Config
,
GPT2Model
,
load_tf_weights_in_gpt2
from
...utils
import
logging
from
transformers.file_utils
import
CONFIG_NAME
,
WEIGHTS_NAME
from
.
import
GPT2Config
,
GPT2Model
,
load_tf_weights_in_gpt2
from
transformers.utils
import
logging
logging
.
set_verbosity_info
()
logging
.
set_verbosity_info
()
...
...
src/transformers/models/longformer/convert_longformer_original_pytorch_lightning_to_pytorch.py
View file @
053efc5d
...
@@ -20,7 +20,7 @@ import argparse
...
@@ -20,7 +20,7 @@ import argparse
import
pytorch_lightning
as
pl
import
pytorch_lightning
as
pl
import
torch
import
torch
from
.
import
LongformerForQuestionAnswering
,
LongformerModel
from
transformers
import
LongformerForQuestionAnswering
,
LongformerModel
class
LightningModel
(
pl
.
LightningModule
):
class
LightningModel
(
pl
.
LightningModule
):
...
...
src/transformers/models/lxmert/convert_lxmert_original_tf_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -16,14 +16,14 @@
...
@@ -16,14 +16,14 @@
import
argparse
import
argparse
import
logging
import
torch
import
torch
from
.
import
LxmertConfig
,
LxmertForPreTraining
,
load_tf_weights_in_lxmert
from
transformers
import
LxmertConfig
,
LxmertForPreTraining
,
load_tf_weights_in_lxmert
from
transformers.utils
import
logging
logging
.
basicConfig
(
level
=
logging
.
INFO
)
logging
.
set_verbosity_info
(
)
def
convert_tf_checkpoint_to_pytorch
(
tf_checkpoint_path
,
config_file
,
pytorch_dump_path
):
def
convert_tf_checkpoint_to_pytorch
(
tf_checkpoint_path
,
config_file
,
pytorch_dump_path
):
...
...
src/transformers/models/marian/convert_marian_tatoeba_to_pytorch.py
View file @
053efc5d
...
@@ -17,7 +17,7 @@ import os
...
@@ -17,7 +17,7 @@ import os
from
pathlib
import
Path
from
pathlib
import
Path
from
typing
import
List
,
Tuple
from
typing
import
List
,
Tuple
from
.convert_marian_to_pytorch
import
(
from
transformers.models.marian
.convert_marian_to_pytorch
import
(
FRONT_MATTER_TEMPLATE
,
FRONT_MATTER_TEMPLATE
,
_parse_readme
,
_parse_readme
,
convert_all_sentencepiece_models
,
convert_all_sentencepiece_models
,
...
...
src/transformers/models/marian/convert_marian_to_pytorch.py
View file @
053efc5d
...
@@ -26,8 +26,8 @@ import numpy as np
...
@@ -26,8 +26,8 @@ import numpy as np
import
torch
import
torch
from
tqdm
import
tqdm
from
tqdm
import
tqdm
from
...hf_api
import
HfApi
from
transformers
import
MarianConfig
,
MarianMTModel
,
MarianTokenizer
from
.
import
MarianConfig
,
MarianMTModel
,
MarianTokenizer
from
transformers.hf_api
import
HfApi
def
remove_suffix
(
text
:
str
,
suffix
:
str
):
def
remove_suffix
(
text
:
str
,
suffix
:
str
):
...
...
src/transformers/models/mbart/convert_mbart_original_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -16,9 +16,8 @@ import argparse
...
@@ -16,9 +16,8 @@ import argparse
import
torch
import
torch
from
..bart
import
BartForConditionalGeneration
from
transformers
import
BartForConditionalGeneration
,
MBartConfig
from
..bart.convert_bart_original_pytorch_checkpoint_to_pytorch
import
remove_ignore_keys_
from
transformers.models.bart.convert_bart_original_pytorch_checkpoint_to_pytorch
import
remove_ignore_keys_
from
.
import
MBartConfig
def
convert_fairseq_mbart_checkpoint_from_disk
(
checkpoint_path
,
hf_config_path
=
"facebook/mbart-large-en-ro"
):
def
convert_fairseq_mbart_checkpoint_from_disk
(
checkpoint_path
,
hf_config_path
=
"facebook/mbart-large-en-ro"
):
...
...
src/transformers/models/mobilebert/convert_mobilebert_original_tf_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -16,8 +16,8 @@ import argparse
...
@@ -16,8 +16,8 @@ import argparse
import
torch
import
torch
from
...utils
import
logging
from
transformers
import
MobileBertConfig
,
MobileBertForPreTraining
,
load_tf_weights_in_mobilebert
from
.
import
MobileBertConfig
,
MobileBertForPreTraining
,
load_tf_weights_in_mobilebert
from
transformers.utils
import
logging
logging
.
set_verbosity_info
()
logging
.
set_verbosity_info
()
...
...
src/transformers/models/pegasus/convert_pegasus_tf_to_pytorch.py
View file @
053efc5d
...
@@ -22,8 +22,8 @@ import tensorflow as tf
...
@@ -22,8 +22,8 @@ import tensorflow as tf
import
torch
import
torch
from
tqdm
import
tqdm
from
tqdm
import
tqdm
from
.
import
PegasusConfig
,
PegasusForConditionalGeneration
,
PegasusTokenizer
from
transformers
import
PegasusConfig
,
PegasusForConditionalGeneration
,
PegasusTokenizer
from
.configuration_pegasus
import
DEFAULTS
,
task_specific_params
from
transformers.models.pegasus
.configuration_pegasus
import
DEFAULTS
,
task_specific_params
PATTERNS
=
[
PATTERNS
=
[
...
...
Prev
1
2
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment