Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
053efc5d
"...git@developer.sourcefind.cn:chenpangpang/open-webui.git" did not exist on "8110a872d5b8fb88cd28c694486781f2dcb9a130"
Unverified
Commit
053efc5d
authored
Jan 19, 2021
by
Sylvain Gugger
Committed by
GitHub
Jan 19, 2021
Browse files
Fix imports in conversion scripts (#9674)
parent
2390c16f
Changes
28
Show whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
47 additions
and
41 deletions
+47
-41
src/transformers/models/albert/convert_albert_original_tf_checkpoint_to_pytorch.py
...lbert/convert_albert_original_tf_checkpoint_to_pytorch.py
+2
-2
src/transformers/models/bart/convert_bart_original_pytorch_checkpoint_to_pytorch.py
...rt/convert_bart_original_pytorch_checkpoint_to_pytorch.py
+9
-3
src/transformers/models/bert/convert_bert_original_tf2_checkpoint_to_pytorch.py
...s/bert/convert_bert_original_tf2_checkpoint_to_pytorch.py
+2
-2
src/transformers/models/bert/convert_bert_original_tf_checkpoint_to_pytorch.py
...ls/bert/convert_bert_original_tf_checkpoint_to_pytorch.py
+2
-2
src/transformers/models/bert/convert_bert_pytorch_checkpoint_to_original_tf.py
...ls/bert/convert_bert_pytorch_checkpoint_to_original_tf.py
+1
-1
src/transformers/models/blenderbot/convert_blenderbot_original_pytorch_checkpoint_to_pytorch.py
...vert_blenderbot_original_pytorch_checkpoint_to_pytorch.py
+2
-2
src/transformers/models/dialogpt/convert_dialogpt_original_pytorch_checkpoint_to_pytorch.py
...onvert_dialogpt_original_pytorch_checkpoint_to_pytorch.py
+1
-1
src/transformers/models/dpr/convert_dpr_original_checkpoint_to_pytorch.py
.../models/dpr/convert_dpr_original_checkpoint_to_pytorch.py
+1
-2
src/transformers/models/electra/convert_electra_original_tf_checkpoint_to_pytorch.py
...ctra/convert_electra_original_tf_checkpoint_to_pytorch.py
+2
-2
src/transformers/models/fsmt/convert_fsmt_original_pytorch_checkpoint_to_pytorch.py
...mt/convert_fsmt_original_pytorch_checkpoint_to_pytorch.py
+5
-4
src/transformers/models/funnel/__init__.py
src/transformers/models/funnel/__init__.py
+1
-0
src/transformers/models/funnel/convert_funnel_original_tf_checkpoint_to_pytorch.py
...unnel/convert_funnel_original_tf_checkpoint_to_pytorch.py
+3
-3
src/transformers/models/gpt2/convert_gpt2_original_tf_checkpoint_to_pytorch.py
...ls/gpt2/convert_gpt2_original_tf_checkpoint_to_pytorch.py
+3
-3
src/transformers/models/longformer/convert_longformer_original_pytorch_lightning_to_pytorch.py
...nvert_longformer_original_pytorch_lightning_to_pytorch.py
+1
-1
src/transformers/models/lxmert/convert_lxmert_original_tf_checkpoint_to_pytorch.py
...xmert/convert_lxmert_original_tf_checkpoint_to_pytorch.py
+3
-3
src/transformers/models/marian/convert_marian_tatoeba_to_pytorch.py
...ormers/models/marian/convert_marian_tatoeba_to_pytorch.py
+1
-1
src/transformers/models/marian/convert_marian_to_pytorch.py
src/transformers/models/marian/convert_marian_to_pytorch.py
+2
-2
src/transformers/models/mbart/convert_mbart_original_checkpoint_to_pytorch.py
...els/mbart/convert_mbart_original_checkpoint_to_pytorch.py
+2
-3
src/transformers/models/mobilebert/convert_mobilebert_original_tf_checkpoint_to_pytorch.py
...t/convert_mobilebert_original_tf_checkpoint_to_pytorch.py
+2
-2
src/transformers/models/pegasus/convert_pegasus_tf_to_pytorch.py
...nsformers/models/pegasus/convert_pegasus_tf_to_pytorch.py
+2
-2
No files found.
src/transformers/models/albert/convert_albert_original_tf_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -19,8 +19,8 @@ import argparse
...
@@ -19,8 +19,8 @@ import argparse
import
torch
import
torch
from
...utils
import
logging
from
transformers
import
AlbertConfig
,
AlbertForPreTraining
,
load_tf_weights_in_albert
from
.
import
AlbertConfig
,
AlbertForPreTraining
,
load_tf_weights_in_albert
from
transformers.utils
import
logging
logging
.
set_verbosity_info
()
logging
.
set_verbosity_info
()
...
...
src/transformers/models/bart/convert_bart_original_pytorch_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -23,9 +23,15 @@ import fairseq
...
@@ -23,9 +23,15 @@ import fairseq
import
torch
import
torch
from
packaging
import
version
from
packaging
import
version
from
...utils
import
logging
from
transformers
import
(
from
.
import
BartConfig
,
BartForConditionalGeneration
,
BartForSequenceClassification
,
BartModel
,
BartTokenizer
BartConfig
,
from
.modeling_bart
import
_make_linear_from_emb
BartForConditionalGeneration
,
BartForSequenceClassification
,
BartModel
,
BartTokenizer
,
)
from
transformers.models.bart.modeling_bart
import
_make_linear_from_emb
from
transformers.utils
import
logging
FAIRSEQ_MODELS
=
[
"bart.large"
,
"bart.large.mnli"
,
"bart.large.cnn"
,
"bart_xsum/model.pt"
]
FAIRSEQ_MODELS
=
[
"bart.large"
,
"bart.large.mnli"
,
"bart.large.cnn"
,
"bart_xsum/model.pt"
]
...
...
src/transformers/models/bert/convert_bert_original_tf2_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -28,8 +28,8 @@ import re
...
@@ -28,8 +28,8 @@ import re
import
tensorflow
as
tf
import
tensorflow
as
tf
import
torch
import
torch
from
...utils
import
logging
from
transformers
import
BertConfig
,
BertModel
from
.
import
BertConfig
,
BertModel
from
transformers.utils
import
logging
logging
.
set_verbosity_info
()
logging
.
set_verbosity_info
()
...
...
src/transformers/models/bert/convert_bert_original_tf_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -19,8 +19,8 @@ import argparse
...
@@ -19,8 +19,8 @@ import argparse
import
torch
import
torch
from
...utils
import
logging
from
transformers
import
BertConfig
,
BertForPreTraining
,
load_tf_weights_in_bert
from
.
import
BertConfig
,
BertForPreTraining
,
load_tf_weights_in_bert
from
transformers.utils
import
logging
logging
.
set_verbosity_info
()
logging
.
set_verbosity_info
()
...
...
src/transformers/models/bert/convert_bert_pytorch_checkpoint_to_original_tf.py
View file @
053efc5d
...
@@ -22,7 +22,7 @@ import numpy as np
...
@@ -22,7 +22,7 @@ import numpy as np
import
tensorflow
as
tf
import
tensorflow
as
tf
import
torch
import
torch
from
.
import
BertModel
from
transformers
import
BertModel
def
convert_pytorch_checkpoint_to_tf
(
model
:
BertModel
,
ckpt_dir
:
str
,
model_name
:
str
):
def
convert_pytorch_checkpoint_to_tf
(
model
:
BertModel
,
ckpt_dir
:
str
,
model_name
:
str
):
...
...
src/transformers/models/blenderbot/convert_blenderbot_original_pytorch_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -18,8 +18,8 @@ import argparse
...
@@ -18,8 +18,8 @@ import argparse
import
torch
import
torch
from
...models.bart
import
BartConfig
,
BartForConditionalGeneration
from
transformers
import
BartConfig
,
BartForConditionalGeneration
from
..
.utils
import
logging
from
transformers
.utils
import
logging
logging
.
set_verbosity_info
()
logging
.
set_verbosity_info
()
...
...
src/transformers/models/dialogpt/convert_dialogpt_original_pytorch_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -17,7 +17,7 @@ import os
...
@@ -17,7 +17,7 @@ import os
import
torch
import
torch
from
..
.file_utils
import
WEIGHTS_NAME
from
transformers
.file_utils
import
WEIGHTS_NAME
DIALOGPT_MODELS
=
[
"small"
,
"medium"
,
"large"
]
DIALOGPT_MODELS
=
[
"small"
,
"medium"
,
"large"
]
...
...
src/transformers/models/dpr/convert_dpr_original_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -19,8 +19,7 @@ from pathlib import Path
...
@@ -19,8 +19,7 @@ from pathlib import Path
import
torch
import
torch
from
torch.serialization
import
default_restore_location
from
torch.serialization
import
default_restore_location
from
...models.bert
import
BertConfig
from
.transformers
import
BertConfig
,
DPRConfig
,
DPRContextEncoder
,
DPRQuestionEncoder
,
DPRReader
from
.
import
DPRConfig
,
DPRContextEncoder
,
DPRQuestionEncoder
,
DPRReader
CheckpointState
=
collections
.
namedtuple
(
CheckpointState
=
collections
.
namedtuple
(
...
...
src/transformers/models/electra/convert_electra_original_tf_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -19,8 +19,8 @@ import argparse
...
@@ -19,8 +19,8 @@ import argparse
import
torch
import
torch
from
...utils
import
logging
from
transformers
import
ElectraConfig
,
ElectraForMaskedLM
,
ElectraForPreTraining
,
load_tf_weights_in_electra
from
.
import
ElectraConfig
,
ElectraForMaskedLM
,
ElectraForPreTraining
,
load_tf_weights_in_electra
from
transformers.utils
import
logging
logging
.
set_verbosity_info
()
logging
.
set_verbosity_info
()
...
...
src/transformers/models/fsmt/convert_fsmt_original_pytorch_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -31,10 +31,11 @@ import torch
...
@@ -31,10 +31,11 @@ import torch
from
fairseq
import
hub_utils
from
fairseq
import
hub_utils
from
fairseq.data.dictionary
import
Dictionary
from
fairseq.data.dictionary
import
Dictionary
from
...file_utils
import
WEIGHTS_NAME
from
transfomers.models.fsmt.tokenization_fsmt
import
VOCAB_FILES_NAMES
from
...tokenization_utils_base
import
TOKENIZER_CONFIG_FILE
from
transformers
import
FSMTConfig
,
FSMTForConditionalGeneration
from
...utils
import
logging
from
transformers.file_utils
import
WEIGHTS_NAME
from
.
import
VOCAB_FILES_NAMES
,
FSMTConfig
,
FSMTForConditionalGeneration
from
transformers.tokenization_utils_base
import
TOKENIZER_CONFIG_FILE
from
transformers.utils
import
logging
logging
.
set_verbosity_warning
()
logging
.
set_verbosity_warning
()
...
...
src/transformers/models/funnel/__init__.py
View file @
053efc5d
...
@@ -23,6 +23,7 @@ from ...file_utils import _BaseLazyModule, is_tf_available, is_tokenizers_availa
...
@@ -23,6 +23,7 @@ from ...file_utils import _BaseLazyModule, is_tf_available, is_tokenizers_availa
_import_structure
=
{
_import_structure
=
{
"configuration_funnel"
:
[
"FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP"
,
"FunnelConfig"
],
"configuration_funnel"
:
[
"FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP"
,
"FunnelConfig"
],
"convert_funnel_original_tf_checkpoint_to_pytorch"
:
[],
"tokenization_funnel"
:
[
"FunnelTokenizer"
],
"tokenization_funnel"
:
[
"FunnelTokenizer"
],
}
}
...
...
src/transformers/models/funnel/convert_funnel_original_tf_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -16,14 +16,14 @@
...
@@ -16,14 +16,14 @@
import
argparse
import
argparse
import
logging
import
torch
import
torch
from
.
import
FunnelConfig
,
FunnelForPreTraining
,
load_tf_weights_in_funnel
from
transformers
import
FunnelConfig
,
FunnelForPreTraining
,
load_tf_weights_in_funnel
from
transformers.utils
import
logging
logging
.
basicConfig
(
level
=
logging
.
INFO
)
logging
.
set_verbosity_info
(
)
def
convert_tf_checkpoint_to_pytorch
(
tf_checkpoint_path
,
config_file
,
pytorch_dump_path
):
def
convert_tf_checkpoint_to_pytorch
(
tf_checkpoint_path
,
config_file
,
pytorch_dump_path
):
...
...
src/transformers/models/gpt2/convert_gpt2_original_tf_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -19,9 +19,9 @@ import argparse
...
@@ -19,9 +19,9 @@ import argparse
import
torch
import
torch
from
...file_utils
import
CONFIG_NAME
,
WEIGHTS_NAME
from
transformers
import
GPT2Config
,
GPT2Model
,
load_tf_weights_in_gpt2
from
...utils
import
logging
from
transformers.file_utils
import
CONFIG_NAME
,
WEIGHTS_NAME
from
.
import
GPT2Config
,
GPT2Model
,
load_tf_weights_in_gpt2
from
transformers.utils
import
logging
logging
.
set_verbosity_info
()
logging
.
set_verbosity_info
()
...
...
src/transformers/models/longformer/convert_longformer_original_pytorch_lightning_to_pytorch.py
View file @
053efc5d
...
@@ -20,7 +20,7 @@ import argparse
...
@@ -20,7 +20,7 @@ import argparse
import
pytorch_lightning
as
pl
import
pytorch_lightning
as
pl
import
torch
import
torch
from
.
import
LongformerForQuestionAnswering
,
LongformerModel
from
transformers
import
LongformerForQuestionAnswering
,
LongformerModel
class
LightningModel
(
pl
.
LightningModule
):
class
LightningModel
(
pl
.
LightningModule
):
...
...
src/transformers/models/lxmert/convert_lxmert_original_tf_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -16,14 +16,14 @@
...
@@ -16,14 +16,14 @@
import
argparse
import
argparse
import
logging
import
torch
import
torch
from
.
import
LxmertConfig
,
LxmertForPreTraining
,
load_tf_weights_in_lxmert
from
transformers
import
LxmertConfig
,
LxmertForPreTraining
,
load_tf_weights_in_lxmert
from
transformers.utils
import
logging
logging
.
basicConfig
(
level
=
logging
.
INFO
)
logging
.
set_verbosity_info
(
)
def
convert_tf_checkpoint_to_pytorch
(
tf_checkpoint_path
,
config_file
,
pytorch_dump_path
):
def
convert_tf_checkpoint_to_pytorch
(
tf_checkpoint_path
,
config_file
,
pytorch_dump_path
):
...
...
src/transformers/models/marian/convert_marian_tatoeba_to_pytorch.py
View file @
053efc5d
...
@@ -17,7 +17,7 @@ import os
...
@@ -17,7 +17,7 @@ import os
from
pathlib
import
Path
from
pathlib
import
Path
from
typing
import
List
,
Tuple
from
typing
import
List
,
Tuple
from
.convert_marian_to_pytorch
import
(
from
transformers.models.marian
.convert_marian_to_pytorch
import
(
FRONT_MATTER_TEMPLATE
,
FRONT_MATTER_TEMPLATE
,
_parse_readme
,
_parse_readme
,
convert_all_sentencepiece_models
,
convert_all_sentencepiece_models
,
...
...
src/transformers/models/marian/convert_marian_to_pytorch.py
View file @
053efc5d
...
@@ -26,8 +26,8 @@ import numpy as np
...
@@ -26,8 +26,8 @@ import numpy as np
import
torch
import
torch
from
tqdm
import
tqdm
from
tqdm
import
tqdm
from
...hf_api
import
HfApi
from
transformers
import
MarianConfig
,
MarianMTModel
,
MarianTokenizer
from
.
import
MarianConfig
,
MarianMTModel
,
MarianTokenizer
from
transformers.hf_api
import
HfApi
def
remove_suffix
(
text
:
str
,
suffix
:
str
):
def
remove_suffix
(
text
:
str
,
suffix
:
str
):
...
...
src/transformers/models/mbart/convert_mbart_original_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -16,9 +16,8 @@ import argparse
...
@@ -16,9 +16,8 @@ import argparse
import
torch
import
torch
from
..bart
import
BartForConditionalGeneration
from
transformers
import
BartForConditionalGeneration
,
MBartConfig
from
..bart.convert_bart_original_pytorch_checkpoint_to_pytorch
import
remove_ignore_keys_
from
transformers.models.bart.convert_bart_original_pytorch_checkpoint_to_pytorch
import
remove_ignore_keys_
from
.
import
MBartConfig
def
convert_fairseq_mbart_checkpoint_from_disk
(
checkpoint_path
,
hf_config_path
=
"facebook/mbart-large-en-ro"
):
def
convert_fairseq_mbart_checkpoint_from_disk
(
checkpoint_path
,
hf_config_path
=
"facebook/mbart-large-en-ro"
):
...
...
src/transformers/models/mobilebert/convert_mobilebert_original_tf_checkpoint_to_pytorch.py
View file @
053efc5d
...
@@ -16,8 +16,8 @@ import argparse
...
@@ -16,8 +16,8 @@ import argparse
import
torch
import
torch
from
...utils
import
logging
from
transformers
import
MobileBertConfig
,
MobileBertForPreTraining
,
load_tf_weights_in_mobilebert
from
.
import
MobileBertConfig
,
MobileBertForPreTraining
,
load_tf_weights_in_mobilebert
from
transformers.utils
import
logging
logging
.
set_verbosity_info
()
logging
.
set_verbosity_info
()
...
...
src/transformers/models/pegasus/convert_pegasus_tf_to_pytorch.py
View file @
053efc5d
...
@@ -22,8 +22,8 @@ import tensorflow as tf
...
@@ -22,8 +22,8 @@ import tensorflow as tf
import
torch
import
torch
from
tqdm
import
tqdm
from
tqdm
import
tqdm
from
.
import
PegasusConfig
,
PegasusForConditionalGeneration
,
PegasusTokenizer
from
transformers
import
PegasusConfig
,
PegasusForConditionalGeneration
,
PegasusTokenizer
from
.configuration_pegasus
import
DEFAULTS
,
task_specific_params
from
transformers.models.pegasus
.configuration_pegasus
import
DEFAULTS
,
task_specific_params
PATTERNS
=
[
PATTERNS
=
[
...
...
Prev
1
2
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment