Unverified Commit 77cd0e13 authored by Lysandre Debut's avatar Lysandre Debut Committed by GitHub
Browse files

Conversion scripts shouldn't have relative imports (#6991)

parent 1650130b
...@@ -20,8 +20,7 @@ import argparse ...@@ -20,8 +20,7 @@ import argparse
import torch import torch
from transformers import AlbertConfig, AlbertForPreTraining, load_tf_weights_in_albert from transformers import AlbertConfig, AlbertForPreTraining, load_tf_weights_in_albert
from transformers.utils import logging
from .utils import logging
logging.set_verbosity_info() logging.set_verbosity_info()
......
...@@ -31,8 +31,7 @@ from transformers import ( ...@@ -31,8 +31,7 @@ from transformers import (
BartTokenizer, BartTokenizer,
) )
from transformers.modeling_bart import _make_linear_from_emb from transformers.modeling_bart import _make_linear_from_emb
from transformers.utils import logging
from .utils import logging
FAIRSEQ_MODELS = ["bart.large", "bart.large.mnli", "bart.large.cnn", "bart_xsum/model.pt"] FAIRSEQ_MODELS = ["bart.large", "bart.large.mnli", "bart.large.cnn", "bart_xsum/model.pt"]
......
...@@ -15,8 +15,7 @@ import tensorflow as tf ...@@ -15,8 +15,7 @@ import tensorflow as tf
import torch import torch
from transformers import BertConfig, BertModel from transformers import BertConfig, BertModel
from transformers.utils import logging
from .utils import logging
logging.set_verbosity_info() logging.set_verbosity_info()
......
...@@ -20,8 +20,7 @@ import argparse ...@@ -20,8 +20,7 @@ import argparse
import torch import torch
from transformers import BertConfig, BertForPreTraining, load_tf_weights_in_bert from transformers import BertConfig, BertForPreTraining, load_tf_weights_in_bert
from transformers.utils import logging
from .utils import logging
logging.set_verbosity_info() logging.set_verbosity_info()
......
...@@ -20,8 +20,7 @@ import argparse ...@@ -20,8 +20,7 @@ import argparse
import torch import torch
from transformers import ElectraConfig, ElectraForMaskedLM, ElectraForPreTraining, load_tf_weights_in_electra from transformers import ElectraConfig, ElectraForMaskedLM, ElectraForPreTraining, load_tf_weights_in_electra
from transformers.utils import logging
from .utils import logging
logging.set_verbosity_info() logging.set_verbosity_info()
......
...@@ -20,8 +20,7 @@ import argparse ...@@ -20,8 +20,7 @@ import argparse
import torch import torch
from transformers import CONFIG_NAME, WEIGHTS_NAME, GPT2Config, GPT2Model, load_tf_weights_in_gpt2 from transformers import CONFIG_NAME, WEIGHTS_NAME, GPT2Config, GPT2Model, load_tf_weights_in_gpt2
from transformers.utils import logging
from .utils import logging
logging.set_verbosity_info() logging.set_verbosity_info()
......
...@@ -3,8 +3,7 @@ import argparse ...@@ -3,8 +3,7 @@ import argparse
import torch import torch
from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert
from transformers.utils import logging
from .utils import logging
logging.set_verbosity_info() logging.set_verbosity_info()
......
...@@ -20,8 +20,7 @@ import argparse ...@@ -20,8 +20,7 @@ import argparse
import torch import torch
from transformers import CONFIG_NAME, WEIGHTS_NAME, OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers import CONFIG_NAME, WEIGHTS_NAME, OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt
from transformers.utils import logging
from .utils import logging
logging.set_verbosity_info() logging.set_verbosity_info()
......
...@@ -78,8 +78,7 @@ from transformers import ( ...@@ -78,8 +78,7 @@ from transformers import (
load_pytorch_checkpoint_in_tf2_model, load_pytorch_checkpoint_in_tf2_model,
) )
from transformers.file_utils import hf_bucket_url from transformers.file_utils import hf_bucket_url
from transformers.utils import logging
from .utils import logging
if is_torch_available(): if is_torch_available():
......
...@@ -22,8 +22,7 @@ import numpy as np ...@@ -22,8 +22,7 @@ import numpy as np
import torch import torch
from transformers import ReformerConfig, ReformerModelWithLMHead from transformers import ReformerConfig, ReformerModelWithLMHead
from transformers.utils import logging
from .utils import logging
logging.set_verbosity_info() logging.set_verbosity_info()
......
...@@ -26,8 +26,7 @@ from packaging import version ...@@ -26,8 +26,7 @@ from packaging import version
from transformers.modeling_bert import BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput from transformers.modeling_bert import BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput
from transformers.modeling_roberta import RobertaConfig, RobertaForMaskedLM, RobertaForSequenceClassification from transformers.modeling_roberta import RobertaConfig, RobertaForMaskedLM, RobertaForSequenceClassification
from transformers.utils import logging
from .utils import logging
if version.parse(fairseq.__version__) < version.parse("0.9.0"): if version.parse(fairseq.__version__) < version.parse("0.9.0"):
......
...@@ -20,8 +20,7 @@ import argparse ...@@ -20,8 +20,7 @@ import argparse
import torch import torch
from transformers import T5Config, T5Model, load_tf_weights_in_t5 from transformers import T5Config, T5Model, load_tf_weights_in_t5
from transformers.utils import logging
from .utils import logging
logging.set_verbosity_info() logging.set_verbosity_info()
......
...@@ -31,8 +31,7 @@ from transformers import ( ...@@ -31,8 +31,7 @@ from transformers import (
load_tf_weights_in_transfo_xl, load_tf_weights_in_transfo_xl,
) )
from transformers.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES
from transformers.utils import logging
from .utils import logging
logging.set_verbosity_info() logging.set_verbosity_info()
......
...@@ -23,8 +23,7 @@ import torch ...@@ -23,8 +23,7 @@ import torch
from transformers import CONFIG_NAME, WEIGHTS_NAME from transformers import CONFIG_NAME, WEIGHTS_NAME
from transformers.tokenization_xlm import VOCAB_FILES_NAMES from transformers.tokenization_xlm import VOCAB_FILES_NAMES
from transformers.utils import logging
from .utils import logging
logging.set_verbosity_info() logging.set_verbosity_info()
......
...@@ -29,8 +29,7 @@ from transformers import ( ...@@ -29,8 +29,7 @@ from transformers import (
XLNetLMHeadModel, XLNetLMHeadModel,
load_tf_weights_in_xlnet, load_tf_weights_in_xlnet,
) )
from transformers.utils import logging
from .utils import logging
GLUE_TASKS_NUM_LABELS = { GLUE_TASKS_NUM_LABELS = {
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment