Commit 158e82e0 authored by Aymeric Augustin's avatar Aymeric Augustin
Browse files

Sort imports with isort.

This is the result of:

    $ isort --recursive examples templates transformers utils hubconf.py setup.py
parent bc1715c1
...@@ -24,6 +24,7 @@ from io import open ...@@ -24,6 +24,7 @@ from io import open
from .tokenization_utils import PreTrainedTokenizer from .tokenization_utils import PreTrainedTokenizer
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
VOCAB_FILES_NAMES = {"vocab_file": "vocab.txt"} VOCAB_FILES_NAMES = {"vocab_file": "vocab.txt"}
......
...@@ -19,13 +19,15 @@ from __future__ import absolute_import, division, print_function, unicode_litera ...@@ -19,13 +19,15 @@ from __future__ import absolute_import, division, print_function, unicode_litera
import collections import collections
import logging import logging
import os import os
import six
import unicodedata import unicodedata
from io import open from io import open
from .tokenization_bert import BertTokenizer, BasicTokenizer, WordpieceTokenizer, load_vocab import six
from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer, load_vocab
from .tokenization_utils import PreTrainedTokenizer from .tokenization_utils import PreTrainedTokenizer
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
VOCAB_FILES_NAMES = {"vocab_file": "vocab.txt"} VOCAB_FILES_NAMES = {"vocab_file": "vocab.txt"}
......
...@@ -20,9 +20,12 @@ import os ...@@ -20,9 +20,12 @@ import os
from shutil import copyfile from shutil import copyfile
import sentencepiece as spm import sentencepiece as spm
from transformers.tokenization_utils import PreTrainedTokenizer from transformers.tokenization_utils import PreTrainedTokenizer
from .tokenization_xlnet import SPIECE_UNDERLINE from .tokenization_xlnet import SPIECE_UNDERLINE
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
VOCAB_FILES_NAMES = {"vocab_file": "sentencepiece.bpe.model"} VOCAB_FILES_NAMES = {"vocab_file": "sentencepiece.bpe.model"}
......
...@@ -18,11 +18,13 @@ from __future__ import absolute_import, division, print_function, unicode_litera ...@@ -18,11 +18,13 @@ from __future__ import absolute_import, division, print_function, unicode_litera
import json import json
import logging import logging
import os import os
import regex as re
from io import open from io import open
import regex as re
from .tokenization_utils import PreTrainedTokenizer from .tokenization_utils import PreTrainedTokenizer
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
VOCAB_FILES_NAMES = { VOCAB_FILES_NAMES = {
......
...@@ -24,6 +24,7 @@ from io import open ...@@ -24,6 +24,7 @@ from io import open
from .tokenization_bert import BertTokenizer from .tokenization_bert import BertTokenizer
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
VOCAB_FILES_NAMES = {"vocab_file": "vocab.txt"} VOCAB_FILES_NAMES = {"vocab_file": "vocab.txt"}
......
...@@ -15,13 +15,17 @@ ...@@ -15,13 +15,17 @@
"""Tokenization classes for OpenAI GPT.""" """Tokenization classes for OpenAI GPT."""
from __future__ import absolute_import, division, print_function, unicode_literals from __future__ import absolute_import, division, print_function, unicode_literals
import sys
import json import json
import logging import logging
import os import os
import regex as re import sys
from io import open from io import open
import regex as re
from .tokenization_utils import PreTrainedTokenizer
try: try:
from functools import lru_cache from functools import lru_cache
except ImportError: except ImportError:
...@@ -31,8 +35,6 @@ except ImportError: ...@@ -31,8 +35,6 @@ except ImportError:
return lambda func: func return lambda func: func
from .tokenization_utils import PreTrainedTokenizer
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
VOCAB_FILES_NAMES = { VOCAB_FILES_NAMES = {
......
...@@ -21,8 +21,9 @@ import os ...@@ -21,8 +21,9 @@ import os
import re import re
from io import open from io import open
from .tokenization_utils import PreTrainedTokenizer
from .tokenization_bert import BasicTokenizer from .tokenization_bert import BasicTokenizer
from .tokenization_utils import PreTrainedTokenizer
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
......
...@@ -15,15 +15,17 @@ ...@@ -15,15 +15,17 @@
"""Tokenization classes for RoBERTa.""" """Tokenization classes for RoBERTa."""
from __future__ import absolute_import, division, print_function, unicode_literals from __future__ import absolute_import, division, print_function, unicode_literals
import sys
import json import json
import logging import logging
import os import os
import regex as re import sys
from io import open from io import open
import regex as re
from .tokenization_gpt2 import GPT2Tokenizer from .tokenization_gpt2 import GPT2Tokenizer
try: try:
from functools import lru_cache from functools import lru_cache
except ImportError: except ImportError:
......
...@@ -19,11 +19,13 @@ from __future__ import absolute_import, division, print_function, unicode_litera ...@@ -19,11 +19,13 @@ from __future__ import absolute_import, division, print_function, unicode_litera
import logging import logging
import os import os
import re import re
import six
from shutil import copyfile from shutil import copyfile
import six
from .tokenization_utils import PreTrainedTokenizer from .tokenization_utils import PreTrainedTokenizer
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
SPIECE_UNDERLINE = "▁" SPIECE_UNDERLINE = "▁"
......
...@@ -30,6 +30,7 @@ import numpy as np ...@@ -30,6 +30,7 @@ import numpy as np
from .file_utils import cached_path from .file_utils import cached_path
from .tokenization_utils import PreTrainedTokenizer from .tokenization_utils import PreTrainedTokenizer
try: try:
import torch import torch
except ImportError: except ImportError:
......
...@@ -15,16 +15,18 @@ ...@@ -15,16 +15,18 @@
"""Tokenization classes for OpenAI GPT.""" """Tokenization classes for OpenAI GPT."""
from __future__ import absolute_import, division, print_function, unicode_literals from __future__ import absolute_import, division, print_function, unicode_literals
import logging
import os
import json
import six
import copy import copy
import itertools import itertools
import json
import logging
import os
import re import re
from io import open from io import open
from .file_utils import cached_path, is_remote_url, hf_bucket_url, is_tf_available, is_torch_available import six
from .file_utils import cached_path, hf_bucket_url, is_remote_url, is_tf_available, is_torch_available
if is_tf_available(): if is_tf_available():
import tensorflow as tf import tensorflow as tf
......
...@@ -25,8 +25,9 @@ from io import open ...@@ -25,8 +25,9 @@ from io import open
import sacremoses as sm import sacremoses as sm
from .tokenization_utils import PreTrainedTokenizer
from .tokenization_bert import BasicTokenizer from .tokenization_bert import BasicTokenizer
from .tokenization_utils import PreTrainedTokenizer
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
......
...@@ -20,9 +20,12 @@ import os ...@@ -20,9 +20,12 @@ import os
from shutil import copyfile from shutil import copyfile
import sentencepiece as spm import sentencepiece as spm
from transformers.tokenization_utils import PreTrainedTokenizer from transformers.tokenization_utils import PreTrainedTokenizer
from .tokenization_xlnet import SPIECE_UNDERLINE from .tokenization_xlnet import SPIECE_UNDERLINE
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
VOCAB_FILES_NAMES = {"vocab_file": "sentencepiece.bpe.model"} VOCAB_FILES_NAMES = {"vocab_file": "sentencepiece.bpe.model"}
......
...@@ -17,13 +17,14 @@ from __future__ import absolute_import, division, print_function, unicode_litera ...@@ -17,13 +17,14 @@ from __future__ import absolute_import, division, print_function, unicode_litera
import logging import logging
import os import os
import unicodedata
from shutil import copyfile from shutil import copyfile
import unicodedata
import six import six
from .tokenization_utils import PreTrainedTokenizer from .tokenization_utils import PreTrainedTokenizer
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
VOCAB_FILES_NAMES = {"vocab_file": "spiece.model"} VOCAB_FILES_NAMES = {"vocab_file": "spiece.model"}
......
...@@ -18,14 +18,15 @@ rm MSRParaphraseCorpus.msi ...@@ -18,14 +18,15 @@ rm MSRParaphraseCorpus.msi
2/11/19: It looks like SentEval actually *is* hosting the extracted data. Hooray! 2/11/19: It looks like SentEval actually *is* hosting the extracted data. Hooray!
""" """
import argparse
import os import os
import sys
import shutil import shutil
import argparse import sys
import tempfile import tempfile
import urllib.request import urllib.request
import zipfile import zipfile
TASKS = ["CoLA", "SST", "MRPC", "QQP", "STS", "MNLI", "SNLI", "QNLI", "RTE", "WNLI", "diagnostic"] TASKS = ["CoLA", "SST", "MRPC", "QQP", "STS", "MNLI", "SNLI", "QNLI", "RTE", "WNLI", "diagnostic"]
TASK2PATH = { TASK2PATH = {
"CoLA": "https://firebasestorage.googleapis.com/v0/b/mtl-sentence-representations.appspot.com/o/data%2FCoLA.zip?alt=media&token=46d5e637-3411-4188-bc44-5809b5bfb5f4", "CoLA": "https://firebasestorage.googleapis.com/v0/b/mtl-sentence-representations.appspot.com/o/data%2FCoLA.zip?alt=media&token=46d5e637-3411-4188-bc44-5809b5bfb5f4",
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment