"test/srt/git@developer.sourcefind.cn:zhaoyu6/sglang.git" did not exist on "276e7b3e4e762119f4b3a2bd7663e1f19a7c304c"
Unverified Commit a5737779 authored by Sylvain Gugger's avatar Sylvain Gugger Committed by GitHub
Browse files

Update repo to isort v5 (#6686)

* Run new isort

* More changes

* Update CI, CONTRIBUTING and benchmarks
parent d329c9b0
...@@ -28,13 +28,13 @@ if is_tf_available(): ...@@ -28,13 +28,13 @@ if is_tf_available():
import tensorflow as tf import tensorflow as tf
from transformers.modeling_tf_xlnet import ( from transformers.modeling_tf_xlnet import (
TFXLNetModel, TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST,
TFXLNetLMHeadModel, TFXLNetForMultipleChoice,
TFXLNetForQuestionAnsweringSimple,
TFXLNetForSequenceClassification, TFXLNetForSequenceClassification,
TFXLNetForTokenClassification, TFXLNetForTokenClassification,
TFXLNetForQuestionAnsweringSimple, TFXLNetLMHeadModel,
TFXLNetForMultipleChoice, TFXLNetModel,
TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST,
) )
......
...@@ -25,7 +25,8 @@ from .test_modeling_common import ModelTesterMixin, ids_tensor ...@@ -25,7 +25,8 @@ from .test_modeling_common import ModelTesterMixin, ids_tensor
if is_torch_available(): if is_torch_available():
import torch import torch
from transformers import TransfoXLConfig, TransfoXLModel, TransfoXLLMHeadModel
from transformers import TransfoXLConfig, TransfoXLLMHeadModel, TransfoXLModel
from transformers.modeling_transfo_xl import TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.modeling_transfo_xl import TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST
......
...@@ -25,15 +25,16 @@ from .test_modeling_common import ModelTesterMixin, ids_tensor, random_attention ...@@ -25,15 +25,16 @@ from .test_modeling_common import ModelTesterMixin, ids_tensor, random_attention
if is_torch_available(): if is_torch_available():
import torch import torch
from transformers import ( from transformers import (
XLMConfig, XLMConfig,
XLMModel, XLMForMultipleChoice,
XLMWithLMHeadModel,
XLMForTokenClassification,
XLMForQuestionAnswering, XLMForQuestionAnswering,
XLMForSequenceClassification,
XLMForQuestionAnsweringSimple, XLMForQuestionAnsweringSimple,
XLMForMultipleChoice, XLMForSequenceClassification,
XLMForTokenClassification,
XLMModel,
XLMWithLMHeadModel,
) )
from transformers.modeling_xlm import XLM_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.modeling_xlm import XLM_PRETRAINED_MODEL_ARCHIVE_LIST
......
...@@ -22,6 +22,7 @@ from transformers.testing_utils import slow ...@@ -22,6 +22,7 @@ from transformers.testing_utils import slow
if is_torch_available(): if is_torch_available():
import torch import torch
from transformers import XLMRobertaModel from transformers import XLMRobertaModel
......
...@@ -29,13 +29,13 @@ if is_torch_available(): ...@@ -29,13 +29,13 @@ if is_torch_available():
from transformers import ( from transformers import (
XLNetConfig, XLNetConfig,
XLNetModel,
XLNetLMHeadModel,
XLNetForMultipleChoice, XLNetForMultipleChoice,
XLNetForSequenceClassification,
XLNetForTokenClassification,
XLNetForQuestionAnswering, XLNetForQuestionAnswering,
XLNetForQuestionAnsweringSimple, XLNetForQuestionAnsweringSimple,
XLNetForSequenceClassification,
XLNetForTokenClassification,
XLNetLMHeadModel,
XLNetModel,
) )
from transformers.modeling_xlnet import XLNET_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.modeling_xlnet import XLNET_PRETRAINED_MODEL_ARCHIVE_LIST
......
...@@ -8,7 +8,8 @@ if is_tf_available(): ...@@ -8,7 +8,8 @@ if is_tf_available():
import tensorflow as tf import tensorflow as tf
from tensorflow.python.eager import context from tensorflow.python.eager import context
from tensorflow.python.framework import ops from tensorflow.python.framework import ops
from transformers import create_optimizer, GradientAccumulator
from transformers import GradientAccumulator, create_optimizer
@require_tf @require_tf
......
...@@ -28,11 +28,7 @@ from transformers.tokenization_utils import AddedToken ...@@ -28,11 +28,7 @@ from transformers.tokenization_utils import AddedToken
if TYPE_CHECKING: if TYPE_CHECKING:
from transformers import ( from transformers import PretrainedConfig, PreTrainedModel, TFPreTrainedModel
PretrainedConfig,
PreTrainedModel,
TFPreTrainedModel,
)
def merge_model_tokenizer_mappings( def merge_model_tokenizer_mappings(
...@@ -1398,6 +1394,7 @@ class TokenizerTesterMixin: ...@@ -1398,6 +1394,7 @@ class TokenizerTesterMixin:
@require_torch @require_torch
def test_torch_encode_plus_sent_to_model(self): def test_torch_encode_plus_sent_to_model(self):
import torch import torch
from transformers import MODEL_MAPPING, TOKENIZER_MAPPING from transformers import MODEL_MAPPING, TOKENIZER_MAPPING
MODEL_TOKENIZER_MAPPING = merge_model_tokenizer_mappings(MODEL_MAPPING, TOKENIZER_MAPPING) MODEL_TOKENIZER_MAPPING = merge_model_tokenizer_mappings(MODEL_MAPPING, TOKENIZER_MAPPING)
......
...@@ -232,7 +232,8 @@ class ReformerTokenizationTest(TokenizerTesterMixin, unittest.TestCase): ...@@ -232,7 +232,8 @@ class ReformerTokenizationTest(TokenizerTesterMixin, unittest.TestCase):
@require_torch @require_torch
def test_torch_encode_plus_sent_to_model(self): def test_torch_encode_plus_sent_to_model(self):
import torch import torch
from transformers import ReformerModel, ReformerConfig
from transformers import ReformerConfig, ReformerModel
# Build sequence # Build sequence
first_ten_tokens = list(self.big_tokenizer.get_vocab().keys())[:10] first_ten_tokens = list(self.big_tokenizer.get_vocab().keys())[:10]
......
...@@ -24,7 +24,7 @@ from .test_tokenization_common import TokenizerTesterMixin ...@@ -24,7 +24,7 @@ from .test_tokenization_common import TokenizerTesterMixin
if is_torch_available(): if is_torch_available():
from transformers.tokenization_transfo_xl import TransfoXLTokenizer, VOCAB_FILES_NAMES from transformers.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer
@require_torch @require_torch
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment