Unverified Commit 29c10a41 authored by Lysandre Debut's avatar Lysandre Debut Committed by GitHub
Browse files

[Test refactor 1/5] Per-folder tests reorganization (#15725)



* Per-folder tests reorganization
Co-authored-by: default avatarsgugger <sylvain.gugger@gmail.com>
Co-authored-by: default avatarStas Bekman <stas@stason.org>
parent fecb08c2
...@@ -20,9 +20,9 @@ from transformers import BertConfig, is_torch_available ...@@ -20,9 +20,9 @@ from transformers import BertConfig, is_torch_available
from transformers.models.auto import get_values from transformers.models.auto import get_values
from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device
from .test_configuration_common import ConfigTester from ..generation.test_generation_utils import GenerationTesterMixin
from .test_generation_utils import GenerationTesterMixin from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ..test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask
if is_torch_available(): if is_torch_available():
......
...@@ -19,7 +19,7 @@ import numpy as np ...@@ -19,7 +19,7 @@ import numpy as np
from transformers import BertConfig, is_flax_available from transformers import BertConfig, is_flax_available
from transformers.testing_utils import require_flax, slow from transformers.testing_utils import require_flax, slow
from .test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask from ..test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask
if is_flax_available(): if is_flax_available():
......
...@@ -20,9 +20,9 @@ from transformers import BertConfig, is_tf_available ...@@ -20,9 +20,9 @@ from transformers import BertConfig, is_tf_available
from transformers.models.auto import get_values from transformers.models.auto import get_values
from transformers.testing_utils import require_tf, slow from transformers.testing_utils import require_tf, slow
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ..test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor
from .test_modeling_tf_core import TFCoreModelTesterMixin from ..utils.test_modeling_tf_core import TFCoreModelTesterMixin
if is_tf_available(): if is_tf_available():
......
...@@ -29,7 +29,7 @@ from transformers.models.bert.tokenization_bert import ( ...@@ -29,7 +29,7 @@ from transformers.models.bert.tokenization_bert import (
) )
from transformers.testing_utils import require_tokenizers, slow from transformers.testing_utils import require_tokenizers, slow
from .test_tokenization_common import TokenizerTesterMixin, filter_non_english from ..test_tokenization_common import TokenizerTesterMixin, filter_non_english
@require_tokenizers @require_tokenizers
......
...@@ -19,9 +19,9 @@ import unittest ...@@ -19,9 +19,9 @@ import unittest
from transformers import BertGenerationConfig, is_torch_available from transformers import BertGenerationConfig, is_torch_available
from transformers.testing_utils import require_torch, slow, torch_device from transformers.testing_utils import require_torch, slow, torch_device
from .test_configuration_common import ConfigTester from ..generation.test_generation_utils import GenerationTesterMixin
from .test_generation_utils import GenerationTesterMixin from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ..test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask
if is_torch_available(): if is_torch_available():
......
...@@ -15,17 +15,18 @@ ...@@ -15,17 +15,18 @@
import os import os
import unittest import unittest
from os.path import dirname
from transformers import BertGenerationTokenizer from transformers import BertGenerationTokenizer
from transformers.file_utils import cached_property from transformers.file_utils import cached_property
from transformers.testing_utils import require_sentencepiece, require_torch, slow from transformers.testing_utils import require_sentencepiece, require_torch, slow
from .test_tokenization_common import TokenizerTesterMixin from ..test_tokenization_common import TokenizerTesterMixin
SPIECE_UNDERLINE = "▁" SPIECE_UNDERLINE = "▁"
SAMPLE_VOCAB = os.path.join(os.path.dirname(os.path.abspath(__file__)), "fixtures/test_sentencepiece.model") SAMPLE_VOCAB = os.path.join(dirname(dirname(os.path.abspath(__file__))), "fixtures/test_sentencepiece.model")
@require_sentencepiece @require_sentencepiece
......
...@@ -29,7 +29,7 @@ from transformers.models.bert_japanese.tokenization_bert_japanese import ( ...@@ -29,7 +29,7 @@ from transformers.models.bert_japanese.tokenization_bert_japanese import (
) )
from transformers.testing_utils import custom_tokenizers from transformers.testing_utils import custom_tokenizers
from .test_tokenization_common import TokenizerTesterMixin from ..test_tokenization_common import TokenizerTesterMixin
@custom_tokenizers @custom_tokenizers
......
...@@ -18,7 +18,7 @@ import unittest ...@@ -18,7 +18,7 @@ import unittest
from transformers.models.bertweet.tokenization_bertweet import VOCAB_FILES_NAMES, BertweetTokenizer from transformers.models.bertweet.tokenization_bertweet import VOCAB_FILES_NAMES, BertweetTokenizer
from .test_tokenization_common import TokenizerTesterMixin from ..test_tokenization_common import TokenizerTesterMixin
class BertweetTokenizationTest(TokenizerTesterMixin, unittest.TestCase): class BertweetTokenizationTest(TokenizerTesterMixin, unittest.TestCase):
......
...@@ -17,14 +17,13 @@ ...@@ -17,14 +17,13 @@
import unittest import unittest
from tests.test_modeling_common import floats_tensor
from transformers import BigBirdConfig, is_torch_available from transformers import BigBirdConfig, is_torch_available
from transformers.models.auto import get_values from transformers.models.auto import get_values
from transformers.models.big_bird.tokenization_big_bird import BigBirdTokenizer from transformers.models.big_bird.tokenization_big_bird import BigBirdTokenizer
from transformers.testing_utils import require_torch, slow, torch_device from transformers.testing_utils import require_torch, slow, torch_device
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ..test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask
if is_torch_available(): if is_torch_available():
......
...@@ -19,7 +19,7 @@ import numpy as np ...@@ -19,7 +19,7 @@ import numpy as np
from transformers import BigBirdConfig, is_flax_available from transformers import BigBirdConfig, is_flax_available
from transformers.testing_utils import require_flax, slow from transformers.testing_utils import require_flax, slow
from .test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask from ..test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask
if is_flax_available(): if is_flax_available():
......
...@@ -15,17 +15,18 @@ ...@@ -15,17 +15,18 @@
import os import os
import unittest import unittest
from os.path import dirname
from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers import BigBirdTokenizer, BigBirdTokenizerFast
from transformers.file_utils import cached_property from transformers.file_utils import cached_property
from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow
from .test_tokenization_common import TokenizerTesterMixin from ..test_tokenization_common import TokenizerTesterMixin
SPIECE_UNDERLINE = "▁" SPIECE_UNDERLINE = "▁"
SAMPLE_VOCAB = os.path.join(os.path.dirname(os.path.abspath(__file__)), "fixtures/test_sentencepiece.model") SAMPLE_VOCAB = os.path.join(dirname(dirname(os.path.abspath(__file__))), "fixtures/test_sentencepiece.model")
@require_sentencepiece @require_sentencepiece
......
...@@ -22,9 +22,9 @@ import unittest ...@@ -22,9 +22,9 @@ import unittest
from transformers import BigBirdPegasusConfig, is_torch_available from transformers import BigBirdPegasusConfig, is_torch_available
from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device
from .test_configuration_common import ConfigTester from ..generation.test_generation_utils import GenerationTesterMixin
from .test_generation_utils import GenerationTesterMixin from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, ids_tensor from ..test_modeling_common import ModelTesterMixin, ids_tensor
if is_torch_available(): if is_torch_available():
......
...@@ -21,9 +21,9 @@ from transformers import BlenderbotConfig, is_torch_available ...@@ -21,9 +21,9 @@ from transformers import BlenderbotConfig, is_torch_available
from transformers.file_utils import cached_property from transformers.file_utils import cached_property
from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device
from .test_configuration_common import ConfigTester from ..generation.test_generation_utils import GenerationTesterMixin
from .test_generation_utils import GenerationTesterMixin from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, ids_tensor from ..test_modeling_common import ModelTesterMixin, ids_tensor
if is_torch_available(): if is_torch_available():
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment