"vscode:/vscode.git/clone" did not exist on "546e7679e7f692ebeefcfc5063cec271a55bae20"
Unverified Commit 29c10a41 authored by Lysandre Debut's avatar Lysandre Debut Committed by GitHub
Browse files

[Test refactor 1/5] Per-folder tests reorganization (#15725)



* Per-folder tests reorganization
Co-authored-by: default avatarsgugger <sylvain.gugger@gmail.com>
Co-authored-by: default avatarStas Bekman <stas@stason.org>
parent fecb08c2
...@@ -20,8 +20,8 @@ import timeout_decorator # noqa ...@@ -20,8 +20,8 @@ import timeout_decorator # noqa
from transformers import BlenderbotConfig, is_flax_available from transformers import BlenderbotConfig, is_flax_available
from transformers.testing_utils import jax_device, require_flax, slow from transformers.testing_utils import jax_device, require_flax, slow
from .test_generation_flax_utils import FlaxGenerationTesterMixin from ..generation.test_generation_flax_utils import FlaxGenerationTesterMixin
from .test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor from ..test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor
if is_flax_available(): if is_flax_available():
......
...@@ -20,8 +20,8 @@ from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available ...@@ -20,8 +20,8 @@ from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available
from transformers.file_utils import cached_property from transformers.file_utils import cached_property
from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.testing_utils import require_tf, require_tokenizers, slow
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ..test_modeling_tf_common import TFModelTesterMixin, ids_tensor
if is_tf_available(): if is_tf_available():
......
...@@ -21,9 +21,9 @@ from transformers import BlenderbotSmallConfig, is_torch_available ...@@ -21,9 +21,9 @@ from transformers import BlenderbotSmallConfig, is_torch_available
from transformers.file_utils import cached_property from transformers.file_utils import cached_property
from transformers.testing_utils import require_torch, slow, torch_device from transformers.testing_utils import require_torch, slow, torch_device
from .test_configuration_common import ConfigTester from ..generation.test_generation_utils import GenerationTesterMixin
from .test_generation_utils import GenerationTesterMixin from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, ids_tensor from ..test_modeling_common import ModelTesterMixin, ids_tensor
if is_torch_available(): if is_torch_available():
......
...@@ -20,8 +20,8 @@ import timeout_decorator # noqa ...@@ -20,8 +20,8 @@ import timeout_decorator # noqa
from transformers import BlenderbotSmallConfig, is_flax_available from transformers import BlenderbotSmallConfig, is_flax_available
from transformers.testing_utils import require_flax, slow from transformers.testing_utils import require_flax, slow
from .test_generation_flax_utils import FlaxGenerationTesterMixin from ..generation.test_generation_flax_utils import FlaxGenerationTesterMixin
from .test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor from ..test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor
if is_flax_available(): if is_flax_available():
......
...@@ -20,8 +20,8 @@ from transformers import BlenderbotSmallConfig, BlenderbotSmallTokenizer, is_tf_ ...@@ -20,8 +20,8 @@ from transformers import BlenderbotSmallConfig, BlenderbotSmallTokenizer, is_tf_
from transformers.file_utils import cached_property from transformers.file_utils import cached_property
from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.testing_utils import require_tf, require_tokenizers, slow
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ..test_modeling_tf_common import TFModelTesterMixin, ids_tensor
if is_tf_available(): if is_tf_available():
......
...@@ -23,7 +23,7 @@ from transformers.models.blenderbot_small.tokenization_blenderbot_small import ( ...@@ -23,7 +23,7 @@ from transformers.models.blenderbot_small.tokenization_blenderbot_small import (
BlenderbotSmallTokenizer, BlenderbotSmallTokenizer,
) )
from .test_tokenization_common import TokenizerTesterMixin from ..test_tokenization_common import TokenizerTesterMixin
class BlenderbotSmallTokenizerTest(TokenizerTesterMixin, unittest.TestCase): class BlenderbotSmallTokenizerTest(TokenizerTesterMixin, unittest.TestCase):
......
...@@ -24,7 +24,7 @@ from typing import Tuple ...@@ -24,7 +24,7 @@ from typing import Tuple
from transformers import AddedToken, BatchEncoding, ByT5Tokenizer from transformers import AddedToken, BatchEncoding, ByT5Tokenizer
from transformers.file_utils import cached_property, is_tf_available, is_torch_available from transformers.file_utils import cached_property, is_tf_available, is_torch_available
from .test_tokenization_common import TokenizerTesterMixin from ..test_tokenization_common import TokenizerTesterMixin
if is_torch_available(): if is_torch_available():
......
...@@ -15,16 +15,17 @@ ...@@ -15,16 +15,17 @@
import os import os
import unittest import unittest
from os.path import dirname
from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers import CamembertTokenizer, CamembertTokenizerFast
from transformers.file_utils import is_torch_available from transformers.file_utils import is_torch_available
from transformers.testing_utils import require_sentencepiece, require_tokenizers, slow from transformers.testing_utils import require_sentencepiece, require_tokenizers, slow
from .test_tokenization_common import TokenizerTesterMixin from ..test_tokenization_common import TokenizerTesterMixin
SAMPLE_VOCAB = os.path.join(os.path.dirname(os.path.abspath(__file__)), "fixtures/test_sentencepiece.model") SAMPLE_VOCAB = os.path.join(dirname(dirname(os.path.abspath(__file__))), "fixtures/test_sentencepiece.model")
SAMPLE_BPE_VOCAB = os.path.join(os.path.dirname(os.path.abspath(__file__)), "fixtures/test_sentencepiece_bpe.model") SAMPLE_BPE_VOCAB = os.path.join(dirname(dirname(os.path.abspath(__file__))), "fixtures/test_sentencepiece_bpe.model")
FRAMEWORK = "pt" if is_torch_available() else "tf" FRAMEWORK = "pt" if is_torch_available() else "tf"
......
...@@ -21,8 +21,8 @@ from typing import List, Tuple ...@@ -21,8 +21,8 @@ from typing import List, Tuple
from transformers import CanineConfig, is_torch_available from transformers import CanineConfig, is_torch_available
from transformers.testing_utils import require_torch, slow, torch_device from transformers.testing_utils import require_torch, slow, torch_device
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, _config_zero_init, global_rng, ids_tensor, random_attention_mask from ..test_modeling_common import ModelTesterMixin, _config_zero_init, global_rng, ids_tensor, random_attention_mask
if is_torch_available(): if is_torch_available():
......
...@@ -24,7 +24,7 @@ from transformers.file_utils import cached_property ...@@ -24,7 +24,7 @@ from transformers.file_utils import cached_property
from transformers.testing_utils import require_tokenizers, require_torch from transformers.testing_utils import require_tokenizers, require_torch
from transformers.tokenization_utils import AddedToken from transformers.tokenization_utils import AddedToken
from .test_tokenization_common import TokenizerTesterMixin from ..test_tokenization_common import TokenizerTesterMixin
class CanineTokenizationTest(TokenizerTesterMixin, unittest.TestCase): class CanineTokenizationTest(TokenizerTesterMixin, unittest.TestCase):
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment