Unverified Commit 29c10a41 authored by Lysandre Debut's avatar Lysandre Debut Committed by GitHub
Browse files

[Test refactor 1/5] Per-folder tests reorganization (#15725)



* Per-folder tests reorganization
Co-authored-by: default avatarsgugger <sylvain.gugger@gmail.com>
Co-authored-by: default avatarStas Bekman <stas@stason.org>
parent fecb08c2
...@@ -26,7 +26,7 @@ from transformers.testing_utils import ( ...@@ -26,7 +26,7 @@ from transformers.testing_utils import (
slow, slow,
) )
from .test_modeling_bert import BertModelTester from ..bert.test_modeling_bert import BertModelTester
if is_tf_available(): if is_tf_available():
......
...@@ -41,7 +41,7 @@ from transformers.testing_utils import PASS, USER, is_staging_test ...@@ -41,7 +41,7 @@ from transformers.testing_utils import PASS, USER, is_staging_test
from transformers.tokenization_utils import TOKENIZER_CONFIG_FILE from transformers.tokenization_utils import TOKENIZER_CONFIG_FILE
sys.path.append(str(Path(__file__).parent.parent / "utils")) sys.path.append(str(Path(__file__).parent.parent.parent / "utils"))
from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_configuration import CustomConfig # noqa E402
from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402
...@@ -50,11 +50,11 @@ from test_module.custom_tokenization import CustomTokenizer # noqa E402 ...@@ -50,11 +50,11 @@ from test_module.custom_tokenization import CustomTokenizer # noqa E402
SAMPLE_PROCESSOR_CONFIG = os.path.join( SAMPLE_PROCESSOR_CONFIG = os.path.join(
os.path.dirname(os.path.abspath(__file__)), "fixtures/dummy_feature_extractor_config.json" os.path.dirname(os.path.abspath(__file__)), "../fixtures/dummy_feature_extractor_config.json"
) )
SAMPLE_VOCAB = os.path.join(os.path.dirname(os.path.abspath(__file__)), "fixtures/vocab.json") SAMPLE_VOCAB = os.path.join(os.path.dirname(os.path.abspath(__file__)), "../fixtures/vocab.json")
SAMPLE_PROCESSOR_CONFIG_DIR = os.path.join(os.path.dirname(os.path.abspath(__file__)), "fixtures") SAMPLE_PROCESSOR_CONFIG_DIR = os.path.join(os.path.dirname(os.path.abspath(__file__)), "../fixtures")
class AutoFeatureExtractorTest(unittest.TestCase): class AutoFeatureExtractorTest(unittest.TestCase):
......
...@@ -25,9 +25,9 @@ from transformers import BartConfig, is_torch_available ...@@ -25,9 +25,9 @@ from transformers import BartConfig, is_torch_available
from transformers.file_utils import cached_property from transformers.file_utils import cached_property
from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device
from .test_configuration_common import ConfigTester from ..generation.test_generation_utils import GenerationTesterMixin
from .test_generation_utils import GenerationTesterMixin from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ..test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor
if is_torch_available(): if is_torch_available():
......
...@@ -20,8 +20,8 @@ import timeout_decorator # noqa ...@@ -20,8 +20,8 @@ import timeout_decorator # noqa
from transformers import BartConfig, is_flax_available from transformers import BartConfig, is_flax_available
from transformers.testing_utils import require_flax, slow from transformers.testing_utils import require_flax, slow
from .test_generation_flax_utils import FlaxGenerationTesterMixin from ..generation.test_generation_flax_utils import FlaxGenerationTesterMixin
from .test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor from ..test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor
if is_flax_available(): if is_flax_available():
......
...@@ -21,9 +21,9 @@ from transformers import BartConfig, BartTokenizer, is_tf_available ...@@ -21,9 +21,9 @@ from transformers import BartConfig, BartTokenizer, is_tf_available
from transformers.file_utils import cached_property from transformers.file_utils import cached_property
from transformers.testing_utils import require_tf, slow from transformers.testing_utils import require_tf, slow
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ..test_modeling_tf_common import TFModelTesterMixin, ids_tensor
from .test_modeling_tf_core import TFCoreModelTesterMixin from ..utils.test_modeling_tf_core import TFCoreModelTesterMixin
if is_tf_available(): if is_tf_available():
......
...@@ -20,7 +20,7 @@ from transformers.file_utils import cached_property ...@@ -20,7 +20,7 @@ from transformers.file_utils import cached_property
from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES
from transformers.testing_utils import require_tokenizers, require_torch from transformers.testing_utils import require_tokenizers, require_torch
from .test_tokenization_common import TokenizerTesterMixin, filter_roberta_detectors from ..test_tokenization_common import TokenizerTesterMixin, filter_roberta_detectors
@require_tokenizers @require_tokenizers
......
...@@ -18,7 +18,7 @@ import unittest ...@@ -18,7 +18,7 @@ import unittest
from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding
from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow
from .test_tokenization_common import TokenizerTesterMixin from ..test_tokenization_common import TokenizerTesterMixin
@require_tokenizers @require_tokenizers
......
...@@ -15,13 +15,14 @@ ...@@ -15,13 +15,14 @@
import os import os
import unittest import unittest
from os.path import dirname
from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer
from .test_tokenization_common import TokenizerTesterMixin from ..test_tokenization_common import TokenizerTesterMixin
SAMPLE_VOCAB = os.path.join(os.path.dirname(os.path.abspath(__file__)), "fixtures/test_sentencepiece_bpe.model") SAMPLE_VOCAB = os.path.join(dirname(dirname(os.path.abspath(__file__))), "fixtures/test_sentencepiece_bpe.model")
class BartphoTokenizerTest(TokenizerTesterMixin, unittest.TestCase): class BartphoTokenizerTest(TokenizerTesterMixin, unittest.TestCase):
......
...@@ -22,7 +22,7 @@ from datasets import load_dataset ...@@ -22,7 +22,7 @@ from datasets import load_dataset
from transformers.file_utils import is_torch_available, is_vision_available from transformers.file_utils import is_torch_available, is_vision_available
from transformers.testing_utils import require_torch, require_vision from transformers.testing_utils import require_torch, require_vision
from .test_feature_extraction_common import FeatureExtractionSavingTestMixin, prepare_image_inputs from ..test_feature_extraction_common import FeatureExtractionSavingTestMixin, prepare_image_inputs
if is_torch_available(): if is_torch_available():
......
...@@ -25,8 +25,8 @@ from transformers.file_utils import cached_property, is_torch_available, is_visi ...@@ -25,8 +25,8 @@ from transformers.file_utils import cached_property, is_torch_available, is_visi
from transformers.models.auto import get_values from transformers.models.auto import get_values
from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.testing_utils import require_torch, require_vision, slow, torch_device
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ..test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor
if is_torch_available(): if is_torch_available():
......
...@@ -21,8 +21,8 @@ from transformers import BeitConfig ...@@ -21,8 +21,8 @@ from transformers import BeitConfig
from transformers.file_utils import cached_property, is_flax_available, is_vision_available from transformers.file_utils import cached_property, is_flax_available, is_vision_available
from transformers.testing_utils import require_flax, require_vision, slow from transformers.testing_utils import require_flax, require_vision, slow
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor from ..test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor
if is_flax_available(): if is_flax_available():
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment