Unverified Commit 29c10a41 authored by Lysandre Debut's avatar Lysandre Debut Committed by GitHub
Browse files

[Test refactor 1/5] Per-folder tests reorganization (#15725)



* Per-folder tests reorganization
Co-authored-by: default avatarsgugger <sylvain.gugger@gmail.com>
Co-authored-by: default avatarStas Bekman <stas@stason.org>
parent fecb08c2
...@@ -35,8 +35,8 @@ from transformers.file_utils import cached_property ...@@ -35,8 +35,8 @@ from transformers.file_utils import cached_property
from transformers.models.auto import get_values from transformers.models.auto import get_values
from transformers.testing_utils import require_scatter, require_torch, slow, torch_device from transformers.testing_utils import require_scatter, require_torch, slow, torch_device
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ..test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask
if is_torch_available(): if is_torch_available():
......
...@@ -37,8 +37,8 @@ from transformers.file_utils import cached_property ...@@ -37,8 +37,8 @@ from transformers.file_utils import cached_property
from transformers.models.auto import get_values from transformers.models.auto import get_values
from transformers.testing_utils import require_tensorflow_probability, require_tf, slow from transformers.testing_utils import require_tensorflow_probability, require_tf, slow
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ..test_modeling_tf_common import TFModelTesterMixin, ids_tensor
if is_tf_available(): if is_tf_available():
......
...@@ -41,7 +41,7 @@ from transformers.testing_utils import ( ...@@ -41,7 +41,7 @@ from transformers.testing_utils import (
slow, slow,
) )
from .test_tokenization_common import TokenizerTesterMixin, filter_non_english, merge_model_tokenizer_mappings from ..test_tokenization_common import TokenizerTesterMixin, filter_non_english, merge_model_tokenizer_mappings
@require_tokenizers @require_tokenizers
......
...@@ -23,7 +23,7 @@ import unittest ...@@ -23,7 +23,7 @@ import unittest
from transformers import AutoTokenizer, PreTrainedTokenizerFast from transformers import AutoTokenizer, PreTrainedTokenizerFast
from transformers.testing_utils import require_tokenizers from transformers.testing_utils import require_tokenizers
from .test_tokenization_common import TokenizerTesterMixin from ..test_tokenization_common import TokenizerTesterMixin
@require_tokenizers @require_tokenizers
......
...@@ -370,7 +370,7 @@ class TrainerIntegrationPrerunTest(TestCasePlus, TrainerIntegrationCommon): ...@@ -370,7 +370,7 @@ class TrainerIntegrationPrerunTest(TestCasePlus, TrainerIntegrationCommon):
def setUp(self): def setUp(self):
super().setUp() super().setUp()
args = TrainingArguments(".") args = TrainingArguments("..")
self.n_epochs = args.num_train_epochs self.n_epochs = args.num_train_epochs
self.batch_size = args.train_batch_size self.batch_size = args.train_batch_size
trainer = get_regression_trainer(learning_rate=0.1) trainer = get_regression_trainer(learning_rate=0.1)
...@@ -536,7 +536,7 @@ class TrainerIntegrationPrerunTest(TestCasePlus, TrainerIntegrationCommon): ...@@ -536,7 +536,7 @@ class TrainerIntegrationPrerunTest(TestCasePlus, TrainerIntegrationCommon):
class TrainerIntegrationTest(TestCasePlus, TrainerIntegrationCommon): class TrainerIntegrationTest(TestCasePlus, TrainerIntegrationCommon):
def setUp(self): def setUp(self):
super().setUp() super().setUp()
args = TrainingArguments(".") args = TrainingArguments("..")
self.n_epochs = args.num_train_epochs self.n_epochs = args.num_train_epochs
self.batch_size = args.train_batch_size self.batch_size = args.train_batch_size
...@@ -1148,7 +1148,7 @@ class TrainerIntegrationTest(TestCasePlus, TrainerIntegrationCommon): ...@@ -1148,7 +1148,7 @@ class TrainerIntegrationTest(TestCasePlus, TrainerIntegrationCommon):
data = FiniteIterableDataset(length=num_samples) data = FiniteIterableDataset(length=num_samples)
train_args = TrainingArguments( train_args = TrainingArguments(
".", "..",
max_steps=available_steps + 1, # set a higher number than actually available max_steps=available_steps + 1, # set a higher number than actually available
per_device_train_batch_size=batch_size, per_device_train_batch_size=batch_size,
) )
...@@ -1589,7 +1589,7 @@ class TrainerIntegrationWithHubTester(unittest.TestCase): ...@@ -1589,7 +1589,7 @@ class TrainerIntegrationWithHubTester(unittest.TestCase):
@require_optuna @require_optuna
class TrainerHyperParameterOptunaIntegrationTest(unittest.TestCase): class TrainerHyperParameterOptunaIntegrationTest(unittest.TestCase):
def setUp(self): def setUp(self):
args = TrainingArguments(".") args = TrainingArguments("..")
self.n_epochs = args.num_train_epochs self.n_epochs = args.num_train_epochs
self.batch_size = args.train_batch_size self.batch_size = args.train_batch_size
...@@ -1635,7 +1635,7 @@ class TrainerHyperParameterOptunaIntegrationTest(unittest.TestCase): ...@@ -1635,7 +1635,7 @@ class TrainerHyperParameterOptunaIntegrationTest(unittest.TestCase):
@require_ray @require_ray
class TrainerHyperParameterRayIntegrationTest(unittest.TestCase): class TrainerHyperParameterRayIntegrationTest(unittest.TestCase):
def setUp(self): def setUp(self):
args = TrainingArguments(".") args = TrainingArguments("..")
self.n_epochs = args.num_train_epochs self.n_epochs = args.num_train_epochs
self.batch_size = args.train_batch_size self.batch_size = args.train_batch_size
...@@ -1699,7 +1699,7 @@ class TrainerHyperParameterRayIntegrationTest(unittest.TestCase): ...@@ -1699,7 +1699,7 @@ class TrainerHyperParameterRayIntegrationTest(unittest.TestCase):
@require_sigopt @require_sigopt
class TrainerHyperParameterSigOptIntegrationTest(unittest.TestCase): class TrainerHyperParameterSigOptIntegrationTest(unittest.TestCase):
def setUp(self): def setUp(self):
args = TrainingArguments(".") args = TrainingArguments("..")
self.n_epochs = args.num_train_epochs self.n_epochs = args.num_train_epochs
self.batch_size = args.train_batch_size self.batch_size = args.train_batch_size
...@@ -1845,7 +1845,7 @@ class TrainerOptimizerChoiceTest(unittest.TestCase): ...@@ -1845,7 +1845,7 @@ class TrainerOptimizerChoiceTest(unittest.TestCase):
@require_wandb @require_wandb
class TrainerHyperParameterWandbIntegrationTest(unittest.TestCase): class TrainerHyperParameterWandbIntegrationTest(unittest.TestCase):
def setUp(self): def setUp(self):
args = TrainingArguments(".") args = TrainingArguments("..")
self.n_epochs = args.num_train_epochs self.n_epochs = args.num_train_epochs
self.batch_size = args.train_batch_size self.batch_size = args.train_batch_size
......
...@@ -20,8 +20,8 @@ import unittest ...@@ -20,8 +20,8 @@ import unittest
from transformers import TransfoXLConfig, is_tf_available from transformers import TransfoXLConfig, is_tf_available
from transformers.testing_utils import require_tf, slow from transformers.testing_utils import require_tf, slow
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ..test_modeling_tf_common import TFModelTesterMixin, ids_tensor
if is_tf_available(): if is_tf_available():
......
...@@ -20,9 +20,9 @@ import unittest ...@@ -20,9 +20,9 @@ import unittest
from transformers import TransfoXLConfig, is_torch_available from transformers import TransfoXLConfig, is_torch_available
from transformers.testing_utils import require_torch, require_torch_multi_gpu, slow, torch_device from transformers.testing_utils import require_torch, require_torch_multi_gpu, slow, torch_device
from .test_configuration_common import ConfigTester from ..generation.test_generation_utils import GenerationTesterMixin
from .test_generation_utils import GenerationTesterMixin from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, ids_tensor from ..test_modeling_common import ModelTesterMixin, ids_tensor
if is_torch_available(): if is_torch_available():
......
...@@ -19,7 +19,7 @@ import unittest ...@@ -19,7 +19,7 @@ import unittest
from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer
from .test_tokenization_common import TokenizerTesterMixin from ..test_tokenization_common import TokenizerTesterMixin
class TransfoXLTokenizationTest(TokenizerTesterMixin, unittest.TestCase): class TransfoXLTokenizationTest(TokenizerTesterMixin, unittest.TestCase):
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment