Unverified Commit 29c10a41 authored by Lysandre Debut's avatar Lysandre Debut Committed by GitHub
Browse files

[Test refactor 1/5] Per-folder tests reorganization (#15725)



* Per-folder tests reorganization
Co-authored-by: default avatarsgugger <sylvain.gugger@gmail.com>
Co-authored-by: default avatarStas Bekman <stas@stason.org>
parent fecb08c2
...@@ -22,7 +22,7 @@ import unittest ...@@ -22,7 +22,7 @@ import unittest
import black import black
git_repo_path = os.path.abspath(os.path.dirname(os.path.dirname(__file__))) git_repo_path = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__))))
sys.path.append(os.path.join(git_repo_path, "utils")) sys.path.append(os.path.join(git_repo_path, "utils"))
import check_copies # noqa: E402 import check_copies # noqa: E402
......
...@@ -21,7 +21,7 @@ import numpy as np ...@@ -21,7 +21,7 @@ import numpy as np
from transformers.file_utils import is_torch_available, is_vision_available from transformers.file_utils import is_torch_available, is_vision_available
from transformers.testing_utils import require_torch, require_vision from transformers.testing_utils import require_torch, require_vision
from .test_feature_extraction_common import FeatureExtractionSavingTestMixin, prepare_image_inputs from ..test_feature_extraction_common import FeatureExtractionSavingTestMixin, prepare_image_inputs
if is_torch_available(): if is_torch_available():
......
...@@ -23,8 +23,8 @@ from transformers.file_utils import cached_property ...@@ -23,8 +23,8 @@ from transformers.file_utils import cached_property
from transformers.models.auto import get_values from transformers.models.auto import get_values
from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.testing_utils import require_torch, require_vision, slow, torch_device
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ..test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask
if is_torch_available(): if is_torch_available():
......
...@@ -22,9 +22,9 @@ import numpy as np ...@@ -22,9 +22,9 @@ import numpy as np
from transformers import is_flax_available, is_torch_available, is_vision_available from transformers import is_flax_available, is_torch_available, is_vision_available
from transformers.testing_utils import is_pt_flax_cross_test, require_flax, require_vision, slow, torch_device from transformers.testing_utils import is_pt_flax_cross_test, require_flax, require_vision, slow, torch_device
from .test_modeling_flax_common import floats_tensor, ids_tensor from ..gpt2.test_modeling_flax_gpt2 import FlaxGPT2ModelTester
from .test_modeling_flax_gpt2 import FlaxGPT2ModelTester from ..test_modeling_flax_common import floats_tensor, ids_tensor
from .test_modeling_flax_vit import FlaxViTModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester
if is_flax_available(): if is_flax_available():
......
...@@ -32,9 +32,9 @@ from transformers.testing_utils import ( ...@@ -32,9 +32,9 @@ from transformers.testing_utils import (
torch_device, torch_device,
) )
from .test_modeling_tf_common import floats_tensor, ids_tensor from ..gpt2.test_modeling_tf_gpt2 import TFGPT2ModelTester
from .test_modeling_tf_gpt2 import TFGPT2ModelTester from ..test_modeling_tf_common import floats_tensor, ids_tensor
from .test_modeling_tf_vit import TFViTModelTester from ..vit.test_modeling_tf_vit import TFViTModelTester
if is_tf_available(): if is_tf_available():
...@@ -527,7 +527,9 @@ class TFVisionEncoderDecoderMixin: ...@@ -527,7 +527,9 @@ class TFVisionEncoderDecoderMixin:
@require_tf @require_tf
class TFViT2GPT2EncoderDecoderModelTest(TFVisionEncoderDecoderMixin, unittest.TestCase): class TFViT2GPT2EncoderDecoderModelTest(TFVisionEncoderDecoderMixin, unittest.TestCase):
def get_pretrained_model(self): def get_pretrained_model(self):
return TFVisionEncoderDecoderModel.from_encoder_decoder_pretrained("google/vit-base-patch16-224-in21k", "gpt2") return TFVisionEncoderDecoderModel.from_encoder_decoder_pretrained(
"google/vit-base-patch16-224-in21k", "../gpt2"
)
def get_encoder_decoder_model(self, config, decoder_config): def get_encoder_decoder_model(self, config, decoder_config):
encoder_model = TFViTModel(config, name="encoder") encoder_model = TFViTModel(config, name="encoder")
...@@ -572,10 +574,12 @@ class TFViT2GPT2EncoderDecoderModelTest(TFVisionEncoderDecoderMixin, unittest.Te ...@@ -572,10 +574,12 @@ class TFViT2GPT2EncoderDecoderModelTest(TFVisionEncoderDecoderMixin, unittest.Te
@require_tf @require_tf
class TFVisionEncoderDecoderModelTest(unittest.TestCase): class TFVisionEncoderDecoderModelTest(unittest.TestCase):
def get_from_encoderdecoder_pretrained_model(self): def get_from_encoderdecoder_pretrained_model(self):
return TFVisionEncoderDecoderModel.from_encoder_decoder_pretrained("google/vit-base-patch16-224-in21k", "gpt2") return TFVisionEncoderDecoderModel.from_encoder_decoder_pretrained(
"google/vit-base-patch16-224-in21k", "../gpt2"
)
def get_decoder_config(self): def get_decoder_config(self):
config = AutoConfig.from_pretrained("gpt2") config = AutoConfig.from_pretrained("../gpt2")
config.is_decoder = True config.is_decoder = True
config.add_cross_attention = True config.add_cross_attention = True
return config return config
...@@ -585,7 +589,7 @@ class TFVisionEncoderDecoderModelTest(unittest.TestCase): ...@@ -585,7 +589,7 @@ class TFVisionEncoderDecoderModelTest(unittest.TestCase):
def get_encoder_decoder_models(self): def get_encoder_decoder_models(self):
encoder_model = TFViTModel.from_pretrained("google/vit-base-patch16-224-in21k", name="encoder") encoder_model = TFViTModel.from_pretrained("google/vit-base-patch16-224-in21k", name="encoder")
decoder_model = TFGPT2LMHeadModel.from_pretrained("gpt2", config=self.get_decoder_config(), name="decoder") decoder_model = TFGPT2LMHeadModel.from_pretrained("../gpt2", config=self.get_decoder_config(), name="decoder")
return {"encoder": encoder_model, "decoder": decoder_model} return {"encoder": encoder_model, "decoder": decoder_model}
def _check_configuration_tie(self, model): def _check_configuration_tie(self, model):
...@@ -614,7 +618,7 @@ def prepare_img(): ...@@ -614,7 +618,7 @@ def prepare_img():
class TFVisionEncoderDecoderModelSaveLoadTests(unittest.TestCase): class TFVisionEncoderDecoderModelSaveLoadTests(unittest.TestCase):
def get_encoder_decoder_config(self): def get_encoder_decoder_config(self):
encoder_config = AutoConfig.from_pretrained("google/vit-base-patch16-224-in21k") encoder_config = AutoConfig.from_pretrained("google/vit-base-patch16-224-in21k")
decoder_config = AutoConfig.from_pretrained("gpt2", is_decoder=True, add_cross_attention=True) decoder_config = AutoConfig.from_pretrained("../gpt2", is_decoder=True, add_cross_attention=True)
return VisionEncoderDecoderConfig.from_encoder_decoder_configs(encoder_config, decoder_config) return VisionEncoderDecoderConfig.from_encoder_decoder_configs(encoder_config, decoder_config)
def get_encoder_decoder_config_small(self): def get_encoder_decoder_config_small(self):
...@@ -729,7 +733,7 @@ class TFVisionEncoderDecoderModelSaveLoadTests(unittest.TestCase): ...@@ -729,7 +733,7 @@ class TFVisionEncoderDecoderModelSaveLoadTests(unittest.TestCase):
config = self.get_encoder_decoder_config() config = self.get_encoder_decoder_config()
feature_extractor = AutoFeatureExtractor.from_pretrained("google/vit-base-patch16-224-in21k") feature_extractor = AutoFeatureExtractor.from_pretrained("google/vit-base-patch16-224-in21k")
decoder_tokenizer = AutoTokenizer.from_pretrained("gpt2") decoder_tokenizer = AutoTokenizer.from_pretrained("../gpt2")
img = prepare_img() img = prepare_img()
pixel_values = feature_extractor(images=img, return_tensors="tf").pixel_values pixel_values = feature_extractor(images=img, return_tensors="tf").pixel_values
...@@ -746,7 +750,7 @@ class TFVisionEncoderDecoderModelSaveLoadTests(unittest.TestCase): ...@@ -746,7 +750,7 @@ class TFVisionEncoderDecoderModelSaveLoadTests(unittest.TestCase):
encoder = TFAutoModel.from_pretrained("google/vit-base-patch16-224-in21k", name="encoder") encoder = TFAutoModel.from_pretrained("google/vit-base-patch16-224-in21k", name="encoder")
# It's necessary to specify `add_cross_attention=True` here. # It's necessary to specify `add_cross_attention=True` here.
decoder = TFAutoModelForCausalLM.from_pretrained( decoder = TFAutoModelForCausalLM.from_pretrained(
"gpt2", is_decoder=True, add_cross_attention=True, name="decoder" "../gpt2", is_decoder=True, add_cross_attention=True, name="decoder"
) )
pretrained_encoder_dir = os.path.join(tmp_dirname, "pretrained_encoder") pretrained_encoder_dir = os.path.join(tmp_dirname, "pretrained_encoder")
pretrained_decoder_dir = os.path.join(tmp_dirname, "pretrained_decoder") pretrained_decoder_dir = os.path.join(tmp_dirname, "pretrained_decoder")
......
...@@ -22,11 +22,11 @@ from datasets import load_dataset ...@@ -22,11 +22,11 @@ from datasets import load_dataset
from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.file_utils import cached_property, is_torch_available, is_vision_available
from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.testing_utils import require_torch, require_vision, slow, torch_device
from .test_modeling_bert import BertModelTester from ..bert.test_modeling_bert import BertModelTester
from .test_modeling_common import floats_tensor, ids_tensor, random_attention_mask from ..deit.test_modeling_deit import DeiTModelTester
from .test_modeling_deit import DeiTModelTester from ..test_modeling_common import floats_tensor, ids_tensor, random_attention_mask
from .test_modeling_trocr import TrOCRStandaloneDecoderModelTester from ..trocr.test_modeling_trocr import TrOCRStandaloneDecoderModelTester
from .test_modeling_vit import ViTModelTester from ..vit.test_modeling_vit import ViTModelTester
if is_torch_available(): if is_torch_available():
......
...@@ -31,10 +31,10 @@ from transformers.testing_utils import ( ...@@ -31,10 +31,10 @@ from transformers.testing_utils import (
torch_device, torch_device,
) )
from .test_modeling_flax_bert import FlaxBertModelTester from ..bert.test_modeling_flax_bert import FlaxBertModelTester
from .test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester
from .test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask
from .test_modeling_flax_vit import FlaxViTModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester
if is_flax_available(): if is_flax_available():
......
...@@ -24,12 +24,12 @@ import numpy as np ...@@ -24,12 +24,12 @@ import numpy as np
from transformers.file_utils import is_flax_available, is_torch_available, is_vision_available from transformers.file_utils import is_flax_available, is_torch_available, is_vision_available
from transformers.testing_utils import is_pt_flax_cross_test, require_torch, require_vision, slow, torch_device from transformers.testing_utils import is_pt_flax_cross_test, require_torch, require_vision, slow, torch_device
from .test_modeling_bert import BertModelTester from ..bert.test_modeling_bert import BertModelTester
from .test_modeling_clip import CLIPVisionModelTester from ..clip.test_modeling_clip import CLIPVisionModelTester
from .test_modeling_common import floats_tensor, ids_tensor, random_attention_mask from ..deit.test_modeling_deit import DeiTModelTester
from .test_modeling_deit import DeiTModelTester from ..roberta.test_modeling_roberta import RobertaModelTester
from .test_modeling_roberta import RobertaModelTester from ..test_modeling_common import floats_tensor, ids_tensor, random_attention_mask
from .test_modeling_vit import ViTModelTester from ..vit.test_modeling_vit import ViTModelTester
if is_torch_available(): if is_torch_available():
......
...@@ -17,12 +17,11 @@ ...@@ -17,12 +17,11 @@
import copy import copy
import unittest import unittest
from tests.test_modeling_common import floats_tensor
from transformers import VisualBertConfig, is_torch_available from transformers import VisualBertConfig, is_torch_available
from transformers.testing_utils import require_torch, slow, torch_device from transformers.testing_utils import require_torch, slow, torch_device
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, ids_tensor from ..test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor
if is_torch_available(): if is_torch_available():
......
...@@ -21,7 +21,7 @@ import numpy as np ...@@ -21,7 +21,7 @@ import numpy as np
from transformers.file_utils import is_torch_available, is_vision_available from transformers.file_utils import is_torch_available, is_vision_available
from transformers.testing_utils import require_torch, require_vision from transformers.testing_utils import require_torch, require_vision
from .test_feature_extraction_common import FeatureExtractionSavingTestMixin, prepare_image_inputs from ..test_feature_extraction_common import FeatureExtractionSavingTestMixin, prepare_image_inputs
if is_torch_available(): if is_torch_available():
......
...@@ -20,8 +20,8 @@ import numpy as np ...@@ -20,8 +20,8 @@ import numpy as np
from transformers import ViTConfig, is_flax_available from transformers import ViTConfig, is_flax_available
from transformers.testing_utils import require_flax, slow from transformers.testing_utils import require_flax, slow
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor from ..test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor
if is_flax_available(): if is_flax_available():
......
...@@ -24,8 +24,8 @@ from transformers import ViTConfig ...@@ -24,8 +24,8 @@ from transformers import ViTConfig
from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.file_utils import cached_property, is_tf_available, is_vision_available
from transformers.testing_utils import require_tf, require_vision, slow, tooslow from transformers.testing_utils import require_tf, require_vision, slow, tooslow
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ..test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor
if is_tf_available(): if is_tf_available():
......
...@@ -22,8 +22,8 @@ from transformers import ViTConfig ...@@ -22,8 +22,8 @@ from transformers import ViTConfig
from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.file_utils import cached_property, is_torch_available, is_vision_available
from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.testing_utils import require_torch, require_vision, slow, torch_device
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ..test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor
if is_torch_available(): if is_torch_available():
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment