Unverified Commit 29c10a41 authored by Lysandre Debut's avatar Lysandre Debut Committed by GitHub
Browse files

[Test refactor 1/5] Per-folder tests reorganization (#15725)



* Per-folder tests reorganization
Co-authored-by: default avatarsgugger <sylvain.gugger@gmail.com>
Co-authored-by: default avatarStas Bekman <stas@stason.org>
parent fecb08c2
...@@ -26,8 +26,8 @@ from transformers import ViTMAEConfig ...@@ -26,8 +26,8 @@ from transformers import ViTMAEConfig
from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.file_utils import cached_property, is_torch_available, is_vision_available
from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.testing_utils import require_torch, require_vision, slow, torch_device
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ..test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor
if is_torch_available(): if is_torch_available():
......
...@@ -23,7 +23,7 @@ import numpy as np ...@@ -23,7 +23,7 @@ import numpy as np
from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, Wav2Vec2Config, Wav2Vec2FeatureExtractor from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, Wav2Vec2Config, Wav2Vec2FeatureExtractor
from transformers.testing_utils import require_torch, slow from transformers.testing_utils import require_torch, slow
from .test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin from ..test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin
global_rng = random.Random() global_rng = random.Random()
......
...@@ -30,7 +30,7 @@ from transformers.testing_utils import ( ...@@ -30,7 +30,7 @@ from transformers.testing_utils import (
slow, slow,
) )
from .test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, random_attention_mask from ..test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, random_attention_mask
if is_flax_available(): if is_flax_available():
......
...@@ -29,8 +29,8 @@ from transformers import Wav2Vec2Config, is_tf_available ...@@ -29,8 +29,8 @@ from transformers import Wav2Vec2Config, is_tf_available
from transformers.file_utils import is_librosa_available, is_pyctcdecode_available from transformers.file_utils import is_librosa_available, is_pyctcdecode_available
from transformers.testing_utils import require_librosa, require_pyctcdecode, require_tf, slow from transformers.testing_utils import require_librosa, require_pyctcdecode, require_tf, slow
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ..test_modeling_tf_common import TFModelTesterMixin, ids_tensor
if is_tf_available(): if is_tf_available():
......
...@@ -20,7 +20,6 @@ import unittest ...@@ -20,7 +20,6 @@ import unittest
import numpy as np import numpy as np
from datasets import load_dataset from datasets import load_dataset
from tests.test_modeling_common import floats_tensor, ids_tensor, random_attention_mask
from transformers import Wav2Vec2Config, is_torch_available from transformers import Wav2Vec2Config, is_torch_available
from transformers.testing_utils import ( from transformers.testing_utils import (
is_pt_flax_cross_test, is_pt_flax_cross_test,
...@@ -34,8 +33,14 @@ from transformers.testing_utils import ( ...@@ -34,8 +33,14 @@ from transformers.testing_utils import (
torch_device, torch_device,
) )
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, _config_zero_init from ..test_modeling_common import (
ModelTesterMixin,
_config_zero_init,
floats_tensor,
ids_tensor,
random_attention_mask,
)
if is_torch_available(): if is_torch_available():
......
...@@ -32,7 +32,7 @@ from transformers import ( ...@@ -32,7 +32,7 @@ from transformers import (
from transformers.models.wav2vec2.tokenization_wav2vec2 import VOCAB_FILES_NAMES, Wav2Vec2CTCTokenizerOutput from transformers.models.wav2vec2.tokenization_wav2vec2 import VOCAB_FILES_NAMES, Wav2Vec2CTCTokenizerOutput
from transformers.testing_utils import require_torch, slow from transformers.testing_utils import require_torch, slow
from .test_tokenization_common import TokenizerTesterMixin from ..test_tokenization_common import TokenizerTesterMixin
global_rng = random.Random() global_rng = random.Random()
......
...@@ -23,7 +23,7 @@ from transformers.models.wav2vec2.tokenization_wav2vec2 import VOCAB_FILES_NAMES ...@@ -23,7 +23,7 @@ from transformers.models.wav2vec2.tokenization_wav2vec2 import VOCAB_FILES_NAMES
from transformers.models.wav2vec2_phoneme.tokenization_wav2vec2_phoneme import Wav2Vec2PhonemeCTCTokenizerOutput from transformers.models.wav2vec2_phoneme.tokenization_wav2vec2_phoneme import Wav2Vec2PhonemeCTCTokenizerOutput
from transformers.testing_utils import require_phonemizer from transformers.testing_utils import require_phonemizer
from .test_tokenization_common import TokenizerTesterMixin from ..test_tokenization_common import TokenizerTesterMixin
@require_phonemizer @require_phonemizer
......
...@@ -28,7 +28,7 @@ from transformers.models.wav2vec2 import Wav2Vec2CTCTokenizer, Wav2Vec2FeatureEx ...@@ -28,7 +28,7 @@ from transformers.models.wav2vec2 import Wav2Vec2CTCTokenizer, Wav2Vec2FeatureEx
from transformers.models.wav2vec2.tokenization_wav2vec2 import VOCAB_FILES_NAMES from transformers.models.wav2vec2.tokenization_wav2vec2 import VOCAB_FILES_NAMES
from transformers.testing_utils import require_pyctcdecode from transformers.testing_utils import require_pyctcdecode
from .test_feature_extraction_wav2vec2 import floats_list from ..wav2vec2.test_feature_extraction_wav2vec2 import floats_list
if is_pyctcdecode_available(): if is_pyctcdecode_available():
......
...@@ -20,12 +20,17 @@ import unittest ...@@ -20,12 +20,17 @@ import unittest
import pytest import pytest
from datasets import load_dataset from datasets import load_dataset
from tests.test_modeling_common import floats_tensor, ids_tensor, random_attention_mask
from transformers import WavLMConfig, is_torch_available from transformers import WavLMConfig, is_torch_available
from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device
from .test_configuration_common import ConfigTester from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, _config_zero_init from ..test_modeling_common import (
ModelTesterMixin,
_config_zero_init,
floats_tensor,
ids_tensor,
random_attention_mask,
)
if is_torch_available(): if is_torch_available():
......
...@@ -21,8 +21,8 @@ import transformers ...@@ -21,8 +21,8 @@ import transformers
from transformers import XGLMConfig, XGLMTokenizer, is_flax_available, is_torch_available from transformers import XGLMConfig, XGLMTokenizer, is_flax_available, is_torch_available
from transformers.testing_utils import is_pt_flax_cross_test, require_flax, require_sentencepiece, slow from transformers.testing_utils import is_pt_flax_cross_test, require_flax, require_sentencepiece, slow
from .test_generation_flax_utils import FlaxGenerationTesterMixin from ..generation.test_generation_flax_utils import FlaxGenerationTesterMixin
from .test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ..test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask
if is_flax_available(): if is_flax_available():
......
...@@ -21,9 +21,9 @@ import unittest ...@@ -21,9 +21,9 @@ import unittest
from transformers import XGLMConfig, is_torch_available from transformers import XGLMConfig, is_torch_available
from transformers.testing_utils import require_torch, slow, torch_device from transformers.testing_utils import require_torch, slow, torch_device
from .test_configuration_common import ConfigTester from ..generation.test_generation_utils import GenerationTesterMixin
from .test_generation_utils import GenerationTesterMixin from ..test_configuration_common import ConfigTester
from .test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ..test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask
if is_torch_available(): if is_torch_available():
......
...@@ -23,10 +23,10 @@ from transformers import SPIECE_UNDERLINE, XGLMTokenizer, XGLMTokenizerFast ...@@ -23,10 +23,10 @@ from transformers import SPIECE_UNDERLINE, XGLMTokenizer, XGLMTokenizerFast
from transformers.file_utils import cached_property from transformers.file_utils import cached_property
from transformers.testing_utils import require_sentencepiece, require_tokenizers, slow from transformers.testing_utils import require_sentencepiece, require_tokenizers, slow
from .test_tokenization_common import TokenizerTesterMixin from ..test_tokenization_common import TokenizerTesterMixin
SAMPLE_VOCAB = os.path.join(os.path.dirname(os.path.abspath(__file__)), "fixtures/test_sentencepiece.model") SAMPLE_VOCAB = os.path.join(os.path.dirname(os.path.abspath(__file__)), "../fixtures/test_sentencepiece.model")
@require_sentencepiece @require_sentencepiece
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment