"server/vscode:/vscode.git/clone" did not exist on "853d4eb9cf51fced975a428de15428fb4860a449"
Unverified Commit 7aa6af11 authored by Dhruv Nair's avatar Dhruv Nair Committed by GitHub
Browse files

[Refactor] Move testing utils out of src (#12238)

* update

* update

* update

* update

* update

* merge main

* Revert "merge main"

This reverts commit 65efbcead58644b31596ed2d714f7cee0e0238d3.
parent 87b800e1
...@@ -32,7 +32,8 @@ from diffusers import ( ...@@ -32,7 +32,8 @@ from diffusers import (
StableDiffusionPipeline, StableDiffusionPipeline,
) )
from diffusers.utils.import_utils import is_accelerate_available from diffusers.utils.import_utils import is_accelerate_available
from diffusers.utils.testing_utils import (
from ..testing_utils import (
Expectations, Expectations,
backend_empty_cache, backend_empty_cache,
load_image, load_image,
...@@ -47,7 +48,7 @@ from diffusers.utils.testing_utils import ( ...@@ -47,7 +48,7 @@ from diffusers.utils.testing_utils import (
sys.path.append(".") sys.path.append(".")
from utils import PeftLoraLoaderMixinTests, check_if_lora_correctly_set # noqa: E402 from .utils import PeftLoraLoaderMixinTests, check_if_lora_correctly_set # noqa: E402
if is_accelerate_available(): if is_accelerate_available():
......
...@@ -28,7 +28,8 @@ from diffusers import ( ...@@ -28,7 +28,8 @@ from diffusers import (
) )
from diffusers.utils import load_image from diffusers.utils import load_image
from diffusers.utils.import_utils import is_accelerate_available from diffusers.utils.import_utils import is_accelerate_available
from diffusers.utils.testing_utils import (
from ..testing_utils import (
backend_empty_cache, backend_empty_cache,
is_flaky, is_flaky,
nightly, nightly,
...@@ -42,7 +43,7 @@ from diffusers.utils.testing_utils import ( ...@@ -42,7 +43,7 @@ from diffusers.utils.testing_utils import (
sys.path.append(".") sys.path.append(".")
from utils import PeftLoraLoaderMixinTests # noqa: E402 from .utils import PeftLoraLoaderMixinTests # noqa: E402
if is_accelerate_available(): if is_accelerate_available():
......
...@@ -35,7 +35,8 @@ from diffusers import ( ...@@ -35,7 +35,8 @@ from diffusers import (
) )
from diffusers.utils import logging from diffusers.utils import logging
from diffusers.utils.import_utils import is_accelerate_available from diffusers.utils.import_utils import is_accelerate_available
from diffusers.utils.testing_utils import (
from ..testing_utils import (
CaptureLogger, CaptureLogger,
backend_empty_cache, backend_empty_cache,
is_flaky, is_flaky,
...@@ -51,7 +52,7 @@ from diffusers.utils.testing_utils import ( ...@@ -51,7 +52,7 @@ from diffusers.utils.testing_utils import (
sys.path.append(".") sys.path.append(".")
from utils import PeftLoraLoaderMixinTests, check_if_lora_correctly_set, state_dicts_almost_equal # noqa: E402 from .utils import PeftLoraLoaderMixinTests, check_if_lora_correctly_set, state_dicts_almost_equal # noqa: E402
if is_accelerate_available(): if is_accelerate_available():
......
...@@ -24,7 +24,8 @@ from diffusers import ( ...@@ -24,7 +24,8 @@ from diffusers import (
WanPipeline, WanPipeline,
WanTransformer3DModel, WanTransformer3DModel,
) )
from diffusers.utils.testing_utils import (
from ..testing_utils import (
floats_tensor, floats_tensor,
require_peft_backend, require_peft_backend,
skip_mps, skip_mps,
...@@ -33,7 +34,7 @@ from diffusers.utils.testing_utils import ( ...@@ -33,7 +34,7 @@ from diffusers.utils.testing_utils import (
sys.path.append(".") sys.path.append(".")
from utils import PeftLoraLoaderMixinTests # noqa: E402 from .utils import PeftLoraLoaderMixinTests # noqa: E402
@require_peft_backend @require_peft_backend
......
...@@ -25,7 +25,8 @@ from transformers import AutoTokenizer, T5EncoderModel ...@@ -25,7 +25,8 @@ from transformers import AutoTokenizer, T5EncoderModel
from diffusers import AutoencoderKLWan, FlowMatchEulerDiscreteScheduler, WanVACEPipeline, WanVACETransformer3DModel from diffusers import AutoencoderKLWan, FlowMatchEulerDiscreteScheduler, WanVACEPipeline, WanVACETransformer3DModel
from diffusers.utils.import_utils import is_peft_available from diffusers.utils.import_utils import is_peft_available
from diffusers.utils.testing_utils import (
from ..testing_utils import (
floats_tensor, floats_tensor,
is_flaky, is_flaky,
require_peft_backend, require_peft_backend,
...@@ -40,7 +41,7 @@ if is_peft_available(): ...@@ -40,7 +41,7 @@ if is_peft_available():
sys.path.append(".") sys.path.append(".")
from utils import PeftLoraLoaderMixinTests # noqa: E402 from .utils import PeftLoraLoaderMixinTests # noqa: E402
@require_peft_backend @require_peft_backend
......
...@@ -32,7 +32,8 @@ from diffusers import ( ...@@ -32,7 +32,8 @@ from diffusers import (
) )
from diffusers.utils import logging from diffusers.utils import logging
from diffusers.utils.import_utils import is_peft_available from diffusers.utils.import_utils import is_peft_available
from diffusers.utils.testing_utils import (
from ..testing_utils import (
CaptureLogger, CaptureLogger,
check_if_dicts_are_equal, check_if_dicts_are_equal,
floats_tensor, floats_tensor,
......
...@@ -21,7 +21,8 @@ from parameterized import parameterized ...@@ -21,7 +21,8 @@ from parameterized import parameterized
from diffusers import AsymmetricAutoencoderKL from diffusers import AsymmetricAutoencoderKL
from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.import_utils import is_xformers_available
from diffusers.utils.testing_utils import (
from ...testing_utils import (
Expectations, Expectations,
backend_empty_cache, backend_empty_cache,
enable_full_determinism, enable_full_determinism,
...@@ -34,7 +35,6 @@ from diffusers.utils.testing_utils import ( ...@@ -34,7 +35,6 @@ from diffusers.utils.testing_utils import (
torch_all_close, torch_all_close,
torch_device, torch_device,
) )
from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin
......
...@@ -15,8 +15,8 @@ ...@@ -15,8 +15,8 @@
import unittest import unittest
from diffusers import AutoencoderKLCosmos from diffusers import AutoencoderKLCosmos
from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, torch_device
from ...testing_utils import enable_full_determinism, floats_tensor, torch_device
from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin
......
...@@ -16,12 +16,12 @@ ...@@ -16,12 +16,12 @@
import unittest import unittest
from diffusers import AutoencoderDC from diffusers import AutoencoderDC
from diffusers.utils.testing_utils import (
from ...testing_utils import (
enable_full_determinism, enable_full_determinism,
floats_tensor, floats_tensor,
torch_device, torch_device,
) )
from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin
......
...@@ -19,12 +19,12 @@ import torch ...@@ -19,12 +19,12 @@ import torch
from diffusers import AutoencoderKLHunyuanVideo from diffusers import AutoencoderKLHunyuanVideo
from diffusers.models.autoencoders.autoencoder_kl_hunyuan_video import prepare_causal_attention_mask from diffusers.models.autoencoders.autoencoder_kl_hunyuan_video import prepare_causal_attention_mask
from diffusers.utils.testing_utils import (
from ...testing_utils import (
enable_full_determinism, enable_full_determinism,
floats_tensor, floats_tensor,
torch_device, torch_device,
) )
from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin
......
...@@ -21,7 +21,8 @@ from parameterized import parameterized ...@@ -21,7 +21,8 @@ from parameterized import parameterized
from diffusers import AutoencoderKL from diffusers import AutoencoderKL
from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.import_utils import is_xformers_available
from diffusers.utils.testing_utils import (
from ...testing_utils import (
backend_empty_cache, backend_empty_cache,
enable_full_determinism, enable_full_determinism,
floats_tensor, floats_tensor,
...@@ -34,7 +35,6 @@ from diffusers.utils.testing_utils import ( ...@@ -34,7 +35,6 @@ from diffusers.utils.testing_utils import (
torch_all_close, torch_all_close,
torch_device, torch_device,
) )
from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin
......
...@@ -18,12 +18,12 @@ import unittest ...@@ -18,12 +18,12 @@ import unittest
import torch import torch
from diffusers import AutoencoderKLCogVideoX from diffusers import AutoencoderKLCogVideoX
from diffusers.utils.testing_utils import (
from ...testing_utils import (
enable_full_determinism, enable_full_determinism,
floats_tensor, floats_tensor,
torch_device, torch_device,
) )
from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin
......
...@@ -16,12 +16,12 @@ ...@@ -16,12 +16,12 @@
import unittest import unittest
from diffusers import AutoencoderKLTemporalDecoder from diffusers import AutoencoderKLTemporalDecoder
from diffusers.utils.testing_utils import (
from ...testing_utils import (
enable_full_determinism, enable_full_determinism,
floats_tensor, floats_tensor,
torch_device, torch_device,
) )
from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin
......
...@@ -18,12 +18,12 @@ import unittest ...@@ -18,12 +18,12 @@ import unittest
import torch import torch
from diffusers import AutoencoderKLLTXVideo from diffusers import AutoencoderKLLTXVideo
from diffusers.utils.testing_utils import (
from ...testing_utils import (
enable_full_determinism, enable_full_determinism,
floats_tensor, floats_tensor,
torch_device, torch_device,
) )
from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin
......
...@@ -16,8 +16,8 @@ ...@@ -16,8 +16,8 @@
import unittest import unittest
from diffusers import AutoencoderKLMagvit from diffusers import AutoencoderKLMagvit
from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, torch_device
from ...testing_utils import enable_full_determinism, floats_tensor, torch_device
from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin
......
...@@ -16,12 +16,12 @@ ...@@ -16,12 +16,12 @@
import unittest import unittest
from diffusers import AutoencoderKLMochi from diffusers import AutoencoderKLMochi
from diffusers.utils.testing_utils import (
from ...testing_utils import (
enable_full_determinism, enable_full_determinism,
floats_tensor, floats_tensor,
torch_device, torch_device,
) )
from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin
......
...@@ -21,7 +21,8 @@ from datasets import load_dataset ...@@ -21,7 +21,8 @@ from datasets import load_dataset
from parameterized import parameterized from parameterized import parameterized
from diffusers import AutoencoderOobleck from diffusers import AutoencoderOobleck
from diffusers.utils.testing_utils import (
from ...testing_utils import (
backend_empty_cache, backend_empty_cache,
enable_full_determinism, enable_full_determinism,
floats_tensor, floats_tensor,
...@@ -29,7 +30,6 @@ from diffusers.utils.testing_utils import ( ...@@ -29,7 +30,6 @@ from diffusers.utils.testing_utils import (
torch_all_close, torch_all_close,
torch_device, torch_device,
) )
from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin
......
...@@ -21,7 +21,8 @@ import torch ...@@ -21,7 +21,8 @@ import torch
from parameterized import parameterized from parameterized import parameterized
from diffusers import AutoencoderTiny from diffusers import AutoencoderTiny
from diffusers.utils.testing_utils import (
from ...testing_utils import (
backend_empty_cache, backend_empty_cache,
enable_full_determinism, enable_full_determinism,
floats_tensor, floats_tensor,
...@@ -30,7 +31,6 @@ from diffusers.utils.testing_utils import ( ...@@ -30,7 +31,6 @@ from diffusers.utils.testing_utils import (
torch_all_close, torch_all_close,
torch_device, torch_device,
) )
from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin
......
...@@ -18,8 +18,8 @@ import unittest ...@@ -18,8 +18,8 @@ import unittest
import torch import torch
from diffusers import AutoencoderKLWan from diffusers import AutoencoderKLWan
from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, torch_device
from ...testing_utils import enable_full_determinism, floats_tensor, torch_device
from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin from ..test_modeling_common import ModelTesterMixin, UNetTesterMixin
......
...@@ -20,7 +20,9 @@ import numpy as np ...@@ -20,7 +20,9 @@ import numpy as np
import torch import torch
from diffusers import ConsistencyDecoderVAE, StableDiffusionPipeline from diffusers import ConsistencyDecoderVAE, StableDiffusionPipeline
from diffusers.utils.testing_utils import ( from diffusers.utils.torch_utils import randn_tensor
from ...testing_utils import (
backend_empty_cache, backend_empty_cache,
enable_full_determinism, enable_full_determinism,
load_image, load_image,
...@@ -28,8 +30,6 @@ from diffusers.utils.testing_utils import ( ...@@ -28,8 +30,6 @@ from diffusers.utils.testing_utils import (
torch_all_close, torch_all_close,
torch_device, torch_device,
) )
from diffusers.utils.torch_utils import randn_tensor
from ..test_modeling_common import ModelTesterMixin from ..test_modeling_common import ModelTesterMixin
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment