Unverified Commit a7ca03aa authored by Patrick von Platen's avatar Patrick von Platen Committed by GitHub
Browse files

Replace flake8 with ruff and update black (#2279)

* before running make style

* remove left overs from flake8

* finish

* make fix-copies

* final fix

* more fixes
parent f5ccffec
...@@ -15,7 +15,6 @@ ...@@ -15,7 +15,6 @@
from typing import Callable, List, Optional, Tuple, Union from typing import Callable, List, Optional, Tuple, Union
import torch import torch
from transformers import CLIPTextModel, CLIPTokenizer from transformers import CLIPTextModel, CLIPTokenizer
from ...configuration_utils import ConfigMixin, register_to_config from ...configuration_utils import ConfigMixin, register_to_config
......
...@@ -17,7 +17,6 @@ from typing import List, Optional, Tuple, Union ...@@ -17,7 +17,6 @@ from typing import List, Optional, Tuple, Union
import numpy as np import numpy as np
import torch import torch
from scipy import integrate from scipy import integrate
from ..configuration_utils import ConfigMixin, register_to_config from ..configuration_utils import ConfigMixin, register_to_config
......
# This file is autogenerated by the command `make fix-copies`, do not edit. # This file is autogenerated by the command `make fix-copies`, do not edit.
# flake8: noqa
from ..utils import DummyObject, requires_backends from ..utils import DummyObject, requires_backends
......
# This file is autogenerated by the command `make fix-copies`, do not edit. # This file is autogenerated by the command `make fix-copies`, do not edit.
# flake8: noqa
from ..utils import DummyObject, requires_backends from ..utils import DummyObject, requires_backends
......
# This file is autogenerated by the command `make fix-copies`, do not edit. # This file is autogenerated by the command `make fix-copies`, do not edit.
# flake8: noqa
from ..utils import DummyObject, requires_backends from ..utils import DummyObject, requires_backends
......
# This file is autogenerated by the command `make fix-copies`, do not edit. # This file is autogenerated by the command `make fix-copies`, do not edit.
# flake8: noqa
from ..utils import DummyObject, requires_backends from ..utils import DummyObject, requires_backends
......
# This file is autogenerated by the command `make fix-copies`, do not edit. # This file is autogenerated by the command `make fix-copies`, do not edit.
# flake8: noqa
from ..utils import DummyObject, requires_backends from ..utils import DummyObject, requires_backends
......
# This file is autogenerated by the command `make fix-copies`, do not edit. # This file is autogenerated by the command `make fix-copies`, do not edit.
# flake8: noqa
from ..utils import DummyObject, requires_backends from ..utils import DummyObject, requires_backends
......
# This file is autogenerated by the command `make fix-copies`, do not edit. # This file is autogenerated by the command `make fix-copies`, do not edit.
# flake8: noqa
from ..utils import DummyObject, requires_backends from ..utils import DummyObject, requires_backends
......
# This file is autogenerated by the command `make fix-copies`, do not edit. # This file is autogenerated by the command `make fix-copies`, do not edit.
# flake8: noqa
from ..utils import DummyObject, requires_backends from ..utils import DummyObject, requires_backends
......
# This file is autogenerated by the command `make fix-copies`, do not edit. # This file is autogenerated by the command `make fix-copies`, do not edit.
# flake8: noqa
from ..utils import DummyObject, requires_backends from ..utils import DummyObject, requires_backends
......
...@@ -18,14 +18,16 @@ import logging ...@@ -18,14 +18,16 @@ import logging
import os import os
import sys import sys
import threading import threading
from logging import CRITICAL # NOQA from logging import (
from logging import DEBUG # NOQA CRITICAL, # NOQA
from logging import ERROR # NOQA DEBUG, # NOQA
from logging import FATAL # NOQA ERROR, # NOQA
from logging import INFO # NOQA FATAL, # NOQA
from logging import NOTSET # NOQA INFO, # NOQA
from logging import WARN # NOQA NOTSET, # NOQA
from logging import WARNING # NOQA WARN, # NOQA
WARNING, # NOQA
)
from typing import Optional from typing import Optional
from tqdm import auto as tqdm_lib from tqdm import auto as tqdm_lib
......
...@@ -11,7 +11,6 @@ from pathlib import Path ...@@ -11,7 +11,6 @@ from pathlib import Path
from typing import Optional, Union from typing import Optional, Union
import numpy as np import numpy as np
import PIL.Image import PIL.Image
import PIL.ImageOps import PIL.ImageOps
import requests import requests
......
...@@ -18,6 +18,7 @@ import tempfile ...@@ -18,6 +18,7 @@ import tempfile
import unittest import unittest
import torch import torch
from parameterized import parameterized
from diffusers import UNet2DConditionModel from diffusers import UNet2DConditionModel
from diffusers.models.cross_attention import CrossAttnProcessor, LoRACrossAttnProcessor from diffusers.models.cross_attention import CrossAttnProcessor, LoRACrossAttnProcessor
...@@ -31,7 +32,6 @@ from diffusers.utils import ( ...@@ -31,7 +32,6 @@ from diffusers.utils import (
torch_device, torch_device,
) )
from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.import_utils import is_xformers_available
from parameterized import parameterized
from ..test_modeling_common import ModelTesterMixin from ..test_modeling_common import ModelTesterMixin
......
import gc import gc
import unittest import unittest
from parameterized import parameterized
from diffusers import FlaxUNet2DConditionModel from diffusers import FlaxUNet2DConditionModel
from diffusers.utils import is_flax_available from diffusers.utils import is_flax_available
from diffusers.utils.testing_utils import load_hf_numpy, require_flax, slow from diffusers.utils.testing_utils import load_hf_numpy, require_flax, slow
from parameterized import parameterized
if is_flax_available(): if is_flax_available():
......
...@@ -17,11 +17,11 @@ import gc ...@@ -17,11 +17,11 @@ import gc
import unittest import unittest
import torch import torch
from parameterized import parameterized
from diffusers import AutoencoderKL from diffusers import AutoencoderKL
from diffusers.models import ModelMixin from diffusers.models import ModelMixin
from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device
from parameterized import parameterized
from ..test_modeling_common import ModelTesterMixin from ..test_modeling_common import ModelTesterMixin
......
...@@ -18,6 +18,7 @@ import unittest ...@@ -18,6 +18,7 @@ import unittest
import numpy as np import numpy as np
import torch import torch
from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer
from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNet2DConditionModel from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNet2DConditionModel
from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( from diffusers.pipelines.alt_diffusion.modeling_roberta_series import (
...@@ -26,7 +27,6 @@ from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( ...@@ -26,7 +27,6 @@ from diffusers.pipelines.alt_diffusion.modeling_roberta_series import (
) )
from diffusers.utils import slow, torch_device from diffusers.utils import slow, torch_device
from diffusers.utils.testing_utils import require_torch_gpu from diffusers.utils.testing_utils import require_torch_gpu
from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer
from ...test_pipelines_common import PipelineTesterMixin from ...test_pipelines_common import PipelineTesterMixin
......
...@@ -19,6 +19,7 @@ import unittest ...@@ -19,6 +19,7 @@ import unittest
import numpy as np import numpy as np
import torch import torch
from transformers import XLMRobertaTokenizer
from diffusers import AltDiffusionImg2ImgPipeline, AutoencoderKL, PNDMScheduler, UNet2DConditionModel from diffusers import AltDiffusionImg2ImgPipeline, AutoencoderKL, PNDMScheduler, UNet2DConditionModel
from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( from diffusers.pipelines.alt_diffusion.modeling_roberta_series import (
...@@ -27,7 +28,6 @@ from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( ...@@ -27,7 +28,6 @@ from diffusers.pipelines.alt_diffusion.modeling_roberta_series import (
) )
from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device
from diffusers.utils.testing_utils import require_torch_gpu from diffusers.utils.testing_utils import require_torch_gpu
from transformers import XLMRobertaTokenizer
torch.backends.cuda.matmul.allow_tf32 = False torch.backends.cuda.matmul.allow_tf32 = False
......
...@@ -18,10 +18,10 @@ import unittest ...@@ -18,10 +18,10 @@ import unittest
import numpy as np import numpy as np
import torch import torch
from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer
from diffusers import AutoencoderKL, DDIMScheduler, LDMTextToImagePipeline, UNet2DConditionModel from diffusers import AutoencoderKL, DDIMScheduler, LDMTextToImagePipeline, UNet2DConditionModel
from diffusers.utils.testing_utils import load_numpy, nightly, require_torch_gpu, slow, torch_device from diffusers.utils.testing_utils import load_numpy, nightly, require_torch_gpu, slow, torch_device
from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer
from ...test_pipelines_common import PipelineTesterMixin from ...test_pipelines_common import PipelineTesterMixin
......
...@@ -17,10 +17,10 @@ import unittest ...@@ -17,10 +17,10 @@ import unittest
import numpy as np import numpy as np
import torch import torch
from transformers import CLIPTextConfig, CLIPTextModel
from diffusers import DDIMScheduler, LDMPipeline, UNet2DModel, VQModel from diffusers import DDIMScheduler, LDMPipeline, UNet2DModel, VQModel
from diffusers.utils.testing_utils import require_torch, slow, torch_device from diffusers.utils.testing_utils import require_torch, slow, torch_device
from transformers import CLIPTextConfig, CLIPTextModel
torch.backends.cuda.matmul.allow_tf32 = False torch.backends.cuda.matmul.allow_tf32 = False
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment