Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
renzhc
diffusers_dcu
Commits
aa1f00fd
Unverified
Commit
aa1f00fd
authored
Apr 10, 2024
by
YiYi Xu
Committed by
GitHub
Apr 10, 2024
Browse files
Fix cpu offload related slow tests (#7618)
* fix * up --------- Co-authored-by:
yiyixuxu
<
yixu310@gmail,com
>
parent
d95b9934
Changes
10
Hide whitespace changes
Inline
Side-by-side
Showing
10 changed files
with
12 additions
and
145 deletions
+12
-145
src/diffusers/pipelines/deepfloyd_if/pipeline_if.py
src/diffusers/pipelines/deepfloyd_if/pipeline_if.py
+1
-18
src/diffusers/pipelines/deepfloyd_if/pipeline_if_img2img.py
src/diffusers/pipelines/deepfloyd_if/pipeline_if_img2img.py
+1
-19
src/diffusers/pipelines/deepfloyd_if/pipeline_if_img2img_superresolution.py
...lines/deepfloyd_if/pipeline_if_img2img_superresolution.py
+1
-21
src/diffusers/pipelines/deepfloyd_if/pipeline_if_inpainting.py
...iffusers/pipelines/deepfloyd_if/pipeline_if_inpainting.py
+1
-19
src/diffusers/pipelines/deepfloyd_if/pipeline_if_inpainting_superresolution.py
...es/deepfloyd_if/pipeline_if_inpainting_superresolution.py
+1
-19
src/diffusers/pipelines/deepfloyd_if/pipeline_if_superresolution.py
...ers/pipelines/deepfloyd_if/pipeline_if_superresolution.py
+1
-19
src/diffusers/pipelines/kandinsky/pipeline_kandinsky_combined.py
...fusers/pipelines/kandinsky/pipeline_kandinsky_combined.py
+3
-0
src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_combined.py
.../pipelines/kandinsky2_2/pipeline_kandinsky2_2_combined.py
+3
-0
src/diffusers/pipelines/kandinsky3/pipeline_kandinsky3.py
src/diffusers/pipelines/kandinsky3/pipeline_kandinsky3.py
+0
-15
src/diffusers/pipelines/kandinsky3/pipeline_kandinsky3_img2img.py
...users/pipelines/kandinsky3/pipeline_kandinsky3_img2img.py
+0
-15
No files found.
src/diffusers/pipelines/deepfloyd_if/pipeline_if.py
View file @
aa1f00fd
...
@@ -12,7 +12,6 @@ from ...models import UNet2DConditionModel
...
@@ -12,7 +12,6 @@ from ...models import UNet2DConditionModel
from
...schedulers
import
DDPMScheduler
from
...schedulers
import
DDPMScheduler
from
...utils
import
(
from
...utils
import
(
BACKENDS_MAPPING
,
BACKENDS_MAPPING
,
is_accelerate_available
,
is_bs4_available
,
is_bs4_available
,
is_ftfy_available
,
is_ftfy_available
,
logging
,
logging
,
...
@@ -115,6 +114,7 @@ class IFPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -115,6 +114,7 @@ class IFPipeline(DiffusionPipeline, LoraLoaderMixin):
_optional_components
=
[
"tokenizer"
,
"text_encoder"
,
"safety_checker"
,
"feature_extractor"
,
"watermarker"
]
_optional_components
=
[
"tokenizer"
,
"text_encoder"
,
"safety_checker"
,
"feature_extractor"
,
"watermarker"
]
model_cpu_offload_seq
=
"text_encoder->unet"
model_cpu_offload_seq
=
"text_encoder->unet"
_exclude_from_cpu_offload
=
[
"watermarker"
]
def
__init__
(
def
__init__
(
self
,
self
,
...
@@ -156,20 +156,6 @@ class IFPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -156,20 +156,6 @@ class IFPipeline(DiffusionPipeline, LoraLoaderMixin):
)
)
self
.
register_to_config
(
requires_safety_checker
=
requires_safety_checker
)
self
.
register_to_config
(
requires_safety_checker
=
requires_safety_checker
)
def
remove_all_hooks
(
self
):
if
is_accelerate_available
():
from
accelerate.hooks
import
remove_hook_from_module
else
:
raise
ImportError
(
"Please install accelerate via `pip install accelerate`"
)
for
model
in
[
self
.
text_encoder
,
self
.
unet
,
self
.
safety_checker
]:
if
model
is
not
None
:
remove_hook_from_module
(
model
,
recurse
=
True
)
self
.
unet_offload_hook
=
None
self
.
text_encoder_offload_hook
=
None
self
.
final_offload_hook
=
None
@
torch
.
no_grad
()
@
torch
.
no_grad
()
def
encode_prompt
(
def
encode_prompt
(
self
,
self
,
...
@@ -335,9 +321,6 @@ class IFPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -335,9 +321,6 @@ class IFPipeline(DiffusionPipeline, LoraLoaderMixin):
nsfw_detected
=
None
nsfw_detected
=
None
watermark_detected
=
None
watermark_detected
=
None
if
hasattr
(
self
,
"unet_offload_hook"
)
and
self
.
unet_offload_hook
is
not
None
:
self
.
unet_offload_hook
.
offload
()
return
image
,
nsfw_detected
,
watermark_detected
return
image
,
nsfw_detected
,
watermark_detected
# Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline.prepare_extra_step_kwargs
# Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline.prepare_extra_step_kwargs
...
...
src/diffusers/pipelines/deepfloyd_if/pipeline_if_img2img.py
View file @
aa1f00fd
...
@@ -15,7 +15,6 @@ from ...schedulers import DDPMScheduler
...
@@ -15,7 +15,6 @@ from ...schedulers import DDPMScheduler
from
...utils
import
(
from
...utils
import
(
BACKENDS_MAPPING
,
BACKENDS_MAPPING
,
PIL_INTERPOLATION
,
PIL_INTERPOLATION
,
is_accelerate_available
,
is_bs4_available
,
is_bs4_available
,
is_ftfy_available
,
is_ftfy_available
,
logging
,
logging
,
...
@@ -139,6 +138,7 @@ class IFImg2ImgPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -139,6 +138,7 @@ class IFImg2ImgPipeline(DiffusionPipeline, LoraLoaderMixin):
_optional_components
=
[
"tokenizer"
,
"text_encoder"
,
"safety_checker"
,
"feature_extractor"
,
"watermarker"
]
_optional_components
=
[
"tokenizer"
,
"text_encoder"
,
"safety_checker"
,
"feature_extractor"
,
"watermarker"
]
model_cpu_offload_seq
=
"text_encoder->unet"
model_cpu_offload_seq
=
"text_encoder->unet"
_exclude_from_cpu_offload
=
[
"watermarker"
]
def
__init__
(
def
__init__
(
self
,
self
,
...
@@ -180,21 +180,6 @@ class IFImg2ImgPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -180,21 +180,6 @@ class IFImg2ImgPipeline(DiffusionPipeline, LoraLoaderMixin):
)
)
self
.
register_to_config
(
requires_safety_checker
=
requires_safety_checker
)
self
.
register_to_config
(
requires_safety_checker
=
requires_safety_checker
)
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.remove_all_hooks
def
remove_all_hooks
(
self
):
if
is_accelerate_available
():
from
accelerate.hooks
import
remove_hook_from_module
else
:
raise
ImportError
(
"Please install accelerate via `pip install accelerate`"
)
for
model
in
[
self
.
text_encoder
,
self
.
unet
,
self
.
safety_checker
]:
if
model
is
not
None
:
remove_hook_from_module
(
model
,
recurse
=
True
)
self
.
unet_offload_hook
=
None
self
.
text_encoder_offload_hook
=
None
self
.
final_offload_hook
=
None
@
torch
.
no_grad
()
@
torch
.
no_grad
()
def
encode_prompt
(
def
encode_prompt
(
self
,
self
,
...
@@ -361,9 +346,6 @@ class IFImg2ImgPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -361,9 +346,6 @@ class IFImg2ImgPipeline(DiffusionPipeline, LoraLoaderMixin):
nsfw_detected
=
None
nsfw_detected
=
None
watermark_detected
=
None
watermark_detected
=
None
if
hasattr
(
self
,
"unet_offload_hook"
)
and
self
.
unet_offload_hook
is
not
None
:
self
.
unet_offload_hook
.
offload
()
return
image
,
nsfw_detected
,
watermark_detected
return
image
,
nsfw_detected
,
watermark_detected
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.prepare_extra_step_kwargs
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.prepare_extra_step_kwargs
...
...
src/diffusers/pipelines/deepfloyd_if/pipeline_if_img2img_superresolution.py
View file @
aa1f00fd
...
@@ -16,7 +16,6 @@ from ...schedulers import DDPMScheduler
...
@@ -16,7 +16,6 @@ from ...schedulers import DDPMScheduler
from
...utils
import
(
from
...utils
import
(
BACKENDS_MAPPING
,
BACKENDS_MAPPING
,
PIL_INTERPOLATION
,
PIL_INTERPOLATION
,
is_accelerate_available
,
is_bs4_available
,
is_bs4_available
,
is_ftfy_available
,
is_ftfy_available
,
logging
,
logging
,
...
@@ -143,6 +142,7 @@ class IFImg2ImgSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -143,6 +142,7 @@ class IFImg2ImgSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
_optional_components
=
[
"tokenizer"
,
"text_encoder"
,
"safety_checker"
,
"feature_extractor"
]
_optional_components
=
[
"tokenizer"
,
"text_encoder"
,
"safety_checker"
,
"feature_extractor"
]
model_cpu_offload_seq
=
"text_encoder->unet"
model_cpu_offload_seq
=
"text_encoder->unet"
_exclude_from_cpu_offload
=
[
"watermarker"
]
def
__init__
(
def
__init__
(
self
,
self
,
...
@@ -191,21 +191,6 @@ class IFImg2ImgSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -191,21 +191,6 @@ class IFImg2ImgSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
)
)
self
.
register_to_config
(
requires_safety_checker
=
requires_safety_checker
)
self
.
register_to_config
(
requires_safety_checker
=
requires_safety_checker
)
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.remove_all_hooks
def
remove_all_hooks
(
self
):
if
is_accelerate_available
():
from
accelerate.hooks
import
remove_hook_from_module
else
:
raise
ImportError
(
"Please install accelerate via `pip install accelerate`"
)
for
model
in
[
self
.
text_encoder
,
self
.
unet
,
self
.
safety_checker
]:
if
model
is
not
None
:
remove_hook_from_module
(
model
,
recurse
=
True
)
self
.
unet_offload_hook
=
None
self
.
text_encoder_offload_hook
=
None
self
.
final_offload_hook
=
None
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline._text_preprocessing
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline._text_preprocessing
def
_text_preprocessing
(
self
,
text
,
clean_caption
=
False
):
def
_text_preprocessing
(
self
,
text
,
clean_caption
=
False
):
if
clean_caption
and
not
is_bs4_available
():
if
clean_caption
and
not
is_bs4_available
():
...
@@ -513,9 +498,6 @@ class IFImg2ImgSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -513,9 +498,6 @@ class IFImg2ImgSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
nsfw_detected
=
None
nsfw_detected
=
None
watermark_detected
=
None
watermark_detected
=
None
if
hasattr
(
self
,
"unet_offload_hook"
)
and
self
.
unet_offload_hook
is
not
None
:
self
.
unet_offload_hook
.
offload
()
return
image
,
nsfw_detected
,
watermark_detected
return
image
,
nsfw_detected
,
watermark_detected
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.prepare_extra_step_kwargs
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.prepare_extra_step_kwargs
...
@@ -1012,8 +994,6 @@ class IFImg2ImgSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -1012,8 +994,6 @@ class IFImg2ImgSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
nsfw_detected
=
None
nsfw_detected
=
None
watermark_detected
=
None
watermark_detected
=
None
if
hasattr
(
self
,
"unet_offload_hook"
)
and
self
.
unet_offload_hook
is
not
None
:
self
.
unet_offload_hook
.
offload
()
else
:
else
:
# 10. Post-processing
# 10. Post-processing
image
=
(
image
/
2
+
0.5
).
clamp
(
0
,
1
)
image
=
(
image
/
2
+
0.5
).
clamp
(
0
,
1
)
...
...
src/diffusers/pipelines/deepfloyd_if/pipeline_if_inpainting.py
View file @
aa1f00fd
...
@@ -15,7 +15,6 @@ from ...schedulers import DDPMScheduler
...
@@ -15,7 +15,6 @@ from ...schedulers import DDPMScheduler
from
...utils
import
(
from
...utils
import
(
BACKENDS_MAPPING
,
BACKENDS_MAPPING
,
PIL_INTERPOLATION
,
PIL_INTERPOLATION
,
is_accelerate_available
,
is_bs4_available
,
is_bs4_available
,
is_ftfy_available
,
is_ftfy_available
,
logging
,
logging
,
...
@@ -142,6 +141,7 @@ class IFInpaintingPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -142,6 +141,7 @@ class IFInpaintingPipeline(DiffusionPipeline, LoraLoaderMixin):
_optional_components
=
[
"tokenizer"
,
"text_encoder"
,
"safety_checker"
,
"feature_extractor"
,
"watermarker"
]
_optional_components
=
[
"tokenizer"
,
"text_encoder"
,
"safety_checker"
,
"feature_extractor"
,
"watermarker"
]
model_cpu_offload_seq
=
"text_encoder->unet"
model_cpu_offload_seq
=
"text_encoder->unet"
_exclude_from_cpu_offload
=
[
"watermarker"
]
def
__init__
(
def
__init__
(
self
,
self
,
...
@@ -183,21 +183,6 @@ class IFInpaintingPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -183,21 +183,6 @@ class IFInpaintingPipeline(DiffusionPipeline, LoraLoaderMixin):
)
)
self
.
register_to_config
(
requires_safety_checker
=
requires_safety_checker
)
self
.
register_to_config
(
requires_safety_checker
=
requires_safety_checker
)
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.remove_all_hooks
def
remove_all_hooks
(
self
):
if
is_accelerate_available
():
from
accelerate.hooks
import
remove_hook_from_module
else
:
raise
ImportError
(
"Please install accelerate via `pip install accelerate`"
)
for
model
in
[
self
.
text_encoder
,
self
.
unet
,
self
.
safety_checker
]:
if
model
is
not
None
:
remove_hook_from_module
(
model
,
recurse
=
True
)
self
.
unet_offload_hook
=
None
self
.
text_encoder_offload_hook
=
None
self
.
final_offload_hook
=
None
@
torch
.
no_grad
()
@
torch
.
no_grad
()
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.encode_prompt
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.encode_prompt
def
encode_prompt
(
def
encode_prompt
(
...
@@ -365,9 +350,6 @@ class IFInpaintingPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -365,9 +350,6 @@ class IFInpaintingPipeline(DiffusionPipeline, LoraLoaderMixin):
nsfw_detected
=
None
nsfw_detected
=
None
watermark_detected
=
None
watermark_detected
=
None
if
hasattr
(
self
,
"unet_offload_hook"
)
and
self
.
unet_offload_hook
is
not
None
:
self
.
unet_offload_hook
.
offload
()
return
image
,
nsfw_detected
,
watermark_detected
return
image
,
nsfw_detected
,
watermark_detected
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.prepare_extra_step_kwargs
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.prepare_extra_step_kwargs
...
...
src/diffusers/pipelines/deepfloyd_if/pipeline_if_inpainting_superresolution.py
View file @
aa1f00fd
...
@@ -16,7 +16,6 @@ from ...schedulers import DDPMScheduler
...
@@ -16,7 +16,6 @@ from ...schedulers import DDPMScheduler
from
...utils
import
(
from
...utils
import
(
BACKENDS_MAPPING
,
BACKENDS_MAPPING
,
PIL_INTERPOLATION
,
PIL_INTERPOLATION
,
is_accelerate_available
,
is_bs4_available
,
is_bs4_available
,
is_ftfy_available
,
is_ftfy_available
,
logging
,
logging
,
...
@@ -145,6 +144,7 @@ class IFInpaintingSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -145,6 +144,7 @@ class IFInpaintingSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
model_cpu_offload_seq
=
"text_encoder->unet"
model_cpu_offload_seq
=
"text_encoder->unet"
_optional_components
=
[
"tokenizer"
,
"text_encoder"
,
"safety_checker"
,
"feature_extractor"
,
"watermarker"
]
_optional_components
=
[
"tokenizer"
,
"text_encoder"
,
"safety_checker"
,
"feature_extractor"
,
"watermarker"
]
_exclude_from_cpu_offload
=
[
"watermarker"
]
def
__init__
(
def
__init__
(
self
,
self
,
...
@@ -193,21 +193,6 @@ class IFInpaintingSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -193,21 +193,6 @@ class IFInpaintingSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
)
)
self
.
register_to_config
(
requires_safety_checker
=
requires_safety_checker
)
self
.
register_to_config
(
requires_safety_checker
=
requires_safety_checker
)
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.remove_all_hooks
def
remove_all_hooks
(
self
):
if
is_accelerate_available
():
from
accelerate.hooks
import
remove_hook_from_module
else
:
raise
ImportError
(
"Please install accelerate via `pip install accelerate`"
)
for
model
in
[
self
.
text_encoder
,
self
.
unet
,
self
.
safety_checker
]:
if
model
is
not
None
:
remove_hook_from_module
(
model
,
recurse
=
True
)
self
.
unet_offload_hook
=
None
self
.
text_encoder_offload_hook
=
None
self
.
final_offload_hook
=
None
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline._text_preprocessing
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline._text_preprocessing
def
_text_preprocessing
(
self
,
text
,
clean_caption
=
False
):
def
_text_preprocessing
(
self
,
text
,
clean_caption
=
False
):
if
clean_caption
and
not
is_bs4_available
():
if
clean_caption
and
not
is_bs4_available
():
...
@@ -515,9 +500,6 @@ class IFInpaintingSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -515,9 +500,6 @@ class IFInpaintingSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
nsfw_detected
=
None
nsfw_detected
=
None
watermark_detected
=
None
watermark_detected
=
None
if
hasattr
(
self
,
"unet_offload_hook"
)
and
self
.
unet_offload_hook
is
not
None
:
self
.
unet_offload_hook
.
offload
()
return
image
,
nsfw_detected
,
watermark_detected
return
image
,
nsfw_detected
,
watermark_detected
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.prepare_extra_step_kwargs
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.prepare_extra_step_kwargs
...
...
src/diffusers/pipelines/deepfloyd_if/pipeline_if_superresolution.py
View file @
aa1f00fd
...
@@ -15,7 +15,6 @@ from ...models import UNet2DConditionModel
...
@@ -15,7 +15,6 @@ from ...models import UNet2DConditionModel
from
...schedulers
import
DDPMScheduler
from
...schedulers
import
DDPMScheduler
from
...utils
import
(
from
...utils
import
(
BACKENDS_MAPPING
,
BACKENDS_MAPPING
,
is_accelerate_available
,
is_bs4_available
,
is_bs4_available
,
is_ftfy_available
,
is_ftfy_available
,
logging
,
logging
,
...
@@ -101,6 +100,7 @@ class IFSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -101,6 +100,7 @@ class IFSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
_optional_components
=
[
"tokenizer"
,
"text_encoder"
,
"safety_checker"
,
"feature_extractor"
,
"watermarker"
]
_optional_components
=
[
"tokenizer"
,
"text_encoder"
,
"safety_checker"
,
"feature_extractor"
,
"watermarker"
]
model_cpu_offload_seq
=
"text_encoder->unet"
model_cpu_offload_seq
=
"text_encoder->unet"
_exclude_from_cpu_offload
=
[
"watermarker"
]
def
__init__
(
def
__init__
(
self
,
self
,
...
@@ -149,21 +149,6 @@ class IFSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -149,21 +149,6 @@ class IFSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
)
)
self
.
register_to_config
(
requires_safety_checker
=
requires_safety_checker
)
self
.
register_to_config
(
requires_safety_checker
=
requires_safety_checker
)
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.remove_all_hooks
def
remove_all_hooks
(
self
):
if
is_accelerate_available
():
from
accelerate.hooks
import
remove_hook_from_module
else
:
raise
ImportError
(
"Please install accelerate via `pip install accelerate`"
)
for
model
in
[
self
.
text_encoder
,
self
.
unet
,
self
.
safety_checker
]:
if
model
is
not
None
:
remove_hook_from_module
(
model
,
recurse
=
True
)
self
.
unet_offload_hook
=
None
self
.
text_encoder_offload_hook
=
None
self
.
final_offload_hook
=
None
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline._text_preprocessing
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline._text_preprocessing
def
_text_preprocessing
(
self
,
text
,
clean_caption
=
False
):
def
_text_preprocessing
(
self
,
text
,
clean_caption
=
False
):
if
clean_caption
and
not
is_bs4_available
():
if
clean_caption
and
not
is_bs4_available
():
...
@@ -471,9 +456,6 @@ class IFSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -471,9 +456,6 @@ class IFSuperResolutionPipeline(DiffusionPipeline, LoraLoaderMixin):
nsfw_detected
=
None
nsfw_detected
=
None
watermark_detected
=
None
watermark_detected
=
None
if
hasattr
(
self
,
"unet_offload_hook"
)
and
self
.
unet_offload_hook
is
not
None
:
self
.
unet_offload_hook
.
offload
()
return
image
,
nsfw_detected
,
watermark_detected
return
image
,
nsfw_detected
,
watermark_detected
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.prepare_extra_step_kwargs
# Copied from diffusers.pipelines.deepfloyd_if.pipeline_if.IFPipeline.prepare_extra_step_kwargs
...
...
src/diffusers/pipelines/kandinsky/pipeline_kandinsky_combined.py
View file @
aa1f00fd
...
@@ -143,6 +143,7 @@ class KandinskyCombinedPipeline(DiffusionPipeline):
...
@@ -143,6 +143,7 @@ class KandinskyCombinedPipeline(DiffusionPipeline):
_load_connected_pipes
=
True
_load_connected_pipes
=
True
model_cpu_offload_seq
=
"text_encoder->unet->movq->prior_prior->prior_image_encoder->prior_text_encoder"
model_cpu_offload_seq
=
"text_encoder->unet->movq->prior_prior->prior_image_encoder->prior_text_encoder"
_exclude_from_cpu_offload
=
[
"prior_prior"
]
def
__init__
(
def
__init__
(
self
,
self
,
...
@@ -360,6 +361,7 @@ class KandinskyImg2ImgCombinedPipeline(DiffusionPipeline):
...
@@ -360,6 +361,7 @@ class KandinskyImg2ImgCombinedPipeline(DiffusionPipeline):
_load_connected_pipes
=
True
_load_connected_pipes
=
True
model_cpu_offload_seq
=
"prior_text_encoder->prior_image_encoder->prior_prior->"
"text_encoder->unet->movq"
model_cpu_offload_seq
=
"prior_text_encoder->prior_image_encoder->prior_prior->"
"text_encoder->unet->movq"
_exclude_from_cpu_offload
=
[
"prior_prior"
]
def
__init__
(
def
__init__
(
self
,
self
,
...
@@ -600,6 +602,7 @@ class KandinskyInpaintCombinedPipeline(DiffusionPipeline):
...
@@ -600,6 +602,7 @@ class KandinskyInpaintCombinedPipeline(DiffusionPipeline):
_load_connected_pipes
=
True
_load_connected_pipes
=
True
model_cpu_offload_seq
=
"prior_text_encoder->prior_image_encoder->prior_prior->text_encoder->unet->movq"
model_cpu_offload_seq
=
"prior_text_encoder->prior_image_encoder->prior_prior->text_encoder->unet->movq"
_exclude_from_cpu_offload
=
[
"prior_prior"
]
def
__init__
(
def
__init__
(
self
,
self
,
...
...
src/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_combined.py
View file @
aa1f00fd
...
@@ -135,6 +135,7 @@ class KandinskyV22CombinedPipeline(DiffusionPipeline):
...
@@ -135,6 +135,7 @@ class KandinskyV22CombinedPipeline(DiffusionPipeline):
model_cpu_offload_seq
=
"prior_text_encoder->prior_image_encoder->unet->movq"
model_cpu_offload_seq
=
"prior_text_encoder->prior_image_encoder->unet->movq"
_load_connected_pipes
=
True
_load_connected_pipes
=
True
_exclude_from_cpu_offload
=
[
"prior_prior"
]
def
__init__
(
def
__init__
(
self
,
self
,
...
@@ -362,6 +363,7 @@ class KandinskyV22Img2ImgCombinedPipeline(DiffusionPipeline):
...
@@ -362,6 +363,7 @@ class KandinskyV22Img2ImgCombinedPipeline(DiffusionPipeline):
model_cpu_offload_seq
=
"prior_text_encoder->prior_image_encoder->unet->movq"
model_cpu_offload_seq
=
"prior_text_encoder->prior_image_encoder->unet->movq"
_load_connected_pipes
=
True
_load_connected_pipes
=
True
_exclude_from_cpu_offload
=
[
"prior_prior"
]
def
__init__
(
def
__init__
(
self
,
self
,
...
@@ -610,6 +612,7 @@ class KandinskyV22InpaintCombinedPipeline(DiffusionPipeline):
...
@@ -610,6 +612,7 @@ class KandinskyV22InpaintCombinedPipeline(DiffusionPipeline):
model_cpu_offload_seq
=
"prior_text_encoder->prior_image_encoder->unet->movq"
model_cpu_offload_seq
=
"prior_text_encoder->prior_image_encoder->unet->movq"
_load_connected_pipes
=
True
_load_connected_pipes
=
True
_exclude_from_cpu_offload
=
[
"prior_prior"
]
def
__init__
(
def
__init__
(
self
,
self
,
...
...
src/diffusers/pipelines/kandinsky3/pipeline_kandinsky3.py
View file @
aa1f00fd
...
@@ -8,7 +8,6 @@ from ...models import Kandinsky3UNet, VQModel
...
@@ -8,7 +8,6 @@ from ...models import Kandinsky3UNet, VQModel
from
...schedulers
import
DDPMScheduler
from
...schedulers
import
DDPMScheduler
from
...utils
import
(
from
...utils
import
(
deprecate
,
deprecate
,
is_accelerate_available
,
logging
,
logging
,
replace_example_docstring
,
replace_example_docstring
,
)
)
...
@@ -72,20 +71,6 @@ class Kandinsky3Pipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -72,20 +71,6 @@ class Kandinsky3Pipeline(DiffusionPipeline, LoraLoaderMixin):
tokenizer
=
tokenizer
,
text_encoder
=
text_encoder
,
unet
=
unet
,
scheduler
=
scheduler
,
movq
=
movq
tokenizer
=
tokenizer
,
text_encoder
=
text_encoder
,
unet
=
unet
,
scheduler
=
scheduler
,
movq
=
movq
)
)
def
remove_all_hooks
(
self
):
if
is_accelerate_available
():
from
accelerate.hooks
import
remove_hook_from_module
else
:
raise
ImportError
(
"Please install accelerate via `pip install accelerate`"
)
for
model
in
[
self
.
text_encoder
,
self
.
unet
,
self
.
movq
]:
if
model
is
not
None
:
remove_hook_from_module
(
model
,
recurse
=
True
)
self
.
unet_offload_hook
=
None
self
.
text_encoder_offload_hook
=
None
self
.
final_offload_hook
=
None
def
process_embeds
(
self
,
embeddings
,
attention_mask
,
cut_context
):
def
process_embeds
(
self
,
embeddings
,
attention_mask
,
cut_context
):
if
cut_context
:
if
cut_context
:
embeddings
[
attention_mask
==
0
]
=
torch
.
zeros_like
(
embeddings
[
attention_mask
==
0
])
embeddings
[
attention_mask
==
0
]
=
torch
.
zeros_like
(
embeddings
[
attention_mask
==
0
])
...
...
src/diffusers/pipelines/kandinsky3/pipeline_kandinsky3_img2img.py
View file @
aa1f00fd
...
@@ -12,7 +12,6 @@ from ...models import Kandinsky3UNet, VQModel
...
@@ -12,7 +12,6 @@ from ...models import Kandinsky3UNet, VQModel
from
...schedulers
import
DDPMScheduler
from
...schedulers
import
DDPMScheduler
from
...utils
import
(
from
...utils
import
(
deprecate
,
deprecate
,
is_accelerate_available
,
logging
,
logging
,
replace_example_docstring
,
replace_example_docstring
,
)
)
...
@@ -96,20 +95,6 @@ class Kandinsky3Img2ImgPipeline(DiffusionPipeline, LoraLoaderMixin):
...
@@ -96,20 +95,6 @@ class Kandinsky3Img2ImgPipeline(DiffusionPipeline, LoraLoaderMixin):
return
timesteps
,
num_inference_steps
-
t_start
return
timesteps
,
num_inference_steps
-
t_start
def
remove_all_hooks
(
self
):
if
is_accelerate_available
():
from
accelerate.hooks
import
remove_hook_from_module
else
:
raise
ImportError
(
"Please install accelerate via `pip install accelerate`"
)
for
model
in
[
self
.
text_encoder
,
self
.
unet
]:
if
model
is
not
None
:
remove_hook_from_module
(
model
,
recurse
=
True
)
self
.
unet_offload_hook
=
None
self
.
text_encoder_offload_hook
=
None
self
.
final_offload_hook
=
None
def
_process_embeds
(
self
,
embeddings
,
attention_mask
,
cut_context
):
def
_process_embeds
(
self
,
embeddings
,
attention_mask
,
cut_context
):
# return embeddings, attention_mask
# return embeddings, attention_mask
if
cut_context
:
if
cut_context
:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment