"git@developer.sourcefind.cn:OpenDAS/apex.git" did not exist on "dc41c5ce17688283aa103ac753b247744d64ed96"
Unverified Commit c9a219b3 authored by fancydaddy's avatar fancydaddy Committed by GitHub
Browse files

add from_single_file to animatediff (#10924)



* Update pipeline_animatediff.py

* Update pipeline_animatediff_controlnet.py

* Update pipeline_animatediff_sparsectrl.py

* Update pipeline_animatediff_video2video.py

* Update pipeline_animatediff_video2video_controlnet.py

---------
Co-authored-by: default avatarDhruv Nair <dhruv.nair@gmail.com>
parent 9e910c46
...@@ -19,7 +19,7 @@ import torch ...@@ -19,7 +19,7 @@ import torch
from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, CLIPVisionModelWithProjection from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, CLIPVisionModelWithProjection
from ...image_processor import PipelineImageInput from ...image_processor import PipelineImageInput
from ...loaders import IPAdapterMixin, StableDiffusionLoraLoaderMixin, TextualInversionLoaderMixin from ...loaders import FromSingleFileMixin, IPAdapterMixin, StableDiffusionLoraLoaderMixin, TextualInversionLoaderMixin
from ...models import AutoencoderKL, ImageProjection, UNet2DConditionModel, UNetMotionModel from ...models import AutoencoderKL, ImageProjection, UNet2DConditionModel, UNetMotionModel
from ...models.lora import adjust_lora_scale_text_encoder from ...models.lora import adjust_lora_scale_text_encoder
from ...models.unets.unet_motion_model import MotionAdapter from ...models.unets.unet_motion_model import MotionAdapter
...@@ -83,6 +83,7 @@ class AnimateDiffPipeline( ...@@ -83,6 +83,7 @@ class AnimateDiffPipeline(
StableDiffusionLoraLoaderMixin, StableDiffusionLoraLoaderMixin,
FreeInitMixin, FreeInitMixin,
AnimateDiffFreeNoiseMixin, AnimateDiffFreeNoiseMixin,
FromSingleFileMixin,
): ):
r""" r"""
Pipeline for text-to-video generation. Pipeline for text-to-video generation.
......
...@@ -20,7 +20,7 @@ import torch.nn.functional as F ...@@ -20,7 +20,7 @@ import torch.nn.functional as F
from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, CLIPVisionModelWithProjection from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, CLIPVisionModelWithProjection
from ...image_processor import PipelineImageInput from ...image_processor import PipelineImageInput
from ...loaders import IPAdapterMixin, StableDiffusionLoraLoaderMixin, TextualInversionLoaderMixin from ...loaders import FromSingleFileMixin, IPAdapterMixin, StableDiffusionLoraLoaderMixin, TextualInversionLoaderMixin
from ...models import ( from ...models import (
AutoencoderKL, AutoencoderKL,
ControlNetModel, ControlNetModel,
...@@ -125,6 +125,7 @@ class AnimateDiffControlNetPipeline( ...@@ -125,6 +125,7 @@ class AnimateDiffControlNetPipeline(
StableDiffusionLoraLoaderMixin, StableDiffusionLoraLoaderMixin,
FreeInitMixin, FreeInitMixin,
AnimateDiffFreeNoiseMixin, AnimateDiffFreeNoiseMixin,
FromSingleFileMixin,
): ):
r""" r"""
Pipeline for text-to-video generation with ControlNet guidance. Pipeline for text-to-video generation with ControlNet guidance.
......
...@@ -22,7 +22,7 @@ import torch.nn.functional as F ...@@ -22,7 +22,7 @@ import torch.nn.functional as F
from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, CLIPVisionModelWithProjection from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, CLIPVisionModelWithProjection
from ...image_processor import PipelineImageInput, VaeImageProcessor from ...image_processor import PipelineImageInput, VaeImageProcessor
from ...loaders import IPAdapterMixin, StableDiffusionLoraLoaderMixin, TextualInversionLoaderMixin from ...loaders import FromSingleFileMixin, IPAdapterMixin, StableDiffusionLoraLoaderMixin, TextualInversionLoaderMixin
from ...models import AutoencoderKL, ImageProjection, UNet2DConditionModel, UNetMotionModel from ...models import AutoencoderKL, ImageProjection, UNet2DConditionModel, UNetMotionModel
from ...models.controlnets.controlnet_sparsectrl import SparseControlNetModel from ...models.controlnets.controlnet_sparsectrl import SparseControlNetModel
from ...models.lora import adjust_lora_scale_text_encoder from ...models.lora import adjust_lora_scale_text_encoder
...@@ -136,6 +136,7 @@ class AnimateDiffSparseControlNetPipeline( ...@@ -136,6 +136,7 @@ class AnimateDiffSparseControlNetPipeline(
IPAdapterMixin, IPAdapterMixin,
StableDiffusionLoraLoaderMixin, StableDiffusionLoraLoaderMixin,
FreeInitMixin, FreeInitMixin,
FromSingleFileMixin,
): ):
r""" r"""
Pipeline for controlled text-to-video generation using the method described in [SparseCtrl: Adding Sparse Controls Pipeline for controlled text-to-video generation using the method described in [SparseCtrl: Adding Sparse Controls
......
...@@ -19,7 +19,7 @@ import torch ...@@ -19,7 +19,7 @@ import torch
from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, CLIPVisionModelWithProjection from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, CLIPVisionModelWithProjection
from ...image_processor import PipelineImageInput from ...image_processor import PipelineImageInput
from ...loaders import IPAdapterMixin, StableDiffusionLoraLoaderMixin, TextualInversionLoaderMixin from ...loaders import FromSingleFileMixin, IPAdapterMixin, StableDiffusionLoraLoaderMixin, TextualInversionLoaderMixin
from ...models import AutoencoderKL, ImageProjection, UNet2DConditionModel, UNetMotionModel from ...models import AutoencoderKL, ImageProjection, UNet2DConditionModel, UNetMotionModel
from ...models.lora import adjust_lora_scale_text_encoder from ...models.lora import adjust_lora_scale_text_encoder
from ...models.unets.unet_motion_model import MotionAdapter from ...models.unets.unet_motion_model import MotionAdapter
...@@ -186,6 +186,7 @@ class AnimateDiffVideoToVideoPipeline( ...@@ -186,6 +186,7 @@ class AnimateDiffVideoToVideoPipeline(
StableDiffusionLoraLoaderMixin, StableDiffusionLoraLoaderMixin,
FreeInitMixin, FreeInitMixin,
AnimateDiffFreeNoiseMixin, AnimateDiffFreeNoiseMixin,
FromSingleFileMixin,
): ):
r""" r"""
Pipeline for video-to-video generation. Pipeline for video-to-video generation.
......
...@@ -20,7 +20,7 @@ import torch.nn.functional as F ...@@ -20,7 +20,7 @@ import torch.nn.functional as F
from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, CLIPVisionModelWithProjection from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, CLIPVisionModelWithProjection
from ...image_processor import PipelineImageInput from ...image_processor import PipelineImageInput
from ...loaders import IPAdapterMixin, StableDiffusionLoraLoaderMixin, TextualInversionLoaderMixin from ...loaders import FromSingleFileMixin, IPAdapterMixin, StableDiffusionLoraLoaderMixin, TextualInversionLoaderMixin
from ...models import ( from ...models import (
AutoencoderKL, AutoencoderKL,
ControlNetModel, ControlNetModel,
...@@ -204,6 +204,7 @@ class AnimateDiffVideoToVideoControlNetPipeline( ...@@ -204,6 +204,7 @@ class AnimateDiffVideoToVideoControlNetPipeline(
StableDiffusionLoraLoaderMixin, StableDiffusionLoraLoaderMixin,
FreeInitMixin, FreeInitMixin,
AnimateDiffFreeNoiseMixin, AnimateDiffFreeNoiseMixin,
FromSingleFileMixin,
): ):
r""" r"""
Pipeline for video-to-video generation with ControlNet guidance. Pipeline for video-to-video generation with ControlNet guidance.
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment