Unverified Commit c0f5346a authored by co63oc's avatar co63oc Committed by GitHub
Browse files

Fix procecss process (#6591)



* Fix words

* Fix

---------
Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
parent 087daee2
...@@ -848,7 +848,7 @@ class IPAdapterFaceIDStableDiffusionPipeline( ...@@ -848,7 +848,7 @@ class IPAdapterFaceIDStableDiffusionPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -930,7 +930,7 @@ class IPAdapterFaceIDStableDiffusionPipeline( ...@@ -930,7 +930,7 @@ class IPAdapterFaceIDStableDiffusionPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -395,7 +395,7 @@ class LatentConsistencyModelWalkPipeline( ...@@ -395,7 +395,7 @@ class LatentConsistencyModelWalkPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -477,7 +477,7 @@ class LatentConsistencyModelWalkPipeline( ...@@ -477,7 +477,7 @@ class LatentConsistencyModelWalkPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -1307,7 +1307,7 @@ class LLMGroundedDiffusionPipeline( ...@@ -1307,7 +1307,7 @@ class LLMGroundedDiffusionPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -1391,7 +1391,7 @@ class LLMGroundedDiffusionPipeline( ...@@ -1391,7 +1391,7 @@ class LLMGroundedDiffusionPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -789,7 +789,7 @@ class SDXLLongPromptWeightingPipeline( ...@@ -789,7 +789,7 @@ class SDXLLongPromptWeightingPipeline(
if prompt_embeds is None: if prompt_embeds is None:
prompt_2 = prompt_2 or prompt prompt_2 = prompt_2 or prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
prompt_embeds_list = [] prompt_embeds_list = []
prompts = [prompt, prompt_2] prompts = [prompt, prompt_2]
for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders): for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders):
......
...@@ -247,7 +247,7 @@ class AnimateDiffControlNetPipeline(DiffusionPipeline, TextualInversionLoaderMix ...@@ -247,7 +247,7 @@ class AnimateDiffControlNetPipeline(DiffusionPipeline, TextualInversionLoaderMix
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -329,7 +329,7 @@ class AnimateDiffControlNetPipeline(DiffusionPipeline, TextualInversionLoaderMix ...@@ -329,7 +329,7 @@ class AnimateDiffControlNetPipeline(DiffusionPipeline, TextualInversionLoaderMix
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -289,7 +289,7 @@ class DemoFusionSDXLPipeline(DiffusionPipeline, FromSingleFileMixin, LoraLoaderM ...@@ -289,7 +289,7 @@ class DemoFusionSDXLPipeline(DiffusionPipeline, FromSingleFileMixin, LoraLoaderM
if prompt_embeds is None: if prompt_embeds is None:
prompt_2 = prompt_2 or prompt prompt_2 = prompt_2 or prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
prompt_embeds_list = [] prompt_embeds_list = []
prompts = [prompt, prompt_2] prompts = [prompt, prompt_2]
for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders): for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders):
......
...@@ -233,7 +233,7 @@ class FabricPipeline(DiffusionPipeline): ...@@ -233,7 +233,7 @@ class FabricPipeline(DiffusionPipeline):
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -304,7 +304,7 @@ class FabricPipeline(DiffusionPipeline): ...@@ -304,7 +304,7 @@ class FabricPipeline(DiffusionPipeline):
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -632,7 +632,7 @@ class StyleAlignedSDXLPipeline( ...@@ -632,7 +632,7 @@ class StyleAlignedSDXLPipeline(
prompt_2 = prompt_2 or prompt prompt_2 = prompt_2 or prompt
prompt_2 = [prompt_2] if isinstance(prompt_2, str) else prompt_2 prompt_2 = [prompt_2] if isinstance(prompt_2, str) else prompt_2
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
prompt_embeds_list = [] prompt_embeds_list = []
prompts = [prompt, prompt_2] prompts = [prompt, prompt_2]
for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders): for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders):
......
...@@ -250,7 +250,7 @@ class StableDiffusionUpscaleLDM3DPipeline( ...@@ -250,7 +250,7 @@ class StableDiffusionUpscaleLDM3DPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -332,7 +332,7 @@ class StableDiffusionUpscaleLDM3DPipeline( ...@@ -332,7 +332,7 @@ class StableDiffusionUpscaleLDM3DPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -363,7 +363,7 @@ class StableDiffusionXLControlNetAdapterPipeline( ...@@ -363,7 +363,7 @@ class StableDiffusionXLControlNetAdapterPipeline(
prompt_2 = prompt_2 or prompt prompt_2 = prompt_2 or prompt
prompt_2 = [prompt_2] if isinstance(prompt_2, str) else prompt_2 prompt_2 = [prompt_2] if isinstance(prompt_2, str) else prompt_2
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
prompt_embeds_list = [] prompt_embeds_list = []
prompts = [prompt, prompt_2] prompts = [prompt, prompt_2]
for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders): for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders):
......
...@@ -512,7 +512,7 @@ class StableDiffusionXLControlNetAdapterInpaintPipeline(DiffusionPipeline, FromS ...@@ -512,7 +512,7 @@ class StableDiffusionXLControlNetAdapterInpaintPipeline(DiffusionPipeline, FromS
prompt_2 = prompt_2 or prompt prompt_2 = prompt_2 or prompt
prompt_2 = [prompt_2] if isinstance(prompt_2, str) else prompt_2 prompt_2 = [prompt_2] if isinstance(prompt_2, str) else prompt_2
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
prompt_embeds_list = [] prompt_embeds_list = []
prompts = [prompt, prompt_2] prompts = [prompt, prompt_2]
for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders): for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders):
......
...@@ -449,7 +449,7 @@ class StableDiffusionIPEXPipeline(DiffusionPipeline, TextualInversionLoaderMixin ...@@ -449,7 +449,7 @@ class StableDiffusionIPEXPipeline(DiffusionPipeline, TextualInversionLoaderMixin
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -513,7 +513,7 @@ class StableDiffusionIPEXPipeline(DiffusionPipeline, TextualInversionLoaderMixin ...@@ -513,7 +513,7 @@ class StableDiffusionIPEXPipeline(DiffusionPipeline, TextualInversionLoaderMixin
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -392,7 +392,7 @@ class StableDiffusionRepaintPipeline(DiffusionPipeline, TextualInversionLoaderMi ...@@ -392,7 +392,7 @@ class StableDiffusionRepaintPipeline(DiffusionPipeline, TextualInversionLoaderMi
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -456,7 +456,7 @@ class StableDiffusionRepaintPipeline(DiffusionPipeline, TextualInversionLoaderMi ...@@ -456,7 +456,7 @@ class StableDiffusionRepaintPipeline(DiffusionPipeline, TextualInversionLoaderMi
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -266,7 +266,7 @@ class StableDiffusionControlNetXSPipeline( ...@@ -266,7 +266,7 @@ class StableDiffusionControlNetXSPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -348,7 +348,7 @@ class StableDiffusionControlNetXSPipeline( ...@@ -348,7 +348,7 @@ class StableDiffusionControlNetXSPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -274,7 +274,7 @@ class StableDiffusionXLControlNetXSPipeline( ...@@ -274,7 +274,7 @@ class StableDiffusionXLControlNetXSPipeline(
prompt_2 = prompt_2 or prompt prompt_2 = prompt_2 or prompt
prompt_2 = [prompt_2] if isinstance(prompt_2, str) else prompt_2 prompt_2 = [prompt_2] if isinstance(prompt_2, str) else prompt_2
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
prompt_embeds_list = [] prompt_embeds_list = []
prompts = [prompt, prompt_2] prompts = [prompt, prompt_2]
for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders): for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders):
......
...@@ -280,7 +280,7 @@ class AnimateDiffPipeline(DiffusionPipeline, TextualInversionLoaderMixin, IPAdap ...@@ -280,7 +280,7 @@ class AnimateDiffPipeline(DiffusionPipeline, TextualInversionLoaderMixin, IPAdap
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -362,7 +362,7 @@ class AnimateDiffPipeline(DiffusionPipeline, TextualInversionLoaderMixin, IPAdap ...@@ -362,7 +362,7 @@ class AnimateDiffPipeline(DiffusionPipeline, TextualInversionLoaderMixin, IPAdap
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -291,7 +291,7 @@ class AnimateDiffVideoToVideoPipeline(DiffusionPipeline, TextualInversionLoaderM ...@@ -291,7 +291,7 @@ class AnimateDiffVideoToVideoPipeline(DiffusionPipeline, TextualInversionLoaderM
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -373,7 +373,7 @@ class AnimateDiffVideoToVideoPipeline(DiffusionPipeline, TextualInversionLoaderM ...@@ -373,7 +373,7 @@ class AnimateDiffVideoToVideoPipeline(DiffusionPipeline, TextualInversionLoaderM
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -360,7 +360,7 @@ class StableDiffusionControlNetPipeline( ...@@ -360,7 +360,7 @@ class StableDiffusionControlNetPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -442,7 +442,7 @@ class StableDiffusionControlNetPipeline( ...@@ -442,7 +442,7 @@ class StableDiffusionControlNetPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -353,7 +353,7 @@ class StableDiffusionControlNetImg2ImgPipeline( ...@@ -353,7 +353,7 @@ class StableDiffusionControlNetImg2ImgPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -435,7 +435,7 @@ class StableDiffusionControlNetImg2ImgPipeline( ...@@ -435,7 +435,7 @@ class StableDiffusionControlNetImg2ImgPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -478,7 +478,7 @@ class StableDiffusionControlNetInpaintPipeline( ...@@ -478,7 +478,7 @@ class StableDiffusionControlNetInpaintPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -560,7 +560,7 @@ class StableDiffusionControlNetInpaintPipeline( ...@@ -560,7 +560,7 @@ class StableDiffusionControlNetInpaintPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment