Unverified Commit c0f5346a authored by co63oc's avatar co63oc Committed by GitHub
Browse files

Fix procecss process (#6591)



* Fix words

* Fix

---------
Co-authored-by: default avatarYiYi Xu <yixu310@gmail.com>
parent 087daee2
...@@ -358,7 +358,7 @@ class StableDiffusionXLControlNetInpaintPipeline( ...@@ -358,7 +358,7 @@ class StableDiffusionXLControlNetInpaintPipeline(
prompt_2 = prompt_2 or prompt prompt_2 = prompt_2 or prompt
prompt_2 = [prompt_2] if isinstance(prompt_2, str) else prompt_2 prompt_2 = [prompt_2] if isinstance(prompt_2, str) else prompt_2
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
prompt_embeds_list = [] prompt_embeds_list = []
prompts = [prompt, prompt_2] prompts = [prompt, prompt_2]
for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders): for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders):
......
...@@ -351,7 +351,7 @@ class StableDiffusionXLControlNetPipeline( ...@@ -351,7 +351,7 @@ class StableDiffusionXLControlNetPipeline(
prompt_2 = prompt_2 or prompt prompt_2 = prompt_2 or prompt
prompt_2 = [prompt_2] if isinstance(prompt_2, str) else prompt_2 prompt_2 = [prompt_2] if isinstance(prompt_2, str) else prompt_2
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
prompt_embeds_list = [] prompt_embeds_list = []
prompts = [prompt, prompt_2] prompts = [prompt, prompt_2]
for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders): for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders):
......
...@@ -400,7 +400,7 @@ class StableDiffusionXLControlNetImg2ImgPipeline( ...@@ -400,7 +400,7 @@ class StableDiffusionXLControlNetImg2ImgPipeline(
prompt_2 = prompt_2 or prompt prompt_2 = prompt_2 or prompt
prompt_2 = [prompt_2] if isinstance(prompt_2, str) else prompt_2 prompt_2 = [prompt_2] if isinstance(prompt_2, str) else prompt_2
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
prompt_embeds_list = [] prompt_embeds_list = []
prompts = [prompt, prompt_2] prompts = [prompt, prompt_2]
for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders): for prompt, tokenizer, text_encoder in zip(prompts, tokenizers, text_encoders):
......
...@@ -373,7 +373,7 @@ class AltDiffusionPipeline( ...@@ -373,7 +373,7 @@ class AltDiffusionPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -455,7 +455,7 @@ class AltDiffusionPipeline( ...@@ -455,7 +455,7 @@ class AltDiffusionPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -384,7 +384,7 @@ class AltDiffusionImg2ImgPipeline( ...@@ -384,7 +384,7 @@ class AltDiffusionImg2ImgPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -466,7 +466,7 @@ class AltDiffusionImg2ImgPipeline( ...@@ -466,7 +466,7 @@ class AltDiffusionImg2ImgPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -341,7 +341,7 @@ class CycleDiffusionPipeline(DiffusionPipeline, TextualInversionLoaderMixin, Lor ...@@ -341,7 +341,7 @@ class CycleDiffusionPipeline(DiffusionPipeline, TextualInversionLoaderMixin, Lor
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -423,7 +423,7 @@ class CycleDiffusionPipeline(DiffusionPipeline, TextualInversionLoaderMixin, Lor ...@@ -423,7 +423,7 @@ class CycleDiffusionPipeline(DiffusionPipeline, TextualInversionLoaderMixin, Lor
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -311,7 +311,7 @@ class StableDiffusionInpaintPipelineLegacy( ...@@ -311,7 +311,7 @@ class StableDiffusionInpaintPipelineLegacy(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -393,7 +393,7 @@ class StableDiffusionInpaintPipelineLegacy( ...@@ -393,7 +393,7 @@ class StableDiffusionInpaintPipelineLegacy(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -263,7 +263,7 @@ class StableDiffusionModelEditingPipeline(DiffusionPipeline, TextualInversionLoa ...@@ -263,7 +263,7 @@ class StableDiffusionModelEditingPipeline(DiffusionPipeline, TextualInversionLoa
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -345,7 +345,7 @@ class StableDiffusionModelEditingPipeline(DiffusionPipeline, TextualInversionLoa ...@@ -345,7 +345,7 @@ class StableDiffusionModelEditingPipeline(DiffusionPipeline, TextualInversionLoa
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -273,7 +273,7 @@ class StableDiffusionParadigmsPipeline( ...@@ -273,7 +273,7 @@ class StableDiffusionParadigmsPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -355,7 +355,7 @@ class StableDiffusionParadigmsPipeline( ...@@ -355,7 +355,7 @@ class StableDiffusionParadigmsPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -463,7 +463,7 @@ class StableDiffusionPix2PixZeroPipeline(DiffusionPipeline): ...@@ -463,7 +463,7 @@ class StableDiffusionPix2PixZeroPipeline(DiffusionPipeline):
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -545,7 +545,7 @@ class StableDiffusionPix2PixZeroPipeline(DiffusionPipeline): ...@@ -545,7 +545,7 @@ class StableDiffusionPix2PixZeroPipeline(DiffusionPipeline):
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -331,7 +331,7 @@ class LatentConsistencyModelImg2ImgPipeline( ...@@ -331,7 +331,7 @@ class LatentConsistencyModelImg2ImgPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -413,7 +413,7 @@ class LatentConsistencyModelImg2ImgPipeline( ...@@ -413,7 +413,7 @@ class LatentConsistencyModelImg2ImgPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -315,7 +315,7 @@ class LatentConsistencyModelPipeline( ...@@ -315,7 +315,7 @@ class LatentConsistencyModelPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -397,7 +397,7 @@ class LatentConsistencyModelPipeline( ...@@ -397,7 +397,7 @@ class LatentConsistencyModelPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -340,7 +340,7 @@ class PIAPipeline( ...@@ -340,7 +340,7 @@ class PIAPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -422,7 +422,7 @@ class PIAPipeline( ...@@ -422,7 +422,7 @@ class PIAPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -369,7 +369,7 @@ class StableDiffusionPipeline( ...@@ -369,7 +369,7 @@ class StableDiffusionPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -451,7 +451,7 @@ class StableDiffusionPipeline( ...@@ -451,7 +451,7 @@ class StableDiffusionPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -242,7 +242,7 @@ class StableDiffusionDepth2ImgPipeline(DiffusionPipeline, TextualInversionLoader ...@@ -242,7 +242,7 @@ class StableDiffusionDepth2ImgPipeline(DiffusionPipeline, TextualInversionLoader
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -324,7 +324,7 @@ class StableDiffusionDepth2ImgPipeline(DiffusionPipeline, TextualInversionLoader ...@@ -324,7 +324,7 @@ class StableDiffusionDepth2ImgPipeline(DiffusionPipeline, TextualInversionLoader
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -415,7 +415,7 @@ class StableDiffusionImg2ImgPipeline( ...@@ -415,7 +415,7 @@ class StableDiffusionImg2ImgPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -497,7 +497,7 @@ class StableDiffusionImg2ImgPipeline( ...@@ -497,7 +497,7 @@ class StableDiffusionImg2ImgPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -487,7 +487,7 @@ class StableDiffusionInpaintPipeline( ...@@ -487,7 +487,7 @@ class StableDiffusionInpaintPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -569,7 +569,7 @@ class StableDiffusionInpaintPipeline( ...@@ -569,7 +569,7 @@ class StableDiffusionInpaintPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -523,7 +523,7 @@ class StableDiffusionInstructPix2PixPipeline( ...@@ -523,7 +523,7 @@ class StableDiffusionInstructPix2PixPipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -587,7 +587,7 @@ class StableDiffusionInstructPix2PixPipeline( ...@@ -587,7 +587,7 @@ class StableDiffusionInstructPix2PixPipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -262,7 +262,7 @@ class StableDiffusionUpscalePipeline( ...@@ -262,7 +262,7 @@ class StableDiffusionUpscalePipeline(
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -344,7 +344,7 @@ class StableDiffusionUpscalePipeline( ...@@ -344,7 +344,7 @@ class StableDiffusionUpscalePipeline(
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
...@@ -359,7 +359,7 @@ class StableUnCLIPPipeline(DiffusionPipeline, TextualInversionLoaderMixin, LoraL ...@@ -359,7 +359,7 @@ class StableUnCLIPPipeline(DiffusionPipeline, TextualInversionLoaderMixin, LoraL
batch_size = prompt_embeds.shape[0] batch_size = prompt_embeds.shape[0]
if prompt_embeds is None: if prompt_embeds is None:
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
prompt = self.maybe_convert_prompt(prompt, self.tokenizer) prompt = self.maybe_convert_prompt(prompt, self.tokenizer)
...@@ -441,7 +441,7 @@ class StableUnCLIPPipeline(DiffusionPipeline, TextualInversionLoaderMixin, LoraL ...@@ -441,7 +441,7 @@ class StableUnCLIPPipeline(DiffusionPipeline, TextualInversionLoaderMixin, LoraL
else: else:
uncond_tokens = negative_prompt uncond_tokens = negative_prompt
# textual inversion: procecss multi-vector tokens if necessary # textual inversion: process multi-vector tokens if necessary
if isinstance(self, TextualInversionLoaderMixin): if isinstance(self, TextualInversionLoaderMixin):
uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer) uncond_tokens = self.maybe_convert_prompt(uncond_tokens, self.tokenizer)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment