Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
renzhc
diffusers_dcu
Commits
db33af06
Unverified
Commit
db33af06
authored
May 24, 2024
by
Tolga Cangöz
Committed by
GitHub
May 24, 2024
Browse files
Fix a grammatical error in the `raise` messages (#8272)
Fix grammatical error
parent
1096f88e
Changes
27
Hide whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
29 additions
and
29 deletions
+29
-29
examples/community/latent_consistency_img2img.py
examples/community/latent_consistency_img2img.py
+1
-1
examples/community/latent_consistency_txt2img.py
examples/community/latent_consistency_txt2img.py
+1
-1
examples/community/scheduling_ufogen.py
examples/community/scheduling_ufogen.py
+1
-1
src/diffusers/schedulers/scheduling_ddim.py
src/diffusers/schedulers/scheduling_ddim.py
+1
-1
src/diffusers/schedulers/scheduling_ddim_inverse.py
src/diffusers/schedulers/scheduling_ddim_inverse.py
+1
-1
src/diffusers/schedulers/scheduling_ddim_parallel.py
src/diffusers/schedulers/scheduling_ddim_parallel.py
+1
-1
src/diffusers/schedulers/scheduling_ddpm.py
src/diffusers/schedulers/scheduling_ddpm.py
+1
-1
src/diffusers/schedulers/scheduling_ddpm_parallel.py
src/diffusers/schedulers/scheduling_ddpm_parallel.py
+1
-1
src/diffusers/schedulers/scheduling_deis_multistep.py
src/diffusers/schedulers/scheduling_deis_multistep.py
+3
-3
src/diffusers/schedulers/scheduling_dpmsolver_multistep.py
src/diffusers/schedulers/scheduling_dpmsolver_multistep.py
+3
-3
src/diffusers/schedulers/scheduling_dpmsolver_multistep_flax.py
...ffusers/schedulers/scheduling_dpmsolver_multistep_flax.py
+2
-2
src/diffusers/schedulers/scheduling_dpmsolver_multistep_inverse.py
...sers/schedulers/scheduling_dpmsolver_multistep_inverse.py
+3
-3
src/diffusers/schedulers/scheduling_dpmsolver_sde.py
src/diffusers/schedulers/scheduling_dpmsolver_sde.py
+1
-1
src/diffusers/schedulers/scheduling_dpmsolver_singlestep.py
src/diffusers/schedulers/scheduling_dpmsolver_singlestep.py
+3
-3
src/diffusers/schedulers/scheduling_edm_dpmsolver_multistep.py
...iffusers/schedulers/scheduling_edm_dpmsolver_multistep.py
+1
-1
src/diffusers/schedulers/scheduling_euler_ancestral_discrete.py
...ffusers/schedulers/scheduling_euler_ancestral_discrete.py
+1
-1
src/diffusers/schedulers/scheduling_euler_discrete.py
src/diffusers/schedulers/scheduling_euler_discrete.py
+1
-1
src/diffusers/schedulers/scheduling_heun_discrete.py
src/diffusers/schedulers/scheduling_heun_discrete.py
+1
-1
src/diffusers/schedulers/scheduling_k_dpm_2_ancestral_discrete.py
...users/schedulers/scheduling_k_dpm_2_ancestral_discrete.py
+1
-1
src/diffusers/schedulers/scheduling_k_dpm_2_discrete.py
src/diffusers/schedulers/scheduling_k_dpm_2_discrete.py
+1
-1
No files found.
examples/community/latent_consistency_img2img.py
View file @
db33af06
...
...
@@ -565,7 +565,7 @@ class LCMSchedulerWithTimestamp(SchedulerMixin, ConfigMixin):
# Glide cosine schedule
self
.
betas
=
betas_for_alpha_bar
(
num_train_timesteps
)
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
# Rescale for zero SNR
if
rescale_betas_zero_snr
:
...
...
examples/community/latent_consistency_txt2img.py
View file @
db33af06
...
...
@@ -477,7 +477,7 @@ class LCMScheduler(SchedulerMixin, ConfigMixin):
# Glide cosine schedule
self
.
betas
=
betas_for_alpha_bar
(
num_train_timesteps
)
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
# Rescale for zero SNR
if
rescale_betas_zero_snr
:
...
...
examples/community/scheduling_ufogen.py
View file @
db33af06
...
...
@@ -218,7 +218,7 @@ class UFOGenScheduler(SchedulerMixin, ConfigMixin):
betas
=
torch
.
linspace
(
-
6
,
6
,
num_train_timesteps
)
self
.
betas
=
torch
.
sigmoid
(
betas
)
*
(
beta_end
-
beta_start
)
+
beta_start
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
# Rescale for zero SNR
if
rescale_betas_zero_snr
:
...
...
src/diffusers/schedulers/scheduling_ddim.py
View file @
db33af06
...
...
@@ -211,7 +211,7 @@ class DDIMScheduler(SchedulerMixin, ConfigMixin):
# Glide cosine schedule
self
.
betas
=
betas_for_alpha_bar
(
num_train_timesteps
)
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
# Rescale for zero SNR
if
rescale_betas_zero_snr
:
...
...
src/diffusers/schedulers/scheduling_ddim_inverse.py
View file @
db33af06
...
...
@@ -207,7 +207,7 @@ class DDIMInverseScheduler(SchedulerMixin, ConfigMixin):
# Glide cosine schedule
self
.
betas
=
betas_for_alpha_bar
(
num_train_timesteps
)
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
# Rescale for zero SNR
if
rescale_betas_zero_snr
:
...
...
src/diffusers/schedulers/scheduling_ddim_parallel.py
View file @
db33af06
...
...
@@ -218,7 +218,7 @@ class DDIMParallelScheduler(SchedulerMixin, ConfigMixin):
# Glide cosine schedule
self
.
betas
=
betas_for_alpha_bar
(
num_train_timesteps
)
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
# Rescale for zero SNR
if
rescale_betas_zero_snr
:
...
...
src/diffusers/schedulers/scheduling_ddpm.py
View file @
db33af06
...
...
@@ -211,7 +211,7 @@ class DDPMScheduler(SchedulerMixin, ConfigMixin):
betas
=
torch
.
linspace
(
-
6
,
6
,
num_train_timesteps
)
self
.
betas
=
torch
.
sigmoid
(
betas
)
*
(
beta_end
-
beta_start
)
+
beta_start
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
# Rescale for zero SNR
if
rescale_betas_zero_snr
:
...
...
src/diffusers/schedulers/scheduling_ddpm_parallel.py
View file @
db33af06
...
...
@@ -219,7 +219,7 @@ class DDPMParallelScheduler(SchedulerMixin, ConfigMixin):
betas
=
torch
.
linspace
(
-
6
,
6
,
num_train_timesteps
)
self
.
betas
=
torch
.
sigmoid
(
betas
)
*
(
beta_end
-
beta_start
)
+
beta_start
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
# Rescale for zero SNR
if
rescale_betas_zero_snr
:
...
...
src/diffusers/schedulers/scheduling_deis_multistep.py
View file @
db33af06
...
...
@@ -152,7 +152,7 @@ class DEISMultistepScheduler(SchedulerMixin, ConfigMixin):
# Glide cosine schedule
self
.
betas
=
betas_for_alpha_bar
(
num_train_timesteps
)
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
self
.
alphas
=
1.0
-
self
.
betas
self
.
alphas_cumprod
=
torch
.
cumprod
(
self
.
alphas
,
dim
=
0
)
...
...
@@ -170,13 +170,13 @@ class DEISMultistepScheduler(SchedulerMixin, ConfigMixin):
if
algorithm_type
in
[
"dpmsolver"
,
"dpmsolver++"
]:
self
.
register_to_config
(
algorithm_type
=
"deis"
)
else
:
raise
NotImplementedError
(
f
"
{
algorithm_type
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
algorithm_type
}
is not implemented for
{
self
.
__class__
}
"
)
if
solver_type
not
in
[
"logrho"
]:
if
solver_type
in
[
"midpoint"
,
"heun"
,
"bh1"
,
"bh2"
]:
self
.
register_to_config
(
solver_type
=
"logrho"
)
else
:
raise
NotImplementedError
(
f
"solver type
{
solver_type
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"solver type
{
solver_type
}
is not implemented for
{
self
.
__class__
}
"
)
# setable values
self
.
num_inference_steps
=
None
...
...
src/diffusers/schedulers/scheduling_dpmsolver_multistep.py
View file @
db33af06
...
...
@@ -229,7 +229,7 @@ class DPMSolverMultistepScheduler(SchedulerMixin, ConfigMixin):
# Glide cosine schedule
self
.
betas
=
betas_for_alpha_bar
(
num_train_timesteps
)
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
if
rescale_betas_zero_snr
:
self
.
betas
=
rescale_zero_terminal_snr
(
self
.
betas
)
...
...
@@ -256,13 +256,13 @@ class DPMSolverMultistepScheduler(SchedulerMixin, ConfigMixin):
if
algorithm_type
==
"deis"
:
self
.
register_to_config
(
algorithm_type
=
"dpmsolver++"
)
else
:
raise
NotImplementedError
(
f
"
{
algorithm_type
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
algorithm_type
}
is not implemented for
{
self
.
__class__
}
"
)
if
solver_type
not
in
[
"midpoint"
,
"heun"
]:
if
solver_type
in
[
"logrho"
,
"bh1"
,
"bh2"
]:
self
.
register_to_config
(
solver_type
=
"midpoint"
)
else
:
raise
NotImplementedError
(
f
"
{
solver_type
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
solver_type
}
is not implemented for
{
self
.
__class__
}
"
)
if
algorithm_type
not
in
[
"dpmsolver++"
,
"sde-dpmsolver++"
]
and
final_sigmas_type
==
"zero"
:
raise
ValueError
(
...
...
src/diffusers/schedulers/scheduling_dpmsolver_multistep_flax.py
View file @
db33af06
...
...
@@ -182,9 +182,9 @@ class FlaxDPMSolverMultistepScheduler(FlaxSchedulerMixin, ConfigMixin):
# settings for DPM-Solver
if
self
.
config
.
algorithm_type
not
in
[
"dpmsolver"
,
"dpmsolver++"
]:
raise
NotImplementedError
(
f
"
{
self
.
config
.
algorithm_type
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
self
.
config
.
algorithm_type
}
is not implemented for
{
self
.
__class__
}
"
)
if
self
.
config
.
solver_type
not
in
[
"midpoint"
,
"heun"
]:
raise
NotImplementedError
(
f
"
{
self
.
config
.
solver_type
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
self
.
config
.
solver_type
}
is not implemented for
{
self
.
__class__
}
"
)
# standard deviation of the initial noise distribution
init_noise_sigma
=
jnp
.
array
(
1.0
,
dtype
=
self
.
dtype
)
...
...
src/diffusers/schedulers/scheduling_dpmsolver_multistep_inverse.py
View file @
db33af06
...
...
@@ -178,7 +178,7 @@ class DPMSolverMultistepInverseScheduler(SchedulerMixin, ConfigMixin):
# Glide cosine schedule
self
.
betas
=
betas_for_alpha_bar
(
num_train_timesteps
)
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
self
.
alphas
=
1.0
-
self
.
betas
self
.
alphas_cumprod
=
torch
.
cumprod
(
self
.
alphas
,
dim
=
0
)
...
...
@@ -196,13 +196,13 @@ class DPMSolverMultistepInverseScheduler(SchedulerMixin, ConfigMixin):
if
algorithm_type
==
"deis"
:
self
.
register_to_config
(
algorithm_type
=
"dpmsolver++"
)
else
:
raise
NotImplementedError
(
f
"
{
algorithm_type
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
algorithm_type
}
is not implemented for
{
self
.
__class__
}
"
)
if
solver_type
not
in
[
"midpoint"
,
"heun"
]:
if
solver_type
in
[
"logrho"
,
"bh1"
,
"bh2"
]:
self
.
register_to_config
(
solver_type
=
"midpoint"
)
else
:
raise
NotImplementedError
(
f
"
{
solver_type
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
solver_type
}
is not implemented for
{
self
.
__class__
}
"
)
# setable values
self
.
num_inference_steps
=
None
...
...
src/diffusers/schedulers/scheduling_dpmsolver_sde.py
View file @
db33af06
...
...
@@ -184,7 +184,7 @@ class DPMSolverSDEScheduler(SchedulerMixin, ConfigMixin):
# Glide cosine schedule
self
.
betas
=
betas_for_alpha_bar
(
num_train_timesteps
)
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
self
.
alphas
=
1.0
-
self
.
betas
self
.
alphas_cumprod
=
torch
.
cumprod
(
self
.
alphas
,
dim
=
0
)
...
...
src/diffusers/schedulers/scheduling_dpmsolver_singlestep.py
View file @
db33af06
...
...
@@ -172,7 +172,7 @@ class DPMSolverSinglestepScheduler(SchedulerMixin, ConfigMixin):
# Glide cosine schedule
self
.
betas
=
betas_for_alpha_bar
(
num_train_timesteps
)
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
self
.
alphas
=
1.0
-
self
.
betas
self
.
alphas_cumprod
=
torch
.
cumprod
(
self
.
alphas
,
dim
=
0
)
...
...
@@ -190,12 +190,12 @@ class DPMSolverSinglestepScheduler(SchedulerMixin, ConfigMixin):
if
algorithm_type
==
"deis"
:
self
.
register_to_config
(
algorithm_type
=
"dpmsolver++"
)
else
:
raise
NotImplementedError
(
f
"
{
algorithm_type
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
algorithm_type
}
is not implemented for
{
self
.
__class__
}
"
)
if
solver_type
not
in
[
"midpoint"
,
"heun"
]:
if
solver_type
in
[
"logrho"
,
"bh1"
,
"bh2"
]:
self
.
register_to_config
(
solver_type
=
"midpoint"
)
else
:
raise
NotImplementedError
(
f
"
{
solver_type
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
solver_type
}
is not implemented for
{
self
.
__class__
}
"
)
if
algorithm_type
!=
"dpmsolver++"
and
final_sigmas_type
==
"zero"
:
raise
ValueError
(
...
...
src/diffusers/schedulers/scheduling_edm_dpmsolver_multistep.py
View file @
db33af06
...
...
@@ -119,7 +119,7 @@ class EDMDPMSolverMultistepScheduler(SchedulerMixin, ConfigMixin):
if
solver_type
in
[
"logrho"
,
"bh1"
,
"bh2"
]:
self
.
register_to_config
(
solver_type
=
"midpoint"
)
else
:
raise
NotImplementedError
(
f
"
{
solver_type
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
solver_type
}
is not implemented for
{
self
.
__class__
}
"
)
if
algorithm_type
not
in
[
"dpmsolver++"
,
"sde-dpmsolver++"
]
and
final_sigmas_type
==
"zero"
:
raise
ValueError
(
...
...
src/diffusers/schedulers/scheduling_euler_ancestral_discrete.py
View file @
db33af06
...
...
@@ -190,7 +190,7 @@ class EulerAncestralDiscreteScheduler(SchedulerMixin, ConfigMixin):
# Glide cosine schedule
self
.
betas
=
betas_for_alpha_bar
(
num_train_timesteps
)
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
if
rescale_betas_zero_snr
:
self
.
betas
=
rescale_zero_terminal_snr
(
self
.
betas
)
...
...
src/diffusers/schedulers/scheduling_euler_discrete.py
View file @
db33af06
...
...
@@ -205,7 +205,7 @@ class EulerDiscreteScheduler(SchedulerMixin, ConfigMixin):
# Glide cosine schedule
self
.
betas
=
betas_for_alpha_bar
(
num_train_timesteps
)
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
if
rescale_betas_zero_snr
:
self
.
betas
=
rescale_zero_terminal_snr
(
self
.
betas
)
...
...
src/diffusers/schedulers/scheduling_heun_discrete.py
View file @
db33af06
...
...
@@ -135,7 +135,7 @@ class HeunDiscreteScheduler(SchedulerMixin, ConfigMixin):
elif
beta_schedule
==
"exp"
:
self
.
betas
=
betas_for_alpha_bar
(
num_train_timesteps
,
alpha_transform_type
=
"exp"
)
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
self
.
alphas
=
1.0
-
self
.
betas
self
.
alphas_cumprod
=
torch
.
cumprod
(
self
.
alphas
,
dim
=
0
)
...
...
src/diffusers/schedulers/scheduling_k_dpm_2_ancestral_discrete.py
View file @
db33af06
...
...
@@ -129,7 +129,7 @@ class KDPM2AncestralDiscreteScheduler(SchedulerMixin, ConfigMixin):
# Glide cosine schedule
self
.
betas
=
betas_for_alpha_bar
(
num_train_timesteps
)
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
self
.
alphas
=
1.0
-
self
.
betas
self
.
alphas_cumprod
=
torch
.
cumprod
(
self
.
alphas
,
dim
=
0
)
...
...
src/diffusers/schedulers/scheduling_k_dpm_2_discrete.py
View file @
db33af06
...
...
@@ -128,7 +128,7 @@ class KDPM2DiscreteScheduler(SchedulerMixin, ConfigMixin):
# Glide cosine schedule
self
.
betas
=
betas_for_alpha_bar
(
num_train_timesteps
)
else
:
raise
NotImplementedError
(
f
"
{
beta_schedule
}
does
is not implemented for
{
self
.
__class__
}
"
)
raise
NotImplementedError
(
f
"
{
beta_schedule
}
is not implemented for
{
self
.
__class__
}
"
)
self
.
alphas
=
1.0
-
self
.
betas
self
.
alphas_cumprod
=
torch
.
cumprod
(
self
.
alphas
,
dim
=
0
)
...
...
Prev
1
2
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment