Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
renzhc
diffusers_dcu
Commits
393aefcd
Unverified
Commit
393aefcd
authored
May 08, 2025
by
Sayak Paul
Committed by
GitHub
May 08, 2025
Browse files
[tests] fix audioldm2 for transformers main. (#11522)
fix audioldm2 for transformers main.
parent
6674a515
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
13 additions
and
1 deletion
+13
-1
src/diffusers/pipelines/audioldm2/pipeline_audioldm2.py
src/diffusers/pipelines/audioldm2/pipeline_audioldm2.py
+13
-1
No files found.
src/diffusers/pipelines/audioldm2/pipeline_audioldm2.py
View file @
393aefcd
...
@@ -40,6 +40,7 @@ from ...utils import (
...
@@ -40,6 +40,7 @@ from ...utils import (
logging
,
logging
,
replace_example_docstring
,
replace_example_docstring
,
)
)
from
...utils.import_utils
import
is_transformers_version
from
...utils.torch_utils
import
randn_tensor
from
...utils.torch_utils
import
randn_tensor
from
..pipeline_utils
import
AudioPipelineOutput
,
DiffusionPipeline
from
..pipeline_utils
import
AudioPipelineOutput
,
DiffusionPipeline
from
.modeling_audioldm2
import
AudioLDM2ProjectionModel
,
AudioLDM2UNet2DConditionModel
from
.modeling_audioldm2
import
AudioLDM2ProjectionModel
,
AudioLDM2UNet2DConditionModel
...
@@ -312,8 +313,19 @@ class AudioLDM2Pipeline(DiffusionPipeline):
...
@@ -312,8 +313,19 @@ class AudioLDM2Pipeline(DiffusionPipeline):
`inputs_embeds (`torch.Tensor` of shape `(batch_size, sequence_length, hidden_size)`):
`inputs_embeds (`torch.Tensor` of shape `(batch_size, sequence_length, hidden_size)`):
The sequence of generated hidden-states.
The sequence of generated hidden-states.
"""
"""
cache_position_kwargs
=
{}
if
is_transformers_version
(
"<"
,
"4.52.0.dev0"
):
cache_position_kwargs
[
"input_ids"
]
=
inputs_embeds
cache_position_kwargs
[
"model_kwargs"
]
=
model_kwargs
else
:
cache_position_kwargs
[
"seq_length"
]
=
inputs_embeds
.
shape
[
0
]
cache_position_kwargs
[
"device"
]
=
(
self
.
language_model
.
device
if
getattr
(
self
,
"language_model"
,
None
)
is
not
None
else
self
.
device
)
cache_position_kwargs
[
"model_kwargs"
]
=
model_kwargs
max_new_tokens
=
max_new_tokens
if
max_new_tokens
is
not
None
else
self
.
language_model
.
config
.
max_new_tokens
max_new_tokens
=
max_new_tokens
if
max_new_tokens
is
not
None
else
self
.
language_model
.
config
.
max_new_tokens
model_kwargs
=
self
.
language_model
.
_get_initial_cache_position
(
inputs_embeds
,
model_kwargs
)
model_kwargs
=
self
.
language_model
.
_get_initial_cache_position
(
**
cache_position_kwargs
)
for
_
in
range
(
max_new_tokens
):
for
_
in
range
(
max_new_tokens
):
# prepare model inputs
# prepare model inputs
model_inputs
=
prepare_inputs_for_generation
(
inputs_embeds
,
**
model_kwargs
)
model_inputs
=
prepare_inputs_for_generation
(
inputs_embeds
,
**
model_kwargs
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment