Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
renzhc
diffusers_dcu
Commits
f6f16a0c
Unverified
Commit
f6f16a0c
authored
Sep 03, 2024
by
Dhruv Nair
Committed by
GitHub
Sep 03, 2024
Browse files
[CI] More Fast GPU Test Fixes (#9346)
* update * update * update * update
parent
1c1ccaa0
Changes
5
Hide whitespace changes
Inline
Side-by-side
Showing
5 changed files
with
9 additions
and
1 deletion
+9
-1
tests/lora/test_lora_layers_sd.py
tests/lora/test_lora_layers_sd.py
+2
-1
tests/models/autoencoders/test_models_vae.py
tests/models/autoencoders/test_models_vae.py
+4
-0
tests/models/test_modeling_common.py
tests/models/test_modeling_common.py
+1
-0
tests/models/transformers/test_models_transformer_cogvideox.py
.../models/transformers/test_models_transformer_cogvideox.py
+1
-0
tests/models/transformers/test_models_transformer_lumina.py
tests/models/transformers/test_models_transformer_lumina.py
+1
-0
No files found.
tests/lora/test_lora_layers_sd.py
View file @
f6f16a0c
...
...
@@ -157,11 +157,12 @@ class StableDiffusionLoRATests(PeftLoraLoaderMixinTests, unittest.TestCase):
if
(
"adapter-1"
in
n
or
"adapter-2"
in
n
)
and
not
isinstance
(
m
,
(
nn
.
Dropout
,
nn
.
Identity
)):
self
.
assertTrue
(
m
.
weight
.
device
!=
torch
.
device
(
"cpu"
))
@
slow
@
require_torch_gpu
def
test_integration_move_lora_dora_cpu
(
self
):
from
peft
import
LoraConfig
path
=
"
runwayml/stable-diffusion-v1-5
"
path
=
"
Lykon/dreamshaper-8
"
unet_lora_config
=
LoraConfig
(
init_lora_weights
=
"gaussian"
,
target_modules
=
[
"to_k"
,
"to_q"
,
"to_v"
,
"to_out.0"
],
...
...
tests/models/autoencoders/test_models_vae.py
View file @
f6f16a0c
...
...
@@ -528,6 +528,10 @@ class AutoencoderOobleckTests(ModelTesterMixin, UNetTesterMixin, unittest.TestCa
def
test_forward_with_norm_groups
(
self
):
pass
@
unittest
.
skip
(
"No attention module used in this model"
)
def
test_set_attn_processor_for_determinism
(
self
):
return
@
slow
class
AutoencoderTinyIntegrationTests
(
unittest
.
TestCase
):
...
...
tests/models/test_modeling_common.py
View file @
f6f16a0c
...
...
@@ -220,6 +220,7 @@ class ModelTesterMixin:
base_precision
=
1e-3
forward_requires_fresh_args
=
False
model_split_percents
=
[
0.5
,
0.7
,
0.9
]
uses_custom_attn_processor
=
False
def
check_device_map_is_respected
(
self
,
model
,
device_map
):
for
param_name
,
param
in
model
.
named_parameters
():
...
...
tests/models/transformers/test_models_transformer_cogvideox.py
View file @
f6f16a0c
...
...
@@ -32,6 +32,7 @@ enable_full_determinism()
class
CogVideoXTransformerTests
(
ModelTesterMixin
,
unittest
.
TestCase
):
model_class
=
CogVideoXTransformer3DModel
main_input_name
=
"hidden_states"
uses_custom_attn_processor
=
True
@
property
def
dummy_input
(
self
):
...
...
tests/models/transformers/test_models_transformer_lumina.py
View file @
f6f16a0c
...
...
@@ -32,6 +32,7 @@ enable_full_determinism()
class
LuminaNextDiT2DModelTransformerTests
(
ModelTesterMixin
,
unittest
.
TestCase
):
model_class
=
LuminaNextDiT2DModel
main_input_name
=
"hidden_states"
uses_custom_attn_processor
=
True
@
property
def
dummy_input
(
self
):
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment