Unverified Commit 6cf0be5d authored by Sayak Paul's avatar Sayak Paul Committed by GitHub
Browse files

fix warning log for Transformer SD3 (#8496)

fix warning log
parent ec068f9b
...@@ -282,9 +282,10 @@ class SD3Transformer2DModel(ModelMixin, ConfigMixin, PeftAdapterMixin, FromOrigi ...@@ -282,9 +282,10 @@ class SD3Transformer2DModel(ModelMixin, ConfigMixin, PeftAdapterMixin, FromOrigi
# weight the lora layers by setting `lora_scale` for each PEFT layer # weight the lora layers by setting `lora_scale` for each PEFT layer
scale_lora_layers(self, lora_scale) scale_lora_layers(self, lora_scale)
else: else:
logger.warning( if joint_attention_kwargs is not None and joint_attention_kwargs.get("scale", None) is not None:
"Passing `scale` via `joint_attention_kwargs` when not using the PEFT backend is ineffective." logger.warning(
) "Passing `scale` via `joint_attention_kwargs` when not using the PEFT backend is ineffective."
)
height, width = hidden_states.shape[-2:] height, width = hidden_states.shape[-2:]
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment