Unverified Commit 78031c29 authored by SahilCarterr's avatar SahilCarterr Committed by GitHub
Browse files

[Fix] enable_xformers_memory_efficient_attention() in Flux Pipeline (#12337)

* FIxes enable_xformers_memory_efficient_attention()

* Update attention.py
parent d83d35c1
......@@ -241,7 +241,7 @@ class AttentionModuleMixin:
op_fw, op_bw = attention_op
dtype, *_ = op_fw.SUPPORTED_DTYPES
q = torch.randn((1, 2, 40), device="cuda", dtype=dtype)
_ = xops.memory_efficient_attention(q, q, q)
_ = xops.ops.memory_efficient_attention(q, q, q)
except Exception as e:
raise e
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment