Refactor the attention functions.
There's no reason for the whole CrossAttention object to be repeated when only the operation in the middle changes.
Showing
Please register or sign in to comment
There's no reason for the whole CrossAttention object to be repeated when only the operation in the middle changes.