[reuse] Add layers used in [Leveraging redundancy in attention with Reuse...
[reuse] Add layers used in [Leveraging redundancy in attention with Reuse Transformers](https://arxiv.org/abs/2110.06821). PiperOrigin-RevId: 408969659
Showing
This diff is collapsed.
Please register or sign in to comment