"git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "9e5c4d39ab4300a22f2baa329e57e302748a1eec"
Remove the need for `einsum` in Albert's attention computation (#12394)
* debug albert einsum * Fix matmul computation * Let's use torch linear layer. * Style.
Showing
Please register or sign in to comment