"...git@developer.sourcefind.cn:renzhc/diffusers_dcu.git" did not exist on "8957324363d8b239d82db4909fbf8c0875683e3d"
Fixing key padding mask during transformer generation
Summary: https://github.com/pytorch/fairseq/pull/1097 added key padding mask history in TransformerDecoderLayer, but during an edge case where only the current or only the previous key_padding_mask exists, the resulting key_padding_mask is the wrong size. This diff adds empty columns in such a case to ensure key_padding_mask is a usable size. Reviewed By: myleott Differential Revision: D18224313 fbshipit-source-id: c9fb7266baf0a2d79a66704e00a5ea8bd2987ff6
Showing
Please register or sign in to comment