"test/assets/git@developer.sourcefind.cn:hehl2/torchaudio.git" did not exist on "9bd633e3f92f4bbfb6bc2eb46f407018703ef75e"
Add global_attention_mask to gen_kwargs (#16485)
If global_attention_mask is found in the models inputs (used by certain models, like LED) in the prediction_step method of Seq2SeqTrainer, it is added to the gen_kwargs, which are passed to model.decode(). This allows us to properly set the global attention when decoding.
Showing
Please register or sign in to comment