"examples/token-classification/utils_ner.py" did not exist on "81d6841b4be25a164235975e5ebdcf99d7a26633"
Add global_attention_mask to gen_kwargs (#16485)
If global_attention_mask is found in the models inputs (used by certain models, like LED) in the prediction_step method of Seq2SeqTrainer, it is added to the gen_kwargs, which are passed to model.decode(). This allows us to properly set the global attention when decoding.
Showing
Please register or sign in to comment