"git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "d3adb985d1fb97f68acc2e4b8ae54f2bfca05fda"
Unverified Commit 2b8b6c92 authored by Sam Shleifer's avatar Sam Shleifer Committed by GitHub
Browse files

[cleanup] PretrainedModel.generate: remove unused kwargs (#4761)

parent 5bf9afbf
...@@ -1143,9 +1143,7 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin): ...@@ -1143,9 +1143,7 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin):
repetition_penalty=repetition_penalty, repetition_penalty=repetition_penalty,
no_repeat_ngram_size=no_repeat_ngram_size, no_repeat_ngram_size=no_repeat_ngram_size,
bad_words_ids=bad_words_ids, bad_words_ids=bad_words_ids,
bos_token_id=bos_token_id,
pad_token_id=pad_token_id, pad_token_id=pad_token_id,
decoder_start_token_id=decoder_start_token_id,
eos_token_id=eos_token_id, eos_token_id=eos_token_id,
batch_size=effective_batch_size, batch_size=effective_batch_size,
num_return_sequences=num_return_sequences, num_return_sequences=num_return_sequences,
...@@ -1170,9 +1168,7 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin): ...@@ -1170,9 +1168,7 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin):
repetition_penalty=repetition_penalty, repetition_penalty=repetition_penalty,
no_repeat_ngram_size=no_repeat_ngram_size, no_repeat_ngram_size=no_repeat_ngram_size,
bad_words_ids=bad_words_ids, bad_words_ids=bad_words_ids,
bos_token_id=bos_token_id,
pad_token_id=pad_token_id, pad_token_id=pad_token_id,
decoder_start_token_id=decoder_start_token_id,
eos_token_id=eos_token_id, eos_token_id=eos_token_id,
batch_size=effective_batch_size, batch_size=effective_batch_size,
encoder_outputs=encoder_outputs, encoder_outputs=encoder_outputs,
...@@ -1196,10 +1192,8 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin): ...@@ -1196,10 +1192,8 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin):
repetition_penalty, repetition_penalty,
no_repeat_ngram_size, no_repeat_ngram_size,
bad_words_ids, bad_words_ids,
bos_token_id,
pad_token_id, pad_token_id,
eos_token_id, eos_token_id,
decoder_start_token_id,
batch_size, batch_size,
encoder_outputs, encoder_outputs,
attention_mask, attention_mask,
...@@ -1318,10 +1312,8 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin): ...@@ -1318,10 +1312,8 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin):
repetition_penalty, repetition_penalty,
no_repeat_ngram_size, no_repeat_ngram_size,
bad_words_ids, bad_words_ids,
bos_token_id,
pad_token_id, pad_token_id,
eos_token_id, eos_token_id,
decoder_start_token_id,
batch_size, batch_size,
num_return_sequences, num_return_sequences,
length_penalty, length_penalty,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment