Unverified Commit 60d5f8f9 authored by Zach Mueller's avatar Zach Mueller Committed by GitHub
Browse files

🚨🚨🚨Deprecate `evaluation_strategy` to `eval_strategy`🚨🚨🚨 (#30190)

* Alias

* Note alias

* Tests and src

* Rest

* Clean

* Change typing?

* Fix tests

* Deprecation versions
parent c86d020e
...@@ -128,12 +128,12 @@ Rufen Sie [`~evaluate.compute`] auf `metric` auf, um die Genauigkeit Ihrer Vorhe ...@@ -128,12 +128,12 @@ Rufen Sie [`~evaluate.compute`] auf `metric` auf, um die Genauigkeit Ihrer Vorhe
... return metric.compute(predictions=predictions, references=labels) ... return metric.compute(predictions=predictions, references=labels)
``` ```
Wenn Sie Ihre Bewertungsmetriken während der Feinabstimmung überwachen möchten, geben Sie den Parameter `evaluation_strategy` in Ihren Trainingsargumenten an, um die Bewertungsmetrik am Ende jeder Epoche zu ermitteln: Wenn Sie Ihre Bewertungsmetriken während der Feinabstimmung überwachen möchten, geben Sie den Parameter `eval_strategy` in Ihren Trainingsargumenten an, um die Bewertungsmetrik am Ende jeder Epoche zu ermitteln:
```py ```py
>>> from transformers import TrainingArguments, Trainer >>> from transformers import TrainingArguments, Trainer
>>> training_args = TrainingArguments(output_dir="test_trainer", evaluation_strategy="epoch") >>> training_args = TrainingArguments(output_dir="test_trainer", eval_strategy="epoch")
``` ```
### Trainer ### Trainer
......
...@@ -145,7 +145,7 @@ arguments: ...@@ -145,7 +145,7 @@ arguments:
```py ```py
default_args = { default_args = {
"output_dir": "tmp", "output_dir": "tmp",
"evaluation_strategy": "steps", "eval_strategy": "steps",
"num_train_epochs": 1, "num_train_epochs": 1,
"log_level": "error", "log_level": "error",
"report_to": "none", "report_to": "none",
......
...@@ -270,7 +270,7 @@ At this point, only three steps remain: ...@@ -270,7 +270,7 @@ At this point, only three steps remain:
... gradient_checkpointing=True, ... gradient_checkpointing=True,
... fp16=True, ... fp16=True,
... group_by_length=True, ... group_by_length=True,
... evaluation_strategy="steps", ... eval_strategy="steps",
... per_device_eval_batch_size=8, ... per_device_eval_batch_size=8,
... save_steps=1000, ... save_steps=1000,
... eval_steps=1000, ... eval_steps=1000,
......
...@@ -221,7 +221,7 @@ At this point, only three steps remain: ...@@ -221,7 +221,7 @@ At this point, only three steps remain:
```py ```py
>>> training_args = TrainingArguments( >>> training_args = TrainingArguments(
... output_dir="my_awesome_mind_model", ... output_dir="my_awesome_mind_model",
... evaluation_strategy="epoch", ... eval_strategy="epoch",
... save_strategy="epoch", ... save_strategy="epoch",
... learning_rate=3e-5, ... learning_rate=3e-5,
... per_device_train_batch_size=32, ... per_device_train_batch_size=32,
......
...@@ -399,7 +399,7 @@ In this case the `output_dir` will also be the name of the repo where your model ...@@ -399,7 +399,7 @@ In this case the `output_dir` will also be the name of the repo where your model
... num_train_epochs=20, ... num_train_epochs=20,
... save_steps=200, ... save_steps=200,
... logging_steps=50, ... logging_steps=50,
... evaluation_strategy="steps", ... eval_strategy="steps",
... learning_rate=5e-5, ... learning_rate=5e-5,
... save_total_limit=2, ... save_total_limit=2,
... remove_unused_columns=False, ... remove_unused_columns=False,
......
...@@ -196,7 +196,7 @@ training_args = TrainingArguments( ...@@ -196,7 +196,7 @@ training_args = TrainingArguments(
per_device_eval_batch_size=32, per_device_eval_batch_size=32,
gradient_accumulation_steps=2, gradient_accumulation_steps=2,
save_total_limit=3, save_total_limit=3,
evaluation_strategy="steps", eval_strategy="steps",
eval_steps=50, eval_steps=50,
save_strategy="steps", save_strategy="steps",
save_steps=50, save_steps=50,
......
...@@ -302,7 +302,7 @@ At this point, only three steps remain: ...@@ -302,7 +302,7 @@ At this point, only three steps remain:
>>> training_args = TrainingArguments( >>> training_args = TrainingArguments(
... output_dir="my_awesome_food_model", ... output_dir="my_awesome_food_model",
... remove_unused_columns=False, ... remove_unused_columns=False,
... evaluation_strategy="epoch", ... eval_strategy="epoch",
... save_strategy="epoch", ... save_strategy="epoch",
... learning_rate=5e-5, ... learning_rate=5e-5,
... per_device_train_batch_size=16, ... per_device_train_batch_size=16,
......
...@@ -112,7 +112,7 @@ training_args = TrainingArguments( ...@@ -112,7 +112,7 @@ training_args = TrainingArguments(
fp16=True, fp16=True,
logging_dir=f"{repo_name}/logs", logging_dir=f"{repo_name}/logs",
logging_strategy="epoch", logging_strategy="epoch",
evaluation_strategy="epoch", eval_strategy="epoch",
save_strategy="epoch", save_strategy="epoch",
load_best_model_at_end=True, load_best_model_at_end=True,
metric_for_best_model="accuracy", metric_for_best_model="accuracy",
......
...@@ -249,7 +249,7 @@ At this point, only three steps remain: ...@@ -249,7 +249,7 @@ At this point, only three steps remain:
```py ```py
>>> training_args = TrainingArguments( >>> training_args = TrainingArguments(
... output_dir="my_awesome_eli5_clm-model", ... output_dir="my_awesome_eli5_clm-model",
... evaluation_strategy="epoch", ... eval_strategy="epoch",
... learning_rate=2e-5, ... learning_rate=2e-5,
... weight_decay=0.01, ... weight_decay=0.01,
... push_to_hub=True, ... push_to_hub=True,
......
...@@ -238,7 +238,7 @@ At this point, only three steps remain: ...@@ -238,7 +238,7 @@ At this point, only three steps remain:
```py ```py
>>> training_args = TrainingArguments( >>> training_args = TrainingArguments(
... output_dir="my_awesome_eli5_mlm_model", ... output_dir="my_awesome_eli5_mlm_model",
... evaluation_strategy="epoch", ... eval_strategy="epoch",
... learning_rate=2e-5, ... learning_rate=2e-5,
... num_train_epochs=3, ... num_train_epochs=3,
... weight_decay=0.01, ... weight_decay=0.01,
......
...@@ -265,7 +265,7 @@ At this point, only three steps remain: ...@@ -265,7 +265,7 @@ At this point, only three steps remain:
```py ```py
>>> training_args = TrainingArguments( >>> training_args = TrainingArguments(
... output_dir="my_awesome_swag_model", ... output_dir="my_awesome_swag_model",
... evaluation_strategy="epoch", ... eval_strategy="epoch",
... save_strategy="epoch", ... save_strategy="epoch",
... load_best_model_at_end=True, ... load_best_model_at_end=True,
... learning_rate=5e-5, ... learning_rate=5e-5,
......
...@@ -218,7 +218,7 @@ At this point, only three steps remain: ...@@ -218,7 +218,7 @@ At this point, only three steps remain:
```py ```py
>>> training_args = TrainingArguments( >>> training_args = TrainingArguments(
... output_dir="my_awesome_qa_model", ... output_dir="my_awesome_qa_model",
... evaluation_strategy="epoch", ... eval_strategy="epoch",
... learning_rate=2e-5, ... learning_rate=2e-5,
... per_device_train_batch_size=16, ... per_device_train_batch_size=16,
... per_device_eval_batch_size=16, ... per_device_eval_batch_size=16,
......
...@@ -535,7 +535,7 @@ At this point, only three steps remain: ...@@ -535,7 +535,7 @@ At this point, only three steps remain:
... per_device_train_batch_size=2, ... per_device_train_batch_size=2,
... per_device_eval_batch_size=2, ... per_device_eval_batch_size=2,
... save_total_limit=3, ... save_total_limit=3,
... evaluation_strategy="steps", ... eval_strategy="steps",
... save_strategy="steps", ... save_strategy="steps",
... save_steps=20, ... save_steps=20,
... eval_steps=20, ... eval_steps=20,
......
...@@ -187,7 +187,7 @@ At this point, only three steps remain: ...@@ -187,7 +187,7 @@ At this point, only three steps remain:
... per_device_eval_batch_size=16, ... per_device_eval_batch_size=16,
... num_train_epochs=2, ... num_train_epochs=2,
... weight_decay=0.01, ... weight_decay=0.01,
... evaluation_strategy="epoch", ... eval_strategy="epoch",
... save_strategy="epoch", ... save_strategy="epoch",
... load_best_model_at_end=True, ... load_best_model_at_end=True,
... push_to_hub=True, ... push_to_hub=True,
......
...@@ -202,7 +202,7 @@ At this point, only three steps remain: ...@@ -202,7 +202,7 @@ At this point, only three steps remain:
```py ```py
>>> training_args = Seq2SeqTrainingArguments( >>> training_args = Seq2SeqTrainingArguments(
... output_dir="my_awesome_billsum_model", ... output_dir="my_awesome_billsum_model",
... evaluation_strategy="epoch", ... eval_strategy="epoch",
... learning_rate=2e-5, ... learning_rate=2e-5,
... per_device_train_batch_size=16, ... per_device_train_batch_size=16,
... per_device_eval_batch_size=16, ... per_device_eval_batch_size=16,
......
...@@ -477,7 +477,7 @@ only look at the loss: ...@@ -477,7 +477,7 @@ only look at the loss:
... max_steps=4000, ... max_steps=4000,
... gradient_checkpointing=True, ... gradient_checkpointing=True,
... fp16=True, ... fp16=True,
... evaluation_strategy="steps", ... eval_strategy="steps",
... per_device_eval_batch_size=2, ... per_device_eval_batch_size=2,
... save_steps=1000, ... save_steps=1000,
... eval_steps=1000, ... eval_steps=1000,
......
...@@ -290,7 +290,7 @@ At this point, only three steps remain: ...@@ -290,7 +290,7 @@ At this point, only three steps remain:
... per_device_eval_batch_size=16, ... per_device_eval_batch_size=16,
... num_train_epochs=2, ... num_train_epochs=2,
... weight_decay=0.01, ... weight_decay=0.01,
... evaluation_strategy="epoch", ... eval_strategy="epoch",
... save_strategy="epoch", ... save_strategy="epoch",
... load_best_model_at_end=True, ... load_best_model_at_end=True,
... push_to_hub=True, ... push_to_hub=True,
......
...@@ -209,7 +209,7 @@ At this point, only three steps remain: ...@@ -209,7 +209,7 @@ At this point, only three steps remain:
```py ```py
>>> training_args = Seq2SeqTrainingArguments( >>> training_args = Seq2SeqTrainingArguments(
... output_dir="my_awesome_opus_books_model", ... output_dir="my_awesome_opus_books_model",
... evaluation_strategy="epoch", ... eval_strategy="epoch",
... learning_rate=2e-5, ... learning_rate=2e-5,
... per_device_train_batch_size=16, ... per_device_train_batch_size=16,
... per_device_eval_batch_size=16, ... per_device_eval_batch_size=16,
......
...@@ -354,7 +354,7 @@ Most of the training arguments are self-explanatory, but one that is quite impor ...@@ -354,7 +354,7 @@ Most of the training arguments are self-explanatory, but one that is quite impor
>>> args = TrainingArguments( >>> args = TrainingArguments(
... new_model_name, ... new_model_name,
... remove_unused_columns=False, ... remove_unused_columns=False,
... evaluation_strategy="epoch", ... eval_strategy="epoch",
... save_strategy="epoch", ... save_strategy="epoch",
... learning_rate=5e-5, ... learning_rate=5e-5,
... per_device_train_batch_size=batch_size, ... per_device_train_batch_size=batch_size,
......
...@@ -62,7 +62,7 @@ training_args = TrainingArguments( ...@@ -62,7 +62,7 @@ training_args = TrainingArguments(
per_device_eval_batch_size=16, per_device_eval_batch_size=16,
num_train_epochs=2, num_train_epochs=2,
weight_decay=0.01, weight_decay=0.01,
evaluation_strategy="epoch", eval_strategy="epoch",
save_strategy="epoch", save_strategy="epoch",
load_best_model_at_end=True, load_best_model_at_end=True,
push_to_hub=True, push_to_hub=True,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment