Unverified Commit adb785b0 authored by Sylvain Gugger's avatar Sylvain Gugger Committed by GitHub
Browse files

Switch mem metrics flag (#11851)



* Switch mem metrics flag

* Update src/transformers/training_args.py
Co-authored-by: default avatarStas Bekman <stas00@users.noreply.github.com>
Co-authored-by: default avatarStas Bekman <stas00@users.noreply.github.com>
parent fcdb85e9
...@@ -303,8 +303,9 @@ class TrainingArguments: ...@@ -303,8 +303,9 @@ class TrainingArguments:
otherwise. otherwise.
dataloader_pin_memory (:obj:`bool`, `optional`, defaults to :obj:`True`): dataloader_pin_memory (:obj:`bool`, `optional`, defaults to :obj:`True`):
Whether you want to pin memory in data loaders or not. Will default to :obj:`True`. Whether you want to pin memory in data loaders or not. Will default to :obj:`True`.
skip_memory_metrics (:obj:`bool`, `optional`, defaults to :obj:`False`): skip_memory_metrics (:obj:`bool`, `optional`, defaults to :obj:`True`):
Whether to skip adding of memory profiler reports to metrics. Defaults to :obj:`False`. Whether to skip adding of memory profiler reports to metrics. This is skipped by default because it slows
down the training and evaluation speed.
push_to_hub (:obj:`bool`, `optional`, defaults to :obj:`False`): push_to_hub (:obj:`bool`, `optional`, defaults to :obj:`False`):
Whether or not to upload the trained model to the hub after training. This argument is not directly used by Whether or not to upload the trained model to the hub after training. This argument is not directly used by
:class:`~transformers.Trainer`, it's intended to be used by your training/evaluation scripts instead. See :class:`~transformers.Trainer`, it's intended to be used by your training/evaluation scripts instead. See
...@@ -546,7 +547,7 @@ class TrainingArguments: ...@@ -546,7 +547,7 @@ class TrainingArguments:
default=True, metadata={"help": "Whether or not to pin memory for DataLoader."} default=True, metadata={"help": "Whether or not to pin memory for DataLoader."}
) )
skip_memory_metrics: bool = field( skip_memory_metrics: bool = field(
default=False, metadata={"help": "Whether or not to skip adding of memory profiler reports to metrics."} default=True, metadata={"help": "Whether or not to skip adding of memory profiler reports to metrics."}
) )
use_legacy_prediction_loop: bool = field( use_legacy_prediction_loop: bool = field(
default=False, metadata={"help": "Whether or not to use the legacy prediction_loop in the Trainer."} default=False, metadata={"help": "Whether or not to use the legacy prediction_loop in the Trainer."}
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment