Unverified Commit f1b938fd authored by Sylvain Gugger's avatar Sylvain Gugger Committed by GitHub
Browse files

Update to use datasets remove_cloumns method (#11343)

* Update to use datasets remove_cloumns method

* Quality
parent cfd2eaa8
...@@ -16,13 +16,10 @@ ...@@ -16,13 +16,10 @@
A subclass of `Trainer` specific to Question-Answering tasks A subclass of `Trainer` specific to Question-Answering tasks
""" """
from transformers import Trainer, is_datasets_available, is_torch_tpu_available from transformers import Trainer, is_torch_tpu_available
from transformers.trainer_utils import PredictionOutput from transformers.trainer_utils import PredictionOutput
if is_datasets_available():
import datasets
if is_torch_tpu_available(): if is_torch_tpu_available():
import torch_xla.core.xla_model as xm import torch_xla.core.xla_model as xm
import torch_xla.debug.metrics as met import torch_xla.debug.metrics as met
...@@ -54,10 +51,6 @@ class QuestionAnsweringTrainer(Trainer): ...@@ -54,10 +51,6 @@ class QuestionAnsweringTrainer(Trainer):
finally: finally:
self.compute_metrics = compute_metrics self.compute_metrics = compute_metrics
# We might have removed columns from the dataset so we put them back.
if isinstance(eval_dataset, datasets.Dataset):
eval_dataset.set_format(type=eval_dataset.format["type"], columns=list(eval_dataset.features.keys()))
if self.post_process_function is not None and self.compute_metrics is not None: if self.post_process_function is not None and self.compute_metrics is not None:
eval_preds = self.post_process_function(eval_examples, eval_dataset, output.predictions) eval_preds = self.post_process_function(eval_examples, eval_dataset, output.predictions)
metrics = self.compute_metrics(eval_preds) metrics = self.compute_metrics(eval_preds)
...@@ -94,10 +87,6 @@ class QuestionAnsweringTrainer(Trainer): ...@@ -94,10 +87,6 @@ class QuestionAnsweringTrainer(Trainer):
if self.post_process_function is None or self.compute_metrics is None: if self.post_process_function is None or self.compute_metrics is None:
return output return output
# We might have removed columns from the dataset so we put them back.
if isinstance(test_dataset, datasets.Dataset):
test_dataset.set_format(type=test_dataset.format["type"], columns=list(test_dataset.features.keys()))
eval_preds = self.post_process_function(test_examples, test_dataset, output.predictions, "test") eval_preds = self.post_process_function(test_examples, test_dataset, output.predictions, "test")
metrics = self.compute_metrics(eval_preds) metrics = self.compute_metrics(eval_preds)
......
...@@ -394,11 +394,6 @@ class Trainer: ...@@ -394,11 +394,6 @@ class Trainer:
raise ValueError("train_dataset does not implement __len__, max_steps has to be specified") raise ValueError("train_dataset does not implement __len__, max_steps has to be specified")
self._signature_columns = None self._signature_columns = None
if is_datasets_available():
if isinstance(train_dataset, datasets.Dataset):
self._remove_unused_columns(self.train_dataset, description="training")
if isinstance(eval_dataset, datasets.Dataset):
self._remove_unused_columns(self.eval_dataset, description="evaluation")
# Mixed precision setup # Mixed precision setup
self.use_apex = False self.use_apex = False
...@@ -503,7 +498,13 @@ class Trainer: ...@@ -503,7 +498,13 @@ class Trainer:
f"`{self.model.__class__.__name__}.forward` and have been ignored: {', '.join(ignored_columns)}." f"`{self.model.__class__.__name__}.forward` and have been ignored: {', '.join(ignored_columns)}."
) )
dataset.set_format(type=dataset.format["type"], columns=columns, format_kwargs=dataset.format["format_kwargs"]) if version.parse(datasets.__version__) < version.parse("1.4.0"):
dataset.set_format(
type=dataset.format["type"], columns=columns, format_kwargs=dataset.format["format_kwargs"]
)
return dataset
else:
return dataset.remove_columns(ignored_columns)
def _get_train_sampler(self) -> Optional[torch.utils.data.sampler.Sampler]: def _get_train_sampler(self) -> Optional[torch.utils.data.sampler.Sampler]:
if not isinstance(self.train_dataset, collections.abc.Sized): if not isinstance(self.train_dataset, collections.abc.Sized):
...@@ -565,17 +566,20 @@ class Trainer: ...@@ -565,17 +566,20 @@ class Trainer:
if self.train_dataset is None: if self.train_dataset is None:
raise ValueError("Trainer: training requires a train_dataset.") raise ValueError("Trainer: training requires a train_dataset.")
if isinstance(self.train_dataset, torch.utils.data.dataset.IterableDataset): train_dataset = self.train_dataset
if is_datasets_available() and isinstance(train_dataset, datasets.Dataset):
train_dataset = self._remove_unused_columns(train_dataset, description="training")
if isinstance(train_dataset, torch.utils.data.dataset.IterableDataset):
if self.args.world_size > 1: if self.args.world_size > 1:
train_dataset = IterableDatasetShard( train_dataset = IterableDatasetShard(
self.train_dataset, train_dataset,
batch_size=self.args.train_batch_size, batch_size=self.args.train_batch_size,
drop_last=self.args.dataloader_drop_last, drop_last=self.args.dataloader_drop_last,
num_processes=self.args.world_size, num_processes=self.args.world_size,
process_index=self.args.process_index, process_index=self.args.process_index,
) )
else:
train_dataset = self.train_dataset
return DataLoader( return DataLoader(
train_dataset, train_dataset,
batch_size=self.args.train_batch_size, batch_size=self.args.train_batch_size,
...@@ -587,7 +591,7 @@ class Trainer: ...@@ -587,7 +591,7 @@ class Trainer:
train_sampler = self._get_train_sampler() train_sampler = self._get_train_sampler()
return DataLoader( return DataLoader(
self.train_dataset, train_dataset,
batch_size=self.args.train_batch_size, batch_size=self.args.train_batch_size,
sampler=train_sampler, sampler=train_sampler,
collate_fn=self.data_collator, collate_fn=self.data_collator,
...@@ -638,10 +642,11 @@ class Trainer: ...@@ -638,10 +642,11 @@ class Trainer:
""" """
if eval_dataset is None and self.eval_dataset is None: if eval_dataset is None and self.eval_dataset is None:
raise ValueError("Trainer: evaluation requires an eval_dataset.") raise ValueError("Trainer: evaluation requires an eval_dataset.")
elif is_datasets_available() and isinstance(eval_dataset, datasets.Dataset):
self._remove_unused_columns(eval_dataset, description="evaluation")
eval_dataset = eval_dataset if eval_dataset is not None else self.eval_dataset eval_dataset = eval_dataset if eval_dataset is not None else self.eval_dataset
if is_datasets_available() and isinstance(eval_dataset, datasets.Dataset):
eval_dataset = self._remove_unused_columns(eval_dataset, description="evaluation")
if isinstance(eval_dataset, torch.utils.data.dataset.IterableDataset): if isinstance(eval_dataset, torch.utils.data.dataset.IterableDataset):
if self.args.world_size > 1: if self.args.world_size > 1:
eval_dataset = IterableDatasetShard( eval_dataset = IterableDatasetShard(
...@@ -683,7 +688,7 @@ class Trainer: ...@@ -683,7 +688,7 @@ class Trainer:
``model.forward()`` method are automatically removed. It must implement :obj:`__len__`. ``model.forward()`` method are automatically removed. It must implement :obj:`__len__`.
""" """
if is_datasets_available() and isinstance(test_dataset, datasets.Dataset): if is_datasets_available() and isinstance(test_dataset, datasets.Dataset):
self._remove_unused_columns(test_dataset, description="test") test_dataset = self._remove_unused_columns(test_dataset, description="test")
if isinstance(test_dataset, torch.utils.data.dataset.IterableDataset): if isinstance(test_dataset, torch.utils.data.dataset.IterableDataset):
if self.args.world_size > 1: if self.args.world_size > 1:
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment