Unverified Commit 31c351c4 authored by agossard's avatar agossard Committed by GitHub
Browse files

For IterableDataset, return DataLoader using self._train_batch_size. … (#21447)

For IterableDataset, return DataLoader using self._train_batch_size. This is consistent with how we generate a regular DataLoader, and leads to the correct args.per_device_train_batch_size eventually ending up on each GPU.
parent 833174c9
......@@ -868,7 +868,7 @@ class Trainer:
return DataLoader(
train_dataset,
batch_size=self.args.per_device_train_batch_size,
batch_size=self._train_batch_size,
collate_fn=data_collator,
num_workers=self.args.dataloader_num_workers,
pin_memory=self.args.dataloader_pin_memory,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment