Unverified Commit 31c351c4 authored by agossard's avatar agossard Committed by GitHub
Browse files

For IterableDataset, return DataLoader using self._train_batch_size. … (#21447)

For IterableDataset, return DataLoader using self._train_batch_size. This is consistent with how we generate a regular DataLoader, and leads to the correct args.per_device_train_batch_size eventually ending up on each GPU.
parent 833174c9
...@@ -868,7 +868,7 @@ class Trainer: ...@@ -868,7 +868,7 @@ class Trainer:
return DataLoader( return DataLoader(
train_dataset, train_dataset,
batch_size=self.args.per_device_train_batch_size, batch_size=self._train_batch_size,
collate_fn=data_collator, collate_fn=data_collator,
num_workers=self.args.dataloader_num_workers, num_workers=self.args.dataloader_num_workers,
pin_memory=self.args.dataloader_pin_memory, pin_memory=self.args.dataloader_pin_memory,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment