"git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "3bb44662602b06ffd3ca25b585b5eb515bc22995"
Commit 85788bae authored by Kiyoung Kim's avatar Kiyoung Kim Committed by Lysandre
Browse files

Revert "Gradient accumulation for TFTrainer (#9585)"

This reverts commit 3f40070c.
parent 82498cbc
...@@ -638,9 +638,7 @@ class TFTrainer: ...@@ -638,9 +638,7 @@ class TFTrainer:
reduced_features = { reduced_features = {
k: ft[: self.args.train_batch_size // self.args.n_replicas] for k, ft in features.items() k: ft[: self.args.train_batch_size // self.args.n_replicas] for k, ft in features.items()
} }
reduced_labels = { reduced_labels = labels[: self.args.train_batch_size // self.args.n_replicas]
k: lbl[: self.args.train_batch_size // self.args.n_replicas] for k, lbl in labels.items()
}
self.training_step(reduced_features, reduced_labels, nb_instances_in_global_batch) self.training_step(reduced_features, reduced_labels, nb_instances_in_global_batch)
...@@ -652,13 +650,9 @@ class TFTrainer: ...@@ -652,13 +650,9 @@ class TFTrainer:
for k, ft in features.items() for k, ft in features.items()
} }
labels = { labels = tf.concat(
k: tf.concat( [labels[self.args.train_batch_size // self.args.n_replicas :], reduced_labels], axis=0
[lbl[self.args.train_batch_size // self.args.n_replicas :], reduced_labels[k]], )
axis=0,
)
for k, lbl in labels.items()
}
gradients = self.gradient_accumulator.gradients gradients = self.gradient_accumulator.gradients
gradients = [ gradients = [
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment