Commit 08a74a32 authored by Myle Ott's avatar Myle Ott
Browse files

Fix gradient clipping when --clip-norm=0

parent fd28c880
...@@ -265,7 +265,10 @@ class MultiprocessingTrainer(MultiprocessingEventLoop): ...@@ -265,7 +265,10 @@ class MultiprocessingTrainer(MultiprocessingEventLoop):
self._all_reduce_and_rescale_grads(grad_denom) self._all_reduce_and_rescale_grads(grad_denom)
# clip grads # clip grads
if self.args.clip_norm > 0:
grad_norm = torch.nn.utils.clip_grad_norm(self.model.parameters(), self.args.clip_norm) grad_norm = torch.nn.utils.clip_grad_norm(self.model.parameters(), self.args.clip_norm)
else:
grad_norm = math.sqrt(sum([p.grad.data.norm()**2 for p in self.model.parameters()]))
# take an optimization step # take an optimization step
self.optimizer.step() self.optimizer.step()
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment