Commit c81fed46 authored by Myle Ott's avatar Myle Ott Committed by Facebook Github Bot
Browse files

Back out "[fairseq][PR] Fix bug (the returned value has a dimension mismatch)...

Back out "[fairseq][PR] Fix bug (the returned value has a dimension mismatch) in label-smoothed-cross-entropy for MoE" (#837)

Summary:
Pull Request resolved: https://github.com/fairinternal/fairseq-py/pull/837

Original commit changeset: a73bc03d2280

Differential Revision: D16904372

fbshipit-source-id: b4c4047b2686ba47258cdf0783059726134c920a
parent 6ce55e4b
...@@ -16,9 +16,9 @@ def label_smoothed_nll_loss(lprobs, target, epsilon, ignore_index=None, reduce=T ...@@ -16,9 +16,9 @@ def label_smoothed_nll_loss(lprobs, target, epsilon, ignore_index=None, reduce=T
nll_loss = -lprobs.gather(dim=-1, index=target) nll_loss = -lprobs.gather(dim=-1, index=target)
smooth_loss = -lprobs.sum(dim=-1, keepdim=True) smooth_loss = -lprobs.sum(dim=-1, keepdim=True)
if ignore_index is not None: if ignore_index is not None:
pad_mask = target.eq(ignore_index) non_pad_mask = target.ne(ignore_index)
nll_loss[pad_mask] = nll_loss[pad_mask] * 0. nll_loss = nll_loss[non_pad_mask]
smooth_loss[pad_mask] = smooth_loss[pad_mask] * 0. smooth_loss = smooth_loss[non_pad_mask]
else: else:
nll_loss = nll_loss.squeeze(-1) nll_loss = nll_loss.squeeze(-1)
smooth_loss = smooth_loss.squeeze(-1) smooth_loss = smooth_loss.squeeze(-1)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment