- 16 Oct, 2024 1 commit
-
-
Xiaodong Wang authored
Differential Revision: D64184710 Pull Request resolved: https://github.com/pytorch/audio/pull/3840
-
- 12 Oct, 2023 1 commit
-
-
moto-meta authored
Differential Revision: D50086556 Pull Request resolved: https://github.com/pytorch/audio/pull/3648
-
- 09 Oct, 2023 1 commit
-
-
moto-meta authored
Differential Revision: D49965263 Pull Request resolved: https://github.com/pytorch/audio/pull/3639
-
- 30 Aug, 2023 1 commit
-
-
atalman authored
Summary: This reverts commit c5939616. Unblock 2.1.0 rc Pull Request resolved: https://github.com/pytorch/audio/pull/3586 Reviewed By: osalpekar Differential Revision: D48842032 Pulled By: atalman fbshipit-source-id: bbdf9e45c9aa5fde00f315a2ff491ed050bc1707
-
- 19 Aug, 2023 1 commit
-
-
Juan Villamizar authored
Summary: Added HIPIFY code and small changes for ROCm. Targeting RNN-T loss. Pull Request resolved: https://github.com/pytorch/audio/pull/2485 Reviewed By: huangruizhe Differential Revision: D43537864 Pulled By: mthrok fbshipit-source-id: 4bdb1f291dc51a12232ccd072b97ae94ae20cc0c
-
- 08 Nov, 2022 1 commit
-
-
Caroline Chen authored
Summary: Add `fused_log_softmax` argument (default/current behavior = True) to rnnt loss. If setting it to `False`, call `log_softmax` on the logits prior to passing it in to the rnnt loss function. The following should produce the same output: ``` rnnt_loss(logits, targets, logit_lengths, target_lengths, fused_log_softmax=True) ``` ``` log_probs = torch.nn.functional.log_softmax(logits, dim=-1) rnnt_loss(log_probs, targets, logit_lengths, target_lengths, fused_log_softmax=False) ``` testing -- unit tests + get same results on the conformer rnnt recipe Pull Request resolved: https://github.com/pytorch/audio/pull/2798 Reviewed By: xiaohui-zhang Differential Revision: D41083523 Pulled By: carolineechen fbshipit-source-id: e15442ceed1f461bbf06b724aa0561ff8827ad61
-
- 03 Aug, 2021 1 commit
-
-
Caroline Chen authored
-
- 06 May, 2021 1 commit
-
-
Caroline Chen authored
-