1. 23 Dec, 2021 3 commits
  2. 24 Nov, 2021 1 commit
    • hwangjeff's avatar
      Add RNN-T beam search decoder (#2028) · 60a85b50
      hwangjeff authored
      Summary:
      Adds beam search decoder for RNN-T implementation ``torchaudio.prototype.RNNT`` that is TorchScript-able and supports both streaming and non-streaming inference.
      
      Pull Request resolved: https://github.com/pytorch/audio/pull/2028
      
      Reviewed By: mthrok
      
      Differential Revision: D32627919
      
      Pulled By: hwangjeff
      
      fbshipit-source-id: aab99e346d6514a3207a9fb69d4b42978b4cdbbd
      60a85b50
  3. 23 Nov, 2021 1 commit
  4. 19 Nov, 2021 1 commit
    • moto's avatar
      Disable SPHINXOPT=-W for local env (#2013) · 3ff46bfa
      moto authored
      Summary:
      With the introduction of tutorials, the turn around time for doc build
      has become longer. By default, the tutorial is not built but SPHINXOPT=-W
      treats it as error.
      
      This commit disable the option for the local build while keeping it
      for the CI.
      
      Pull Request resolved: https://github.com/pytorch/audio/pull/2013
      
      Reviewed By: carolineechen
      
      Differential Revision: D32538952
      
      Pulled By: mthrok
      
      fbshipit-source-id: eae4ffd87100dff466f91abfe26a82aa702d605a
      3ff46bfa
  5. 18 Nov, 2021 1 commit
    • hwangjeff's avatar
      Add Emformer RNN-T model (#2003) · 78ce7010
      hwangjeff authored
      Summary:
      Adds streaming-capable recurrent neural network transducer (RNN-T) model that uses Emformer for its transcription network. Includes two factory functions — one that allows for building a custom model, and one that builds a preconfigured base model.
      
      Pull Request resolved: https://github.com/pytorch/audio/pull/2003
      
      Reviewed By: nateanl
      
      Differential Revision: D32440879
      
      Pulled By: hwangjeff
      
      fbshipit-source-id: 601cb1de368427f25e3b7d120e185960595d2360
      78ce7010
  6. 10 Nov, 2021 1 commit
  7. 05 Nov, 2021 4 commits
  8. 04 Nov, 2021 6 commits
  9. 03 Nov, 2021 1 commit
  10. 02 Nov, 2021 4 commits
  11. 29 Oct, 2021 1 commit
  12. 28 Oct, 2021 1 commit
  13. 27 Oct, 2021 2 commits
  14. 26 Oct, 2021 1 commit
  15. 25 Oct, 2021 1 commit
  16. 18 Oct, 2021 2 commits
  17. 16 Oct, 2021 1 commit
  18. 15 Oct, 2021 5 commits
  19. 08 Oct, 2021 2 commits
  20. 07 Oct, 2021 1 commit
    • moto's avatar
      Merge factory functions of pre-training model and fine-tuned model (#1830) · 274ada80
      moto authored
      This commit merges wav2vec2/hubert factory functions for pre-training and fine-tuning. In #1829, we added parameters to customize the models that are not part of architecture, and `aux_num_out` falls into this category, so it is no longer necessary to have separate functions. This concludes the wav2vec2/HuBERT API update in release 0.10.
      
      The summary of BC-breaking changes on wav2vec2 APIs between 0.9 and 0.10 (when this commit is incorporated)
      1. `Wav2Vec2Model.extract_features`
      In 0.9, it was returning the output from `FeatureExtractor` module. In 0.10, it returns the list of outputs from the intermediate layers of `TransformerEncoder` block.
      2. `wav2vec2_base(num_out: int)` -> `wav2vec2_base(<dropout_params:float>, aux_num_out: Optional[int]=None)`
          - `num_out` was renamed to `aux_num_out` and optional. If it is omitted, the resulting model does not have the linear layer for fine-tuning.
          - Added dropout parameters.
      274ada80