1. 28 Sep, 2021 1 commit
    • moto's avatar
      Add HuBERT model architectures (#1769) · a7854f33
      moto authored
      This commit adds the following HuBERT model architectures
      
       - `base` (pre-training)
       - `large` (pre-training / fine-tuning)
       - `xlarge` (pre-training / fine-tuning)
      
      Since the internal components are same as `Wav2Vec2Model`, it reuses the existing modules..
      With these models, it is possible to 
      - import the pre-trained model published by `fairseq` and TorchScript it.
      - fine-tune the existing model for downstream task.
      a7854f33
  2. 24 Sep, 2021 1 commit
    • moto's avatar
      [BC-Breaking] Split pretraining and finetuning factory functions (#1783) · b2e9f1e4
      moto authored
      * [BC-Breaking] Split pretraining and finetuning factory functions
      
      Previously, factory functions of wav2vec2 only generated the architecture
      for the fine-tuning architecture used in wav2ve2 paper for ASR task.
      That is, pre-training architecture + Linear module, and it did not
      provide a straightforward way to generate architectures for pre-training.
      
      The goal of the original implementation was to allow the inference of
      wav2vec2 in non-Python environment via TorchScript. Now we would like to
      expand it to pre-training/fine-tuning and HuBERT model as well.
      
      Therefore, we need to have factory functions for both pre-training and
      fine-tuning. This commit introduces new factory functions and separate
      functions for pre-training and fine-tuning.
      
      1. New functions for ASR fine-tuning.
      
      We introdcue `wav2vec2_asr_XXX` functions which generates the architecture
      used for the fine-tuning task in wav2vec2 paper. *1
      
      2. Re-purpse the old functions
      
      The existing functions, `wav2vec2_XXX`, now generates the architecture with
      pre-trainig module only. (no Linear module)
      
      Note
      *1 This architecture is just one way to define architecture for fine-tuning
      and it is not universal definition. The new `wav2vec2_asr_XXX` functions are
      designed to provide these specific fine-tuning configuration and they are not
      meant to support generic architecture for downstream task.
      b2e9f1e4
  3. 17 Sep, 2021 1 commit
  4. 23 Aug, 2021 1 commit
  5. 18 Aug, 2021 1 commit
  6. 20 Jul, 2021 1 commit
  7. 03 Jun, 2021 1 commit
    • moto's avatar
      Update docs (#1550) · 0166a851
      moto authored
      * Use `bibtex` for paper citations.
        * add `override.css` for fixing back reference.
        * wav2vec2
        * wav2letter
        * convtasnet
        * deepspeech
        * rnnt-loss
        * griffinlim
      * Fix broken references in `filtering`.
      * Fix note in soundfile backends.
      * Tweak wav2vec2 example.
      * Removes unused `pytorch_theme.css`
      0166a851
  8. 01 Jun, 2021 1 commit
  9. 27 May, 2021 2 commits
  10. 11 May, 2021 1 commit
  11. 01 Oct, 2020 1 commit
  12. 29 Jul, 2020 1 commit
  13. 28 Apr, 2020 1 commit
    • Tomás Osório's avatar
      Add model Wav2Letter (#462) · d678357f
      Tomás Osório authored
      * add wav2letter model
      
      * add unit_test to model
      
      * add docstrings
      
      * add documentation
      
      * fix minor error, change logic on forward
      
      * update padding same with ceil
      
      * add inline typing and minor fixes to docstrings
      
      * remove python2
      
      * add formula do docstrings, change param name
      
      * add test with mfcc, add pytest
      
      * fix bug, update docstrings
      
      * change parameter name
      d678357f