1. 12 Oct, 2022 1 commit
    • Zhaoheng Ni's avatar
      Improve hubert recipe for pre-training and fine-tuning (#2744) · 928248d7
      Zhaoheng Ni authored
      Summary:
      following pr https://github.com/pytorch/audio/issues/2716
      - For preprocessing
        - The HuBERT feature takes lots of memory which may not fit some machines. Enable to use a subset of feature for training a k-means model.
      
      - For pre-training
        - Normalize the loss based on the total number of masked frames across all GPUs.
        - Use mixed precision training. fp16 is not well supported in pytorch_lightning.
        - Log accuracies of masked/unmasked frames during training.
        - Clip the gradients with norm `10.0`.
      
      - For ASR fine-tuning
        - Normalize the loss based on the total number of batches across all GPUs, same as in the conformer recipe of TorchAudio.
        - Use mixed precision training.
        - Add "|" after the end of transcription to capture the silence/word termination, same as in fairseq recipe.
      
      - Update the WER results on LibriSpeech dev and test sets.
      
      |                   | WER% (Viterbi)|  WER% (KenLM) |
      |:-----------------:|--------------:|--------------:|
      | dev-clean         |       10.9    |       4.2     |
      | dev-other         |       17.5    |       9.4     |
      | test-clean        |       10.9    |       4.4     |
      | test-other        |       17.8    |       9.5     |
      
      Pull Request resolved: https://github.com/pytorch/audio/pull/2744
      
      Reviewed By: carolineechen
      
      Differential Revision: D40282322
      
      Pulled By: nateanl
      
      fbshipit-source-id: 4723584c912e70e8970149fe09de005385eaab90
      928248d7
  2. 23 Jun, 2022 1 commit
  3. 07 Jun, 2022 1 commit
    • Zhaoheng Ni's avatar
      Add HuBERT fine-tuning recipe (#2352) · ab5edfcd
      Zhaoheng Ni authored
      Summary:
      The PR contains the CTC fine-tuning recipe of HuBERT Base model.
      The files include:
      - lightning module
      - training script
      - README and the result table
      - evaluation scripts
      
      Pull Request resolved: https://github.com/pytorch/audio/pull/2352
      
      Reviewed By: hwangjeff
      
      Differential Revision: D36915712
      
      Pulled By: nateanl
      
      fbshipit-source-id: 0249635ad5e81a8aa2d228c1d5fe84d78b62a15b
      ab5edfcd
  4. 23 May, 2022 1 commit
  5. 08 Mar, 2022 1 commit
  6. 26 Jul, 2021 1 commit
  7. 27 May, 2020 1 commit
    • moto's avatar
      Self-contain codecs library (#625) · d3c83eaa
      moto authored
      * Clean up extension build mechanism and extension location
      
      * Add back the switch to depend on external sox
      
      * Remove print
      
      * Fix
      d3c83eaa
  8. 21 Aug, 2019 1 commit
  9. 30 Jul, 2019 1 commit
    • Edward Z. Yang's avatar
      Make test scripts runnable without being modules. (#186) · 07b9b9ba
      Edward Z. Yang authored
      
      
      This makes it easier to test against an installed wheel, as the
      torchaudio folder is no longer preferentially picked up when
      you run a test module.
      
      I had to move all tests in subfolders into the top level test
      directory to make this work, since you can't access .. modules
      without mucking around with sys.path (which I don't want to do.)
      
      NB: this BREAKS the syntax where you can run a test by
      saying `python -m test.test`.  Instead, do `python test/test.py`
      or use the pytest runner.
      Signed-off-by: default avatarEdward Z. Yang <ezyang@fb.com>
      07b9b9ba
  10. 25 Dec, 2018 1 commit