1. 15 Oct, 2021 7 commits
  2. 13 Oct, 2021 3 commits
  3. 12 Oct, 2021 2 commits
  4. 11 Oct, 2021 5 commits
  5. 09 Oct, 2021 1 commit
  6. 08 Oct, 2021 10 commits
    • moto's avatar
      9f9b6537
    • moto's avatar
      Default pretrained weights to eval mode (#1843) · cb77a86c
      moto authored
      cb77a86c
    • hwangjeff's avatar
      Update Tacotron2 docs (#1840) · 486022e9
      hwangjeff authored
      486022e9
    • hwangjeff's avatar
      9bbd4600
    • moto's avatar
      94027791
    • moto's avatar
      b1838cfc
    • moto's avatar
      Add license to pre-trained model doc (#1836) · 01764dee
      moto authored
      01764dee
    • moto's avatar
      a43cee71
    • moto's avatar
      Merge factory functions of pre-training model and fine-tuned model (#1830) · 3e5cbc0a
      moto authored
      This commit merges wav2vec2/hubert factory functions for pre-training and fine-tuning. In #1829, we added parameters to customize the models that are not part of architecture, and `aux_num_out` falls into this category, so it is no longer necessary to have separate functions. This concludes the wav2vec2/HuBERT API update in release 0.10.
      
      The summary of BC-breaking changes on wav2vec2 APIs between 0.9 and 0.10 (when this commit is incorporated)
      1. `Wav2Vec2Model.extract_features`
      In 0.9, it was returning the output from `FeatureExtractor` module. In 0.10, it returns the list of outputs from the intermediate layers of `TransformerEncoder` block.
      2. `wav2vec2_base(num_out: int)` -> `wav2vec2_base(<dropout_params:float>, aux_num_out: Optional[int]=None)`
          - `num_out` was renamed to `aux_num_out` and optional. If it is omitted, the resulting model does not have the linear layer for fine-tuning.
          - Added dropout parameters.
      3e5cbc0a
    • moto's avatar
      Make the core wav2vec2 factory function public (#1829) · 0582e73c
      moto authored
      This commit makes the following changes
      1. Make the factory function with full customizability public.
          i.e. `_get_model(...) -> wav2vec2_model(...)`.
      2. Change the other architecture-specific factory functions so that they accept parameters not related to the model architecture (such as dropout).
          i.e. `wav2vec2_base() -> wav2vec2_base(encoder_projection_dropout, encoder_attention_dropout, encoder_ff_interm_dropout, ...)`
      
      ### Why?
      
      While adding the pre-trained weight support, I realized that separating API for model construction and pre-trained support achieves simple code organization because of the good separation of concern. As mentioned in #1821, in this framework,
        1. Model implementation is responsible for computation logic,
        2. factory functions are responsible for customizability and model construction,
        3. and pre-trained weight API is responsible for constructing a model and loading pre-trained weights along with the complementary information (such as pre-processing and class labels).
      
      (note: for simple models, combining 1 and 2 is also okay.)
      
      This means that factory functions has to support all the customizability required by pre-trained weight API. The current implementation uses the internal function like `from .model import Wav2Vec2Model, _get_model`, which is a bit strange.
      
      This PR rectifies it by making the mother factory function public.
      This also clarifies the purpose of having the other factory functions as public API, which is just a syntax sugar for constructing un-trained model with specific architecture. So this commit also adds supplemental parameters to them.
      0582e73c
  7. 07 Oct, 2021 4 commits
  8. 06 Oct, 2021 3 commits
  9. 05 Oct, 2021 5 commits
    • moto's avatar
      Deprecate data utils (#1809) · 407df37d
      moto authored
      * Deprecate data utils
      
      - The design criteria of diskcache_iterator and bg_iterator are not well-specified
      - The implementation does not improve the performance due to GIL and threading
      407df37d
    • moto's avatar
      Deprecate VCTK (#1810) · 93e7f02f
      moto authored
      93e7f02f
    • moto's avatar
      Fix HuBERT xlarge configuration and test (#1811) · 5b07c33e
      moto authored
      1. Fix the HuBERT xlarge model config
      2. In the 48 transformer layers of HuBERT xlarge model, very few elements deviate from the equivalent model of fairseq, and exceeds the default atol 1e-5. This commit relax it to 3e-5 for the specific test.
      5b07c33e
    • moto's avatar
      Skip hubert_xlarge TS test on Windows (#1807) · 3b292ce3
      moto authored
      Writing scripted HuBERT XLarge models fail on Windows CI.
      3b292ce3
    • moto's avatar
      Rename factory functions `wav2vec2_asr_ARCH` to `wav2vec2_ft_ARCH` (#1804) · dacd3fd4
      moto authored
      * Rename factory functions `wav2vec2_asr_ARCH` to `wav2vec2_ft_ARCH`
      
      In #1783, we split the factory functions of wav2vec2 into ones for pretraining models
      and ones for fine-tuning models (pretraining model + extra Linear module).
      
      I picked the name scheme `wav2vec2_asr_ARCH` for factory functions of fine-tuning models,
      but did not feel right, because the architecture code is more generic.
      Even though the resulting model architecture was used for  ASR fine-tuning in the paper, 
      it does not have to be ASR.
      This became more evident as we add pre-trained parameters support, such as #1799.
      It matters more for the weight files that for which task and on which dataset it was
      trained on. For factory function, ASR task is not relevant.
      
      Therefore renaming the functions by replacing `_asr_` to `_ft_` fine-tuning.
      
      Note: Since the new functions are not release yet, this PR itself is not BC-breaking.
      dacd3fd4