- 28 Sep, 2021 1 commit
-
-
moto authored
This commit adds the following HuBERT model architectures - `base` (pre-training) - `large` (pre-training / fine-tuning) - `xlarge` (pre-training / fine-tuning) Since the internal components are same as `Wav2Vec2Model`, it reuses the existing modules.. With these models, it is possible to - import the pre-trained model published by `fairseq` and TorchScript it. - fine-tune the existing model for downstream task.
-
- 24 Sep, 2021 1 commit
-
-
moto authored
* [BC-Breaking] Split pretraining and finetuning factory functions Previously, factory functions of wav2vec2 only generated the architecture for the fine-tuning architecture used in wav2ve2 paper for ASR task. That is, pre-training architecture + Linear module, and it did not provide a straightforward way to generate architectures for pre-training. The goal of the original implementation was to allow the inference of wav2vec2 in non-Python environment via TorchScript. Now we would like to expand it to pre-training/fine-tuning and HuBERT model as well. Therefore, we need to have factory functions for both pre-training and fine-tuning. This commit introduces new factory functions and separate functions for pre-training and fine-tuning. 1. New functions for ASR fine-tuning. We introdcue `wav2vec2_asr_XXX` functions which generates the architecture used for the fine-tuning task in wav2vec2 paper. *1 2. Re-purpse the old functions The existing functions, `wav2vec2_XXX`, now generates the architecture with pre-trainig module only. (no Linear module) Note *1 This architecture is just one way to define architecture for fine-tuning and it is not universal definition. The new `wav2vec2_asr_XXX` functions are designed to provide these specific fine-tuning configuration and they are not meant to support generic architecture for downstream task.
-
- 17 Sep, 2021 1 commit
-
-
moto authored
-
- 23 Aug, 2021 1 commit
-
-
yangarbiter authored
-
- 18 Aug, 2021 1 commit
-
-
yangarbiter authored
-
- 20 Jul, 2021 1 commit
-
-
yangarbiter authored
-
- 03 Jun, 2021 1 commit
-
-
moto authored
* Use `bibtex` for paper citations. * add `override.css` for fixing back reference. * wav2vec2 * wav2letter * convtasnet * deepspeech * rnnt-loss * griffinlim * Fix broken references in `filtering`. * Fix note in soundfile backends. * Tweak wav2vec2 example. * Removes unused `pytorch_theme.css`
-
- 01 Jun, 2021 1 commit
-
-
moto authored
-
- 27 May, 2021 2 commits
- 11 May, 2021 1 commit
-
-
discort authored
Co-authored-by:Vincent Quenneville-Belair <vincentqb@gmail.com>
-
- 01 Oct, 2020 1 commit
-
-
moto authored
-
- 29 Jul, 2020 1 commit
-
-
jimchen90 authored
Co-authored-by:Ji Chen <jimchen90@devfair0160.h2.fair>
-
- 28 Apr, 2020 1 commit
-
-
Tomás Osório authored
* add wav2letter model * add unit_test to model * add docstrings * add documentation * fix minor error, change logic on forward * update padding same with ceil * add inline typing and minor fixes to docstrings * remove python2 * add formula do docstrings, change param name * add test with mfcc, add pytest * fix bug, update docstrings * change parameter name
-