- 05 Oct, 2021 1 commit
-
-
moto authored
-
- 01 Oct, 2021 1 commit
-
-
moto authored
1. Fix the HuBERT xlarge model config 2. In the 48 transformer layers of HuBERT xlarge model, very few elements deviate from the equivalent model of fairseq, and exceeds the default atol 1e-5. This commit relax it to 3e-5 for the specific test.
-
- 30 Sep, 2021 1 commit
-
-
moto authored
Writing scripted HuBERT XLarge models fail on Windows CI.
-
- 29 Sep, 2021 2 commits
-
-
moto authored
* Rename factory functions `wav2vec2_asr_ARCH` to `wav2vec2_ft_ARCH` In #1783, we split the factory functions of wav2vec2 into ones for pretraining models and ones for fine-tuning models (pretraining model + extra Linear module). I picked the name scheme `wav2vec2_asr_ARCH` for factory functions of fine-tuning models, but did not feel right, because the architecture code is more generic. Even though the resulting model architecture was used for ASR fine-tuning in the paper, it does not have to be ASR. This became more evident as we add pre-trained parameters support, such as #1799. It matters more for the weight files that for which task and on which dataset it was trained on. For factory function, ASR task is not relevant. Therefore renaming the functions by replacing `_asr_` to `_ft_` fine-tuning. Note: Since the new functions are not release yet, this PR itself is not BC-breaking.
-
moto authored
-
- 28 Sep, 2021 1 commit
-
-
moto authored
This commit adds the following HuBERT model architectures - `base` (pre-training) - `large` (pre-training / fine-tuning) - `xlarge` (pre-training / fine-tuning) Since the internal components are same as `Wav2Vec2Model`, it reuses the existing modules.. With these models, it is possible to - import the pre-trained model published by `fairseq` and TorchScript it. - fine-tune the existing model for downstream task.
-
- 24 Sep, 2021 1 commit
-
-
moto authored
* [BC-Breaking] Split pretraining and finetuning factory functions Previously, factory functions of wav2vec2 only generated the architecture for the fine-tuning architecture used in wav2ve2 paper for ASR task. That is, pre-training architecture + Linear module, and it did not provide a straightforward way to generate architectures for pre-training. The goal of the original implementation was to allow the inference of wav2vec2 in non-Python environment via TorchScript. Now we would like to expand it to pre-training/fine-tuning and HuBERT model as well. Therefore, we need to have factory functions for both pre-training and fine-tuning. This commit introduces new factory functions and separate functions for pre-training and fine-tuning. 1. New functions for ASR fine-tuning. We introdcue `wav2vec2_asr_XXX` functions which generates the architecture used for the fine-tuning task in wav2vec2 paper. *1 2. Re-purpse the old functions The existing functions, `wav2vec2_XXX`, now generates the architecture with pre-trainig module only. (no Linear module) Note *1 This architecture is just one way to define architecture for fine-tuning and it is not universal definition. The new `wav2vec2_asr_XXX` functions are designed to provide these specific fine-tuning configuration and they are not meant to support generic architecture for downstream task.
-
- 22 Sep, 2021 3 commits
-
-
moto authored
Previously, the Linear module (called `readout`, which is used only for an ASR fine-tuning task) was placed in encoder module. Conceptually, the encoder has nothing to do with a module specific to fine-tuning / downstream task. The problems here are that; 1. encoder can be also used in pre-training phase, in which such a module should not present 2. The choice of Linear module is arbitral, and it is inconvenient for users to have hard-coded module structure in encoder. Therefore, this commit moves the Linear module out the encoder, and places it as `aux` attribute of `Wav2Vec2Model`. (as a result `Wav2Vec2Model` has `feature_extractor`, `encoder` and `aux` attributes.) An alternative approach is to define another module and place `Wav2Vec2Model` and aux module along each other. But that will introduce a new class we need to maintain. The expected use of `aux` is only for 1. loading the pre-trained parameters published by `fairseq` (and it's variations from HF) and 2. creating the same model architectures for comparison experiment. The newly introduced class will not be general enough for downstream adaptations, where there will be a bunch of different more complicated models. (i.e. s3prl) Therefore, based on the minimalistic approach, we put them inside of `Wav2Vec2Model`.
-
moto authored
* Fix HF model integration Previously, when testing wav2vec models from HF transformers, all the model were instantiated as `Wav2Vec2ForCTC` class, while some of them were supposed to be `Wav2Vec2Model`. Fixing this revealed that model importer cannot correctly handle `Wav2Vec2Model` import. This PR fixes these issues.
-
moto authored
Summary: Update fairseq reference from master to main elsewhere Reviewed By: alexeib Differential Revision: D30938472 fbshipit-source-id: 243b98550207f241c9d3265bf3d4060350aaf0a8 Co-authored-by:Diana Liskovich <dianaml@fb.com>
-
- 21 Sep, 2021 1 commit
-
-
moto authored
Apply tweak around the test names so that it's easier to see which tests are failing. Before: `test_import_finetuned_model_2` After: `test_import_finetuned_model_2_wav2vec2_large_lv60k`
-
- 20 Sep, 2021 2 commits
-
-
moto authored
* [BC-Breaking] Update `extract_features` of Wav2Vec2Model Originally, `extract_features` method was returning the result from the convolutional feature extractor module. The features commonly used in downstream tasks are outputs from intermediate layers of transformer block in encoder. This commit update the behavior of `extract_features` to allow selectively retrieve such features.
-
nateanl authored
-
- 17 Sep, 2021 1 commit
-
-
nateanl authored
-
- 02 Sep, 2021 1 commit
-
-
jayleverett authored
* put output tensor on device in `get_whitenoise()` * Update `get_spectrogram()` so that window uses same device as waveform * put window on proper device in `test_griffinlim()`
-
- 27 Aug, 2021 2 commits
- 26 Aug, 2021 2 commits
- 23 Aug, 2021 1 commit
-
-
yangarbiter authored
-
- 20 Aug, 2021 1 commit
-
-
hwangjeff authored
* Add basic filtfilt implementation * Add filtfilt to functional package; add tests Co-authored-by:V G <vladislav.goncharenko@phystech.edu>
-
- 19 Aug, 2021 1 commit
-
-
Caroline Chen authored
-
- 18 Aug, 2021 1 commit
-
-
yangarbiter authored
-
- 17 Aug, 2021 1 commit
-
-
Caroline Chen authored
-
- 12 Aug, 2021 1 commit
-
-
hwangjeff authored
* Reduce length of waveform in pitch_shift batch_consistency test Summary: To address the test failure in T96406395 Reviewed By: carolineechen Differential Revision: D30163741 fbshipit-source-id: f88d86b3da7b1ee52518934567b0b0a62700ee58 * Fix batch consistency test in transforms Summary: The stress test still fails. Add n_fft to address it. Reviewed By: mthrok Differential Revision: D30218279 fbshipit-source-id: 7858efd3e5ac0073193a7883fd314486efc73814 Co-authored-by:Zhaoheng Ni <zni@fb.com>
-
- 11 Aug, 2021 1 commit
-
-
nateanl authored
- Provide InverseSpectrogram module that corresponds to Spectrogram module - Add length parameter to the forward method in transforms Co-authored-by:
dgenzel <dgenzel@fb.com> Co-authored-by:
Zhaoheng Ni <zni@fb.com>
-
- 10 Aug, 2021 2 commits
-
-
Chin-Yun Yu authored
-
yangarbiter authored
-
- 04 Aug, 2021 1 commit
-
-
moto authored
D30080845
-
- 03 Aug, 2021 2 commits
-
-
Caroline Chen authored
-
Caroline Chen authored
-
- 02 Aug, 2021 2 commits
-
-
yangarbiter authored
-
Joel Frank authored
- Renamed torchaudio.functional.create_fb_matrix to torchaudio.functional.melscale_fbanks. - Added interface with a warning for create_fb_matrix
-
- 29 Jul, 2021 1 commit
-
-
Joel Frank authored
Summary: - Add linear_fbank method - Add LFCC in transforms
-
- 28 Jul, 2021 1 commit
-
-
yangarbiter authored
-
- 26 Jul, 2021 2 commits
-
-
yangarbiter authored
-
moto authored
-
- 22 Jul, 2021 1 commit
-
-
hwangjeff authored
Rebases #1571; addresses #1569: "In 0.9.0 we are deprecating the lazy behavior of MelScale because it can make an invalid TorchScript object and it does not align with the design of torchaudio. Now in master branch, we can remove the implementation." Co-authored-by:
Pankaj Patil <pankaj.patil2099@hotmail.com> Co-authored-by:
moto <855818+mthrok@users.noreply.github.com> Co-authored-by:
hwangjeff <jeffhwang@fb.com>
-
- 21 Jul, 2021 1 commit
-
-
Chin-Yun Yu authored
-
- 20 Jul, 2021 1 commit
-
-
hwangjeff authored
-