- 08 Oct, 2021 1 commit
-
-
moto authored
-
- 07 Oct, 2021 7 commits
-
-
Caroline Chen authored
-
nateanl authored
-
moto authored
-
Caroline Chen authored
-
moto authored
This commit merges wav2vec2/hubert factory functions for pre-training and fine-tuning. In #1829, we added parameters to customize the models that are not part of architecture, and `aux_num_out` falls into this category, so it is no longer necessary to have separate functions. This concludes the wav2vec2/HuBERT API update in release 0.10. The summary of BC-breaking changes on wav2vec2 APIs between 0.9 and 0.10 (when this commit is incorporated) 1. `Wav2Vec2Model.extract_features` In 0.9, it was returning the output from `FeatureExtractor` module. In 0.10, it returns the list of outputs from the intermediate layers of `TransformerEncoder` block. 2. `wav2vec2_base(num_out: int)` -> `wav2vec2_base(<dropout_params:float>, aux_num_out: Optional[int]=None)` - `num_out` was renamed to `aux_num_out` and optional. If it is omitted, the resulting model does not have the linear layer for fine-tuning. - Added dropout parameters. -
moto authored
-
moto authored
This commit makes the following changes 1. Make the factory function with full customizability public. i.e. `_get_model(...) -> wav2vec2_model(...)`. 2. Change the other architecture-specific factory functions so that they accept parameters not related to the model architecture (such as dropout). i.e. `wav2vec2_base() -> wav2vec2_base(encoder_projection_dropout, encoder_attention_dropout, encoder_ff_interm_dropout, ...)` ### Why? While adding the pre-trained weight support, I realized that separating API for model construction and pre-trained support achieves simple code organization because of the good separation of concern. As mentioned in #1821, in this framework, 1. Model implementation is responsible for computation logic, 2. factory functions are responsible for customizability and model construction, 3. and pre-trained weight API is responsible for constructing a model and loading pre-trained weights along with the complementary information (such as pre-processing and class labels). (note: for simple models, combining 1 and 2 is also okay.) This means that factory functions has to support all the customizability required by pre-trained weight API. The current implementation uses the internal function like `from .model import Wav2Vec2Model, _get_model`, which is a bit strange. This PR rectifies it by making the mother factory function public. This also clarifies the purpose of having the other factory functions as public API, which is just a syntax sugar for constructing un-trained model with specific architecture. So this commit also adds supplemental parameters to them.
-
- 06 Oct, 2021 7 commits
-
-
kingyiusuen authored
-
moto authored
Add pretrained weights from https://github.com/pytorch/fairseq/tree/main/examples/wav2vec#pre-trained-models - Wav2Vec 2.0 Base / Large / Large (LV-60) - XLSR-53
-
hwangjeff authored
Adds an implementation of Emformer, a memory-efficient transformer architecture introduced in https://ieeexplore.ieee.org/document/9414560 that targets low-latency streaming speech recognition applications.
-
moto authored
-
moto authored
This commit adds - HUBERT_LARGE - HUBERT_XLARGE - HUBERT_ASR_XLARGE
-
moto authored
-
moto authored
-
- 05 Oct, 2021 5 commits
-
-
moto authored
-
moto authored
-
moto authored
Summary: Pull Request resolved: https://github.com/pytorch/audio/pull/1817 This changes the imports in the `torchaudio` to include the new import locations. ``` codemod -d pytorch/audio --extensions py 'torch.quantization' 'torch.ao.quantization' ``` Reviewed By: mthrok Differential Revision: D31302450 fbshipit-source-id: f31a0d4f453f840ea690edb688555a9d585787b5 Co-authored-by:
Zafar Takhirov <zaf@fb.com>
-
moto authored
-
nateanl authored
-
- 01 Oct, 2021 1 commit
-
-
moto authored
1. Fix the HuBERT xlarge model config 2. In the 48 transformer layers of HuBERT xlarge model, very few elements deviate from the equivalent model of fairseq, and exceeds the default atol 1e-5. This commit relax it to 3e-5 for the specific test.
-
- 30 Sep, 2021 5 commits
-
-
moto authored
* Deprecate data utils - The design criteria of diskcache_iterator and bg_iterator are not well-specified - The implementation does not improve the performance due to GIL and threading
-
moto authored
Writing scripted HuBERT XLarge models fail on Windows CI.
-
moto authored
-
moto authored
-
Nicolas Hug authored
* Add enw issue forms * forgot one torchvision left * audio-specific bug * Apply suggestions from code review Co-authored-by:
moto <855818+mthrok@users.noreply.github.com> Co-authored-by:
moto <855818+mthrok@users.noreply.github.com>
-
- 29 Sep, 2021 4 commits
-
-
moto authored
* Rename factory functions `wav2vec2_asr_ARCH` to `wav2vec2_ft_ARCH` In #1783, we split the factory functions of wav2vec2 into ones for pretraining models and ones for fine-tuning models (pretraining model + extra Linear module). I picked the name scheme `wav2vec2_asr_ARCH` for factory functions of fine-tuning models, but did not feel right, because the architecture code is more generic. Even though the resulting model architecture was used for ASR fine-tuning in the paper, it does not have to be ASR. This became more evident as we add pre-trained parameters support, such as #1799. It matters more for the weight files that for which task and on which dataset it was trained on. For factory function, ASR task is not relevant. Therefore renaming the functions by replacing `_asr_` to `_ft_` fine-tuning. Note: Since the new functions are not release yet, this PR itself is not BC-breaking.
-
Yi Zhang authored
* 11.1.0 to 11.1.1 * Fix typo
-
Caroline Chen authored
-
moto authored
-
- 28 Sep, 2021 1 commit
-
-
moto authored
This commit adds the following HuBERT model architectures - `base` (pre-training) - `large` (pre-training / fine-tuning) - `xlarge` (pre-training / fine-tuning) Since the internal components are same as `Wav2Vec2Model`, it reuses the existing modules.. With these models, it is possible to - import the pre-trained model published by `fairseq` and TorchScript it. - fine-tune the existing model for downstream task.
-
- 27 Sep, 2021 2 commits
-
-
Nikita Shulga authored
Also update README.md to mention 1.9.1
-
Yi Zhang authored
* enable windows cudatests * add this dir * minor change * vs integration * Update cuda_install.bat * add logs * minor change * minor change * cp vision conda activate * mv vc_env_helper.bat * minor change * exit if cuda not avaiable * install numpy * improt CMakeLists * check cuda * minor change * change windows GPU image from previous to stable * set libtorch audio suffix as pyd on Windows * reduce changes * check env settings
-
- 26 Sep, 2021 1 commit
-
-
nateanl authored
-
- 25 Sep, 2021 1 commit
-
-
moto authored
-
- 24 Sep, 2021 4 commits
-
-
moto authored
* [BC-Breaking] Split pretraining and finetuning factory functions Previously, factory functions of wav2vec2 only generated the architecture for the fine-tuning architecture used in wav2ve2 paper for ASR task. That is, pre-training architecture + Linear module, and it did not provide a straightforward way to generate architectures for pre-training. The goal of the original implementation was to allow the inference of wav2vec2 in non-Python environment via TorchScript. Now we would like to expand it to pre-training/fine-tuning and HuBERT model as well. Therefore, we need to have factory functions for both pre-training and fine-tuning. This commit introduces new factory functions and separate functions for pre-training and fine-tuning. 1. New functions for ASR fine-tuning. We introdcue `wav2vec2_asr_XXX` functions which generates the architecture used for the fine-tuning task in wav2vec2 paper. *1 2. Re-purpse the old functions The existing functions, `wav2vec2_XXX`, now generates the architecture with pre-trainig module only. (no Linear module) Note *1 This architecture is just one way to define architecture for fine-tuning and it is not universal definition. The new `wav2vec2_asr_XXX` functions are designed to provide these specific fine-tuning configuration and they are not meant to support generic architecture for downstream task.
-
Yi Zhang authored
This commit fixes the local build on Windows with CUDA.
-
nateanl authored
-
Yi Zhang authored
-
- 23 Sep, 2021 1 commit
-
-
Yi Zhang authored
-