Unverified Commit 1b4b82e0 authored by moto's avatar moto Committed by GitHub
Browse files

Update reference from master to main elsewhere (#1784)



Summary: Update fairseq reference from master to main elsewhere

Reviewed By: alexeib

Differential Revision: D30938472

fbshipit-source-id: 243b98550207f241c9d3265bf3d4060350aaf0a8
Co-authored-by: default avatarDiana Liskovich <dianaml@fb.com>
parent 5aedcab3
...@@ -19,7 +19,7 @@ For building decoder, we borrow the pre-trained weights published by `fairseq` a ...@@ -19,7 +19,7 @@ For building decoder, we borrow the pre-trained weights published by `fairseq` a
### 1.1. From `fairseq` ### 1.1. From `fairseq`
For `fairseq` models, you can download pre-trained weights For `fairseq` models, you can download pre-trained weights
You can download a model from [`fairseq` repository](https://github.com/pytorch/fairseq/tree/master/examples/wav2vec). Here, we will use `Base / 960h` model. You also need to download [the letter dictionary file](https://github.com/pytorch/fairseq/tree/master/examples/wav2vec#evaluating-a-ctc-model). You can download a model from [`fairseq` repository](https://github.com/pytorch/fairseq/tree/main/examples/wav2vec). Here, we will use `Base / 960h` model. You also need to download [the letter dictionary file](https://github.com/pytorch/fairseq/tree/main/examples/wav2vec#evaluating-a-ctc-model).
For the decoder part, we use [simple_ctc](https://github.com/mthrok/ctcdecode), which also supports TorchScript. For the decoder part, we use [simple_ctc](https://github.com/mthrok/ctcdecode), which also supports TorchScript.
...@@ -89,7 +89,7 @@ Let's evaluate this word error rate (WER) of this application using [Librispeech ...@@ -89,7 +89,7 @@ Let's evaluate this word error rate (WER) of this application using [Librispeech
### 4.1. Create a list of audio paths ### 4.1. Create a list of audio paths
For the sake of simplifying our C++ code, we will first parse the Librispeech dataset to get the list of audio path For the sake of simplifying our C++ code, we will first parse the Librispeech dataset to get the list of audio path
```bash ```bash
python parse_librispeech.py <PATH_TO_YOUR_DATASET>/LibriSpeech/test-clean ./flist.txt python parse_librispeech.py <PATH_TO_YOUR_DATASET>/LibriSpeech/test-clean ./flist.txt
...@@ -137,9 +137,9 @@ You can also check `hyp.trn.pra` file to see what errors were made. ...@@ -137,9 +137,9 @@ You can also check `hyp.trn.pra` file to see what errors were made.
``` ```
id: (3528-168669-0005) id: (3528-168669-0005)
Scores: (#C #S #D #I) 7 1 0 0 Scores: (#C #S #D #I) 7 1 0 0
REF: there is a stone to be RAISED heavy REF: there is a stone to be RAISED heavy
HYP: there is a stone to be RACED heavy HYP: there is a stone to be RACED heavy
Eval: S Eval: S
``` ```
## 5. Evaluate the pipeline on VoxForge dataset ## 5. Evaluate the pipeline on VoxForge dataset
......
...@@ -2,7 +2,7 @@ ...@@ -2,7 +2,7 @@
"""Generate the conf JSON from fairseq pretrained weight file, that is consumed by unit tests """Generate the conf JSON from fairseq pretrained weight file, that is consumed by unit tests
Usage: Usage:
1. Download pretrained parameters from https://github.com/pytorch/fairseq/tree/master/examples/wav2vec 1. Download pretrained parameters from https://github.com/pytorch/fairseq/tree/main/examples/wav2vec
2. Download the dict from https://dl.fbaipublicfiles.com/fairseq/wav2vec/dict.ltr.txt 2. Download the dict from https://dl.fbaipublicfiles.com/fairseq/wav2vec/dict.ltr.txt
and put it in the same directory as parameter files. and put it in the same directory as parameter files.
3. Run this script and save the resulting JSON configuration in assets directory. 3. Run this script and save the resulting JSON configuration in assets directory.
...@@ -56,7 +56,7 @@ def _parse_args(): ...@@ -56,7 +56,7 @@ def _parse_args():
required=True, required=True,
help=( help=(
'A point file from ' 'A point file from '
'https://github.com/pytorch/fairseq/tree/master/examples/wav2vec' 'https://github.com/pytorch/fairseq/tree/main/examples/wav2vec'
) )
) )
parser.add_argument( parser.add_argument(
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment