README.md 7.45 KB
Newer Older
Myle Ott's avatar
Myle Ott committed
1
# <img src="fairseq_logo.png" width="30"> Introduction
Sergey Edunov's avatar
Sergey Edunov committed
2

Myle Ott's avatar
Myle Ott committed
3
4
Fairseq(-py) is a sequence modeling toolkit that allows researchers and
developers to train custom models for translation, summarization, language
Myle Ott's avatar
Myle Ott committed
5
6
7
8
modeling and other text generation tasks.

### What's New:

9
- November 2019: [XLM-R models and code released](examples/xlmr/README.md)
10
- September 2019: [Nonautoregressive translation code released](examples/nonautoregressive_translation/README.md)
Myle Ott's avatar
Myle Ott committed
11
- August 2019: [WMT'19 models released](examples/wmt19/README.md)
12
- July 2019: fairseq relicensed under MIT license
Myle Ott's avatar
Myle Ott committed
13
14
- July 2019: [RoBERTa models and code released](examples/roberta/README.md)
- June 2019: [wav2vec models and code released](examples/wav2vec/README.md)
Myle Ott's avatar
Myle Ott committed
15
16
17
18

### Features:

Fairseq provides reference implementations of various sequence-to-sequence models, including:
Myle Ott's avatar
Myle Ott committed
19
- **Convolutional Neural Networks (CNN)**
Myle Ott's avatar
Myle Ott committed
20
21
22
23
24
  - [Language Modeling with Gated Convolutional Networks (Dauphin et al., 2017)](examples/language_model/conv_lm/README.md)
  - [Convolutional Sequence to Sequence Learning (Gehring et al., 2017)](examples/conv_seq2seq/README.md)
  - [Classical Structured Prediction Losses for Sequence to Sequence Learning (Edunov et al., 2018)](https://github.com/pytorch/fairseq/tree/classic_seqlevel)
  - [Hierarchical Neural Story Generation (Fan et al., 2018)](examples/stories/README.md)
  - [wav2vec: Unsupervised Pre-training for Speech Recognition (Schneider et al., 2019)](examples/wav2vec/README.md)
25
- **LightConv and DynamicConv models**
Myle Ott's avatar
Myle Ott committed
26
  - [Pay Less Attention with Lightweight and Dynamic Convolutions (Wu et al., 2019)](examples/pay_less_attention_paper/README.md)
Myle Ott's avatar
Myle Ott committed
27
- **Long Short-Term Memory (LSTM) networks**
Myle Ott's avatar
Myle Ott committed
28
  - Effective Approaches to Attention-based Neural Machine Translation (Luong et al., 2015)
Myle Ott's avatar
Myle Ott committed
29
- **Transformer (self-attention) networks**
Myle Ott's avatar
Myle Ott committed
30
31
32
33
34
35
  - Attention Is All You Need (Vaswani et al., 2017)
  - [Scaling Neural Machine Translation (Ott et al., 2018)](examples/scaling_nmt/README.md)
  - [Understanding Back-Translation at Scale (Edunov et al., 2018)](examples/backtranslation/README.md)
  - [Adaptive Input Representations for Neural Language Modeling (Baevski and Auli, 2018)](examples/language_model/transformer_lm/README.md)
  - [Mixture Models for Diverse Machine Translation: Tricks of the Trade (Shen et al., 2019)](examples/translation_moe/README.md)
  - [RoBERTa: A Robustly Optimized BERT Pretraining Approach (Liu et al., 2019)](examples/roberta/README.md)
Myle Ott's avatar
Myle Ott committed
36
  - [Facebook FAIR's WMT19 News Translation Task Submission (Ng et al., 2019)](examples/wmt19/README.md)
37
  - [Jointly Learning to Align and Translate with Transformer Models (Garg et al., 2019)](examples/joint_alignment_translation/README.md )
38
39
40
41
42
43
44
- **Non-autoregressive Transformers**
  - Non-Autoregressive Neural Machine Translation (Gu et al., 2017)
  - Deterministic Non-Autoregressive Neural Sequence Modeling by Iterative Refinement (Lee et al. 2018)
  - Insertion Transformer: Flexible Sequence Generation via Insertion Operations (Stern et al. 2019)
  - Mask-Predict: Parallel Decoding of Conditional Masked Language Models (Ghazvininejad et al., 2019)
  - [Levenshtein Transformer (Gu et al., 2019)](examples/nonautoregressive_translation/README.md)

45

Myle Ott's avatar
Myle Ott committed
46
**Additionally:**
Myle Ott's avatar
Myle Ott committed
47
- multi-GPU (distributed) training on one machine or across multiple machines
Myle Ott's avatar
Myle Ott committed
48
49
50
- fast generation on both CPU and GPU with multiple search algorithms implemented:
  - beam search
  - Diverse Beam Search ([Vijayakumar et al., 2016](https://arxiv.org/abs/1610.02424))
51
  - sampling (unconstrained, top-k and top-p/nucleus)
Myle Ott's avatar
Myle Ott committed
52
- large mini-batch training even on a single GPU via delayed updates
53
- mixed precision training (trains faster with less GPU memory on [NVIDIA tensor cores](https://developer.nvidia.com/tensor-cores))
Myle Ott's avatar
Myle Ott committed
54
- extensible: easily register new models, criterions, tasks, optimizers and learning rate schedulers
Myle Ott's avatar
Myle Ott committed
55

56
We also provide [pre-trained models](#pre-trained-models-and-examples) for several benchmark
Myle Ott's avatar
Myle Ott committed
57
translation and language modeling datasets.
Sergey Edunov's avatar
Sergey Edunov committed
58
59
60
61

![Model](fairseq.gif)

# Requirements and Installation
Myle Ott's avatar
Myle Ott committed
62

63
* [PyTorch](http://pytorch.org/) version >= 1.2.0
Bairen Yi's avatar
Bairen Yi committed
64
* Python version >= 3.5
Myle Ott's avatar
Myle Ott committed
65
* For training new models, you'll also need an NVIDIA GPU and [NCCL](https://github.com/NVIDIA/nccl)
Myle Ott's avatar
Myle Ott committed
66
* **For faster training** install NVIDIA's [apex](https://github.com/NVIDIA/apex) library with the `--cuda_ext` option
Sergey Edunov's avatar
Sergey Edunov committed
67

Myle Ott's avatar
Myle Ott committed
68
69
To install fairseq:
```bash
Myle Ott's avatar
Myle Ott committed
70
71
pip install fairseq
```
Myle Ott's avatar
Myle Ott committed
72
73
74

On MacOS:
```bash
75
76
CFLAGS="-stdlib=libc++" pip install fairseq
```
Myle Ott's avatar
Myle Ott committed
77
78
79
80

If you use Docker make sure to increase the shared memory size either with
`--ipc=host` or `--shm-size` as command line options to `nvidia-docker run`.

Myle Ott's avatar
Myle Ott committed
81
82
83
**Installing from source**

To install fairseq from source and develop locally:
Myle Ott's avatar
Myle Ott committed
84
```bash
Myle Ott's avatar
Myle Ott committed
85
86
87
git clone https://github.com/pytorch/fairseq
cd fairseq
pip install --editable .
Sergey Edunov's avatar
Sergey Edunov committed
88
89
```

Myle Ott's avatar
Myle Ott committed
90
# Getting Started
91

Myle Ott's avatar
Myle Ott committed
92
93
94
The [full documentation](https://fairseq.readthedocs.io/) contains instructions
for getting started, training new models and extending fairseq with new model
types and tasks.
Sergey Edunov's avatar
Sergey Edunov committed
95

96
# Pre-trained models and examples
Sergey Edunov's avatar
Sergey Edunov committed
97

98
99
We provide pre-trained models and pre-processed, binarized test sets for several tasks listed below,
as well as example training and evaluation commands.
Sergey Edunov's avatar
Sergey Edunov committed
100

101
- [Translation](examples/translation/README.md): convolutional and transformer models are available
Myle Ott's avatar
Myle Ott committed
102
- [Language Modeling](examples/language_model/README.md): convolutional and transformer models are available
alexeib's avatar
alexeib committed
103
- [wav2vec](examples/wav2vec/README.md): wav2vec large model is available
Sergey Edunov's avatar
Sergey Edunov committed
104

105
We also have more detailed READMEs to reproduce results from specific papers:
106
- [Jointly Learning to Align and Translate with Transformer Models (Garg et al., 2019)](examples/joint_alignment_translation/README.md )
107
- [Levenshtein Transformer (Gu et al., 2019)](examples/nonautoregressive_translation/README.md)
Myle Ott's avatar
Myle Ott committed
108
- [Facebook FAIR's WMT19 News Translation Task Submission (Ng et al., 2019)](examples/wmt19/README.md)
Myle Ott's avatar
Myle Ott committed
109
110
111
112
113
114
115
116
117
118
- [RoBERTa: A Robustly Optimized BERT Pretraining Approach (Liu et al., 2019)](examples/roberta/README.md)
- [wav2vec: Unsupervised Pre-training for Speech Recognition (Schneider et al., 2019)](examples/wav2vec/README.md)
- [Mixture Models for Diverse Machine Translation: Tricks of the Trade (Shen et al., 2019)](examples/translation_moe/README.md)
- [Pay Less Attention with Lightweight and Dynamic Convolutions (Wu et al., 2019)](examples/pay_less_attention_paper/README.md)
- [Understanding Back-Translation at Scale (Edunov et al., 2018)](examples/backtranslation/README.md)
- [Classical Structured Prediction Losses for Sequence to Sequence Learning (Edunov et al., 2018)](https://github.com/pytorch/fairseq/tree/classic_seqlevel)
- [Hierarchical Neural Story Generation (Fan et al., 2018)](examples/stories/README.md)
- [Scaling Neural Machine Translation (Ott et al., 2018)](examples/scaling_nmt/README.md)
- [Convolutional Sequence to Sequence Learning (Gehring et al., 2017)](examples/conv_seq2seq/README.md)
- [Language Modeling with Gated Convolutional Networks (Dauphin et al., 2017)](examples/language_model/conv_lm/README.md)
Sergey Edunov's avatar
Sergey Edunov committed
119
120
121
122
123
124
125

# Join the fairseq community

* Facebook page: https://www.facebook.com/groups/fairseq.users
* Google group: https://groups.google.com/forum/#!forum/fairseq-users

# License
126
fairseq(-py) is MIT-licensed.
Sergey Edunov's avatar
Sergey Edunov committed
127
The license applies to the pre-trained models as well.
Myle Ott's avatar
Myle Ott committed
128

Myle Ott's avatar
Myle Ott committed
129
130
131
132
133
134
135
136
137
138
139
140
# Citation

Please cite as:

```bibtex
@inproceedings{ott2019fairseq,
  title = {fairseq: A Fast, Extensible Toolkit for Sequence Modeling},
  author = {Myle Ott and Sergey Edunov and Alexei Baevski and Angela Fan and Sam Gross and Nathan Ng and David Grangier and Michael Auli},
  booktitle = {Proceedings of NAACL-HLT 2019: Demonstrations},
  year = {2019},
}
```