@@ -9,7 +9,7 @@ similar API between the different models.
...
@@ -9,7 +9,7 @@ similar API between the different models.
| [Language Generation](#language-generation) | Conditional text generation using the auto-regressive models of the library: GPT, GPT-2, Transformer-XL and XLNet. |
| [Language Generation](#language-generation) | Conditional text generation using the auto-regressive models of the library: GPT, GPT-2, Transformer-XL and XLNet. |
| [GLUE](#glue) | Examples running BERT/XLM/XLNet/RoBERTa on the 9 GLUE tasks. Examples feature distributed training as well as half-precision. |
| [GLUE](#glue) | Examples running BERT/XLM/XLNet/RoBERTa on the 9 GLUE tasks. Examples feature distributed training as well as half-precision. |
| [SQuAD](#squad) | Using BERT for question answering, examples with distributed training. |
| [SQuAD](#squad) | Using BERT for question answering, examples with distributed training. |
| [Multiple Choice](#multiplechoice) | Examples running BERT/XLNet/RoBERTa on the SWAG/RACE/ARC tasks.
| [Multiple Choice](#multiple-choice) | Examples running BERT/XLNet/RoBERTa on the SWAG/RACE/ARC tasks.
## Language model fine-tuning
## Language model fine-tuning
...
@@ -283,17 +283,17 @@ The results are the following:
...
@@ -283,17 +283,17 @@ The results are the following:
loss = 0.04755385363816904
loss = 0.04755385363816904
```
```
##Multiple Choice
##Multiple Choice
Based on the script [`run_multiple_choice.py`]().
Based on the script [`run_multiple_choice.py`]().
#### Fine-tuning on SWAG
#### Fine-tuning on SWAG
Download [swag](https://github.com/rowanz/swagaf/tree/master/data) data
Download [swag](https://github.com/rowanz/swagaf/tree/master/data) data