"...git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "0a921b64595448d27af7ca66cebd153e8860ce95"
Unverified Commit 5dfd1906 authored by Thomas Wolf's avatar Thomas Wolf Committed by GitHub
Browse files

fix typo in readme

parent fa1aa81f
...@@ -210,12 +210,12 @@ For example, fine-tuning BERT-large on SQuAD can be done on a server with 4 k-80 ...@@ -210,12 +210,12 @@ For example, fine-tuning BERT-large on SQuAD can be done on a server with 4 k-80
```bash ```bash
{"exact_match": 84.56953642384106, "f1": 91.04028647786927} {"exact_match": 84.56953642384106, "f1": 91.04028647786927}
``` ```
To get these results that we used a combination of: To get these results we used a combination of:
- multi-GPU training (automatically activated on a multi-GPU server), - multi-GPU training (automatically activated on a multi-GPU server),
- 2 steps of gradient accumulation and - 2 steps of gradient accumulation and
- perform the optimization step on CPU to store Adam's averages in RAM. - perform the optimization step on CPU to store Adam's averages in RAM.
Here are the full list of hyper-parameters we used for this run: Here are the full list of hyper-parameters for this run:
```bash ```bash
python ./run_squad.py \ python ./run_squad.py \
--vocab_file $BERT_LARGE_DIR/vocab.txt \ --vocab_file $BERT_LARGE_DIR/vocab.txt \
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment