Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
278fd28a
"...git@developer.sourcefind.cn:chenpangpang/ComfyUI.git" did not exist on "f50b1fec695cecc8f7c87ce1f39db3f6b49bb3a1"
Unverified
Commit
278fd28a
authored
Nov 13, 2018
by
Thomas Wolf
Committed by
GitHub
Nov 13, 2018
Browse files
added results for 16-bit fine-tuning in readme
parent
d940eeda
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
28 additions
and
0 deletions
+28
-0
README.md
README.md
+28
-0
No files found.
README.md
View file @
278fd28a
...
@@ -236,3 +236,31 @@ python ./run_squad.py \
...
@@ -236,3 +236,31 @@ python ./run_squad.py \
--gradient_accumulation_steps
2
\
--gradient_accumulation_steps
2
\
--optimize_on_cpu
--optimize_on_cpu
```
```
If you have a recent GPU (starting from NVIDIA Volta series), you should try
**16-bit fine-tuning**
(FP16).
Here is an example of hyper-parameters for a FP16 run we tried:
```
bash
python ./run_squad.py
\
--vocab_file
$BERT_LARGE_DIR
/vocab.txt
\
--bert_config_file
$BERT_LARGE_DIR
/bert_config.json
\
--init_checkpoint
$BERT_LARGE_DIR
/pytorch_model.bin
\
--do_lower_case
\
--do_train
\
--do_predict
\
--train_file
$SQUAD_TRAIN
\
--predict_file
$SQUAD_EVAL
\
--learning_rate
3e-5
\
--num_train_epochs
2
\
--max_seq_length
384
\
--doc_stride
128
\
--output_dir
$OUTPUT_DIR
\
--train_batch_size
24
\
--fp16
\
--loss_scale
128
```
The results were similar to the above FP32 results (actually slightly higher):
```
bash
{
"exact_match"
: 84.65468306527909,
"f1"
: 91.238669287002
}
```
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment