Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
6f877d9d
Commit
6f877d9d
authored
Aug 21, 2019
by
VictorSanh
Browse files
Update dev results on GLUE (bert-base-uncased) w/ median on 5 runs
parent
07681b6b
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
12 additions
and
10 deletions
+12
-10
docs/source/examples.rst
docs/source/examples.rst
+12
-10
No files found.
docs/source/examples.rst
View file @
6f877d9d
...
@@ -68,7 +68,9 @@ GLUE results on dev set
...
@@ -68,7 +68,9 @@ GLUE results on dev set
~~~~~~~~~~~~~~~~~~~~~~~
~~~~~~~~~~~~~~~~~~~~~~~
We get the following results on the dev set of GLUE benchmark with an uncased BERT base
We get the following results on the dev set of GLUE benchmark with an uncased BERT base
model. All experiments were run on a P100 GPU with a batch size of 32.
model (`bert-base-uncased`). All experiments ran on 8 V100 GPUs with a total train batch size of 24. Some of
these tasks have a small dataset and training can lead to high variance in the results between different runs.
We report the median on 5 runs (with different seeds) for each of the metrics.
.. list-table::
.. list-table::
:header-rows: 1
:header-rows: 1
...
@@ -78,31 +80,31 @@ model. All experiments were run on a P100 GPU with a batch size of 32.
...
@@ -78,31 +80,31 @@ model. All experiments were run on a P100 GPU with a batch size of 32.
- Result
- Result
* - CoLA
* - CoLA
- Matthew'
s
corr
.
- Matthew'
s
corr
.
- 5
7.29
-
5
5.75
*
-
SST
-
2
*
-
SST
-
2
-
accuracy
-
accuracy
- 9
3
.0
0
-
9
2
.0
9
*
-
MRPC
*
-
MRPC
-
F1
/
accuracy
-
F1
/
accuracy
-
88.85/83.82
-
90.48
/
86.27
*
-
STS
-
B
*
-
STS
-
B
-
Pearson
/
Spearman
corr
.
-
Pearson
/
Spearman
corr
.
- 89.
70/89.37
-
89.
03
/
88.64
*
-
QQP
*
-
QQP
-
accuracy
/
F1
-
accuracy
/
F1
- 90.
7
2/87.
41
-
90.
9
2
/
87.
72
*
-
MNLI
*
-
MNLI
-
matched
acc
./
mismatched
acc
.
-
matched
acc
./
mismatched
acc
.
- 83.
95
/84.
39
-
83.
74
/
84.
06
*
-
QNLI
*
-
QNLI
-
accuracy
-
accuracy
-
8
9.0
4
-
9
1
.0
7
*
-
RTE
*
-
RTE
-
accuracy
-
accuracy
- 6
1.01
-
6
8.59
*
-
WNLI
*
-
WNLI
-
accuracy
-
accuracy
-
5
3.
52
-
4
3.
66
Some
of
these
results
are
significantly
different
from
the
ones
reported
on
the
test
set
Some
of
these
results
are
significantly
different
from
the
ones
reported
on
the
test
set
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment