Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
lm-evaluation-harness
Commits
7d282b5f
"magic_pdf/vscode:/vscode.git/clone" did not exist on "2e87e649edd73133a871722ca34b55459ee9ffa8"
Commit
7d282b5f
authored
Apr 25, 2022
by
cjlovering
Browse files
Add PromptSourceTask to the updated tasks
parent
88745155
Changes
3
Show whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
6 additions
and
6 deletions
+6
-6
lm_eval/tasks/coqa.py
lm_eval/tasks/coqa.py
+2
-2
lm_eval/tasks/drop.py
lm_eval/tasks/drop.py
+2
-2
lm_eval/tasks/race.py
lm_eval/tasks/race.py
+2
-2
No files found.
lm_eval/tasks/coqa.py
View file @
7d282b5f
...
@@ -12,7 +12,7 @@ Homepage: https://stanfordnlp.github.io/coqa/
...
@@ -12,7 +12,7 @@ Homepage: https://stanfordnlp.github.io/coqa/
import
inspect
import
inspect
import
transformers.data.metrics.squad_metrics
as
squad_metrics
import
transformers.data.metrics.squad_metrics
as
squad_metrics
import
lm_eval.datasets.coqa.coqa
import
lm_eval.datasets.coqa.coqa
from
lm_eval.base
import
Task
,
rf
,
mean
from
lm_eval.base
import
PromptSource
Task
,
rf
,
mean
from
itertools
import
zip_longest
from
itertools
import
zip_longest
...
@@ -28,7 +28,7 @@ _CITATION = """
...
@@ -28,7 +28,7 @@ _CITATION = """
"""
"""
class
CoQA
(
Task
):
class
CoQA
(
PromptSource
Task
):
VERSION
=
1
VERSION
=
1
DATASET_PATH
=
inspect
.
getfile
(
lm_eval
.
datasets
.
coqa
.
coqa
)
DATASET_PATH
=
inspect
.
getfile
(
lm_eval
.
datasets
.
coqa
.
coqa
)
DATASET_NAME
=
None
DATASET_NAME
=
None
...
...
lm_eval/tasks/drop.py
View file @
7d282b5f
...
@@ -18,7 +18,7 @@ import re
...
@@ -18,7 +18,7 @@ import re
import
string
import
string
import
lm_eval.datasets.drop.drop
import
lm_eval.datasets.drop.drop
from
scipy.optimize
import
linear_sum_assignment
from
scipy.optimize
import
linear_sum_assignment
from
lm_eval.base
import
Task
,
rf
from
lm_eval.base
import
PromptSource
Task
,
rf
from
lm_eval.metrics
import
mean
from
lm_eval.metrics
import
mean
...
@@ -37,7 +37,7 @@ _CITATION = """
...
@@ -37,7 +37,7 @@ _CITATION = """
_ARTICLES
=
re
.
compile
(
r
"\b(a|an|the)\b"
,
re
.
UNICODE
)
_ARTICLES
=
re
.
compile
(
r
"\b(a|an|the)\b"
,
re
.
UNICODE
)
class
DROP
(
Task
):
class
DROP
(
PromptSource
Task
):
VERSION
=
1
VERSION
=
1
DATASET_PATH
=
inspect
.
getfile
(
lm_eval
.
datasets
.
drop
.
drop
)
DATASET_PATH
=
inspect
.
getfile
(
lm_eval
.
datasets
.
drop
.
drop
)
DATASET_NAME
=
None
DATASET_NAME
=
None
...
...
lm_eval/tasks/race.py
View file @
7d282b5f
...
@@ -12,7 +12,7 @@ Homepage: https://www.cs.cmu.edu/~glai1/data/race/
...
@@ -12,7 +12,7 @@ Homepage: https://www.cs.cmu.edu/~glai1/data/race/
import
collections
import
collections
import
datasets
import
datasets
import
numpy
as
np
import
numpy
as
np
from
lm_eval.base
import
rf
,
Task
from
lm_eval.base
import
PromptSourceTask
,
rf
from
lm_eval.metrics
import
mean
from
lm_eval.metrics
import
mean
...
@@ -34,7 +34,7 @@ class each:
...
@@ -34,7 +34,7 @@ class each:
return
list
(
map
(
self
.
f
,
other
))
return
list
(
map
(
self
.
f
,
other
))
class
RACE
(
Task
):
class
RACE
(
PromptSource
Task
):
VERSION
=
1
VERSION
=
1
DATASET_PATH
=
"race"
DATASET_PATH
=
"race"
DATASET_NAME
=
"high"
DATASET_NAME
=
"high"
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment