Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
lm-evaluation-harness
Commits
d5d19219
Commit
d5d19219
authored
Mar 26, 2021
by
Jonathan Tow
Browse files
Remove unused imports and format imports
parent
c84a4af4
Changes
25
Hide whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
26 additions
and
61 deletions
+26
-61
lm_eval/metrics.py
lm_eval/metrics.py
+0
-1
lm_eval/tasks/anli.py
lm_eval/tasks/anli.py
+1
-0
lm_eval/tasks/arc.py
lm_eval/tasks/arc.py
+0
-2
lm_eval/tasks/arithmetic.py
lm_eval/tasks/arithmetic.py
+1
-0
lm_eval/tasks/common.py
lm_eval/tasks/common.py
+0
-2
lm_eval/tasks/coqa.py
lm_eval/tasks/coqa.py
+2
-8
lm_eval/tasks/ethics.py
lm_eval/tasks/ethics.py
+5
-5
lm_eval/tasks/glue.py
lm_eval/tasks/glue.py
+0
-2
lm_eval/tasks/lambada.py
lm_eval/tasks/lambada.py
+1
-2
lm_eval/tasks/math.py
lm_eval/tasks/math.py
+1
-2
lm_eval/tasks/mathqa.py
lm_eval/tasks/mathqa.py
+3
-2
lm_eval/tasks/naturalqs.py
lm_eval/tasks/naturalqs.py
+2
-1
lm_eval/tasks/pubmedqa.py
lm_eval/tasks/pubmedqa.py
+1
-3
lm_eval/tasks/qa4mre.py
lm_eval/tasks/qa4mre.py
+3
-5
lm_eval/tasks/quac.py
lm_eval/tasks/quac.py
+0
-1
lm_eval/tasks/race.py
lm_eval/tasks/race.py
+0
-5
lm_eval/tasks/sat.py
lm_eval/tasks/sat.py
+1
-8
lm_eval/tasks/sciq.py
lm_eval/tasks/sciq.py
+1
-3
lm_eval/tasks/squad.py
lm_eval/tasks/squad.py
+2
-5
lm_eval/tasks/storycloze.py
lm_eval/tasks/storycloze.py
+2
-4
No files found.
lm_eval/metrics.py
View file @
d5d19219
import
math
import
math
from
collections
import
Iterable
from
collections
import
Iterable
from
pprint
import
pprint
import
numpy
as
np
import
numpy
as
np
import
sacrebleu
import
sacrebleu
...
...
lm_eval/tasks/anli.py
View file @
d5d19219
...
@@ -3,6 +3,7 @@ from lm_eval.base import rf
...
@@ -3,6 +3,7 @@ from lm_eval.base import rf
from
..metrics
import
mean
from
..metrics
import
mean
from
.
common
import
HFTask
from
.
common
import
HFTask
class
ANLIBase
(
HFTask
):
class
ANLIBase
(
HFTask
):
DATASET_PATH
=
"anli"
DATASET_PATH
=
"anli"
DATASET_NAME
=
None
DATASET_NAME
=
None
...
...
lm_eval/tasks/arc.py
View file @
d5d19219
import
numpy
as
np
from
lm_eval.base
import
MultipleChoiceTask
from
lm_eval.base
import
MultipleChoiceTask
from
..metrics
import
mean
from
.
common
import
HFTask
from
.
common
import
HFTask
...
...
lm_eval/tasks/arithmetic.py
View file @
d5d19219
...
@@ -8,6 +8,7 @@ from best_download import download_file
...
@@ -8,6 +8,7 @@ from best_download import download_file
ArithmeticDoc
=
namedtuple
(
'ArithmeticDoc'
,
[
'context'
,
'completion'
])
ArithmeticDoc
=
namedtuple
(
'ArithmeticDoc'
,
[
'context'
,
'completion'
])
class
Arithmetic
(
Task
):
class
Arithmetic
(
Task
):
directory
=
'data/arithmetic/'
directory
=
'data/arithmetic/'
...
...
lm_eval/tasks/common.py
View file @
d5d19219
import
datasets
import
datasets
import
numpy
as
np
import
lm_eval.metrics
import
lm_eval.metrics
from
..base
import
Task
from
..base
import
Task
...
...
lm_eval/tasks/coqa.py
View file @
d5d19219
import
os
import
os
import
json
import
json
import
transformers.data.metrics.squad_metrics
as
squad_metrics
from
lm_eval.base
import
Task
,
rf
,
mean
from
lm_eval.base
import
Task
,
rf
,
mean
from
..utils
import
sh
from
..utils
import
sh
from
itertools
import
zip_longest
from
itertools
import
zip_longest
import
transformers.data.metrics.squad_metrics
as
squad_metrics
import
collections
import
datasets
import
numpy
as
np
from
lm_eval.base
import
rf
,
mean
from
.
common
import
HFTask
from
tqdm
import
tqdm
import
string
,
re
class
CoQA
(
Task
):
class
CoQA
(
Task
):
...
...
lm_eval/tasks/ethics.py
View file @
d5d19219
from
lm_eval.base
import
Task
,
rf
from
lm_eval.metrics
import
mean
from
lm_eval.utils
import
sh
from
.common
import
yesno
import
abc
import
abc
import
csv
import
csv
import
os
import
os
import
random
import
random
import
numpy
as
np
import
numpy
as
np
from
lm_eval.base
import
Task
,
rf
from
lm_eval.metrics
import
mean
from
lm_eval.utils
import
sh
from
.common
import
yesno
class
Ethics
(
Task
):
class
Ethics
(
Task
):
def
download
(
self
):
def
download
(
self
):
...
...
lm_eval/tasks/glue.py
View file @
d5d19219
import
numpy
as
np
import
numpy
as
np
from
lm_eval.base
import
rf
from
lm_eval.base
import
rf
from
..metrics
import
mean
,
matthews_corrcoef
,
f1_score
from
..metrics
import
mean
,
matthews_corrcoef
,
f1_score
from
scipy.stats
import
pearsonr
,
spearmanr
from
tqdm
import
auto
as
tqdm_lib
from
.
common
import
HFTask
,
yesno
from
.
common
import
HFTask
,
yesno
from
..utils
import
general_detokenize
from
..utils
import
general_detokenize
...
...
lm_eval/tasks/lambada.py
View file @
d5d19219
import
json
from
lm_eval.base
import
Task
,
rf
from
lm_eval.base
import
Task
,
rf
from
lm_eval.metrics
import
mean
,
perplexity
from
lm_eval.metrics
import
mean
,
perplexity
from
lm_eval.utils
import
sh
from
lm_eval.utils
import
sh
import
json
import
math
from
best_download
import
download_file
from
best_download
import
download_file
...
...
lm_eval/tasks/math.py
View file @
d5d19219
import
abc
import
json
import
json
import
random
from
lm_eval.utils
import
sh
from
lm_eval.utils
import
sh
from
lm_eval.metrics
import
mean
from
lm_eval.metrics
import
mean
from
lm_eval.base
import
Task
,
rf
from
lm_eval.base
import
Task
,
rf
from
pathlib
import
Path
from
pathlib
import
Path
import
abc
class
Math
(
Task
):
class
Math
(
Task
):
...
...
lm_eval/tasks/mathqa.py
View file @
d5d19219
from
.
common
import
HFTask
from
lm_eval.base
import
mean
,
rf
,
MultipleChoiceTask
import
re
import
re
from
lm_eval.base
import
MultipleChoiceTask
from
.
common
import
HFTask
class
MathQA
(
HFTask
,
MultipleChoiceTask
):
class
MathQA
(
HFTask
,
MultipleChoiceTask
):
DATASET_PATH
=
"math_qa"
DATASET_PATH
=
"math_qa"
...
...
lm_eval/tasks/naturalqs.py
View file @
d5d19219
import
random
from
.
common
import
HFTask
from
.
common
import
HFTask
from
itertools
import
islice
from
itertools
import
islice
import
random
class
NaturalQs
(
HFTask
):
class
NaturalQs
(
HFTask
):
# TODO: naturalqs has a *really* large train set that huggingface just
# TODO: naturalqs has a *really* large train set that huggingface just
...
...
lm_eval/tasks/pubmedqa.py
View file @
d5d19219
import
numpy
as
np
import
numpy
as
np
import
json
from
.common
import
HFTask
import
random
from
.common
import
HFTask
from
lm_eval.base
import
rf
from
lm_eval.base
import
rf
from
..metrics
import
mean
from
..metrics
import
mean
...
...
lm_eval/tasks/qa4mre.py
View file @
d5d19219
import
os
import
os
import
numpy
as
np
from
best_download
import
download_file
from
lm_eval.base
import
MultipleChoiceTask
,
rf
from
lm_eval.metrics
import
mean
import
xml.etree.ElementTree
as
ET
import
xml.etree.ElementTree
as
ET
import
random
from
best_download
import
download_file
from
lm_eval.base
import
MultipleChoiceTask
class
QA4MRE
(
MultipleChoiceTask
):
class
QA4MRE
(
MultipleChoiceTask
):
YEAR
=
None
YEAR
=
None
...
...
lm_eval/tasks/quac.py
View file @
d5d19219
import
json
import
json
import
random
import
os
import
os
from
lm_eval.base
import
Task
from
lm_eval.base
import
Task
from
..utils
import
sh
from
..utils
import
sh
...
...
lm_eval/tasks/race.py
View file @
d5d19219
...
@@ -5,11 +5,6 @@ from lm_eval.base import rf
...
@@ -5,11 +5,6 @@ from lm_eval.base import rf
from
..metrics
import
mean
from
..metrics
import
mean
from
.
common
import
HFTask
from
.
common
import
HFTask
import
os
from
functools
import
reduce
import
operator
from
tqdm
import
tqdm
import
json
class
each
:
class
each
:
def
__init__
(
self
,
f
):
def
__init__
(
self
,
f
):
...
...
lm_eval/tasks/sat.py
View file @
d5d19219
import
json
import
random
import
os
import
os
from
lm_eval.base
import
MultipleChoiceTask
,
rf
from
lm_eval.base
import
MultipleChoiceTask
from
..metrics
import
mean
from
tqdm
import
auto
as
tqdm_lib
from
.
common
import
simple_accuracy_metric
import
numpy
as
np
from
..utils
import
sh
class
SATAnalogies
(
MultipleChoiceTask
):
class
SATAnalogies
(
MultipleChoiceTask
):
...
...
lm_eval/tasks/sciq.py
View file @
d5d19219
import
os
import
os
import
json
import
json
from
..utils
import
sh
from
lm_eval.base
import
MultipleChoiceTask
,
rf
from
..metrics
import
mean
import
zipfile
import
zipfile
from
lm_eval.base
import
MultipleChoiceTask
from
best_download
import
download_file
from
best_download
import
download_file
...
...
lm_eval/tasks/squad.py
View file @
d5d19219
import
numpy
as
np
from
.
common
import
HFTask
from
scipy.stats
import
pearsonr
,
spearmanr
from
sklearn.metrics
import
f1_score
,
matthews_corrcoef
from
tqdm
import
auto
as
tqdm_lib
from
.
common
import
HFTask
,
simple_accuracy_metric
,
yesno
class
SQuAD
(
HFTask
):
class
SQuAD
(
HFTask
):
DATASET_PATH
=
"squad_v2"
DATASET_PATH
=
"squad_v2"
...
...
lm_eval/tasks/storycloze.py
View file @
d5d19219
import
json
import
random
from
lm_eval.base
import
Task
from
..utils
import
sh
import
csv
import
csv
from
lm_eval.base
import
Task
class
StoryCloze
(
Task
):
class
StoryCloze
(
Task
):
NEEDS_MANUAL_DL
=
True
NEEDS_MANUAL_DL
=
True
...
...
Prev
1
2
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment