Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
lm-evaluation-harness
Commits
7ff58fe1
Unverified
Commit
7ff58fe1
authored
Mar 27, 2021
by
Leo Gao
Committed by
GitHub
Mar 27, 2021
Browse files
Merge pull request #166 from jon-tow/remove-unused-imports
Remove unused imports and format imports
parents
a69ba385
ddc044eb
Changes
31
Hide whitespace changes
Inline
Side-by-side
Showing
11 changed files
with
12 additions
and
39 deletions
+12
-39
lm_eval/tasks/superglue.py
lm_eval/tasks/superglue.py
+2
-2
lm_eval/tasks/translation.py
lm_eval/tasks/translation.py
+1
-9
lm_eval/tasks/triviaqa.py
lm_eval/tasks/triviaqa.py
+1
-1
lm_eval/tasks/unscramble.py
lm_eval/tasks/unscramble.py
+0
-1
lm_eval/tasks/wikitext.py
lm_eval/tasks/wikitext.py
+4
-7
main.py
main.py
+0
-2
scripts/cost_estimate.py
scripts/cost_estimate.py
+2
-10
scripts/fewshot_description_experiment.py
scripts/fewshot_description_experiment.py
+0
-4
scripts/write_out.py
scripts/write_out.py
+0
-1
tests/test_models.py
tests/test_models.py
+1
-1
tests/test_tasks.py
tests/test_tasks.py
+1
-1
No files found.
lm_eval/tasks/superglue.py
View file @
7ff58fe1
...
...
@@ -4,11 +4,11 @@ To-do:
- ReCoRD
"""
import
numpy
as
np
import
sklearn
import
transformers.data.metrics.squad_metrics
as
squad_metrics
from
.
common
import
HFTask
,
yesno
from
lm_eval.base
import
rf
from
..metrics
import
mean
,
acc_all
,
metric_max_over_ground_truths
import
sklearn
import
transformers.data.metrics.squad_metrics
as
squad_metrics
from
..utils
import
general_detokenize
...
...
lm_eval/tasks/translation.py
View file @
7ff58fe1
import
abc
import
json
import
random
import
os
from
collections
import
Iterable
from
pprint
import
pprint
import
pycountry
from
pprint
import
pprint
from
sacrebleu
import
sacrebleu
import
logging
from
lm_eval
import
metrics
from
lm_eval.base
import
Task
,
rf
...
...
lm_eval/tasks/triviaqa.py
View file @
7ff58fe1
import
os
import
json
import
random
from
lm_eval.base
import
Task
,
rf
from
..metrics
import
mean
from
..utils
import
sh
class
TriviaQA
(
Task
):
def
download
(
self
):
if
not
os
.
path
.
exists
(
'data/triviaqa'
):
...
...
lm_eval/tasks/unscramble.py
View file @
7ff58fe1
import
gzip
import
json
import
random
import
shutil
from
pathlib
import
Path
from
best_download
import
download_file
...
...
lm_eval/tasks/wikitext.py
View file @
7ff58fe1
import
numpy
as
np
from
scipy.stats
import
pearsonr
,
spearmanr
from
sklearn.metrics
import
f1_score
,
matthews_corrcoef
from
tqdm
import
auto
as
tqdm_lib
from
.
common
import
NLP_TASK
,
simple_accuracy_metric
,
yesno
from
.
common
import
HFTask
class
WikiText103
(
NLP_TASK
):
class
WikiText103
(
HFTask
):
NLP_PATH
=
"wikitext"
NLP_NAME
=
"wikitext-103-raw-v1"
...
...
@@ -66,7 +63,7 @@ class WikiText103(NLP_TASK):
raise
NotImplementedError
(
'Evaluation not implemented'
)
class
WikiText2
(
NLP_TASK
):
class
WikiText2
(
HFTask
):
NLP_PATH
=
"wikitext"
NLP_NAME
=
"wikitext-2-raw-v1"
...
...
main.py
View file @
7ff58fe1
...
...
@@ -2,8 +2,6 @@ import argparse
import
json
import
numpy
as
np
import
random
import
itertools
import
collections
import
logging
from
lm_eval
import
models
,
tasks
,
evaluator
,
base
...
...
scripts/cost_estimate.py
View file @
7ff58fe1
import
argparse
import
json
import
numpy
as
np
import
random
import
itertools
import
collections
import
logging
from
lm_eval
import
models
,
tasks
,
evaluator
,
base
import
random
from
lm_eval.base
import
LM
import
transformers
from
lm_eval
import
tasks
,
evaluator
from
lm_eval.base
import
LM
class
DryrunLM
(
LM
):
...
...
scripts/fewshot_description_experiment.py
View file @
7ff58fe1
import
argparse
import
json
import
numpy
as
np
import
random
import
itertools
import
collections
import
logging
from
lm_eval
import
models
,
tasks
,
evaluator
,
base
logging
.
getLogger
(
"openai"
).
setLevel
(
logging
.
WARNING
)
...
...
scripts/write_out.py
View file @
7ff58fe1
...
...
@@ -2,7 +2,6 @@ import argparse
import
numpy
as
np
import
os
import
random
from
lm_eval
import
tasks
from
lm_eval.utils
import
join_iters
...
...
tests/test_models.py
View file @
7ff58fe1
import
lm_eval.models
as
models
import
lm_eval.base
as
base
def
test_gpt2
():
gpt2
=
models
.
get_model
(
'gpt2'
).
create_from_arg_string
(
"device=cpu"
)
...
...
tests/test_tasks.py
View file @
7ff58fe1
import
lm_eval.tasks
as
tasks
import
lm_eval.base
as
base
from
itertools
import
islice
import
pytest
from
itertools
import
islice
@
pytest
.
mark
.
parametrize
(
"taskname,Task"
,
tasks
.
TASK_REGISTRY
.
items
())
...
...
Prev
1
2
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment