Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
lm-evaluation-harness
Commits
8737d53a
Commit
8737d53a
authored
Jul 03, 2023
by
lintangsutawika
Browse files
fixed super_glue and removed unused yaml config
parent
2ef2d35f
Changes
22
Hide whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
51 additions
and
105 deletions
+51
-105
lm_eval/tasks/super_glue/boolq/promptsource-00.yaml
lm_eval/tasks/super_glue/boolq/promptsource-00.yaml
+0
-14
lm_eval/tasks/super_glue/boolq/promptsource-01.yaml
lm_eval/tasks/super_glue/boolq/promptsource-01.yaml
+0
-5
lm_eval/tasks/super_glue/boolq/promptsource-02.yaml
lm_eval/tasks/super_glue/boolq/promptsource-02.yaml
+0
-5
lm_eval/tasks/super_glue/boolq/seq2seq.yaml
lm_eval/tasks/super_glue/boolq/seq2seq.yaml
+1
-1
lm_eval/tasks/super_glue/cb/default.yaml
lm_eval/tasks/super_glue/cb/default.yaml
+2
-2
lm_eval/tasks/super_glue/cb/promptsource-00.yaml
lm_eval/tasks/super_glue/cb/promptsource-00.yaml
+0
-14
lm_eval/tasks/super_glue/cb/promptsource-01.yaml
lm_eval/tasks/super_glue/cb/promptsource-01.yaml
+0
-5
lm_eval/tasks/super_glue/cb/promptsource-02.yaml
lm_eval/tasks/super_glue/cb/promptsource-02.yaml
+0
-5
lm_eval/tasks/super_glue/cb/t5-prompt.yaml
lm_eval/tasks/super_glue/cb/t5-prompt.yaml
+1
-2
lm_eval/tasks/super_glue/copa/default.yaml
lm_eval/tasks/super_glue/copa/default.yaml
+14
-0
lm_eval/tasks/super_glue/copa/promptsource-00.yaml
lm_eval/tasks/super_glue/copa/promptsource-00.yaml
+0
-14
lm_eval/tasks/super_glue/copa/promptsource-01.yaml
lm_eval/tasks/super_glue/copa/promptsource-01.yaml
+0
-5
lm_eval/tasks/super_glue/copa/promptsource-02.yaml
lm_eval/tasks/super_glue/copa/promptsource-02.yaml
+0
-5
lm_eval/tasks/super_glue/copa/t5-prompt.yaml
lm_eval/tasks/super_glue/copa/t5-prompt.yaml
+1
-2
lm_eval/tasks/super_glue/copa/utils.py
lm_eval/tasks/super_glue/copa/utils.py
+17
-0
lm_eval/tasks/super_glue/record/t5-prompt.yaml
lm_eval/tasks/super_glue/record/t5-prompt.yaml
+1
-2
lm_eval/tasks/super_glue/wic/default.yaml
lm_eval/tasks/super_glue/wic/default.yaml
+14
-0
lm_eval/tasks/super_glue/wic/promptsource-00.yaml
lm_eval/tasks/super_glue/wic/promptsource-00.yaml
+0
-14
lm_eval/tasks/super_glue/wic/promptsource-01.yaml
lm_eval/tasks/super_glue/wic/promptsource-01.yaml
+0
-5
lm_eval/tasks/super_glue/wic/promptsource-02.yaml
lm_eval/tasks/super_glue/wic/promptsource-02.yaml
+0
-5
No files found.
lm_eval/tasks/super_glue/boolq/promptsource-00.yaml
deleted
100644 → 0
View file @
2ef2d35f
group
:
-
super-glue-promptsource
task
:
"
GPT-3
Style"
dataset_path
:
super_glue
dataset_name
:
boolq
training_split
:
train
validation_split
:
validation
use_prompt
:
"
promptsource:GPT-3
Style"
metric_list
:
-
metric
:
exact_match
aggregation
:
mean
higher_is_better
:
true
ignore_case
:
true
ignore_punctuation
:
true
lm_eval/tasks/super_glue/boolq/promptsource-01.yaml
deleted
100644 → 0
View file @
2ef2d35f
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
based
on
the
previous
passage"
use_prompt
:
"
promptsource:based
on
the
previous
passage"
lm_eval/tasks/super_glue/boolq/promptsource-02.yaml
deleted
100644 → 0
View file @
2ef2d35f
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
based
on
the
following
passage"
use_prompt
:
"
promptsource:based
on
the
following
passage"
lm_eval/tasks/super_glue/boolq/seq2seq.yaml
View file @
8737d53a
group
:
-
super-glue-lm-eval-v1
-
super-glue-lm-eval-v1
-seq2seq
task
:
"
boolq-seq2seq"
dataset_path
:
super_glue
dataset_name
:
boolq
...
...
lm_eval/tasks/super_glue/cb/default.yaml
View file @
8737d53a
group
:
-
super-glue-lm-eval-v1
task
:
"
default
"
task
:
"
cb
"
dataset_path
:
super_glue
dataset_name
:
cb
output_type
:
multiple_choice
training_split
:
train
validation_split
:
validation
doc_to_text
:
"
{{premise}}
\n
Question:
{{hypothesis}}.
True,
False,
or
Neither?
\n
Answer:"
doc_to_target
:
"
{{answer_choices[labe]}}"
doc_to_target
:
"
{{answer_choices[labe
l
]}}"
gold_alias
:
"
{{label}}"
# this will be cast to an int.
template_aliases
:
"
{%
set
answer_choices
=
['True',
'False',
'Neither']
%}"
metric_list
:
...
...
lm_eval/tasks/super_glue/cb/promptsource-00.yaml
deleted
100644 → 0
View file @
2ef2d35f
group
:
-
super-glue-promptsource
task
:
"
GPT-3
style"
dataset_path
:
super_glue
dataset_name
:
cb
training_split
:
train
validation_split
:
validation
use_prompt
:
"
promptsource:GPT-3
style"
metric_list
:
-
metric
:
exact_match
aggregation
:
mean
higher_is_better
:
true
ignore_case
:
true
ignore_punctuation
:
true
lm_eval/tasks/super_glue/cb/promptsource-01.yaml
deleted
100644 → 0
View file @
2ef2d35f
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
MNLI
crowdsource"
use_prompt
:
"
promptsource:MNLI
crowdsource"
lm_eval/tasks/super_glue/cb/promptsource-02.yaml
deleted
100644 → 0
View file @
2ef2d35f
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
based
on
the
previous
passage"
use_prompt
:
"
promptsource:based
on
the
previous
passage"
lm_eval/tasks/super_glue/cb/t5-prompt.yaml
View file @
8737d53a
group
:
-
super-glue-t5-prompt
task
:
t5-prompt
reference
:
"
From
Raffel
et.
al.
2019"
task
:
super_glue-cb-t5-prompt
dataset_path
:
super_glue
dataset_name
:
cb
training_split
:
train
...
...
lm_eval/tasks/super_glue/copa/default.yaml
0 → 100644
View file @
8737d53a
group
:
-
super-glue-lm-eval-v1-
task
:
"
copa"
dataset_path
:
super_glue
dataset_name
:
copa
output_type
:
multiple_choice
training_split
:
train
validation_split
:
validation
doc_to_text
:
!function
utils.doc_to_text
doc_to_target
:
!function
utils.doc_to_target
gold_alias
:
"
{{label}}"
# this will be cast to an int.
template_aliases
:
"
{%
set
answer_choices
=
[{{doc.choice1}},
'b']
%}
{{answer_choices}}"
metric_list
:
-
metric
:
acc
lm_eval/tasks/super_glue/copa/promptsource-00.yaml
deleted
100644 → 0
View file @
2ef2d35f
group
:
-
super-glue-promptsource
task
:
"
C1
or
C2?
premise,
so/because…"
dataset_path
:
super_glue
dataset_name
:
copa
training_split
:
train
validation_split
:
validation
use_prompt
:
"
promptsource:C1
or
C2?
premise,
so/because…"
metric_list
:
-
metric
:
exact_match
aggregation
:
mean
higher_is_better
:
true
ignore_case
:
true
ignore_punctuation
:
true
lm_eval/tasks/super_glue/copa/promptsource-01.yaml
deleted
100644 → 0
View file @
2ef2d35f
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
best_option"
use_prompt
:
"
promptsource:best_option"
lm_eval/tasks/super_glue/copa/promptsource-02.yaml
deleted
100644 → 0
View file @
2ef2d35f
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
cause_effect"
use_prompt
:
"
promptsource:cause_effect"
lm_eval/tasks/super_glue/copa/t5-prompt.yaml
View file @
8737d53a
group
:
-
super-glue-t5-prompt
task
:
t5-prompt
reference
:
"
From
Raffel
et.
al.
2019"
task
:
super_glue-copa-t5-prompt
dataset_path
:
super_glue
dataset_name
:
copa
training_split
:
train
...
...
lm_eval/tasks/super_glue/copa/utils.py
0 → 100644
View file @
8737d53a
def
convert_choice
(
choice
):
return
choice
[
0
].
lower
()
+
choice
[
1
:]
def
doc_to_text
(
doc
):
# Drop the period
connector
=
{
"cause"
:
"because"
,
"effect"
:
"therefore"
,
}[
doc
[
"question"
]]
return
doc
[
"premise"
].
strip
()[:
-
1
]
+
f
"
{
connector
}
"
def
doc_to_target
(
doc
):
correct_choice
=
doc
[
"choice1"
]
if
doc
[
"label"
]
==
0
else
doc
[
"choice2"
]
# Connect the sentences
return
" "
+
convert_choice
(
correct_choice
)
lm_eval/tasks/super_glue/record/t5-prompt.yaml
View file @
8737d53a
group
:
-
super-glue-t5-prompt
task
:
t5-prompt
reference
:
"
From
Raffel
et.
al.
2019"
task
:
super_glue-record-t5-prompt
dataset_path
:
super_glue
dataset_name
:
record
training_split
:
train
...
...
lm_eval/tasks/super_glue/wic/default.yaml
0 → 100644
View file @
8737d53a
group
:
-
super-glue-lm-eval-v1
task
:
"
wic"
dataset_path
:
super_glue
dataset_name
:
wic
output_type
:
multiple_choice
training_split
:
train
validation_split
:
validation
doc_to_text
:
!function
utils.doc_to_text
doc_to_target
:
!function
utils.doc_to_target
gold_alias
:
"
{{label}}"
# this will be cast to an int.
template_aliases
:
"
{%
set
answer_choices
=
['no',
'yes']
%}"
metric_list
:
-
metric
:
acc
lm_eval/tasks/super_glue/wic/promptsource-00.yaml
deleted
100644 → 0
View file @
2ef2d35f
group
:
-
super-glue-promptsource
task
:
"
GPT-3-prompt"
dataset_path
:
super_glue
dataset_name
:
wic
training_split
:
train
validation_split
:
validation
use_prompt
:
"
promptsource:GPT-3-prompt"
metric_list
:
-
metric
:
exact_match
aggregation
:
mean
higher_is_better
:
true
ignore_case
:
true
ignore_punctuation
:
true
lm_eval/tasks/super_glue/wic/promptsource-01.yaml
deleted
100644 → 0
View file @
2ef2d35f
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
GPT-3-prompt-with-label"
use_prompt
:
"
promptsource:GPT-3-prompt-with-label"
lm_eval/tasks/super_glue/wic/promptsource-02.yaml
deleted
100644 → 0
View file @
2ef2d35f
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
affirmation_true_or_false"
use_prompt
:
"
promptsource:affirmation_true_or_false"
Prev
1
2
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment