Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
lm-evaluation-harness
Commits
eb42b01b
Commit
eb42b01b
authored
May 23, 2023
by
lintangsutawika
Browse files
added more promptsource examples
parent
1cf51552
Changes
27
Hide whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
119 additions
and
22 deletions
+119
-22
lm_eval/tasks/pile/pile_enron.yaml
lm_eval/tasks/pile/pile_enron.yaml
+0
-0
lm_eval/tasks/super_glue/boolq/promptsource-00.yaml
lm_eval/tasks/super_glue/boolq/promptsource-00.yaml
+4
-4
lm_eval/tasks/super_glue/boolq/promptsource-01.yaml
lm_eval/tasks/super_glue/boolq/promptsource-01.yaml
+5
-0
lm_eval/tasks/super_glue/boolq/promptsource-02.yaml
lm_eval/tasks/super_glue/boolq/promptsource-02.yaml
+5
-0
lm_eval/tasks/super_glue/cb/can_we_infer.yaml
lm_eval/tasks/super_glue/cb/can_we_infer.yaml
+0
-7
lm_eval/tasks/super_glue/cb/claim_true_false_inconclusive.yaml
...al/tasks/super_glue/cb/claim_true_false_inconclusive.yaml
+0
-7
lm_eval/tasks/super_glue/cb/promptsource-00.yaml
lm_eval/tasks/super_glue/cb/promptsource-00.yaml
+4
-4
lm_eval/tasks/super_glue/cb/promptsource-01.yaml
lm_eval/tasks/super_glue/cb/promptsource-01.yaml
+5
-0
lm_eval/tasks/super_glue/cb/promptsource-02.yaml
lm_eval/tasks/super_glue/cb/promptsource-02.yaml
+5
-0
lm_eval/tasks/super_glue/copa/promptsource-00.yaml
lm_eval/tasks/super_glue/copa/promptsource-00.yaml
+14
-0
lm_eval/tasks/super_glue/copa/promptsource-01.yaml
lm_eval/tasks/super_glue/copa/promptsource-01.yaml
+5
-0
lm_eval/tasks/super_glue/copa/promptsource-02.yaml
lm_eval/tasks/super_glue/copa/promptsource-02.yaml
+5
-0
lm_eval/tasks/super_glue/multirc/promptsource-00.yaml
lm_eval/tasks/super_glue/multirc/promptsource-00.yaml
+14
-0
lm_eval/tasks/super_glue/multirc/promptsource-01.yaml
lm_eval/tasks/super_glue/multirc/promptsource-01.yaml
+5
-0
lm_eval/tasks/super_glue/multirc/promptsource-02.yaml
lm_eval/tasks/super_glue/multirc/promptsource-02.yaml
+5
-0
lm_eval/tasks/super_glue/record/promptsource-00.yaml
lm_eval/tasks/super_glue/record/promptsource-00.yaml
+14
-0
lm_eval/tasks/super_glue/record/promptsource-01.yaml
lm_eval/tasks/super_glue/record/promptsource-01.yaml
+5
-0
lm_eval/tasks/super_glue/record/promptsource-02.yaml
lm_eval/tasks/super_glue/record/promptsource-02.yaml
+5
-0
lm_eval/tasks/super_glue/rte/promptsource-00.yaml
lm_eval/tasks/super_glue/rte/promptsource-00.yaml
+14
-0
lm_eval/tasks/super_glue/rte/promptsource-01.yaml
lm_eval/tasks/super_glue/rte/promptsource-01.yaml
+5
-0
No files found.
lm_eval/tasks/pile_enron.yaml
→
lm_eval/tasks/pile
/pile
_enron.yaml
View file @
eb42b01b
File moved
lm_eval/tasks/super_glue/
wsc.fixed/templat
e-00.yaml
→
lm_eval/tasks/super_glue/
boolq/promptsourc
e-00.yaml
View file @
eb42b01b
group
:
group
:
-
t0-eval
-
super-glue-promptsource
task
:
"
does
the
pronoun
refer
to
"
task
:
"
GPT-3
Style
"
dataset_path
:
super_glue
dataset_path
:
super_glue
dataset_name
:
wsc.fixed
dataset_name
:
boolq
training_split
:
train
training_split
:
train
validation_split
:
validation
validation_split
:
validation
use_prompt
:
"
promptsource:
does
the
pronoun
refer
to
"
use_prompt
:
"
promptsource:
GPT-3
Style
"
metric_list
:
metric_list
:
-
metric
:
exact_match
-
metric
:
exact_match
aggregation
:
mean
aggregation
:
mean
...
...
lm_eval/tasks/super_glue/boolq/promptsource-01.yaml
0 → 100644
View file @
eb42b01b
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
based
on
the
previous
passage"
use_prompt
:
"
promptsource:based
on
the
previous
passage"
lm_eval/tasks/super_glue/boolq/promptsource-02.yaml
0 → 100644
View file @
eb42b01b
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
based
on
the
following
passage"
use_prompt
:
"
promptsource:based
on
the
following
passage"
lm_eval/tasks/super_glue/cb/can_we_infer.yaml
deleted
100644 → 0
View file @
1cf51552
group
:
-
super-glue-cb
include
:
based_on_previous_passage.yaml
task
:
can we infer
reference
:
Webson & Pavlick
2021
doc_to_text
:
"
Suppose
{{premise}}
Can
we
infer
that
\"
{{hypothesis}}
\"
?
Yes,
no,
or
maybe?"
doc_to_target
:
"
{%
set
answer_choices
=
['Yes',
'No',
'Maybe']
%}{{answer_choices[label]}}"
lm_eval/tasks/super_glue/cb/claim_true_false_inconclusive.yaml
deleted
100644 → 0
View file @
1cf51552
group
:
-
super-glue-cb
include
:
based_on_previous_passage.yaml
task
:
claim
true
/false/inconclusive
reference
:
Sanh et al.
2021
doc_to_text
:
"
{{premise}}
Based
on
that
information,
is
the
claim:
\"
{{hypothesis}}
\"
\"
true
\"
,
\"
false
\"
,
or
\"
inconclusive
\"
?"
doc_to_target
:
"
{%
set
answer_choices
=
['True',
'False',
'Inconclusive']
%}{{answer_choices[label]}}"
lm_eval/tasks/super_glue/
wsc.fixed/templat
e-0
1
.yaml
→
lm_eval/tasks/super_glue/
cb/promptsourc
e-0
0
.yaml
View file @
eb42b01b
group
:
group
:
-
t0-eval
-
super-glue-promptsource
task
:
"
by
p
they
mean
"
task
:
"
GPT-3
style
"
dataset_path
:
super_glue
dataset_path
:
super_glue
dataset_name
:
wsc.fixed
dataset_name
:
cb
training_split
:
train
training_split
:
train
validation_split
:
validation
validation_split
:
validation
use_prompt
:
"
promptsource:
by
p
they
mean
"
use_prompt
:
"
promptsource:
GPT-3
style
"
metric_list
:
metric_list
:
-
metric
:
exact_match
-
metric
:
exact_match
aggregation
:
mean
aggregation
:
mean
...
...
lm_eval/tasks/super_glue/cb/promptsource-01.yaml
0 → 100644
View file @
eb42b01b
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
MNLI
crowdsource"
use_prompt
:
"
promptsource:MNLI
crowdsource"
lm_eval/tasks/super_glue/cb/promptsource-02.yaml
0 → 100644
View file @
eb42b01b
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
based
on
the
previous
passage"
use_prompt
:
"
promptsource:based
on
the
previous
passage"
lm_eval/tasks/super_glue/c
b/based_on_previous_passage
.yaml
→
lm_eval/tasks/super_glue/c
opa/promptsource-00
.yaml
View file @
eb42b01b
group
:
group
:
-
super-glue-cb
-
super-glue-promptsource
task
:
based on the previous passage
task
:
"
C1
or
C2?
premise,
so/because…"
reference
:
"
Adapted
from
the
BoolQ
prompts
in
Schick
&
Sch
\xFC
tze
2021."
dataset_path
:
super_glue
dataset_path
:
super_glue
dataset_name
:
c
b
dataset_name
:
c
opa
training_split
:
train
training_split
:
train
validation_split
:
validation
validation_split
:
validation
doc_to_text
:
"
{{premise}}
Based
on
the
previous
passage,
is
it
true
that
\"
{{hypothesis}}
\"
?
Yes,
no,
or
maybe?"
use_prompt
:
"
promptsource:C1
or
C2?
premise,
so/because…"
doc_to_target
:
"
{%
set
answer_choices
=
['Yes',
'No',
'Maybe']
%}{{answer_choices[label]}}"
metric_list
:
metric_list
:
-
metric
:
exact_match
-
metric
:
exact_match
aggregation
:
mean
aggregation
:
mean
...
...
lm_eval/tasks/super_glue/copa/promptsource-01.yaml
0 → 100644
View file @
eb42b01b
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
best_option"
use_prompt
:
"
promptsource:best_option"
lm_eval/tasks/super_glue/copa/promptsource-02.yaml
0 → 100644
View file @
eb42b01b
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
cause_effect"
use_prompt
:
"
promptsource:cause_effect"
lm_eval/tasks/super_glue/multirc/promptsource-00.yaml
0 → 100644
View file @
eb42b01b
group
:
-
super-glue-promptsource
task
:
"
I
was
going
to
say…"
dataset_path
:
super_glue
dataset_name
:
multirc
training_split
:
train
validation_split
:
validation
use_prompt
:
"
promptsource:I
was
going
to
say…"
metric_list
:
-
metric
:
exact_match
aggregation
:
mean
higher_is_better
:
true
ignore_case
:
true
ignore_punctuation
:
true
lm_eval/tasks/super_glue/multirc/promptsource-01.yaml
0 → 100644
View file @
eb42b01b
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
Would
it
be
good
to
answer…"
use_prompt
:
"
promptsource:Would
it
be
good
to
answer…"
lm_eval/tasks/super_glue/multirc/promptsource-02.yaml
0 → 100644
View file @
eb42b01b
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
confirm"
use_prompt
:
"
promptsource:confirm"
lm_eval/tasks/super_glue/record/promptsource-00.yaml
0 → 100644
View file @
eb42b01b
group
:
-
super-glue-promptsource
task
:
"
Add
sentence
after
(continuation
choices)"
dataset_path
:
super_glue
dataset_name
:
record
training_split
:
train
validation_split
:
validation
use_prompt
:
"
promptsource:Add
sentence
after
(continuation
choices)"
metric_list
:
-
metric
:
exact_match
aggregation
:
mean
higher_is_better
:
true
ignore_case
:
true
ignore_punctuation
:
true
lm_eval/tasks/super_glue/record/promptsource-01.yaml
0 → 100644
View file @
eb42b01b
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
Add
sentence
after
after
(continuation
choices)"
use_prompt
:
"
promptsource:Add
sentence
after
after
(continuation
choices)"
lm_eval/tasks/super_glue/record/promptsource-02.yaml
0 → 100644
View file @
eb42b01b
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
Can
you
figure
out…"
use_prompt
:
"
promptsource:Can
you
figure
out…"
lm_eval/tasks/super_glue/rte/promptsource-00.yaml
0 → 100644
View file @
eb42b01b
group
:
-
super-glue-promptsource
task
:
"
GPT-3
style"
dataset_path
:
super_glue
dataset_name
:
rte
training_split
:
train
validation_split
:
validation
use_prompt
:
"
promptsource:GPT-3
style"
metric_list
:
-
metric
:
exact_match
aggregation
:
mean
higher_is_better
:
true
ignore_case
:
true
ignore_punctuation
:
true
lm_eval/tasks/super_glue/rte/promptsource-01.yaml
0 → 100644
View file @
eb42b01b
include
:
promptsource-00.yaml
group
:
-
super-glue-promptsource
task
:
"
MNLI
crowdsource"
use_prompt
:
"
promptsource:MNLI
crowdsource"
Prev
1
2
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment