Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
lm-evaluation-harness
Commits
9822b06e
Unverified
Commit
9822b06e
authored
Mar 01, 2024
by
Lintang Sutawika
Committed by
GitHub
Mar 01, 2024
Browse files
Merge branch 'main' into weight_by_size
parents
51f27158
b177c82c
Changes
656
Hide whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
693 additions
and
62 deletions
+693
-62
lm_eval/tasks/bbh/zeroshot/reasoning_about_colored_objects.yaml
...l/tasks/bbh/zeroshot/reasoning_about_colored_objects.yaml
+12
-0
lm_eval/tasks/bbh/zeroshot/ruin_names.yaml
lm_eval/tasks/bbh/zeroshot/ruin_names.yaml
+12
-0
lm_eval/tasks/bbh/zeroshot/salient_translation_error_detection.yaml
...sks/bbh/zeroshot/salient_translation_error_detection.yaml
+12
-0
lm_eval/tasks/bbh/zeroshot/snarks.yaml
lm_eval/tasks/bbh/zeroshot/snarks.yaml
+12
-0
lm_eval/tasks/bbh/zeroshot/sports_understanding.yaml
lm_eval/tasks/bbh/zeroshot/sports_understanding.yaml
+14
-0
lm_eval/tasks/bbh/zeroshot/temporal_sequences.yaml
lm_eval/tasks/bbh/zeroshot/temporal_sequences.yaml
+12
-0
lm_eval/tasks/bbh/zeroshot/tracking_shuffled_objects_five_objects.yaml
.../bbh/zeroshot/tracking_shuffled_objects_five_objects.yaml
+12
-0
lm_eval/tasks/bbh/zeroshot/tracking_shuffled_objects_seven_objects.yaml
...bbh/zeroshot/tracking_shuffled_objects_seven_objects.yaml
+12
-0
lm_eval/tasks/bbh/zeroshot/tracking_shuffled_objects_three_objects.yaml
...bbh/zeroshot/tracking_shuffled_objects_three_objects.yaml
+12
-0
lm_eval/tasks/bbh/zeroshot/utils.py
lm_eval/tasks/bbh/zeroshot/utils.py
+224
-0
lm_eval/tasks/bbh/zeroshot/web_of_lies.yaml
lm_eval/tasks/bbh/zeroshot/web_of_lies.yaml
+13
-0
lm_eval/tasks/bbh/zeroshot/word_sorting.yaml
lm_eval/tasks/bbh/zeroshot/word_sorting.yaml
+8
-0
lm_eval/tasks/belebele/_generate_configs.py
lm_eval/tasks/belebele/_generate_configs.py
+7
-5
lm_eval/tasks/belebele/belebele_default.yaml
lm_eval/tasks/belebele/belebele_default.yaml
+0
-4
lm_eval/tasks/benchmarks/flan/_held_in_template_yaml
lm_eval/tasks/benchmarks/flan/_held_in_template_yaml
+2
-1
lm_eval/tasks/benchmarks/flan/flan_anli.yaml
lm_eval/tasks/benchmarks/flan/flan_anli.yaml
+0
-17
lm_eval/tasks/benchmarks/flan/flan_arc.yaml
lm_eval/tasks/benchmarks/flan/flan_arc.yaml
+0
-14
lm_eval/tasks/benchmarks/flan/flan_boolq.yaml
lm_eval/tasks/benchmarks/flan/flan_boolq.yaml
+0
-7
lm_eval/tasks/benchmarks/flan/flan_cot.yaml
lm_eval/tasks/benchmarks/flan/flan_cot.yaml
+0
-10
lm_eval/tasks/benchmarks/flan/flan_held_in.yaml
lm_eval/tasks/benchmarks/flan/flan_held_in.yaml
+329
-4
No files found.
lm_eval/tasks/bbh/zeroshot/reasoning_about_colored_objects.yaml
View file @
9822b06e
...
@@ -3,3 +3,15 @@
...
@@ -3,3 +3,15 @@
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
include"
:
"
_zeroshot_template_yaml"
"
include"
:
"
_zeroshot_template_yaml"
"
task"
:
"
bbh_zeroshot_reasoning_about_colored_objects"
"
task"
:
"
bbh_zeroshot_reasoning_about_colored_objects"
filter_list
:
-
name
:
"
strict-match"
filter
:
-
function
:
"
take_first"
-
name
:
"
flexible-extract"
filter
:
-
function
:
!function
utils.MultiChoiceRegexFilter
group_select
:
0
ignore_case
:
true
ignore_punctuation
:
true
regex_pattern
:
"
(
\\
([A-Z]
\\
))"
-
function
:
"
take_first"
lm_eval/tasks/bbh/zeroshot/ruin_names.yaml
View file @
9822b06e
...
@@ -3,3 +3,15 @@
...
@@ -3,3 +3,15 @@
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
include"
:
"
_zeroshot_template_yaml"
"
include"
:
"
_zeroshot_template_yaml"
"
task"
:
"
bbh_zeroshot_ruin_names"
"
task"
:
"
bbh_zeroshot_ruin_names"
filter_list
:
-
name
:
"
strict-match"
filter
:
-
function
:
"
take_first"
-
name
:
"
flexible-extract"
filter
:
-
function
:
!function
utils.MultiChoiceRegexFilter
group_select
:
0
ignore_case
:
true
ignore_punctuation
:
true
regex_pattern
:
"
(
\\
([A-Z]
\\
))"
-
function
:
"
take_first"
lm_eval/tasks/bbh/zeroshot/salient_translation_error_detection.yaml
View file @
9822b06e
...
@@ -3,3 +3,15 @@
...
@@ -3,3 +3,15 @@
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
include"
:
"
_zeroshot_template_yaml"
"
include"
:
"
_zeroshot_template_yaml"
"
task"
:
"
bbh_zeroshot_salient_translation_error_detection"
"
task"
:
"
bbh_zeroshot_salient_translation_error_detection"
filter_list
:
-
name
:
"
strict-match"
filter
:
-
function
:
"
take_first"
-
name
:
"
flexible-extract"
filter
:
-
function
:
!function
utils.MultiChoiceRegexFilter
group_select
:
0
ignore_case
:
true
ignore_punctuation
:
true
regex_pattern
:
"
(
\\
([A-Z]
\\
))"
-
function
:
"
take_first"
lm_eval/tasks/bbh/zeroshot/snarks.yaml
View file @
9822b06e
...
@@ -3,3 +3,15 @@
...
@@ -3,3 +3,15 @@
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
include"
:
"
_zeroshot_template_yaml"
"
include"
:
"
_zeroshot_template_yaml"
"
task"
:
"
bbh_zeroshot_snarks"
"
task"
:
"
bbh_zeroshot_snarks"
filter_list
:
-
name
:
"
strict-match"
filter
:
-
function
:
"
take_first"
-
name
:
"
flexible-extract"
filter
:
-
function
:
!function
utils.MultiChoiceRegexFilter
group_select
:
0
ignore_case
:
true
ignore_punctuation
:
true
regex_pattern
:
"
(
\\
([A-Z]
\\
))"
-
function
:
"
take_first"
lm_eval/tasks/bbh/zeroshot/sports_understanding.yaml
View file @
9822b06e
...
@@ -3,3 +3,17 @@
...
@@ -3,3 +3,17 @@
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
include"
:
"
_zeroshot_template_yaml"
"
include"
:
"
_zeroshot_template_yaml"
"
task"
:
"
bbh_zeroshot_sports_understanding"
"
task"
:
"
bbh_zeroshot_sports_understanding"
filter_list
:
-
name
:
"
strict-match"
filter
:
-
function
:
"
take_first"
-
name
:
"
flexible-extract"
filter
:
-
function
:
!function
utils.MapRegexFilter
group_select
:
0
ignore_case
:
true
regex_pattern_to_value
:
\b(no|not plausible)\b
:
"
no"
\b(yes|plausible)\b
:
"
yes"
-
function
:
"
take_first"
lm_eval/tasks/bbh/zeroshot/temporal_sequences.yaml
View file @
9822b06e
...
@@ -3,3 +3,15 @@
...
@@ -3,3 +3,15 @@
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
include"
:
"
_zeroshot_template_yaml"
"
include"
:
"
_zeroshot_template_yaml"
"
task"
:
"
bbh_zeroshot_temporal_sequences"
"
task"
:
"
bbh_zeroshot_temporal_sequences"
filter_list
:
-
name
:
"
strict-match"
filter
:
-
function
:
"
take_first"
-
name
:
"
flexible-extract"
filter
:
-
function
:
!function
utils.MultiChoiceRegexFilter
group_select
:
0
ignore_case
:
true
ignore_punctuation
:
true
regex_pattern
:
"
(
\\
([A-Z]
\\
))"
-
function
:
"
take_first"
lm_eval/tasks/bbh/zeroshot/tracking_shuffled_objects_five_objects.yaml
View file @
9822b06e
...
@@ -3,3 +3,15 @@
...
@@ -3,3 +3,15 @@
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
include"
:
"
_zeroshot_template_yaml"
"
include"
:
"
_zeroshot_template_yaml"
"
task"
:
"
bbh_zeroshot_tracking_shuffled_objects_five_objects"
"
task"
:
"
bbh_zeroshot_tracking_shuffled_objects_five_objects"
filter_list
:
-
name
:
"
strict-match"
filter
:
-
function
:
"
take_first"
-
name
:
"
flexible-extract"
filter
:
-
function
:
!function
utils.MultiChoiceRegexFilter
group_select
:
0
ignore_case
:
true
ignore_punctuation
:
true
regex_pattern
:
"
(
\\
([A-Z]
\\
))"
-
function
:
"
take_first"
lm_eval/tasks/bbh/zeroshot/tracking_shuffled_objects_seven_objects.yaml
View file @
9822b06e
...
@@ -3,3 +3,15 @@
...
@@ -3,3 +3,15 @@
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
include"
:
"
_zeroshot_template_yaml"
"
include"
:
"
_zeroshot_template_yaml"
"
task"
:
"
bbh_zeroshot_tracking_shuffled_objects_seven_objects"
"
task"
:
"
bbh_zeroshot_tracking_shuffled_objects_seven_objects"
filter_list
:
-
name
:
"
strict-match"
filter
:
-
function
:
"
take_first"
-
name
:
"
flexible-extract"
filter
:
-
function
:
!function
utils.MultiChoiceRegexFilter
group_select
:
0
ignore_case
:
true
ignore_punctuation
:
true
regex_pattern
:
"
(
\\
([A-Z]
\\
))"
-
function
:
"
take_first"
lm_eval/tasks/bbh/zeroshot/tracking_shuffled_objects_three_objects.yaml
View file @
9822b06e
...
@@ -3,3 +3,15 @@
...
@@ -3,3 +3,15 @@
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
include"
:
"
_zeroshot_template_yaml"
"
include"
:
"
_zeroshot_template_yaml"
"
task"
:
"
bbh_zeroshot_tracking_shuffled_objects_three_objects"
"
task"
:
"
bbh_zeroshot_tracking_shuffled_objects_three_objects"
filter_list
:
-
name
:
"
strict-match"
filter
:
-
function
:
"
take_first"
-
name
:
"
flexible-extract"
filter
:
-
function
:
!function
utils.MultiChoiceRegexFilter
group_select
:
0
ignore_case
:
true
ignore_punctuation
:
true
regex_pattern
:
"
(
\\
([A-Z]
\\
))"
-
function
:
"
take_first"
lm_eval/tasks/bbh/zeroshot/utils.py
0 → 100644
View file @
9822b06e
import
collections
import
re
import
sys
import
unicodedata
from
lm_eval.filters.extraction
import
Filter
,
RegexFilter
class
ExtendedRegexFilter
(
RegexFilter
):
punct_tbl
=
dict
.
fromkeys
(
i
for
i
in
range
(
sys
.
maxunicode
)
if
unicodedata
.
category
(
chr
(
i
)).
startswith
(
"P"
)
)
def
__init__
(
self
,
regex_pattern
:
str
=
r
"#### (\-?[0-9\.\,]+)"
,
group_select
=
0
,
fallback
:
str
=
"[invalid]"
,
ignore_case
=
False
,
ignore_punctuation
=
False
,
regexes_to_ignore
=
None
,
)
->
None
:
super
().
__init__
(
regex_pattern
,
group_select
,
fallback
)
self
.
ignore_case
=
ignore_case
self
.
ignore_punctuation
=
ignore_punctuation
self
.
regexes_to_ignore
=
regexes_to_ignore
def
filter_ignores
(
self
,
st
):
if
self
.
regexes_to_ignore
is
not
None
:
for
s
in
self
.
regexes_to_ignore
:
st
=
re
.
sub
(
s
,
""
,
st
)
if
self
.
ignore_case
:
st
=
st
.
lower
()
if
self
.
ignore_punctuation
:
# https://stackoverflow.com/a/266162
st
=
st
.
translate
(
self
.
punct_tbl
)
return
st
def
find_match
(
self
,
regex
,
resp
,
convert_dict
=
{}):
match
=
regex
.
findall
(
resp
)
if
match
:
match
=
match
[
self
.
group_select
]
if
isinstance
(
match
,
tuple
):
match
=
[
m
for
m
in
match
if
m
][
0
]
match
=
match
.
strip
()
if
match
and
match
in
convert_dict
:
match
=
convert_dict
[
match
]
return
match
class
MapRegexFilter
(
ExtendedRegexFilter
):
def
__init__
(
self
,
regex_pattern_to_value
:
dict
=
{},
group_select
=
0
,
fallback
:
str
=
"[invalid]"
,
ignore_case
=
False
,
ignore_punctuation
=
False
,
regexes_to_ignore
=
None
,
)
->
None
:
"""
regex_pattern_to_value: Match the regex pattern and change the result into the value
group_select: Selects the (group_select)th match from the findall result. We use the whole regex_patterns, concatenated by |
ignore_case: Lowers the case of response before matching with the given regex
ignore_punctuation: Remove the punctuation before matching with the given regex
regexes_to_ignore: Remove these regexes before matching with the given regex
"""
super
().
__init__
(
"|"
.
join
(
list
(
regex_pattern_to_value
.
keys
())),
group_select
,
fallback
,
ignore_case
,
ignore_punctuation
,
regexes_to_ignore
,
)
self
.
regex_to_value
=
{
re
.
compile
(
r
):
v
for
r
,
v
in
regex_pattern_to_value
.
items
()
}
def
apply
(
self
,
resps
,
docs
):
filtered_resps
=
[]
for
r
in
resps
:
filtered
=
[]
for
resp
in
r
:
whole_match_considering_group_select
=
self
.
find_match
(
self
.
regex
,
self
.
filter_ignores
(
resp
)
)
if
whole_match_considering_group_select
:
for
regex
,
mapped_value
in
self
.
regex_to_value
.
items
():
match
=
self
.
find_match
(
regex
,
self
.
filter_ignores
(
whole_match_considering_group_select
),
)
if
match
:
match
=
mapped_value
break
if
not
whole_match_considering_group_select
or
not
match
:
match
=
self
.
fallback
filtered
.
append
(
match
)
filtered_resps
.
append
(
filtered
)
return
filtered_resps
class
NumberParseRegexFilter
(
ExtendedRegexFilter
):
def
apply
(
self
,
resps
,
docs
):
# here, we assume we have a list, in which each element is
# a list of model responses for some particular input/target pair.
# so we process each of these (same input/target response sets)
# independently (and keep them a list.)
filtered_resps
=
[]
import
regex
from
word2number
import
w2n
# https://www.reddit.com/r/regex/comments/11a38uk/parsing_numbers_written_out_as_english_words
english_number_regex
=
regex
.
compile
(
"((?:(?:zero|one|two|three|four|five|(?:twen|thir|for|fif|six|seven|nine)(?|teen|ty)|eight(?:|een|y)|ten|eleven|twelve|fourteen|hundred|thousand|(?:m|b|tr)illion)(?:zero|one|two|three|four|five|(?:twen|thir|for|fif|six|seven|nine)(?:|teen|ty)|eight(?|een|y)|ten|eleven|twelve|fourteen|hundred|thousand|(?:m|b|tr)illion|[^\S
\r\n
]|,|and|&)+)?(?:zero|one|two|three|four|five|(?:twen|thir|for|fif|six|seven|nine)(?|teen|ty)|eight(?|een|y)|ten|eleven|twelve|fourteen|hundred|thousand|(?:m|b|tr)illion))"
)
for
r
in
resps
:
filtered
=
[]
for
resp
in
r
:
match
=
self
.
find_match
(
self
.
regex
,
resp
)
if
not
match
:
match
=
self
.
find_match
(
english_number_regex
,
resp
.
lower
())
if
match
:
match
=
str
(
w2n
.
word_to_num
(
match
))
if
not
match
:
match
=
self
.
fallback
filtered
.
append
(
match
)
filtered_resps
.
append
(
filtered
)
return
filtered_resps
class
WordSortFilter
(
Filter
):
""" """
def
apply
(
self
,
resps
,
docs
):
filtered_resps
=
[]
for
r
,
doc
in
zip
(
resps
,
docs
):
words
=
doc
[
"input"
].
split
(
"List:"
)[
1
].
strip
().
split
()
regex
=
re
.
compile
(
"|"
.
join
([
f
"
\\
b
{
w
}
\\
b"
for
w
in
words
]))
filtered
=
[]
for
resp
in
r
:
match
=
regex
.
findall
(
resp
)
match
.
reverse
()
ordered_words
=
reversed
(
collections
.
OrderedDict
(
zip
(
match
,
[
None
]
*
len
(
match
)))
)
filtered
.
append
(
" "
.
join
(
ordered_words
))
filtered_resps
.
append
(
filtered
)
return
filtered_resps
class
MultiChoiceRegexFilter
(
ExtendedRegexFilter
):
def
__init__
(
self
,
*
args
,
**
kwargs
):
"""
regex_pattern: The basic regex pattern to use. If fails to match, we will use the customized match procedure
- step 1 : We parse the choices between ([A-Z])s then try to find these choices in the response.
- step 2 : We parse the choice with regex :[\s]*([A-?]), where ? varies by number of choices.
group_select: Selects the (group_select)th match from the findall result.
ignore_case: Ignores the case during step 1 matching
ignore_punctuation: Remove the punctuation during step 1 matching
regexes_to_ignore: Remove these regexes during step 1 matching
"""
super
().
__init__
(
*
args
,
**
kwargs
)
def
apply
(
self
,
resps
,
docs
):
# here, we assume we have a list, in which each element is
# a list of model responses for some particular input/target pair.
# so we process each of these (same input/target response sets)
# independently (and keep them a list.)
filtered_resps
=
[]
for
r
,
doc
in
zip
(
resps
,
docs
):
fallback_regexes
=
[]
choice_to_alpha
=
{}
next_alpha
=
"A"
without_paren_fallback_regexes
=
[]
without_paren_to_target
=
{}
multiple_choices_regex
=
re
.
compile
(
r
"\([A-Z]\)([^\n^(]*)"
)
match
=
multiple_choices_regex
.
findall
(
doc
[
"input"
])
for
m
in
match
:
m
=
self
.
filter_ignores
(
m
.
strip
())
fallback_regexes
.
append
(
f
"
{
re
.
escape
(
m
)
}
"
)
choice_to_alpha
[
m
]
=
f
"(
{
next_alpha
}
)"
without_paren_fallback_regexes
.
append
(
next_alpha
)
without_paren_to_target
[
next_alpha
]
=
f
"(
{
next_alpha
}
)"
next_alpha
=
chr
(
ord
(
next_alpha
)
+
1
)
fallback_regex
=
re
.
compile
(
"|"
.
join
(
fallback_regexes
))
without_paren_fallback_regex
=
"|"
.
join
(
without_paren_fallback_regexes
)
without_paren_fallback_regex
=
re
.
compile
(
f
":[\s]*(
{
without_paren_fallback_regex
}
)"
)
filtered
=
[]
for
resp
in
r
:
match
=
self
.
find_match
(
self
.
regex
,
resp
)
if
not
match
:
match
=
self
.
find_match
(
fallback_regex
,
self
.
filter_ignores
(
resp
),
choice_to_alpha
)
if
not
match
:
match
=
self
.
find_match
(
without_paren_fallback_regex
,
resp
,
without_paren_to_target
)
if
not
match
:
match
=
self
.
fallback
filtered
.
append
(
match
)
filtered_resps
.
append
(
filtered
)
return
filtered_resps
lm_eval/tasks/bbh/zeroshot/web_of_lies.yaml
View file @
9822b06e
...
@@ -3,3 +3,16 @@
...
@@ -3,3 +3,16 @@
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
include"
:
"
_zeroshot_template_yaml"
"
include"
:
"
_zeroshot_template_yaml"
"
task"
:
"
bbh_zeroshot_web_of_lies"
"
task"
:
"
bbh_zeroshot_web_of_lies"
filter_list
:
-
name
:
"
strict-match"
filter
:
-
function
:
"
take_first"
-
name
:
"
flexible-extract"
filter
:
-
function
:
!function
utils.MapRegexFilter
group_select
:
0
ignore_case
:
true
regex_pattern_to_value
:
\b(no|does not tell the truth|is not telling the truth)\b
:
"
no"
\b(yes|tells the truth|is telling the truth)\b
:
"
yes"
-
function
:
"
take_first"
lm_eval/tasks/bbh/zeroshot/word_sorting.yaml
View file @
9822b06e
...
@@ -3,3 +3,11 @@
...
@@ -3,3 +3,11 @@
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
doc_to_text"
:
"
Q:
{{input}}
\n
A:"
"
include"
:
"
_zeroshot_template_yaml"
"
include"
:
"
_zeroshot_template_yaml"
"
task"
:
"
bbh_zeroshot_word_sorting"
"
task"
:
"
bbh_zeroshot_word_sorting"
filter_list
:
-
name
:
"
strict-match"
filter
:
-
function
:
"
take_first"
-
name
:
"
flexible-extract"
filter
:
-
function
:
!function
utils.WordSortFilter
-
function
:
"
take_first"
lm_eval/tasks/belebele/_generate_configs.py
View file @
9822b06e
"""
"""
Take in a YAML, and output all other splits with this YAML
Take in a YAML, and output all other splits with this YAML
"""
"""
import
os
import
yaml
import
argparse
import
argparse
import
request
s
import
o
s
import
requests
import
yaml
from
tqdm
import
tqdm
from
tqdm
import
tqdm
from
lm_eval.utils
import
logging
from
lm_eval.utils
import
logging
API_URL
=
"https://datasets-server.huggingface.co/splits?dataset=facebook/belebele"
API_URL
=
"https://datasets-server.huggingface.co/splits?dataset=facebook/belebele"
...
@@ -39,17 +40,18 @@ if __name__ == "__main__":
...
@@ -39,17 +40,18 @@ if __name__ == "__main__":
def
query
():
def
query
():
response
=
requests
.
get
(
API_URL
)
response
=
requests
.
get
(
API_URL
)
return
response
.
json
()[
"splits"
]
return
response
.
json
()[
"splits"
]
print
(
query
())
print
(
query
())
languages
=
[
split
[
"split"
]
for
split
in
query
()]
languages
=
[
split
[
"split"
]
for
split
in
query
()]
for
lang
in
tqdm
(
languages
):
for
lang
in
tqdm
(
[
lang
for
lang
in
languages
if
"default"
not
in
lang
]
):
yaml_dict
=
{
yaml_dict
=
{
"include"
:
base_yaml_name
,
"include"
:
base_yaml_name
,
"task"
:
f
"belebele_
{
args
.
task_prefix
}
_
{
lang
}
"
"task"
:
f
"belebele_
{
args
.
task_prefix
}
_
{
lang
}
"
if
args
.
task_prefix
!=
""
if
args
.
task_prefix
!=
""
else
f
"belebele_
{
lang
}
"
,
else
f
"belebele_
{
lang
}
"
,
"test_split"
:
lang
,
"test_split"
:
lang
,
"fewshot_split"
:
lang
,
"fewshot_split"
:
lang
,
}
}
file_save_path
=
args
.
save_prefix_path
+
f
"_
{
lang
}
.yaml"
file_save_path
=
args
.
save_prefix_path
+
f
"_
{
lang
}
.yaml"
...
...
lm_eval/tasks/belebele/belebele_default.yaml
deleted
100644 → 0
View file @
51f27158
"
fewshot_split"
:
"
default"
"
include"
:
"
_default_template_yaml"
"
task"
:
"
belebele_default"
"
test_split"
:
"
default"
lm_eval/tasks/benchmarks/flan/
yaml_templates/
held_in_template_yaml
→
lm_eval/tasks/benchmarks/flan/
_
held_in_template_yaml
View file @
9822b06e
output_type: generate_until
output_type: generate_until
validation_split: validation
test_split: null
doc_to_choice: null
metric_list:
metric_list:
- metric: exact_match
- metric: exact_match
aggregation: mean
aggregation: mean
...
...
lm_eval/tasks/benchmarks/flan/flan_anli.yaml
deleted
100644 → 0
View file @
51f27158
group
:
flan_anli
task
:
-
include
:
yaml_templates/held_in_template_yaml
task
:
anli_r1
dataset_path
:
anli
use_prompt
:
prompt_templates/anli.yaml:*
validation_split
:
dev_r1
-
include
:
yaml_templates/held_in_template_yaml
task
:
anli_r2
dataset_path
:
anli
use_prompt
:
prompt_templates/anli.yaml:*
validation_split
:
dev_r2
-
include
:
yaml_templates/held_in_template_yaml
task
:
anli_r3
dataset_path
:
anli
use_prompt
:
prompt_templates/anli.yaml:*
validation_split
:
dev_r3
lm_eval/tasks/benchmarks/flan/flan_arc.yaml
deleted
100644 → 0
View file @
51f27158
group
:
flan_arc
task
:
-
include
:
yaml_templates/held_in_template_yaml
task
:
arc_easy
dataset_path
:
ai2_arc
dataset_name
:
ARC-Easy
use_prompt
:
prompt_templates/arc.yaml:*
validation_split
:
validation
-
include
:
yaml_templates/held_in_template_yaml
task
:
arc_challenge
dataset_path
:
ai2_arc
dataset_name
:
ARC-Challenge
use_prompt
:
prompt_templates/arc.yaml:*
validation_split
:
validation
lm_eval/tasks/benchmarks/flan/flan_boolq.yaml
deleted
100644 → 0
View file @
51f27158
group
:
flan_boolq
task
:
-
include
:
yaml_templates/held_in_template_yaml
dataset_path
:
super_glue
dataset_name
:
boolq
use_prompt
:
prompt_templates/boolq.yaml:*
validation_split
:
validation
lm_eval/tasks/benchmarks/flan/flan_cot.yaml
deleted
100644 → 0
View file @
51f27158
group
:
flan_cot
task
:
-
include
:
yaml_templates/cot_template_yaml
dataset_path
:
gsm8k
use_prompt
:
promptsource:*
validation_split
:
validation
-
include
:
yaml_templates/cot_template_yaml
dataset_path
:
EleutherAI/asdiv
use_prompt
:
promptsource:*
validation_split
:
validation
lm_eval/tasks/benchmarks/flan/flan_held_in.yaml
View file @
9822b06e
group
:
flan_held_in
group
:
flan_held_in
group_alias
:
Flan (Held-In)
task
:
task
:
-
flan_boolq
# ANLI R1
-
flan_rte
-
group
:
anli_r1_flan
-
flan_anli
group_alias
:
ANLI R1
-
flan_arc
task
:
-
task
:
anli_r1
task_alias
:
prompt-0
include
:
_held_in_template_yaml
doc_to_text
:
"
{{premise}}
\n\n
Choose
your
answer:
based
on
the
paragraph
above
can
we
conclude
that
\"
{{hypothesis}}
\"
?
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No
\n
I
think
the
answer
is"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r1
task_alias
:
prompt-1
include
:
_held_in_template_yaml
doc_to_text
:
"
{{premise}}
\n\n
Based
on
that
paragraph
can
we
conclude
that
this
sentence
is
true?
\n
{{hypothesis}}
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r1
task_alias
:
prompt-2
include
:
_held_in_template_yaml
doc_to_text
:
"
{{premise}}
\n\n
Can
we
draw
the
following
conclusion?
\n
{{hypothesis}}
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r1
task_alias
:
prompt-3
include
:
_held_in_template_yaml
doc_to_text
:
"
{{premise}}
\n
Does
this
next
sentence
follow,
given
the
preceding
text?
\n
{{hypothesis}}
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r1
task_alias
:
prompt-4
include
:
_held_in_template_yaml
doc_to_text
:
"
{{premise}}
\n
Can
we
infer
the
following?
\n
{{hypothesis}}
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No
\n
The
answer
is:"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r1
task_alias
:
prompt-5
include
:
_held_in_template_yaml
doc_to_text
:
"
Read
the
following
paragraph
and
determine
if
the
hypothesis
is
true:
\n\n
{{premise}}
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No
\n
Hypothesis:
{{hypothesis}}
\n\n\n
"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r1
task_alias
:
prompt-6
include
:
_held_in_template_yaml
doc_to_text
:
"
Read
the
text
and
determine
if
the
sentence
is
true
(see
options
at
the
end):
\n\n
{{premise}}
\n\n
Sentence:
{{hypothesis}}
\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r1
task_alias
:
prompt-7
include
:
_held_in_template_yaml
doc_to_text
:
"
Can
we
draw
the
following
hypothesis
from
the
context
(see
options)?
\n\n
Context:
\n\n
{{premise}}
\n\n
Hypothesis:
{{hypothesis}}
\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r1
task_alias
:
prompt-8
include
:
_held_in_template_yaml
doc_to_text
:
"
Choose
from
options:
Determine
if
the
sentence
is
true
based
on
the
text
below:
\n
{{hypothesis}}
\n\n
{{premise}}
\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
# ANLI R2
-
group
:
anli_r2_flan
group_alias
:
ANLI R2
task
:
-
task
:
anli_r2
task_alias
:
prompt-0
include
:
_held_in_template_yaml
doc_to_text
:
"
{{premise}}
\n\n
Choose
your
answer:
based
on
the
paragraph
above
can
we
conclude
that
\"
{{hypothesis}}
\"
?
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No
\n
I
think
the
answer
is"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r2
task_alias
:
prompt-1
include
:
_held_in_template_yaml
doc_to_text
:
"
{{premise}}
\n\n
Based
on
that
paragraph
can
we
conclude
that
this
sentence
is
true?
\n
{{hypothesis}}
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r2
task_alias
:
prompt-2
include
:
_held_in_template_yaml
doc_to_text
:
"
{{premise}}
\n\n
Can
we
draw
the
following
conclusion?
\n
{{hypothesis}}
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r2
task_alias
:
prompt-3
include
:
_held_in_template_yaml
doc_to_text
:
"
{{premise}}
\n
Does
this
next
sentence
follow,
given
the
preceding
text?
\n
{{hypothesis}}
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r2
task_alias
:
prompt-4
include
:
_held_in_template_yaml
doc_to_text
:
"
{{premise}}
\n
Can
we
infer
the
following?
\n
{{hypothesis}}
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No
\n
The
answer
is:"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r2
task_alias
:
prompt-5
include
:
_held_in_template_yaml
doc_to_text
:
"
Read
the
following
paragraph
and
determine
if
the
hypothesis
is
true:
\n\n
{{premise}}
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No
\n
Hypothesis:
{{hypothesis}}
\n\n\n
"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r2
task_alias
:
prompt-6
include
:
_held_in_template_yaml
doc_to_text
:
"
Read
the
text
and
determine
if
the
sentence
is
true
(see
options
at
the
end):
\n\n
{{premise}}
\n\n
Sentence:
{{hypothesis}}
\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r2
task_alias
:
prompt-7
include
:
_held_in_template_yaml
doc_to_text
:
"
Can
we
draw
the
following
hypothesis
from
the
context
(see
options)?
\n\n
Context:
\n\n
{{premise}}
\n\n
Hypothesis:
{{hypothesis}}
\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r2
task_alias
:
prompt-8
include
:
_held_in_template_yaml
doc_to_text
:
"
Choose
from
options:
Determine
if
the
sentence
is
true
based
on
the
text
below:
\n
{{hypothesis}}
\n\n
{{premise}}
\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
# ANLI R3
-
group
:
anli_r3_flan
group_alias
:
ANLI R3
task
:
-
task
:
anli_r3
task_alias
:
prompt-0
include
:
_held_in_template_yaml
doc_to_text
:
"
{{premise}}
\n\n
Choose
your
answer:
based
on
the
paragraph
above
can
we
conclude
that
\"
{{hypothesis}}
\"
?
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No
\n
I
think
the
answer
is"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r3
task_alias
:
prompt-1
include
:
_held_in_template_yaml
doc_to_text
:
"
{{premise}}
\n\n
Based
on
that
paragraph
can
we
conclude
that
this
sentence
is
true?
\n
{{hypothesis}}
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r3
task_alias
:
prompt-2
include
:
_held_in_template_yaml
doc_to_text
:
"
{{premise}}
\n\n
Can
we
draw
the
following
conclusion?
\n
{{hypothesis}}
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r3
task_alias
:
prompt-3
include
:
_held_in_template_yaml
doc_to_text
:
"
{{premise}}
\n
Does
this
next
sentence
follow,
given
the
preceding
text?
\n
{{hypothesis}}
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r3
task_alias
:
prompt-4
include
:
_held_in_template_yaml
doc_to_text
:
"
{{premise}}
\n
Can
we
infer
the
following?
\n
{{hypothesis}}
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No
\n
The
answer
is:"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r3
task_alias
:
prompt-5
include
:
_held_in_template_yaml
doc_to_text
:
"
Read
the
following
paragraph
and
determine
if
the
hypothesis
is
true:
\n\n
{{premise}}
\n\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No
\n
Hypothesis:
{{hypothesis}}
\n\n\n
"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r3
task_alias
:
prompt-6
include
:
_held_in_template_yaml
doc_to_text
:
"
Read
the
text
and
determine
if
the
sentence
is
true
(see
options
at
the
end):
\n\n
{{premise}}
\n\n
Sentence:
{{hypothesis}}
\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r3
task_alias
:
prompt-7
include
:
_held_in_template_yaml
doc_to_text
:
"
Can
we
draw
the
following
hypothesis
from
the
context
(see
options)?
\n\n
Context:
\n\n
{{premise}}
\n\n
Hypothesis:
{{hypothesis}}
\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
-
task
:
anli_r3
task_alias
:
prompt-8
include
:
_held_in_template_yaml
doc_to_text
:
"
Choose
from
options:
Determine
if
the
sentence
is
true
based
on
the
text
below:
\n
{{hypothesis}}
\n\n
{{premise}}
\n
OPTIONS:
\n
-
Yes
\n
-
It's
impossible
to
say
\n
-
No"
doc_to_target
:
"
{{[
\"
Yes
\"
,
\"
It's
impossible
to
say
\"
,
\"
No
\"
][label]}}"
# Arc Easy
-
group
:
arc_easy_flan
group_alias
:
Arc Easy
task
:
-
task
:
arc_easy
task_alias
:
prompt-0
include
:
_held_in_template_yaml
doc_to_text
:
"
{{question}}
\n\n
OPTIONS:
\n
-
{{choices.text|join('
\n
-
')}}"
doc_to_target
:
"
{{choices.text[choices.label.index(answerKey)]}}"
-
task
:
arc_easy
task_alias
:
prompt-1
include
:
_held_in_template_yaml
doc_to_text
:
"
Question:
{{question}}
\n
OPTIONS:
\n
-
{{choices.text|join('
\n
-
')}}
\n
Answer:"
doc_to_target
:
"
{{choices.text[choices.label.index(answerKey)]}}"
-
task
:
arc_easy
task_alias
:
prompt-2
include
:
_held_in_template_yaml
doc_to_text
:
"
Question:
{{question}}
\n\n
What
is
the
correct
answer
to
the
question
from
the
following
choices?
\n
OPTIONS:
\n
-
{{choices.text|join('
\n
-
')}}"
doc_to_target
:
"
{{choices.text[choices.label.index(answerKey)]}}"
-
task
:
arc_easy
task_alias
:
prompt-3
include
:
_held_in_template_yaml
doc_to_text
:
"
Q:
{{question}}
\n
What
is
the
correct
answer
to
this
question?
\n
OPTIONS:
\n
-
{{choices.text|join('
\n
-
')}}...A:"
doc_to_target
:
"
{{choices.text[choices.label.index(answerKey)]}}"
-
task
:
arc_easy
task_alias
:
prompt-4
include
:
_held_in_template_yaml
doc_to_text
:
"
Choose
your
answer?
\n\n
{{question}}
\n\n
OPTIONS:
\n
-
{{choices.text|join('
\n
-
')}}"
doc_to_target
:
"
{{choices.text[choices.label.index(answerKey)]}}"
-
task
:
arc_easy
task_alias
:
prompt-5
include
:
_held_in_template_yaml
doc_to_text
:
"
Answer
the
question
\n\n
{{question}}
\n
OPTIONS:
\n
-
{{choices.text|join('
\n
-
')}}"
doc_to_target
:
"
{{choices.text[choices.label.index(answerKey)]}}"
-
task
:
arc_easy
task_alias
:
prompt-6
include
:
_held_in_template_yaml
doc_to_text
:
"
{{question}}
\n\n
Pick
the
answer
from
these
options
\n\n
OPTIONS:
\n
-
{{choices.text|join('
\n
-
')}}"
doc_to_target
:
"
{{choices.text[choices.label.index(answerKey)]}}"
# Arc Challenge
-
group
:
arc_challenge_flan
group_alias
:
Arc Challenge
task
:
-
task
:
arc_challenge
task_alias
:
prompt-0
include
:
_held_in_template_yaml
doc_to_text
:
"
{{question}}
\n\n
OPTIONS:
\n
-
{{choices.text|join('
\n
-
')}}"
doc_to_target
:
"
{{choices.text[choices.label.index(answerKey)]}}"
-
task
:
arc_challenge
task_alias
:
prompt-1
include
:
_held_in_template_yaml
doc_to_text
:
"
Question:
{{question}}
\n
OPTIONS:
\n
-
{{choices.text|join('
\n
-
')}}
\n
Answer:"
doc_to_target
:
"
{{choices.text[choices.label.index(answerKey)]}}"
-
task
:
arc_challenge
task_alias
:
prompt-2
include
:
_held_in_template_yaml
doc_to_text
:
"
Question:
{{question}}
\n\n
What
is
the
correct
answer
to
the
question
from
the
following
choices?
\n
OPTIONS:
\n
-
{{choices.text|join('
\n
-
')}}"
doc_to_target
:
"
{{choices.text[choices.label.index(answerKey)]}}"
-
task
:
arc_challenge
task_alias
:
prompt-3
include
:
_held_in_template_yaml
doc_to_text
:
"
Q:
{{question}}
\n
What
is
the
correct
answer
to
this
question?
\n
OPTIONS:
\n
-
{{choices.text|join('
\n
-
')}}...A:"
doc_to_target
:
"
{{choices.text[choices.label.index(answerKey)]}}"
-
task
:
arc_challenge
task_alias
:
prompt-4
include
:
_held_in_template_yaml
doc_to_text
:
"
Choose
your
answer?
\n\n
{{question}}
\n\n
OPTIONS:
\n
-
{{choices.text|join('
\n
-
')}}"
doc_to_target
:
"
{{choices.text[choices.label.index(answerKey)]}}"
-
task
:
arc_challenge
task_alias
:
prompt-5
include
:
_held_in_template_yaml
doc_to_text
:
"
Answer
the
question
\n\n
{{question}}
\n
OPTIONS:
\n
-
{{choices.text|join('
\n
-
')}}"
doc_to_target
:
"
{{choices.text[choices.label.index(answerKey)]}}"
-
task
:
arc_challenge
task_alias
:
prompt-6
include
:
_held_in_template_yaml
doc_to_text
:
"
{{question}}
\n\n
Pick
the
answer
from
these
options
\n\n
OPTIONS:
\n
-
{{choices.text|join('
\n
-
')}}"
doc_to_target
:
"
{{choices.text[choices.label.index(answerKey)]}}"
# BoolQ
-
group
:
boolq_flan
group_alias
:
BoolQ
task
:
-
task
:
boolq
task_alias
:
prompt-0
include
:
_held_in_template_yaml
doc_to_text
:
"
{{passage}}
\n\n
Can
we
conclude
that
{{question}}?
\n\n
OPTIONS:
\n
-
no
\n
-
yes"
doc_to_target
:
"
{{['no',
'yes'][label]}}"
-
task
:
boolq
task_alias
:
prompt-1
include
:
_held_in_template_yaml
doc_to_text
:
"
{{passage}}
\n\n
Is
it
true
that
{{question}}?
\n\n
OPTIONS:
\n
-
no
\n
-
yes"
doc_to_target
:
"
{{['no',
'yes'][label]}}"
-
task
:
boolq
task_alias
:
prompt-2
include
:
_held_in_template_yaml
doc_to_text
:
"
{{passage}}
\n\n
{{question}}?
\n\n
OPTIONS:
\n
-
no
\n
-
yes"
doc_to_target
:
"
{{['no',
'yes'][label]}}"
-
task
:
boolq
task_alias
:
prompt-3
include
:
_held_in_template_yaml
doc_to_text
:
"
Text:
{{passage}}
\n\n
Question:
{{question}}?
\n\n
OPTIONS:
\n
-
no
\n
-
yes"
doc_to_target
:
"
{{['no',
'yes'][label]}}"
-
task
:
boolq
task_alias
:
prompt-4
include
:
_held_in_template_yaml
doc_to_text
:
"
{{passage}}
\n\n
What's
the
best
answer
to
this
question:
{{question}}?
\n\n
OPTIONS:
\n
-
no
\n
-
yes"
doc_to_target
:
"
{{['no',
'yes'][label]}}"
-
task
:
boolq
task_alias
:
prompt-5
include
:
_held_in_template_yaml
doc_to_text
:
"
{{passage}}
\n
Based
on
the
above
text
what's
the
best
answer
to
this
question:
{{question}}?
\n\n
OPTIONS:
\n
-
no
\n
-
yes"
doc_to_target
:
"
{{['no',
'yes'][label]}}"
-
task
:
boolq
task_alias
:
prompt-6
include
:
_held_in_template_yaml
doc_to_text
:
"
{{passage}}
\n
Answer
this
question
making
sure
that
the
answer
is
supposed
by
the
text:
{{question}}?
\n\n
OPTIONS:
\n
-
no
\n
-
yes"
doc_to_target
:
"
{{['no',
'yes'][label]}}"
-
task
:
boolq
task_alias
:
prompt-7
include
:
_held_in_template_yaml
doc_to_text
:
"
{{passage}}
\n\n
Is
the
following
statement
correct
based
on
the
text
\n\n
{{question}}
\n\n
OPTIONS:
\n
-
no
\n
-
yes"
doc_to_target
:
"
{{['no',
'yes'][label]}}"
-
task
:
boolq
task_alias
:
prompt-8
include
:
_held_in_template_yaml
doc_to_text
:
"
{{passage}}
\n\n
Is
this
statement
correct
\"
{{question}}
\"
?
\n\n
OPTIONS:
\n
-
no
\n
-
yes"
doc_to_target
:
"
{{['no',
'yes'][label]}}"
-
task
:
boolq
task_alias
:
prompt-9
include
:
_held_in_template_yaml
doc_to_text
:
"
Is
it
true
that
{{question}}
based
on
the
following
text?
\n\n
{{passage}}
\n\n
OPTIONS:
\n
-
no
\n
-
yes"
doc_to_target
:
"
{{['no',
'yes'][label]}}"
# RTE
-
group
:
rte_flan
group_alias
:
RTE
task
:
-
task
:
rte
task_alias
:
prompt-0
include
:
_held_in_template_yaml
doc_to_text
:
"
{{sentence1}}
\n\n
Question
with
options:
Based
on
the
paragraph
above
can
we
conclude
that
\"
{{sentence2}}
\"
?
\n\n
OPTIONS:
\n
-
yes
\n
-
no"
doc_to_target
:
"
{{['yes',
'no'][label]}}"
-
task
:
rte
task_alias
:
prompt-1
include
:
_held_in_template_yaml
doc_to_text
:
"
{{sentence1}}
\n\n
Based
on
that
paragraph
can
we
conclude
that
the
sentence
below
is
true?
\n
{{sentence2}}
\n\n
OPTIONS:
\n
-
yes
\n
-
no"
doc_to_target
:
"
{{['yes',
'no'][label]}}"
-
task
:
rte
task_alias
:
prompt-2
include
:
_held_in_template_yaml
doc_to_text
:
"
{{sentence1}}
\n\n
Q
with
options:
Can
we
draw
the
following
conclusion?
\n
{{sentence2}}
\n\n
OPTIONS:
\n
-
yes
\n
-
no"
doc_to_target
:
"
{{['yes',
'no'][label]}}"
-
task
:
rte
task_alias
:
prompt-3
include
:
_held_in_template_yaml
doc_to_text
:
"
{{sentence1}}
\n
Does
this
next
sentence
follow,
given
the
preceding
text?
\n
{{sentence2}}
\n\n
OPTIONS:
\n
-
yes
\n
-
no"
doc_to_target
:
"
{{['yes',
'no'][label]}}"
-
task
:
rte
task_alias
:
prompt-4
include
:
_held_in_template_yaml
doc_to_text
:
"
{{sentence1}}
\n
OPTIONS:
\n
-
yes
\n
-
no
\n
Question:
Can
we
infer
the
following?
\n
{{sentence2}}"
doc_to_target
:
"
{{['yes',
'no'][label]}}"
-
task
:
rte
task_alias
:
prompt-5
include
:
_held_in_template_yaml
doc_to_text
:
"
Read
the
following
paragraph
and
determine
if
the
hypothesis
is
true.
Select
from
options
at
the
end:
\n\n
{{sentence1}}
\n\n
Hypothesis:
{{sentence2}}
\n
OPTIONS:
\n
-
yes
\n
-
no
\n
The
answer
is"
doc_to_target
:
"
{{['yes',
'no'][label]}}"
-
task
:
rte
task_alias
:
prompt-6
include
:
_held_in_template_yaml
doc_to_text
:
"
Read
the
text
and
determine
if
the
sentence
is
true:
\n\n
{{sentence1}}
\n\n
Sentence:
{{sentence2}}
\n
OPTIONS:
\n
-
yes
\n
-
no
\n
A:"
doc_to_target
:
"
{{['yes',
'no'][label]}}"
-
task
:
rte
task_alias
:
prompt-7
include
:
_held_in_template_yaml
doc_to_text
:
"
Question
with
options:
can
we
draw
the
following
hypothesis
from
the
context?
\n\n
Context:
\n\n
{{sentence1}}
\n\n
Hypothesis:
{{sentence2}}
\n
OPTIONS:
\n
-
yes
\n
-
no
\n
A:"
doc_to_target
:
"
{{['yes',
'no'][label]}}"
-
task
:
rte
task_alias
:
prompt-8
include
:
_held_in_template_yaml
doc_to_text
:
"
Determine
if
the
sentence
is
true
based
on
the
text
below.
Choose
from
options.
\n
{{sentence2}}
\n\n
{{sentence1}}
\n
OPTIONS:
\n
-
yes
\n
-
no"
doc_to_target
:
"
{{['yes',
'no'][label]}}"
Prev
1
…
4
5
6
7
8
9
10
11
12
…
33
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment