Commit 622bdda1 authored by lintangsutawika's avatar lintangsutawika
Browse files

updated yamls

parent 499a4468
...@@ -21,7 +21,7 @@ metric_list: ...@@ -21,7 +21,7 @@ metric_list:
- "," - ","
- "\\$" - "\\$"
- ".*### " - ".*### "
delimiter: "\n\n" fewshot_delimiter: "\n\n"
generation_kwargs: generation_kwargs:
until: until:
- "\n\n" - "\n\n"
......
...@@ -7,11 +7,11 @@ output_type: multiple_choice ...@@ -7,11 +7,11 @@ output_type: multiple_choice
training_split: train training_split: train
validation_split: validation validation_split: validation
test_split: test test_split: test
template_aliases: "{% set answer_choices = choices['text'] %}{% set gold = choices.label.index(answerKey.lstrip()) %}" # set the list of possible answer choices, and set what this doc's gold answer is (set what ds column used, and what) doc_to_text: question_stem
doc_to_text: "{{question_stem}}" doc_to_target: "{{choices['text'][choices.label.index(answerKey.lstrip())]}}"
doc_to_target: "{{gold}}" # this will be cast to an int. doc_to_choice: "{{choices.text}}"
should_decontaminate: true should_decontaminate: true
doc_to_decontamination_query: "{{question_stem}}" doc_to_decontamination_query: question_stem
metric_list: metric_list:
- metric: acc - metric: acc
aggregation: mean aggregation: mean
......
...@@ -7,10 +7,9 @@ output_type: multiple_choice ...@@ -7,10 +7,9 @@ output_type: multiple_choice
training_split: train training_split: train
validation_split: validation validation_split: validation
test_split: null test_split: null
template_aliases: "{% set question = goal %}{% set answer_choices = [sol1, sol2] %}{% set gold = label %}" # set the list of possible answer choices, and set what this doc's gold label idx is
doc_to_text: "Question: {{question}}\nAnswer:" doc_to_text: "Question: {{question}}\nAnswer:"
doc_to_target: "{{answer_choices[gold]}}" doc_to_target: label
gold_alias: "{{gold}}" # this will be cast to an int. doc_to_choice: "{{[sol1, sol2]}}"
metric_list: metric_list:
- metric: acc - metric: acc
aggregation: mean aggregation: mean
......
...@@ -5,10 +5,9 @@ dataset_path: corypaik/prost ...@@ -5,10 +5,9 @@ dataset_path: corypaik/prost
dataset_name: null dataset_name: null
output_type: multiple_choice output_type: multiple_choice
test_split: test test_split: test
template_aliases: "{% set answer_choices = [A, B, C, D] %}{% set gold = label %}" # set the list of possible answer choices, and set what this doc's gold answer is (set what ds column used, and what)
doc_to_text: "{{context}}\nQuestion: {{ex_question}}\nAnswer:" doc_to_text: "{{context}}\nQuestion: {{ex_question}}\nAnswer:"
doc_to_target: "{{answer_choices[gold]}}" doc_to_target: label
gold_alias: "{{gold}}" # this will be cast to an int. doc_to_choice: [A, B, C, D]
should_decontaminate: true should_decontaminate: true
doc_to_decontamination_query: "{{context}}\nQuestion: {{ex_question}}\nAnswer:" doc_to_decontamination_query: "{{context}}\nQuestion: {{ex_question}}\nAnswer:"
metric_list: metric_list:
......
...@@ -7,11 +7,10 @@ output_type: multiple_choice ...@@ -7,11 +7,10 @@ output_type: multiple_choice
training_split: null training_split: null
validation_split: null validation_split: null
test_split: train test_split: train
template_aliases: "{% set answer_choices = ['yes', 'no', 'maybe'] %}{% set gold = final_decision %}"
doc_to_text: !function preprocess_pubmedqa.doc_to_text doc_to_text: !function preprocess_pubmedqa.doc_to_text
doc_to_target: !function preprocess_pubmedqa.doc_to_target doc_to_target: final_decision
gold_alias: !function preprocess_pubmedqa.gold_alias doc_to_choice: ["yes", "no", "maybe"]
metric_list: metric_list:
- metric: acc - metric: acc
aggregation: mean aggregation: mean
higher_is_better: true higher_is_better: true
\ No newline at end of file
...@@ -5,9 +5,9 @@ dataset_path: EleutherAI/race ...@@ -5,9 +5,9 @@ dataset_path: EleutherAI/race
dataset_name: high dataset_name: high
output_type: multiple_choice output_type: multiple_choice
test_split: test test_split: test
create_choices: !function preprocess_race.create_choices
doc_to_text: !function preprocess_race.doc_to_text doc_to_text: !function preprocess_race.doc_to_text
doc_to_target: !function preprocess_race.doc_to_target doc_to_target: !function preprocess_race.doc_to_target
doc_to_choice: !function preprocess_race.doc_to_choice
metric_list: metric_list:
- metric: acc - metric: acc
aggregation: mean aggregation: mean
......
...@@ -7,8 +7,7 @@ output_type: multiple_choice ...@@ -7,8 +7,7 @@ output_type: multiple_choice
training_split: train training_split: train
validation_split: validation validation_split: validation
doc_to_text: !function utils.doc_to_text doc_to_text: !function utils.doc_to_text
doc_to_target: !function utils.doc_to_target doc_to_target: label
gold_alias: "{{label}}" # this will be cast to an int. doc_to_choice: ['no', 'yes']
template_aliases: "{% set answer_choices = ['no', 'yes'] %}"
metric_list: metric_list:
- metric: acc - metric: acc
...@@ -7,7 +7,3 @@ def doc_to_text(doc): ...@@ -7,7 +7,3 @@ def doc_to_text(doc):
doc["sentence1"][doc["start1"] : doc["end1"]], doc["sentence1"][doc["start1"] : doc["end1"]],
) )
) )
def doc_to_target(doc):
return " {}".format({0: "no", 1: "yes"}[doc["label"]])
...@@ -6,7 +6,8 @@ dataset_name: wsc ...@@ -6,7 +6,8 @@ dataset_name: wsc
training_split: train training_split: train
validation_split: validation validation_split: validation
doc_to_text: !function "preprocess_wsc.doc_to_text" doc_to_text: !function "preprocess_wsc.doc_to_text"
doc_to_target: "{% set answer_choices = ['False', 'True'] %}{{answer_choices[label]}}" doc_to_target: label
doc_to_choice: ['False', 'True']
metric_list: metric_list:
- metric: exact_match - metric: exact_match
aggregation: mean aggregation: mean
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment