"tests/models/glpn/test_feature_extraction_glpn.py" did not exist on "29c10a41d04f855c433a6cde7797b325651417d2"
Commit a0787a9f authored by baberabb's avatar baberabb
Browse files

Merge remote-tracking branch 'origin/big-refactor' into big-refactor_dp

parents 6359f083 dc5b3d5d
...@@ -10,3 +10,5 @@ doc_to_target: label ...@@ -10,3 +10,5 @@ doc_to_target: label
doc_to_choice: ["yes", "no"] doc_to_choice: ["yes", "no"]
metric_list: metric_list:
- metric: acc - metric: acc
metadata:
- version: 1.0
...@@ -12,3 +12,5 @@ doc_to_choice: ["no", "yes"] ...@@ -12,3 +12,5 @@ doc_to_choice: ["no", "yes"]
metric_list: metric_list:
- metric: acc - metric: acc
- metric: f1 - metric: f1
metadata:
- version: 1.0
...@@ -10,3 +10,5 @@ doc_to_target: label ...@@ -10,3 +10,5 @@ doc_to_target: label
doc_to_choice: ["True", "False"] doc_to_choice: ["True", "False"]
metric_list: metric_list:
- metric: acc - metric: acc
metadata:
- version: 1.0
...@@ -10,3 +10,5 @@ doc_to_target: label ...@@ -10,3 +10,5 @@ doc_to_target: label
doc_to_choice: ["negative", "positive"] doc_to_choice: ["negative", "positive"]
metric_list: metric_list:
- metric: acc - metric: acc
metadata:
- version: 1.0
...@@ -10,3 +10,5 @@ doc_to_target: label ...@@ -10,3 +10,5 @@ doc_to_target: label
doc_to_choice: ["False", "True"] doc_to_choice: ["False", "True"]
metric_list: metric_list:
- metric: acc - metric: acc
metadata:
- version: 2.0
...@@ -30,3 +30,5 @@ filter_list: ...@@ -30,3 +30,5 @@ filter_list:
regex_pattern: "The answer is (\\-?[0-9\\.\\,]*[0-9]+)" regex_pattern: "The answer is (\\-?[0-9\\.\\,]*[0-9]+)"
- function: "majority_vote" - function: "majority_vote"
- function: "take_first" - function: "take_first"
metadata:
- version: 0.0
...@@ -40,3 +40,5 @@ filter_list: ...@@ -40,3 +40,5 @@ filter_list:
- function: "regex" - function: "regex"
regex_pattern: "The answer is (\\-?[0-9\\.\\,]+)." regex_pattern: "The answer is (\\-?[0-9\\.\\,]+)."
- function: "take_first" - function: "take_first"
metadata:
- version: 0.0
...@@ -33,3 +33,5 @@ filter_list: ...@@ -33,3 +33,5 @@ filter_list:
- function: "regex" - function: "regex"
regex_pattern: "#### (\\-?[0-9\\.\\,]+)" regex_pattern: "#### (\\-?[0-9\\.\\,]+)"
- function: "take_first" - function: "take_first"
metadata:
- version: 1.0
...@@ -19,3 +19,5 @@ metric_list: ...@@ -19,3 +19,5 @@ metric_list:
- metric: acc_norm - metric: acc_norm
aggregation: mean aggregation: mean
higher_is_better: true higher_is_better: true
metadata:
- version: 1.0
...@@ -18,3 +18,5 @@ metric_list: ...@@ -18,3 +18,5 @@ metric_list:
- metric: acc_norm - metric: acc_norm
aggregation: mean aggregation: mean
higher_is_better: true higher_is_better: true
metadata:
- version: 1.0
...@@ -11,3 +11,5 @@ doc_to_target: label ...@@ -11,3 +11,5 @@ doc_to_target: label
doc_to_choice: ['no', 'yes'] doc_to_choice: ['no', 'yes']
metric_list: metric_list:
- metric: acc - metric: acc
metadata:
- version: 1.0
...@@ -4,5 +4,6 @@ dataset_name: deontology ...@@ -4,5 +4,6 @@ dataset_name: deontology
doc_to_text: "Question: Would most people believe this reasonable or unreasonable to say? \"{{scenario}} {{excuse.rstrip()}}\"\nAnswer:" doc_to_text: "Question: Would most people believe this reasonable or unreasonable to say? \"{{scenario}} {{excuse.rstrip()}}\"\nAnswer:"
doc_to_target: label doc_to_target: label
doc_to_choice: ['unreasonable', 'reasonable'] doc_to_choice: ['unreasonable', 'reasonable']
metadata:
- version: 1.0
# TODO: implement exact-match metric for this subset # TODO: implement exact-match metric for this subset
...@@ -5,3 +5,5 @@ task: ethics_justice ...@@ -5,3 +5,5 @@ task: ethics_justice
dataset_name: justice dataset_name: justice
doc_to_text: "Question: Would most people believe this reasonable or unreasonable to say? \"{{scenario}}\"\nAnswer:" doc_to_text: "Question: Would most people believe this reasonable or unreasonable to say? \"{{scenario}}\"\nAnswer:"
# TODO: impl. exact match for this and deontology # TODO: impl. exact match for this and deontology
metadata:
- version: 1.0
...@@ -8,3 +8,5 @@ doc_to_target: !function utils.doc_to_target ...@@ -8,3 +8,5 @@ doc_to_target: !function utils.doc_to_target
doc_to_choice: ['no', 'yes'] doc_to_choice: ['no', 'yes']
metric_list: metric_list:
- metric: acc - metric: acc
metadata:
- version: 1.0
...@@ -12,3 +12,5 @@ ...@@ -12,3 +12,5 @@
# metric_list: # metric_list:
# - metric: acc # - metric: acc
# TODO: we want this to be implemented as a winograd_schema task type, actually # TODO: we want this to be implemented as a winograd_schema task type, actually
# metadata:
# - version: 1.0
...@@ -6,3 +6,5 @@ dataset_name: virtue ...@@ -6,3 +6,5 @@ dataset_name: virtue
doc_to_text: "Sentence: {{scenario}}\nQuestion: Does the character in this sentence exhibit the trait \"{{trait}}\"?\nAnswer:" doc_to_text: "Sentence: {{scenario}}\nQuestion: Does the character in this sentence exhibit the trait \"{{trait}}\"?\nAnswer:"
doc_to_target: label doc_to_target: label
doc_to_choice: ['no', 'yes'] doc_to_choice: ['no', 'yes']
metadata:
- version: 1.0
...@@ -16,3 +16,5 @@ metric_list: ...@@ -16,3 +16,5 @@ metric_list:
- metric: acc - metric: acc
aggregation: mean aggregation: mean
higher_is_better: true higher_is_better: true
metadata:
- version: 1.0
...@@ -17,3 +17,5 @@ metric_list: ...@@ -17,3 +17,5 @@ metric_list:
- metric: acc - metric: acc
aggregation: mean aggregation: mean
higher_is_better: true higher_is_better: true
metadata:
- version: 1.0
...@@ -16,3 +16,5 @@ metric_list: ...@@ -16,3 +16,5 @@ metric_list:
- metric: acc - metric: acc
aggregation: mean aggregation: mean
higher_is_better: true higher_is_better: true
metadata:
- version: 1.0
...@@ -17,3 +17,5 @@ metric_list: ...@@ -17,3 +17,5 @@ metric_list:
- metric: acc - metric: acc
aggregation: mean aggregation: mean
higher_is_better: true higher_is_better: true
metadata:
- version: 1.0
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment