"lm_eval/tasks/hendrycks_math.py" did not exist on "c84a4af4d832f55ac1270fbccdf37ca4691a3a12"
default.yaml 437 Bytes
Newer Older
lintangsutawika's avatar
lintangsutawika committed
1
2
group:
  - super-glue-lm-eval-v1
lintangsutawika's avatar
lintangsutawika committed
3
task: cb
lintangsutawika's avatar
lintangsutawika committed
4
5
6
7
8
9
dataset_path: super_glue
dataset_name: cb
output_type: multiple_choice
training_split: train
validation_split: validation
doc_to_text: "{{premise}}\nQuestion: {{hypothesis}}. True, False, or Neither?\nAnswer:"
lintangsutawika's avatar
lintangsutawika committed
10
11
doc_to_target: label
doc_to_choice: ['True', 'False', 'Neither']
lintangsutawika's avatar
lintangsutawika committed
12
13
14
15
metric_list:
  - metric: acc
  - metric: f1
    aggregation: !function "aggregate.cb_multi_fi"
lintangsutawika's avatar
lintangsutawika committed
16
metadata:
17
  version: 1.0