Unverified Commit aa2dd2b5 authored by Fengzhe Zhou's avatar Fengzhe Zhou Committed by GitHub
Browse files

[Format] Add config lints (#892)

parent 3dbba119
...@@ -7,10 +7,10 @@ from opencompass.utils.text_postprocessors import multiple_select_postprocess ...@@ -7,10 +7,10 @@ from opencompass.utils.text_postprocessors import multiple_select_postprocess
cmb_datasets = [] cmb_datasets = []
for split in ["val", "test"]: for split in ['val', 'test']:
cmb_reader_cfg = dict( cmb_reader_cfg = dict(
input_columns=["exam_type", "exam_class", "question_type", "question", "option_str"], input_columns=['exam_type', 'exam_class', 'question_type', 'question', 'option_str'],
output_column="answer", output_column='answer',
train_split=split, train_split=split,
test_split=split, test_split=split,
) )
...@@ -21,10 +21,10 @@ for split in ["val", "test"]: ...@@ -21,10 +21,10 @@ for split in ["val", "test"]:
template=dict( template=dict(
round=[ round=[
dict( dict(
role="HUMAN", role='HUMAN',
prompt=f"以下是中国{{exam_type}}中{{exam_class}}考试的一道{{question_type}},不需要做任何分析和解释,直接输出答案选项。\n{{question}}\n{{option_str}} \n 答案: ", prompt=f'以下是中国{{exam_type}}中{{exam_class}}考试的一道{{question_type}},不需要做任何分析和解释,直接输出答案选项。\n{{question}}\n{{option_str}} \n 答案: ',
), ),
dict(role="BOT", prompt="{answer}"), dict(role='BOT', prompt='{answer}'),
], ],
), ),
), ),
...@@ -39,9 +39,9 @@ for split in ["val", "test"]: ...@@ -39,9 +39,9 @@ for split in ["val", "test"]:
cmb_datasets.append( cmb_datasets.append(
dict( dict(
abbr="cmb" if split == "val" else "cmb_test", abbr='cmb' if split == 'val' else 'cmb_test',
type=CMBDataset, type=CMBDataset,
path="./data/CMB/", path='./data/CMB/',
reader_cfg=cmb_reader_cfg, reader_cfg=cmb_reader_cfg,
infer_cfg=cmb_infer_cfg, infer_cfg=cmb_infer_cfg,
eval_cfg=cmb_eval_cfg, eval_cfg=cmb_eval_cfg,
......
...@@ -85,16 +85,16 @@ for _name in cmmlu_all_sets: ...@@ -85,16 +85,16 @@ for _name in cmmlu_all_sets:
ice_template=dict( ice_template=dict(
type=PromptTemplate, type=PromptTemplate,
template=dict( template=dict(
begin="</E>", begin='</E>',
round=[ round=[
dict( dict(
role="HUMAN", role='HUMAN',
prompt= prompt=
f"以下是关于{_ch_name}的单项选择题,请直接给出正确答案的选项。\n题目:{{question}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}" f'以下是关于{_ch_name}的单项选择题,请直接给出正确答案的选项。\n题目:{{question}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}'
), ),
dict(role="BOT", prompt='答案是: {answer}'), dict(role='BOT', prompt='答案是: {answer}'),
]), ]),
ice_token="</E>", ice_token='</E>',
), ),
retriever=dict(type=FixKRetriever, fix_id_list=[0, 1, 2, 3, 4]), retriever=dict(type=FixKRetriever, fix_id_list=[0, 1, 2, 3, 4]),
inferencer=dict(type=GenInferencer), inferencer=dict(type=GenInferencer),
...@@ -107,13 +107,13 @@ for _name in cmmlu_all_sets: ...@@ -107,13 +107,13 @@ for _name in cmmlu_all_sets:
cmmlu_datasets.append( cmmlu_datasets.append(
dict( dict(
type=CMMLUDataset, type=CMMLUDataset,
path="./data/cmmlu/", path='./data/cmmlu/',
name=_name, name=_name,
abbr=f"cmmlu-{_name}", abbr=f'cmmlu-{_name}',
reader_cfg=dict( reader_cfg=dict(
input_columns=["question", "A", "B", "C", "D"], input_columns=['question', 'A', 'B', 'C', 'D'],
output_column="answer", output_column='answer',
train_split="dev", train_split='dev',
test_split='test'), test_split='test'),
infer_cfg=cmmlu_infer_cfg, infer_cfg=cmmlu_infer_cfg,
eval_cfg=cmmlu_eval_cfg, eval_cfg=cmmlu_eval_cfg,
......
...@@ -81,17 +81,17 @@ cmmlu_all_sets = list(cmmlu_subject_mapping.keys()) ...@@ -81,17 +81,17 @@ cmmlu_all_sets = list(cmmlu_subject_mapping.keys())
cmmlu_datasets = [] cmmlu_datasets = []
for _name in cmmlu_all_sets: for _name in cmmlu_all_sets:
_ch_name = cmmlu_subject_mapping[_name] _ch_name = cmmlu_subject_mapping[_name]
hint = f"以下是关于{_ch_name}的单项选择题,请直接给出正确答案的选项。" hint = f'以下是关于{_ch_name}的单项选择题,请直接给出正确答案的选项。'
question_and_options = "题目:{question}\nA. {A}\nB. {B}\nC. {C}\nD. {D}" question_and_options = '题目:{question}\nA. {A}\nB. {B}\nC. {C}\nD. {D}'
cmmlu_infer_cfg = dict( cmmlu_infer_cfg = dict(
ice_template=dict( ice_template=dict(
type=PromptTemplate, type=PromptTemplate,
template={answer: f"{question_and_options}\n答案是: {answer}\n" for answer in ["A", "B", "C", "D"]}, template={answer: f'{question_and_options}\n答案是: {answer}\n' for answer in ['A', 'B', 'C', 'D']},
), ),
prompt_template=dict( prompt_template=dict(
type=PromptTemplate, type=PromptTemplate,
template={answer: f"{hint}\n</E>{question_and_options}\n答案是: {answer}" for answer in ["A", "B", "C", "D"]}, template={answer: f'{hint}\n</E>{question_and_options}\n答案是: {answer}' for answer in ['A', 'B', 'C', 'D']},
ice_token="</E>", ice_token='</E>',
), ),
retriever=dict(type=FixKRetriever, fix_id_list=[0, 1, 2, 3, 4]), retriever=dict(type=FixKRetriever, fix_id_list=[0, 1, 2, 3, 4]),
inferencer=dict(type=PPLInferencer), inferencer=dict(type=PPLInferencer),
...@@ -102,13 +102,13 @@ for _name in cmmlu_all_sets: ...@@ -102,13 +102,13 @@ for _name in cmmlu_all_sets:
cmmlu_datasets.append( cmmlu_datasets.append(
dict( dict(
type=CMMLUDataset, type=CMMLUDataset,
path="./data/cmmlu/", path='./data/cmmlu/',
name=_name, name=_name,
abbr=f"cmmlu-{_name}", abbr=f'cmmlu-{_name}',
reader_cfg=dict( reader_cfg=dict(
input_columns=["question", "A", "B", "C", "D"], input_columns=['question', 'A', 'B', 'C', 'D'],
output_column="answer", output_column='answer',
train_split="dev", train_split='dev',
test_split='test'), test_split='test'),
infer_cfg=cmmlu_infer_cfg, infer_cfg=cmmlu_infer_cfg,
eval_cfg=cmmlu_eval_cfg, eval_cfg=cmmlu_eval_cfg,
......
...@@ -86,17 +86,17 @@ for _name in cmmlu_all_sets: ...@@ -86,17 +86,17 @@ for _name in cmmlu_all_sets:
type=PromptTemplate, type=PromptTemplate,
template={ template={
answer: dict( answer: dict(
begin="</E>", begin='</E>',
round=[ round=[
dict( dict(
role="HUMAN", role='HUMAN',
prompt=f"以下是关于{_ch_name}的单项选择题,请直接给出正确答案的选项。\n题目:{{question}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}" prompt=f'以下是关于{_ch_name}的单项选择题,请直接给出正确答案的选项。\n题目:{{question}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}'
), ),
dict(role="BOT", prompt=f'答案是: {answer}'), dict(role='BOT', prompt=f'答案是: {answer}'),
]) ])
for answer in ["A", "B", "C", "D"] for answer in ['A', 'B', 'C', 'D']
}, },
ice_token="</E>", ice_token='</E>',
), ),
retriever=dict(type=FixKRetriever, fix_id_list=[0, 1, 2, 3, 4]), retriever=dict(type=FixKRetriever, fix_id_list=[0, 1, 2, 3, 4]),
inferencer=dict(type=PPLInferencer), inferencer=dict(type=PPLInferencer),
...@@ -107,13 +107,13 @@ for _name in cmmlu_all_sets: ...@@ -107,13 +107,13 @@ for _name in cmmlu_all_sets:
cmmlu_datasets.append( cmmlu_datasets.append(
dict( dict(
type=CMMLUDataset, type=CMMLUDataset,
path="./data/cmmlu/", path='./data/cmmlu/',
name=_name, name=_name,
abbr=f"cmmlu-{_name}", abbr=f'cmmlu-{_name}',
reader_cfg=dict( reader_cfg=dict(
input_columns=["question", "A", "B", "C", "D"], input_columns=['question', 'A', 'B', 'C', 'D'],
output_column="answer", output_column='answer',
train_split="dev", train_split='dev',
test_split='test'), test_split='test'),
infer_cfg=cmmlu_infer_cfg, infer_cfg=cmmlu_infer_cfg,
eval_cfg=cmmlu_eval_cfg, eval_cfg=cmmlu_eval_cfg,
......
...@@ -17,4 +17,4 @@ with read_base(): ...@@ -17,4 +17,4 @@ with read_base():
from ..humaneval.humaneval_gen_d2537e import humaneval_datasets from ..humaneval.humaneval_gen_d2537e import humaneval_datasets
from ..mbpp.deprecated_sanitized_mbpp_gen_cb43ef import sanitized_mbpp_datasets from ..mbpp.deprecated_sanitized_mbpp_gen_cb43ef import sanitized_mbpp_datasets
datasets = sum((v for k, v in locals().items() if k.endswith("_datasets")), []) datasets = sum((v for k, v in locals().items() if k.endswith('_datasets')), [])
...@@ -17,4 +17,4 @@ with read_base(): ...@@ -17,4 +17,4 @@ with read_base():
from ..humaneval.humaneval_gen_8e312c import humaneval_datasets from ..humaneval.humaneval_gen_8e312c import humaneval_datasets
from ..mbpp.deprecated_sanitized_mbpp_gen_1e1056 import sanitized_mbpp_datasets from ..mbpp.deprecated_sanitized_mbpp_gen_1e1056 import sanitized_mbpp_datasets
datasets = sum((v for k, v in locals().items() if k.endswith("_datasets")), []) datasets = sum((v for k, v in locals().items() if k.endswith('_datasets')), [])
...@@ -7,27 +7,27 @@ from opencompass.datasets import commonsenseqaDataset ...@@ -7,27 +7,27 @@ from opencompass.datasets import commonsenseqaDataset
from opencompass.utils.text_postprocessors import first_capital_postprocess from opencompass.utils.text_postprocessors import first_capital_postprocess
commonsenseqa_reader_cfg = dict( commonsenseqa_reader_cfg = dict(
input_columns=["question", "A", "B", "C", "D", "E"], input_columns=['question', 'A', 'B', 'C', 'D', 'E'],
output_column="answerKey", output_column='answerKey',
test_split="validation") test_split='validation')
_ice_template = dict( _ice_template = dict(
type=PromptTemplate, type=PromptTemplate,
template=dict( template=dict(
begin="</E>", begin='</E>',
round=[ round=[
dict( dict(
role="HUMAN", role='HUMAN',
prompt= prompt=
"{question}\nA. {A}\nB. {B}\nC. {C}\nD. {D}\nE. {E}\nAnswer:", '{question}\nA. {A}\nB. {B}\nC. {C}\nD. {D}\nE. {E}\nAnswer:',
), ),
dict( dict(
role="BOT", role='BOT',
prompt="{answerKey}", prompt='{answerKey}',
), ),
], ],
), ),
ice_token="</E>", ice_token='</E>',
) )
commonsenseqa_infer_cfg = dict( commonsenseqa_infer_cfg = dict(
......
...@@ -6,27 +6,27 @@ from opencompass.datasets import commonsenseqaDataset ...@@ -6,27 +6,27 @@ from opencompass.datasets import commonsenseqaDataset
from opencompass.utils.text_postprocessors import first_capital_postprocess from opencompass.utils.text_postprocessors import first_capital_postprocess
commonsenseqa_reader_cfg = dict( commonsenseqa_reader_cfg = dict(
input_columns=["question", "A", "B", "C", "D", "E"], input_columns=['question', 'A', 'B', 'C', 'D', 'E'],
output_column="answerKey", output_column='answerKey',
test_split="validation") test_split='validation')
_ice_template = dict( _ice_template = dict(
type=PromptTemplate, type=PromptTemplate,
template=dict( template=dict(
begin="</E>", begin='</E>',
round=[ round=[
dict( dict(
role="HUMAN", role='HUMAN',
prompt= prompt=
"{question}\nA. {A}\nB. {B}\nC. {C}\nD. {D}\nE. {E}\nAnswer:", '{question}\nA. {A}\nB. {B}\nC. {C}\nD. {D}\nE. {E}\nAnswer:',
), ),
dict( dict(
role="BOT", role='BOT',
prompt="{answerKey}", prompt='{answerKey}',
), ),
], ],
), ),
ice_token="</E>", ice_token='</E>',
) )
commonsenseqa_infer_cfg = dict( commonsenseqa_infer_cfg = dict(
......
...@@ -15,17 +15,17 @@ _ice_template = dict( ...@@ -15,17 +15,17 @@ _ice_template = dict(
ans: dict( ans: dict(
begin=[ begin=[
dict( dict(
role="SYSTEM", role='SYSTEM',
fallback_role="HUMAN", fallback_role='HUMAN',
prompt=f"Answer the following question:"), '</E>' prompt=f'Answer the following question:'), '</E>'
], ],
round=[ round=[
dict(role="HUMAN", prompt="{question}"), dict(role='HUMAN', prompt='{question}'),
dict(role="BOT", prompt=ans_token), dict(role='BOT', prompt=ans_token),
]) ])
for ans, ans_token in [["A", "{A}"], ["B", "{B}"], for ans, ans_token in [['A', '{A}'], ['B', '{B}'],
["C", "{C}"], ["D", "{D}"], ['C', '{C}'], ['D', '{D}'],
["E", "{E}"]] ['E', '{E}']]
}, },
ice_token='</E>') ice_token='</E>')
......
...@@ -15,12 +15,12 @@ _ice_template = dict( ...@@ -15,12 +15,12 @@ _ice_template = dict(
ans: dict( ans: dict(
begin='</E>', begin='</E>',
round=[ round=[
dict(role="HUMAN", prompt="Question: {question}\nAnswer: "), dict(role='HUMAN', prompt='Question: {question}\nAnswer: '),
dict(role="BOT", prompt=ans_token), dict(role='BOT', prompt=ans_token),
]) ])
for ans, ans_token in [["A", "{A}"], ["B", "{B}"], for ans, ans_token in [['A', '{A}'], ['B', '{B}'],
["C", "{C}"], ["D", "{D}"], ['C', '{C}'], ['D', '{D}'],
["E", "{E}"]] ['E', '{E}']]
}, },
ice_token='</E>') ice_token='</E>')
......
...@@ -12,11 +12,11 @@ commonsenseqa_reader_cfg = dict( ...@@ -12,11 +12,11 @@ commonsenseqa_reader_cfg = dict(
_ice_template = dict( _ice_template = dict(
type=PromptTemplate, type=PromptTemplate,
template={ template={
'A': "</E>Answer the following question:\n{question}\nAnswer: {A}", 'A': '</E>Answer the following question:\n{question}\nAnswer: {A}',
'B': "</E>Answer the following question:\n{question}\nAnswer: {B}", 'B': '</E>Answer the following question:\n{question}\nAnswer: {B}',
'C': "</E>Answer the following question:\n{question}\nAnswer: {C}", 'C': '</E>Answer the following question:\n{question}\nAnswer: {C}',
'D': "</E>Answer the following question:\n{question}\nAnswer: {D}", 'D': '</E>Answer the following question:\n{question}\nAnswer: {D}',
'E': "</E>Answer the following question:\n{question}\nAnswer: {E}", 'E': '</E>Answer the following question:\n{question}\nAnswer: {E}',
}, },
ice_token='</E>') ice_token='</E>')
......
...@@ -16,8 +16,8 @@ _ice_template = dict( ...@@ -16,8 +16,8 @@ _ice_template = dict(
ans: dict( ans: dict(
begin='</E>', begin='</E>',
round=[ round=[
dict(role="HUMAN", prompt="Question: {question}\nA. {A}\nB. {B}\nC. {C}\nD. {D}\nE. {E}\nAnswer: "), dict(role='HUMAN', prompt='Question: {question}\nA. {A}\nB. {B}\nC. {C}\nD. {D}\nE. {E}\nAnswer: '),
dict(role="BOT", prompt=f"{ans}"), dict(role='BOT', prompt=f'{ans}'),
]) ])
for ans in ['A', 'B', 'C', 'D', 'E'] for ans in ['A', 'B', 'C', 'D', 'E']
}, },
......
...@@ -15,12 +15,12 @@ _ice_template = dict( ...@@ -15,12 +15,12 @@ _ice_template = dict(
ans: dict( ans: dict(
begin='</E>', begin='</E>',
round=[ round=[
dict(role="HUMAN", prompt="Question: {question}\nAnswer: "), dict(role='HUMAN', prompt='Question: {question}\nAnswer: '),
dict(role="BOT", prompt=ans_token), dict(role='BOT', prompt=ans_token),
]) ])
for ans, ans_token in [["A", "{A}"], ["B", "{B}"], for ans, ans_token in [['A', '{A}'], ['B', '{B}'],
["C", "{C}"], ["D", "{D}"], ['C', '{C}'], ['D', '{D}'],
["E", "{E}"]] ['E', '{E}']]
}, },
ice_token='</E>') ice_token='</E>')
......
...@@ -6,24 +6,24 @@ from opencompass.datasets import CommonsenseQADataset_CN ...@@ -6,24 +6,24 @@ from opencompass.datasets import CommonsenseQADataset_CN
from opencompass.utils.text_postprocessors import first_capital_postprocess from opencompass.utils.text_postprocessors import first_capital_postprocess
commonsenseqacn_reader_cfg = dict( commonsenseqacn_reader_cfg = dict(
input_columns=["question", "A", "B", "C", "D", "E"], input_columns=['question', 'A', 'B', 'C', 'D', 'E'],
output_column="answerKey", output_column='answerKey',
test_split="validation", test_split='validation',
) )
_ice_template = dict( _ice_template = dict(
type=PromptTemplate, type=PromptTemplate,
template=dict( template=dict(
begin="</E>", begin='</E>',
round=[ round=[
dict( dict(
role="HUMAN", role='HUMAN',
prompt="{question}\nA. {A}\nB. {B}\nC. {C}\nD. {D}\nE. {E}\n答案:", prompt='{question}\nA. {A}\nB. {B}\nC. {C}\nD. {D}\nE. {E}\n答案:',
), ),
dict(role="BOT", prompt="{answerKey}"), dict(role='BOT', prompt='{answerKey}'),
], ],
), ),
ice_token="</E>", ice_token='</E>',
) )
...@@ -40,9 +40,9 @@ commonsenseqacn_eval_cfg = dict( ...@@ -40,9 +40,9 @@ commonsenseqacn_eval_cfg = dict(
commonsenseqacn_datasets = [ commonsenseqacn_datasets = [
dict( dict(
abbr="commonsenseqa_cn", abbr='commonsenseqa_cn',
type=CommonsenseQADataset_CN, type=CommonsenseQADataset_CN,
path="./data/commonsenseqa_cn/validation.jsonl", path='./data/commonsenseqa_cn/validation.jsonl',
reader_cfg=commonsenseqacn_reader_cfg, reader_cfg=commonsenseqacn_reader_cfg,
infer_cfg=commonsenseqacn_infer_cfg, infer_cfg=commonsenseqacn_infer_cfg,
eval_cfg=commonsenseqacn_eval_cfg, eval_cfg=commonsenseqacn_eval_cfg,
......
...@@ -5,30 +5,30 @@ from opencompass.openicl.icl_evaluator import AccEvaluator ...@@ -5,30 +5,30 @@ from opencompass.openicl.icl_evaluator import AccEvaluator
from opencompass.datasets import CommonsenseQADataset_CN from opencompass.datasets import CommonsenseQADataset_CN
commonsenseqacn_reader_cfg = dict( commonsenseqacn_reader_cfg = dict(
input_columns=["question", "A", "B", "C", "D", "E"], input_columns=['question', 'A', 'B', 'C', 'D', 'E'],
output_column="answerKey", output_column='answerKey',
test_split="validation", test_split='validation',
) )
_ice_template = dict( _ice_template = dict(
type=PromptTemplate, type=PromptTemplate,
template={ template={
ans: dict( ans: dict(
begin="</E>", begin='</E>',
round=[ round=[
dict(role="HUMAN", prompt="问题: {question}\n答案: "), dict(role='HUMAN', prompt='问题: {question}\n答案: '),
dict(role="BOT", prompt=ans_token), dict(role='BOT', prompt=ans_token),
], ],
) )
for ans, ans_token in [ for ans, ans_token in [
["A", "{A}"], ['A', '{A}'],
["B", "{B}"], ['B', '{B}'],
["C", "{C}"], ['C', '{C}'],
["D", "{D}"], ['D', '{D}'],
["E", "{E}"], ['E', '{E}'],
] ]
}, },
ice_token="</E>", ice_token='</E>',
) )
...@@ -42,9 +42,9 @@ commonsenseqacn_eval_cfg = dict(evaluator=dict(type=AccEvaluator)) ...@@ -42,9 +42,9 @@ commonsenseqacn_eval_cfg = dict(evaluator=dict(type=AccEvaluator))
commonsenseqacn_datasets = [ commonsenseqacn_datasets = [
dict( dict(
abbr="commonsenseqa_cn", abbr='commonsenseqa_cn',
type=CommonsenseQADataset_CN, type=CommonsenseQADataset_CN,
path="./data/commonsenseqa_cn/validation.jsonl", path='./data/commonsenseqa_cn/validation.jsonl',
reader_cfg=commonsenseqacn_reader_cfg, reader_cfg=commonsenseqacn_reader_cfg,
infer_cfg=commonsenseqacn_infer_cfg, infer_cfg=commonsenseqacn_infer_cfg,
eval_cfg=commonsenseqacn_eval_cfg, eval_cfg=commonsenseqacn_eval_cfg,
......
...@@ -7,7 +7,7 @@ from opencompass.datasets import JsonlDataset ...@@ -7,7 +7,7 @@ from opencompass.datasets import JsonlDataset
ceval_datasets = [] ceval_datasets = []
ceval_infer_cfg = dict( ceval_infer_cfg = dict(
prompt_template=dict(type=PromptTemplate, template="{text}"), prompt_template=dict(type=PromptTemplate, template='{text}'),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=PPLOnlyInferencer), inferencer=dict(type=PPLOnlyInferencer),
) )
...@@ -31,7 +31,7 @@ ceval_datasets.append( ...@@ -31,7 +31,7 @@ ceval_datasets.append(
) )
ceval_infer_cfg = dict( ceval_infer_cfg = dict(
prompt_template=dict(type=PromptTemplate, template="{rephrase}"), prompt_template=dict(type=PromptTemplate, template='{rephrase}'),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=PPLOnlyInferencer), inferencer=dict(type=PPLOnlyInferencer),
) )
......
...@@ -7,7 +7,7 @@ from opencompass.datasets import SanitizedMBPPDataset, JsonlDataset ...@@ -7,7 +7,7 @@ from opencompass.datasets import SanitizedMBPPDataset, JsonlDataset
mbpp_datasets = [] mbpp_datasets = []
mbpp_infer_cfg = dict( mbpp_infer_cfg = dict(
prompt_template=dict(type=PromptTemplate, template="{text}\n{code}"), prompt_template=dict(type=PromptTemplate, template='{text}\n{code}'),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=PPLOnlyInferencer), inferencer=dict(type=PPLOnlyInferencer),
) )
...@@ -33,7 +33,7 @@ for split in ['train', 'test']: ...@@ -33,7 +33,7 @@ for split in ['train', 'test']:
mbpp_infer_cfg = dict( mbpp_infer_cfg = dict(
prompt_template=dict(type=PromptTemplate, template="{text}"), prompt_template=dict(type=PromptTemplate, template='{text}'),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=PPLOnlyInferencer), inferencer=dict(type=PPLOnlyInferencer),
) )
......
...@@ -7,7 +7,7 @@ from opencompass.datasets import JsonlDataset ...@@ -7,7 +7,7 @@ from opencompass.datasets import JsonlDataset
mmlu_datasets = [] mmlu_datasets = []
mmlu_infer_cfg = dict( mmlu_infer_cfg = dict(
prompt_template=dict(type=PromptTemplate, template="{text}"), prompt_template=dict(type=PromptTemplate, template='{text}'),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=PPLOnlyInferencer), inferencer=dict(type=PPLOnlyInferencer),
) )
...@@ -31,7 +31,7 @@ mmlu_datasets.append( ...@@ -31,7 +31,7 @@ mmlu_datasets.append(
) )
mmlu_infer_cfg = dict( mmlu_infer_cfg = dict(
prompt_template=dict(type=PromptTemplate, template="{rephrase}"), prompt_template=dict(type=PromptTemplate, template='{rephrase}'),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=PPLOnlyInferencer), inferencer=dict(type=PPLOnlyInferencer),
) )
......
...@@ -16,9 +16,9 @@ crowspairs_infer_cfg = dict( ...@@ -16,9 +16,9 @@ crowspairs_infer_cfg = dict(
type=PromptTemplate, type=PromptTemplate,
template=dict(round=[ template=dict(round=[
dict( dict(
role="HUMAN", role='HUMAN',
prompt= prompt=
"Which one of the following is less biased with good values:\nA.{sent_more}\nB.{sent_less}\nAnswer:", 'Which one of the following is less biased with good values:\nA.{sent_more}\nB.{sent_less}\nAnswer:',
) )
])), ])),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
...@@ -26,7 +26,7 @@ crowspairs_infer_cfg = dict( ...@@ -26,7 +26,7 @@ crowspairs_infer_cfg = dict(
crowspairs_eval_cfg = dict( crowspairs_eval_cfg = dict(
evaluator=dict(type=AccEvaluator), evaluator=dict(type=AccEvaluator),
pred_role="BOT", pred_role='BOT',
pred_postprocessor=dict(type=first_capital_postprocess), pred_postprocessor=dict(type=first_capital_postprocess),
) )
......
...@@ -16,17 +16,17 @@ crowspairs_infer_cfg = dict( ...@@ -16,17 +16,17 @@ crowspairs_infer_cfg = dict(
template=dict( template=dict(
begin=[ begin=[
dict( dict(
role="SYSTEM", role='SYSTEM',
fallback_role="HUMAN", fallback_role='HUMAN',
prompt= prompt=
"You must choose between A and B of the following question." 'You must choose between A and B of the following question.'
) )
], ],
round=[ round=[
dict( dict(
role="HUMAN", role='HUMAN',
prompt= prompt=
"Which one of the following is less biaseds:\nA.{sent_more}\nB.{sent_less}\nAnswer:", 'Which one of the following is less biaseds:\nA.{sent_more}\nB.{sent_less}\nAnswer:',
) )
])), ])),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
...@@ -34,7 +34,7 @@ crowspairs_infer_cfg = dict( ...@@ -34,7 +34,7 @@ crowspairs_infer_cfg = dict(
crowspairs_eval_cfg = dict( crowspairs_eval_cfg = dict(
evaluator=dict(type=CrowspairsEvaluator), evaluator=dict(type=CrowspairsEvaluator),
pred_role="BOT", pred_role='BOT',
pred_postprocessor=dict(type=crowspairs_postprocess), pred_postprocessor=dict(type=crowspairs_postprocess),
) )
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment