Unverified Commit aa2dd2b5 authored by Fengzhe Zhou's avatar Fengzhe Zhou Committed by GitHub
Browse files

[Format] Add config lints (#892)

parent 3dbba119
...@@ -12,10 +12,10 @@ ARC_c_infer_cfg = dict( ...@@ -12,10 +12,10 @@ ARC_c_infer_cfg = dict(
prompt_template=dict( prompt_template=dict(
type=PromptTemplate, type=PromptTemplate,
template={ template={
"A": "Question: {question}\nAnswer: {textA}", 'A': 'Question: {question}\nAnswer: {textA}',
"B": "Question: {question}\nAnswer: {textB}", 'B': 'Question: {question}\nAnswer: {textB}',
"C": "Question: {question}\nAnswer: {textC}", 'C': 'Question: {question}\nAnswer: {textC}',
"D": "Question: {question}\nAnswer: {textD}" 'D': 'Question: {question}\nAnswer: {textD}'
}), }),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=PPLInferencer)) inferencer=dict(type=PPLInferencer))
......
...@@ -6,8 +6,8 @@ from opencompass.datasets import ARCDataset ...@@ -6,8 +6,8 @@ from opencompass.datasets import ARCDataset
from opencompass.utils.text_postprocessors import first_option_postprocess from opencompass.utils.text_postprocessors import first_option_postprocess
ARC_e_reader_cfg = dict( ARC_e_reader_cfg = dict(
input_columns=["question", "textA", "textB", "textC", "textD"], input_columns=['question', 'textA', 'textB', 'textC', 'textD'],
output_column="answerKey") output_column='answerKey')
ARC_e_infer_cfg = dict( ARC_e_infer_cfg = dict(
prompt_template=dict( prompt_template=dict(
...@@ -15,9 +15,9 @@ ARC_e_infer_cfg = dict( ...@@ -15,9 +15,9 @@ ARC_e_infer_cfg = dict(
template=dict( template=dict(
round=[ round=[
dict( dict(
role="HUMAN", role='HUMAN',
prompt= prompt=
"Question: {question}\nA. {textA}\nB. {textB}\nC. {textC}\nD. {textD}\nAnswer:" 'Question: {question}\nA. {textA}\nB. {textB}\nC. {textC}\nD. {textD}\nAnswer:'
) )
], ), ], ),
), ),
...@@ -27,15 +27,15 @@ ARC_e_infer_cfg = dict( ...@@ -27,15 +27,15 @@ ARC_e_infer_cfg = dict(
ARC_e_eval_cfg = dict( ARC_e_eval_cfg = dict(
evaluator=dict(type=AccEvaluator), evaluator=dict(type=AccEvaluator),
pred_role="BOT", pred_role='BOT',
pred_postprocessor=dict(type=first_option_postprocess, options='ABCD'), pred_postprocessor=dict(type=first_option_postprocess, options='ABCD'),
) )
ARC_e_datasets = [ ARC_e_datasets = [
dict( dict(
abbr="ARC-e", abbr='ARC-e',
type=ARCDataset, type=ARCDataset,
path="./data/ARC/ARC-e/ARC-Easy-Dev.jsonl", path='./data/ARC/ARC-e/ARC-Easy-Dev.jsonl',
reader_cfg=ARC_e_reader_cfg, reader_cfg=ARC_e_reader_cfg,
infer_cfg=ARC_e_infer_cfg, infer_cfg=ARC_e_infer_cfg,
eval_cfg=ARC_e_eval_cfg, eval_cfg=ARC_e_eval_cfg,
......
...@@ -14,10 +14,10 @@ ARC_e_infer_cfg = dict( ...@@ -14,10 +14,10 @@ ARC_e_infer_cfg = dict(
template={ template={
opt: dict( opt: dict(
round=[ round=[
dict(role="HUMAN", prompt=f"{{question}}\nA. {{textA}}\nB. {{textB}}\nC. {{textC}}\nD. {{textD}}"), dict(role='HUMAN', prompt=f'{{question}}\nA. {{textA}}\nB. {{textB}}\nC. {{textC}}\nD. {{textD}}'),
dict(role="BOT", prompt=f"Answer: {opt}"), dict(role='BOT', prompt=f'Answer: {opt}'),
] ]
) for opt in ["A", "B", "C", "D"] ) for opt in ['A', 'B', 'C', 'D']
}, },
), ),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
......
...@@ -12,29 +12,29 @@ ARC_e_infer_cfg = dict( ...@@ -12,29 +12,29 @@ ARC_e_infer_cfg = dict(
prompt_template=dict( prompt_template=dict(
type=PromptTemplate, type=PromptTemplate,
template={ template={
"A": 'A':
dict( dict(
round=[ round=[
dict(role="HUMAN", prompt="Question: {question}\nAnswer: "), dict(role='HUMAN', prompt='Question: {question}\nAnswer: '),
dict(role="BOT", prompt="{textA}") dict(role='BOT', prompt='{textA}')
], ), ], ),
"B": 'B':
dict( dict(
round=[ round=[
dict(role="HUMAN", prompt="Question: {question}\nAnswer: "), dict(role='HUMAN', prompt='Question: {question}\nAnswer: '),
dict(role="BOT", prompt="{textB}") dict(role='BOT', prompt='{textB}')
], ), ], ),
"C": 'C':
dict( dict(
round=[ round=[
dict(role="HUMAN", prompt="Question: {question}\nAnswer: "), dict(role='HUMAN', prompt='Question: {question}\nAnswer: '),
dict(role="BOT", prompt="{textC}") dict(role='BOT', prompt='{textC}')
], ), ], ),
"D": 'D':
dict( dict(
round=[ round=[
dict(role="HUMAN", prompt="Question: {question}\nAnswer: "), dict(role='HUMAN', prompt='Question: {question}\nAnswer: '),
dict(role="BOT", prompt="{textD}") dict(role='BOT', prompt='{textD}')
], ), ], ),
}), }),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
......
...@@ -12,10 +12,10 @@ ARC_e_infer_cfg = dict( ...@@ -12,10 +12,10 @@ ARC_e_infer_cfg = dict(
prompt_template=dict( prompt_template=dict(
type=PromptTemplate, type=PromptTemplate,
template={ template={
"A": "Question: {question}\nAnswer: {textA}", 'A': 'Question: {question}\nAnswer: {textA}',
"B": "Question: {question}\nAnswer: {textB}", 'B': 'Question: {question}\nAnswer: {textB}',
"C": "Question: {question}\nAnswer: {textC}", 'C': 'Question: {question}\nAnswer: {textC}',
"D": "Question: {question}\nAnswer: {textD}" 'D': 'Question: {question}\nAnswer: {textD}'
}), }),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=PPLInferencer)) inferencer=dict(type=PPLInferencer))
......
...@@ -5,8 +5,8 @@ from opencompass.openicl.icl_inferencer import AgentInferencer ...@@ -5,8 +5,8 @@ from opencompass.openicl.icl_inferencer import AgentInferencer
from opencompass.datasets import CIBenchDataset, CIBenchEvaluator from opencompass.datasets import CIBenchDataset, CIBenchEvaluator
cibench_reader_cfg = dict( cibench_reader_cfg = dict(
input_columns=["questions"], input_columns=['questions'],
output_column="references", output_column='references',
train_split='test', train_split='test',
test_split='test') test_split='test')
...@@ -20,13 +20,13 @@ cibench_infer_cfg = dict( ...@@ -20,13 +20,13 @@ cibench_infer_cfg = dict(
) )
libs = ['matplotlib', 'opencv', 'pandas', 'pytorch', 'scipy', 'seaborn'] libs = ['matplotlib', 'opencv', 'pandas', 'pytorch', 'scipy', 'seaborn']
cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role="BOT") cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role='BOT')
cibench_datasets = [ cibench_datasets = [
dict( dict(
abbr=f"cibench_generation/{lib}", abbr=f'cibench_generation/{lib}',
type=CIBenchDataset, type=CIBenchDataset,
path=f"./data/cibench_dataset/cibench_generation/{lib}", path=f'./data/cibench_dataset/cibench_generation/{lib}',
internet_check=False, internet_check=False,
reader_cfg=cibench_reader_cfg, reader_cfg=cibench_reader_cfg,
infer_cfg=cibench_infer_cfg, infer_cfg=cibench_infer_cfg,
......
...@@ -5,8 +5,8 @@ from opencompass.openicl.icl_inferencer import AgentInferencer ...@@ -5,8 +5,8 @@ from opencompass.openicl.icl_inferencer import AgentInferencer
from opencompass.datasets import CIBenchDataset, CIBenchEvaluator from opencompass.datasets import CIBenchDataset, CIBenchEvaluator
cibench_reader_cfg = dict( cibench_reader_cfg = dict(
input_columns=["questions"], input_columns=['questions'],
output_column="references", output_column='references',
train_split='test', train_split='test',
test_split='test') test_split='test')
...@@ -20,13 +20,13 @@ cibench_infer_cfg = dict( ...@@ -20,13 +20,13 @@ cibench_infer_cfg = dict(
) )
libs = ['matplotlib', 'opencv', 'pandas', 'pytorch', 'scipy', 'seaborn'] libs = ['matplotlib', 'opencv', 'pandas', 'pytorch', 'scipy', 'seaborn']
cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role="BOT") cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role='BOT')
cibench_datasets = [ cibench_datasets = [
dict( dict(
abbr=f"cibench_generation_oracle/{lib}", abbr=f'cibench_generation_oracle/{lib}',
type=CIBenchDataset, type=CIBenchDataset,
path=f"./data/cibench_dataset/cibench_generation/{lib}", path=f'./data/cibench_dataset/cibench_generation/{lib}',
internet_check=False, internet_check=False,
reader_cfg=cibench_reader_cfg, reader_cfg=cibench_reader_cfg,
infer_cfg=cibench_infer_cfg, infer_cfg=cibench_infer_cfg,
......
...@@ -5,8 +5,8 @@ from opencompass.openicl.icl_inferencer import AgentInferencer ...@@ -5,8 +5,8 @@ from opencompass.openicl.icl_inferencer import AgentInferencer
from opencompass.datasets import CIBenchDataset, CIBenchEvaluator from opencompass.datasets import CIBenchDataset, CIBenchEvaluator
cibench_reader_cfg = dict( cibench_reader_cfg = dict(
input_columns=["questions"], input_columns=['questions'],
output_column="references", output_column='references',
train_split='test', train_split='test',
test_split='test') test_split='test')
...@@ -25,12 +25,12 @@ libs = ['/lightgbm', '/matplotlib', '/nltk', '/opencv', '/pandas', '/pytorch', ...@@ -25,12 +25,12 @@ libs = ['/lightgbm', '/matplotlib', '/nltk', '/opencv', '/pandas', '/pytorch',
'_chinese/lightgbm', '_chinese/matplotlib', '_chinese/nltk', '_chinese/lightgbm', '_chinese/matplotlib', '_chinese/nltk',
'_chinese/opencv', '_chinese/pandas', '_chinese/pytorch', '_chinese/opencv', '_chinese/pandas', '_chinese/pytorch',
'_chinese/scipy', '_chinese/seaborn', '_chinese/sklearn', '_chinese/tensorflow'] '_chinese/scipy', '_chinese/seaborn', '_chinese/sklearn', '_chinese/tensorflow']
cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role="BOT") cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role='BOT')
cibench_datasets = [ cibench_datasets = [
dict( dict(
abbr=f"cibench_template{lib}", abbr=f'cibench_template{lib}',
type=CIBenchDataset, type=CIBenchDataset,
path=f"./data/cibench_dataset/cibench_template{lib}", path=f'./data/cibench_dataset/cibench_template{lib}',
internet_check=False, internet_check=False,
reader_cfg=cibench_reader_cfg, reader_cfg=cibench_reader_cfg,
infer_cfg=cibench_infer_cfg, infer_cfg=cibench_infer_cfg,
......
...@@ -5,8 +5,8 @@ from opencompass.openicl.icl_inferencer import AgentInferencer ...@@ -5,8 +5,8 @@ from opencompass.openicl.icl_inferencer import AgentInferencer
from opencompass.datasets import CIBenchDataset, CIBenchEvaluator from opencompass.datasets import CIBenchDataset, CIBenchEvaluator
cibench_reader_cfg = dict( cibench_reader_cfg = dict(
input_columns=["questions"], input_columns=['questions'],
output_column="references", output_column='references',
train_split='test', train_split='test',
test_split='test') test_split='test')
...@@ -25,12 +25,12 @@ libs = ['/lightgbm', '/matplotlib', '/nltk', '/opencv', '/pandas', '/pytorch', ...@@ -25,12 +25,12 @@ libs = ['/lightgbm', '/matplotlib', '/nltk', '/opencv', '/pandas', '/pytorch',
'_chinese/lightgbm', '_chinese/matplotlib', '_chinese/nltk', '_chinese/lightgbm', '_chinese/matplotlib', '_chinese/nltk',
'_chinese/opencv', '_chinese/pandas', '_chinese/pytorch', '_chinese/opencv', '_chinese/pandas', '_chinese/pytorch',
'_chinese/scipy', '_chinese/seaborn', '_chinese/sklearn', '_chinese/tensorflow'] '_chinese/scipy', '_chinese/seaborn', '_chinese/sklearn', '_chinese/tensorflow']
cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role="BOT") cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role='BOT')
cibench_datasets = [ cibench_datasets = [
dict( dict(
abbr=f"cibench_template_oracle{lib}", abbr=f'cibench_template_oracle{lib}',
type=CIBenchDataset, type=CIBenchDataset,
path=f"./data/cibench_dataset/cibench_template{lib}", path=f'./data/cibench_dataset/cibench_template{lib}',
internet_check=False, internet_check=False,
reader_cfg=cibench_reader_cfg, reader_cfg=cibench_reader_cfg,
infer_cfg=cibench_infer_cfg, infer_cfg=cibench_infer_cfg,
......
...@@ -7,15 +7,15 @@ from opencompass.utils.text_postprocessors import first_capital_postprocess ...@@ -7,15 +7,15 @@ from opencompass.utils.text_postprocessors import first_capital_postprocess
C3_reader_cfg = dict( C3_reader_cfg = dict(
input_columns=[ input_columns=[
"question", 'question',
"content", 'content',
"choice0", 'choice0',
"choice1", 'choice1',
"choice2", 'choice2',
"choice3", 'choice3',
"choices", 'choices',
], ],
output_column="label", output_column='label',
) )
C3_infer_cfg = dict( C3_infer_cfg = dict(
...@@ -23,9 +23,9 @@ C3_infer_cfg = dict( ...@@ -23,9 +23,9 @@ C3_infer_cfg = dict(
type=PromptTemplate, type=PromptTemplate,
template=dict(round=[ template=dict(round=[
dict( dict(
role="HUMAN", role='HUMAN',
prompt= prompt=
"{content}\n问:{question}\nA. {choice0}\nB. {choice1}\nC. {choice2}\nD. {choice3}\n请从“A”,“B”,“C”,“D”中进行选择。\n答:", '{content}\n问:{question}\nA. {choice0}\nB. {choice1}\nC. {choice2}\nD. {choice3}\n请从“A”,“B”,“C”,“D”中进行选择。\n答:',
), ),
]), ]),
), ),
...@@ -35,15 +35,15 @@ C3_infer_cfg = dict( ...@@ -35,15 +35,15 @@ C3_infer_cfg = dict(
C3_eval_cfg = dict( C3_eval_cfg = dict(
evaluator=dict(type=AccEvaluator), evaluator=dict(type=AccEvaluator),
pred_role="BOT", pred_role='BOT',
pred_postprocessor=dict(type=first_capital_postprocess), pred_postprocessor=dict(type=first_capital_postprocess),
) )
C3_datasets = [ C3_datasets = [
dict( dict(
abbr="C3", abbr='C3',
type=C3Dataset_V2, type=C3Dataset_V2,
path="./data/CLUE/C3/dev_0.json", path='./data/CLUE/C3/dev_0.json',
reader_cfg=C3_reader_cfg, reader_cfg=C3_reader_cfg,
infer_cfg=C3_infer_cfg, infer_cfg=C3_infer_cfg,
eval_cfg=C3_eval_cfg, eval_cfg=C3_eval_cfg,
......
...@@ -15,10 +15,10 @@ C3_infer_cfg = dict( ...@@ -15,10 +15,10 @@ C3_infer_cfg = dict(
prompt_template=dict( prompt_template=dict(
type=PromptTemplate, type=PromptTemplate,
template={ template={
0: "文章:{content}\n问题:{question}\n答案:{choice0}", 0: '文章:{content}\n问题:{question}\n答案:{choice0}',
1: "文章:{content}\n问题:{question}\n答案:{choice1}", 1: '文章:{content}\n问题:{question}\n答案:{choice1}',
2: "文章:{content}\n问题:{question}\n答案:{choice2}", 2: '文章:{content}\n问题:{question}\n答案:{choice2}',
3: "文章:{content}\n问题:{question}\n答案:{choice3}" 3: '文章:{content}\n问题:{question}\n答案:{choice3}'
}), }),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=PPLInferencer)) inferencer=dict(type=PPLInferencer))
......
...@@ -16,8 +16,8 @@ C3_infer_cfg = dict( ...@@ -16,8 +16,8 @@ C3_infer_cfg = dict(
type=PromptTemplate, type=PromptTemplate,
template={ template={
i: dict(round=[ i: dict(round=[
dict(role="HUMAN", prompt="文章:{content}\n问题:{question}"), dict(role='HUMAN', prompt='文章:{content}\n问题:{question}'),
dict(role="BOT", prompt=f"答案:{{choice{i}}}") dict(role='BOT', prompt=f'答案:{{choice{i}}}')
]) ])
for i in range(4) for i in range(4)
}), }),
......
...@@ -12,15 +12,15 @@ CMRC_infer_cfg = dict( ...@@ -12,15 +12,15 @@ CMRC_infer_cfg = dict(
type=PromptTemplate, type=PromptTemplate,
template=dict(round=[ template=dict(round=[
dict( dict(
role="HUMAN", role='HUMAN',
prompt="根据文章回答问题。你的答案应该尽可能简练,请以 ‘答案是’ 开头的句式作答。\n文章:{context}\n问:{question}\n答:"), prompt='根据文章回答问题。你的答案应该尽可能简练,请以 ‘答案是’ 开头的句式作答。\n文章:{context}\n问:{question}\n答:'),
])), ])),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer)) inferencer=dict(type=GenInferencer))
CMRC_eval_cfg = dict( CMRC_eval_cfg = dict(
evaluator=dict(type=EMEvaluator), evaluator=dict(type=EMEvaluator),
pred_role="BOT", pred_role='BOT',
pred_postprocessor=dict(type=cmrc_postprocess), pred_postprocessor=dict(type=cmrc_postprocess),
) )
......
...@@ -11,15 +11,15 @@ CMRC_infer_cfg = dict( ...@@ -11,15 +11,15 @@ CMRC_infer_cfg = dict(
prompt_template=dict( prompt_template=dict(
type=PromptTemplate, type=PromptTemplate,
template=dict(round=[ template=dict(round=[
dict(role="HUMAN", prompt="文章:{context}\n根据上文,回答如下问题:{question}"), dict(role='HUMAN', prompt='文章:{context}\n根据上文,回答如下问题:{question}'),
dict(role="BOT", prompt="答:"), dict(role='BOT', prompt='答:'),
])), ])),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer)) inferencer=dict(type=GenInferencer))
CMRC_eval_cfg = dict( CMRC_eval_cfg = dict(
evaluator=dict(type=EMEvaluator), evaluator=dict(type=EMEvaluator),
pred_role="BOT", pred_role='BOT',
) )
CMRC_datasets = [ CMRC_datasets = [
......
...@@ -10,7 +10,7 @@ CMRC_reader_cfg = dict( ...@@ -10,7 +10,7 @@ CMRC_reader_cfg = dict(
CMRC_infer_cfg = dict( CMRC_infer_cfg = dict(
prompt_template=dict( prompt_template=dict(
type=PromptTemplate, type=PromptTemplate,
template="文章:{context}\n根据上文,回答如下问题: {question}\n答:"), template='文章:{context}\n根据上文,回答如下问题: {question}\n答:'),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer)) inferencer=dict(type=GenInferencer))
......
...@@ -12,15 +12,15 @@ CMRC_infer_cfg = dict( ...@@ -12,15 +12,15 @@ CMRC_infer_cfg = dict(
type=PromptTemplate, type=PromptTemplate,
template=dict(round=[ template=dict(round=[
dict( dict(
role="HUMAN", role='HUMAN',
prompt="文章:{context}\n根据上文,回答如下问题:\n{question}\n答:"), prompt='文章:{context}\n根据上文,回答如下问题:\n{question}\n答:'),
])), ])),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer)) inferencer=dict(type=GenInferencer))
CMRC_eval_cfg = dict( CMRC_eval_cfg = dict(
evaluator=dict(type=EMEvaluator), evaluator=dict(type=EMEvaluator),
pred_role="BOT", pred_role='BOT',
) )
CMRC_datasets = [ CMRC_datasets = [
......
...@@ -12,15 +12,15 @@ DRCD_infer_cfg = dict( ...@@ -12,15 +12,15 @@ DRCD_infer_cfg = dict(
type=PromptTemplate, type=PromptTemplate,
template=dict(round=[ template=dict(round=[
dict( dict(
role="HUMAN", role='HUMAN',
prompt="根据文章回答问题。你的答案应该尽可能简练,请以 ‘答案是’ 开头的句式作答。\n文章:{context}\n问:{question}\n答:"), prompt='根据文章回答问题。你的答案应该尽可能简练,请以 ‘答案是’ 开头的句式作答。\n文章:{context}\n问:{question}\n答:'),
])), ])),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer)) inferencer=dict(type=GenInferencer))
DRCD_eval_cfg = dict( DRCD_eval_cfg = dict(
evaluator=dict(type=EMEvaluator), evaluator=dict(type=EMEvaluator),
pred_role="BOT", pred_role='BOT',
pred_postprocessor=dict(type=drcd_postprocess), pred_postprocessor=dict(type=drcd_postprocess),
) )
......
...@@ -11,15 +11,15 @@ DRCD_infer_cfg = dict( ...@@ -11,15 +11,15 @@ DRCD_infer_cfg = dict(
prompt_template=dict( prompt_template=dict(
type=PromptTemplate, type=PromptTemplate,
template=dict(round=[ template=dict(round=[
dict(role="HUMAN", prompt="文章:{context}\n根据上文,回答如下问题:{question}"), dict(role='HUMAN', prompt='文章:{context}\n根据上文,回答如下问题:{question}'),
dict(role="BOT", prompt="答:"), dict(role='BOT', prompt='答:'),
])), ])),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer)) inferencer=dict(type=GenInferencer))
DRCD_eval_cfg = dict( DRCD_eval_cfg = dict(
evaluator=dict(type=EMEvaluator), evaluator=dict(type=EMEvaluator),
pred_role="BOT", pred_role='BOT',
) )
DRCD_datasets = [ DRCD_datasets = [
......
...@@ -10,7 +10,7 @@ DRCD_reader_cfg = dict( ...@@ -10,7 +10,7 @@ DRCD_reader_cfg = dict(
DRCD_infer_cfg = dict( DRCD_infer_cfg = dict(
prompt_template=dict( prompt_template=dict(
type=PromptTemplate, type=PromptTemplate,
template="文章:{context}\n根据上文,回答如下问题: {question}\n答:"), template='文章:{context}\n根据上文,回答如下问题: {question}\n答:'),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer)) inferencer=dict(type=GenInferencer))
......
...@@ -12,15 +12,15 @@ DRCD_infer_cfg = dict( ...@@ -12,15 +12,15 @@ DRCD_infer_cfg = dict(
type=PromptTemplate, type=PromptTemplate,
template=dict(round=[ template=dict(round=[
dict( dict(
role="HUMAN", role='HUMAN',
prompt="文章:{context}\n根据上文,回答如下问题:\n{question}\n答:"), prompt='文章:{context}\n根据上文,回答如下问题:\n{question}\n答:'),
])), ])),
retriever=dict(type=ZeroRetriever), retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer)) inferencer=dict(type=GenInferencer))
DRCD_eval_cfg = dict( DRCD_eval_cfg = dict(
evaluator=dict(type=EMEvaluator), evaluator=dict(type=EMEvaluator),
pred_role="BOT", pred_role='BOT',
) )
DRCD_datasets = [ DRCD_datasets = [
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment