Unverified Commit aa2dd2b5 authored by Fengzhe Zhou's avatar Fengzhe Zhou Committed by GitHub
Browse files

[Format] Add config lints (#892)

parent 3dbba119
......@@ -12,10 +12,10 @@ ARC_c_infer_cfg = dict(
prompt_template=dict(
type=PromptTemplate,
template={
"A": "Question: {question}\nAnswer: {textA}",
"B": "Question: {question}\nAnswer: {textB}",
"C": "Question: {question}\nAnswer: {textC}",
"D": "Question: {question}\nAnswer: {textD}"
'A': 'Question: {question}\nAnswer: {textA}',
'B': 'Question: {question}\nAnswer: {textB}',
'C': 'Question: {question}\nAnswer: {textC}',
'D': 'Question: {question}\nAnswer: {textD}'
}),
retriever=dict(type=ZeroRetriever),
inferencer=dict(type=PPLInferencer))
......
......@@ -6,8 +6,8 @@ from opencompass.datasets import ARCDataset
from opencompass.utils.text_postprocessors import first_option_postprocess
ARC_e_reader_cfg = dict(
input_columns=["question", "textA", "textB", "textC", "textD"],
output_column="answerKey")
input_columns=['question', 'textA', 'textB', 'textC', 'textD'],
output_column='answerKey')
ARC_e_infer_cfg = dict(
prompt_template=dict(
......@@ -15,9 +15,9 @@ ARC_e_infer_cfg = dict(
template=dict(
round=[
dict(
role="HUMAN",
role='HUMAN',
prompt=
"Question: {question}\nA. {textA}\nB. {textB}\nC. {textC}\nD. {textD}\nAnswer:"
'Question: {question}\nA. {textA}\nB. {textB}\nC. {textC}\nD. {textD}\nAnswer:'
)
], ),
),
......@@ -27,15 +27,15 @@ ARC_e_infer_cfg = dict(
ARC_e_eval_cfg = dict(
evaluator=dict(type=AccEvaluator),
pred_role="BOT",
pred_role='BOT',
pred_postprocessor=dict(type=first_option_postprocess, options='ABCD'),
)
ARC_e_datasets = [
dict(
abbr="ARC-e",
abbr='ARC-e',
type=ARCDataset,
path="./data/ARC/ARC-e/ARC-Easy-Dev.jsonl",
path='./data/ARC/ARC-e/ARC-Easy-Dev.jsonl',
reader_cfg=ARC_e_reader_cfg,
infer_cfg=ARC_e_infer_cfg,
eval_cfg=ARC_e_eval_cfg,
......
......@@ -14,10 +14,10 @@ ARC_e_infer_cfg = dict(
template={
opt: dict(
round=[
dict(role="HUMAN", prompt=f"{{question}}\nA. {{textA}}\nB. {{textB}}\nC. {{textC}}\nD. {{textD}}"),
dict(role="BOT", prompt=f"Answer: {opt}"),
dict(role='HUMAN', prompt=f'{{question}}\nA. {{textA}}\nB. {{textB}}\nC. {{textC}}\nD. {{textD}}'),
dict(role='BOT', prompt=f'Answer: {opt}'),
]
) for opt in ["A", "B", "C", "D"]
) for opt in ['A', 'B', 'C', 'D']
},
),
retriever=dict(type=ZeroRetriever),
......
......@@ -12,29 +12,29 @@ ARC_e_infer_cfg = dict(
prompt_template=dict(
type=PromptTemplate,
template={
"A":
'A':
dict(
round=[
dict(role="HUMAN", prompt="Question: {question}\nAnswer: "),
dict(role="BOT", prompt="{textA}")
dict(role='HUMAN', prompt='Question: {question}\nAnswer: '),
dict(role='BOT', prompt='{textA}')
], ),
"B":
'B':
dict(
round=[
dict(role="HUMAN", prompt="Question: {question}\nAnswer: "),
dict(role="BOT", prompt="{textB}")
dict(role='HUMAN', prompt='Question: {question}\nAnswer: '),
dict(role='BOT', prompt='{textB}')
], ),
"C":
'C':
dict(
round=[
dict(role="HUMAN", prompt="Question: {question}\nAnswer: "),
dict(role="BOT", prompt="{textC}")
dict(role='HUMAN', prompt='Question: {question}\nAnswer: '),
dict(role='BOT', prompt='{textC}')
], ),
"D":
'D':
dict(
round=[
dict(role="HUMAN", prompt="Question: {question}\nAnswer: "),
dict(role="BOT", prompt="{textD}")
dict(role='HUMAN', prompt='Question: {question}\nAnswer: '),
dict(role='BOT', prompt='{textD}')
], ),
}),
retriever=dict(type=ZeroRetriever),
......
......@@ -12,10 +12,10 @@ ARC_e_infer_cfg = dict(
prompt_template=dict(
type=PromptTemplate,
template={
"A": "Question: {question}\nAnswer: {textA}",
"B": "Question: {question}\nAnswer: {textB}",
"C": "Question: {question}\nAnswer: {textC}",
"D": "Question: {question}\nAnswer: {textD}"
'A': 'Question: {question}\nAnswer: {textA}',
'B': 'Question: {question}\nAnswer: {textB}',
'C': 'Question: {question}\nAnswer: {textC}',
'D': 'Question: {question}\nAnswer: {textD}'
}),
retriever=dict(type=ZeroRetriever),
inferencer=dict(type=PPLInferencer))
......
......@@ -5,8 +5,8 @@ from opencompass.openicl.icl_inferencer import AgentInferencer
from opencompass.datasets import CIBenchDataset, CIBenchEvaluator
cibench_reader_cfg = dict(
input_columns=["questions"],
output_column="references",
input_columns=['questions'],
output_column='references',
train_split='test',
test_split='test')
......@@ -20,13 +20,13 @@ cibench_infer_cfg = dict(
)
libs = ['matplotlib', 'opencv', 'pandas', 'pytorch', 'scipy', 'seaborn']
cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role="BOT")
cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role='BOT')
cibench_datasets = [
dict(
abbr=f"cibench_generation/{lib}",
abbr=f'cibench_generation/{lib}',
type=CIBenchDataset,
path=f"./data/cibench_dataset/cibench_generation/{lib}",
path=f'./data/cibench_dataset/cibench_generation/{lib}',
internet_check=False,
reader_cfg=cibench_reader_cfg,
infer_cfg=cibench_infer_cfg,
......
......@@ -5,8 +5,8 @@ from opencompass.openicl.icl_inferencer import AgentInferencer
from opencompass.datasets import CIBenchDataset, CIBenchEvaluator
cibench_reader_cfg = dict(
input_columns=["questions"],
output_column="references",
input_columns=['questions'],
output_column='references',
train_split='test',
test_split='test')
......@@ -20,13 +20,13 @@ cibench_infer_cfg = dict(
)
libs = ['matplotlib', 'opencv', 'pandas', 'pytorch', 'scipy', 'seaborn']
cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role="BOT")
cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role='BOT')
cibench_datasets = [
dict(
abbr=f"cibench_generation_oracle/{lib}",
abbr=f'cibench_generation_oracle/{lib}',
type=CIBenchDataset,
path=f"./data/cibench_dataset/cibench_generation/{lib}",
path=f'./data/cibench_dataset/cibench_generation/{lib}',
internet_check=False,
reader_cfg=cibench_reader_cfg,
infer_cfg=cibench_infer_cfg,
......
......@@ -5,8 +5,8 @@ from opencompass.openicl.icl_inferencer import AgentInferencer
from opencompass.datasets import CIBenchDataset, CIBenchEvaluator
cibench_reader_cfg = dict(
input_columns=["questions"],
output_column="references",
input_columns=['questions'],
output_column='references',
train_split='test',
test_split='test')
......@@ -25,12 +25,12 @@ libs = ['/lightgbm', '/matplotlib', '/nltk', '/opencv', '/pandas', '/pytorch',
'_chinese/lightgbm', '_chinese/matplotlib', '_chinese/nltk',
'_chinese/opencv', '_chinese/pandas', '_chinese/pytorch',
'_chinese/scipy', '_chinese/seaborn', '_chinese/sklearn', '_chinese/tensorflow']
cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role="BOT")
cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role='BOT')
cibench_datasets = [
dict(
abbr=f"cibench_template{lib}",
abbr=f'cibench_template{lib}',
type=CIBenchDataset,
path=f"./data/cibench_dataset/cibench_template{lib}",
path=f'./data/cibench_dataset/cibench_template{lib}',
internet_check=False,
reader_cfg=cibench_reader_cfg,
infer_cfg=cibench_infer_cfg,
......
......@@ -5,8 +5,8 @@ from opencompass.openicl.icl_inferencer import AgentInferencer
from opencompass.datasets import CIBenchDataset, CIBenchEvaluator
cibench_reader_cfg = dict(
input_columns=["questions"],
output_column="references",
input_columns=['questions'],
output_column='references',
train_split='test',
test_split='test')
......@@ -25,12 +25,12 @@ libs = ['/lightgbm', '/matplotlib', '/nltk', '/opencv', '/pandas', '/pytorch',
'_chinese/lightgbm', '_chinese/matplotlib', '_chinese/nltk',
'_chinese/opencv', '_chinese/pandas', '_chinese/pytorch',
'_chinese/scipy', '_chinese/seaborn', '_chinese/sklearn', '_chinese/tensorflow']
cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role="BOT")
cibench_eval_cfg = dict(evaluator=dict(type=CIBenchEvaluator), pred_role='BOT')
cibench_datasets = [
dict(
abbr=f"cibench_template_oracle{lib}",
abbr=f'cibench_template_oracle{lib}',
type=CIBenchDataset,
path=f"./data/cibench_dataset/cibench_template{lib}",
path=f'./data/cibench_dataset/cibench_template{lib}',
internet_check=False,
reader_cfg=cibench_reader_cfg,
infer_cfg=cibench_infer_cfg,
......
......@@ -7,15 +7,15 @@ from opencompass.utils.text_postprocessors import first_capital_postprocess
C3_reader_cfg = dict(
input_columns=[
"question",
"content",
"choice0",
"choice1",
"choice2",
"choice3",
"choices",
'question',
'content',
'choice0',
'choice1',
'choice2',
'choice3',
'choices',
],
output_column="label",
output_column='label',
)
C3_infer_cfg = dict(
......@@ -23,9 +23,9 @@ C3_infer_cfg = dict(
type=PromptTemplate,
template=dict(round=[
dict(
role="HUMAN",
role='HUMAN',
prompt=
"{content}\n问:{question}\nA. {choice0}\nB. {choice1}\nC. {choice2}\nD. {choice3}\n请从“A”,“B”,“C”,“D”中进行选择。\n答:",
'{content}\n问:{question}\nA. {choice0}\nB. {choice1}\nC. {choice2}\nD. {choice3}\n请从“A”,“B”,“C”,“D”中进行选择。\n答:',
),
]),
),
......@@ -35,15 +35,15 @@ C3_infer_cfg = dict(
C3_eval_cfg = dict(
evaluator=dict(type=AccEvaluator),
pred_role="BOT",
pred_role='BOT',
pred_postprocessor=dict(type=first_capital_postprocess),
)
C3_datasets = [
dict(
abbr="C3",
abbr='C3',
type=C3Dataset_V2,
path="./data/CLUE/C3/dev_0.json",
path='./data/CLUE/C3/dev_0.json',
reader_cfg=C3_reader_cfg,
infer_cfg=C3_infer_cfg,
eval_cfg=C3_eval_cfg,
......
......@@ -15,10 +15,10 @@ C3_infer_cfg = dict(
prompt_template=dict(
type=PromptTemplate,
template={
0: "文章:{content}\n问题:{question}\n答案:{choice0}",
1: "文章:{content}\n问题:{question}\n答案:{choice1}",
2: "文章:{content}\n问题:{question}\n答案:{choice2}",
3: "文章:{content}\n问题:{question}\n答案:{choice3}"
0: '文章:{content}\n问题:{question}\n答案:{choice0}',
1: '文章:{content}\n问题:{question}\n答案:{choice1}',
2: '文章:{content}\n问题:{question}\n答案:{choice2}',
3: '文章:{content}\n问题:{question}\n答案:{choice3}'
}),
retriever=dict(type=ZeroRetriever),
inferencer=dict(type=PPLInferencer))
......
......@@ -16,8 +16,8 @@ C3_infer_cfg = dict(
type=PromptTemplate,
template={
i: dict(round=[
dict(role="HUMAN", prompt="文章:{content}\n问题:{question}"),
dict(role="BOT", prompt=f"答案:{{choice{i}}}")
dict(role='HUMAN', prompt='文章:{content}\n问题:{question}'),
dict(role='BOT', prompt=f'答案:{{choice{i}}}')
])
for i in range(4)
}),
......
......@@ -12,15 +12,15 @@ CMRC_infer_cfg = dict(
type=PromptTemplate,
template=dict(round=[
dict(
role="HUMAN",
prompt="根据文章回答问题。你的答案应该尽可能简练,请以 ‘答案是’ 开头的句式作答。\n文章:{context}\n问:{question}\n答:"),
role='HUMAN',
prompt='根据文章回答问题。你的答案应该尽可能简练,请以 ‘答案是’ 开头的句式作答。\n文章:{context}\n问:{question}\n答:'),
])),
retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer))
CMRC_eval_cfg = dict(
evaluator=dict(type=EMEvaluator),
pred_role="BOT",
pred_role='BOT',
pred_postprocessor=dict(type=cmrc_postprocess),
)
......
......@@ -11,15 +11,15 @@ CMRC_infer_cfg = dict(
prompt_template=dict(
type=PromptTemplate,
template=dict(round=[
dict(role="HUMAN", prompt="文章:{context}\n根据上文,回答如下问题:{question}"),
dict(role="BOT", prompt="答:"),
dict(role='HUMAN', prompt='文章:{context}\n根据上文,回答如下问题:{question}'),
dict(role='BOT', prompt='答:'),
])),
retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer))
CMRC_eval_cfg = dict(
evaluator=dict(type=EMEvaluator),
pred_role="BOT",
pred_role='BOT',
)
CMRC_datasets = [
......
......@@ -10,7 +10,7 @@ CMRC_reader_cfg = dict(
CMRC_infer_cfg = dict(
prompt_template=dict(
type=PromptTemplate,
template="文章:{context}\n根据上文,回答如下问题: {question}\n答:"),
template='文章:{context}\n根据上文,回答如下问题: {question}\n答:'),
retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer))
......
......@@ -12,15 +12,15 @@ CMRC_infer_cfg = dict(
type=PromptTemplate,
template=dict(round=[
dict(
role="HUMAN",
prompt="文章:{context}\n根据上文,回答如下问题:\n{question}\n答:"),
role='HUMAN',
prompt='文章:{context}\n根据上文,回答如下问题:\n{question}\n答:'),
])),
retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer))
CMRC_eval_cfg = dict(
evaluator=dict(type=EMEvaluator),
pred_role="BOT",
pred_role='BOT',
)
CMRC_datasets = [
......
......@@ -12,15 +12,15 @@ DRCD_infer_cfg = dict(
type=PromptTemplate,
template=dict(round=[
dict(
role="HUMAN",
prompt="根据文章回答问题。你的答案应该尽可能简练,请以 ‘答案是’ 开头的句式作答。\n文章:{context}\n问:{question}\n答:"),
role='HUMAN',
prompt='根据文章回答问题。你的答案应该尽可能简练,请以 ‘答案是’ 开头的句式作答。\n文章:{context}\n问:{question}\n答:'),
])),
retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer))
DRCD_eval_cfg = dict(
evaluator=dict(type=EMEvaluator),
pred_role="BOT",
pred_role='BOT',
pred_postprocessor=dict(type=drcd_postprocess),
)
......
......@@ -11,15 +11,15 @@ DRCD_infer_cfg = dict(
prompt_template=dict(
type=PromptTemplate,
template=dict(round=[
dict(role="HUMAN", prompt="文章:{context}\n根据上文,回答如下问题:{question}"),
dict(role="BOT", prompt="答:"),
dict(role='HUMAN', prompt='文章:{context}\n根据上文,回答如下问题:{question}'),
dict(role='BOT', prompt='答:'),
])),
retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer))
DRCD_eval_cfg = dict(
evaluator=dict(type=EMEvaluator),
pred_role="BOT",
pred_role='BOT',
)
DRCD_datasets = [
......
......@@ -10,7 +10,7 @@ DRCD_reader_cfg = dict(
DRCD_infer_cfg = dict(
prompt_template=dict(
type=PromptTemplate,
template="文章:{context}\n根据上文,回答如下问题: {question}\n答:"),
template='文章:{context}\n根据上文,回答如下问题: {question}\n答:'),
retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer))
......
......@@ -12,15 +12,15 @@ DRCD_infer_cfg = dict(
type=PromptTemplate,
template=dict(round=[
dict(
role="HUMAN",
prompt="文章:{context}\n根据上文,回答如下问题:\n{question}\n答:"),
role='HUMAN',
prompt='文章:{context}\n根据上文,回答如下问题:\n{question}\n答:'),
])),
retriever=dict(type=ZeroRetriever),
inferencer=dict(type=GenInferencer))
DRCD_eval_cfg = dict(
evaluator=dict(type=EMEvaluator),
pred_role="BOT",
pred_role='BOT',
)
DRCD_datasets = [
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment