Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
wangsen
MinerU
Commits
b5b58d64
Unverified
Commit
b5b58d64
authored
Apr 10, 2024
by
Kaiwen Liu
Committed by
GitHub
Apr 10, 2024
Browse files
Merge branch 'magicpdf:master' into master
parents
6bc9df82
db54796a
Changes
3
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
269 additions
and
16 deletions
+269
-16
.github/workflows/benchmark.yml
.github/workflows/benchmark.yml
+5
-3
tools/base_data.json
tools/base_data.json
+87
-0
tools/ocr_badcase.py
tools/ocr_badcase.py
+177
-13
No files found.
.github/workflows/benchmark.yml
View file @
b5b58d64
...
@@ -18,14 +18,16 @@ jobs:
...
@@ -18,14 +18,16 @@ jobs:
fail-fast
:
true
fail-fast
:
true
steps
:
steps
:
-
name
:
config-net
run
:
|
export http_proxy=http://bigdata_open_proxy:H89k5qwQRDYfz@10.140.90.20:10811
export https_proxy=http://bigdata_open_proxy:H89k5qwQRDYfz@10.140.90.20:10811
-
name
:
PDF benchmark
-
name
:
PDF benchmark
uses
:
actions/checkout@v3
uses
:
actions/checkout@v3
with
:
with
:
fetch-depth
:
2
fetch-depth
:
2
-
name
:
check-requirements
-
name
:
check-requirements
run
:
|
run
:
|
export http_proxy=http://bigdata_open_proxy:H89k5qwQRDYfz@10.140.90.20:10811
export https_proxy=http://bigdata_open_proxy:H89k5qwQRDYfz@10.140.90.20:10811
changed_files=$(git diff --name-only -r HEAD~1 HEAD)
changed_files=$(git diff --name-only -r HEAD~1 HEAD)
echo $changed_files
echo $changed_files
if [[ $changed_files =~ "requirements.txt" ]]; then
if [[ $changed_files =~ "requirements.txt" ]]; then
...
@@ -36,7 +38,7 @@ jobs:
...
@@ -36,7 +38,7 @@ jobs:
-
name
:
benchmark
-
name
:
benchmark
run
:
|
run
:
|
echo "start test"
echo "start test"
cd tools && python ocr_badcase.py pdf_json_label_0306.json ocr_dataset.json json_files.zip
output
.json
cd tools && python ocr_badcase.py pdf_json_label_0306.json ocr_dataset.json json_files.zip
badcase.json overall.json base_data
.json
notify_to_feishu
:
notify_to_feishu
:
if
:
${{ always() && !cancelled() && contains(needs.*.result, 'failure') && (github.ref_name == 'master') }}
if
:
${{ always() && !cancelled() && contains(needs.*.result, 'failure') && (github.ref_name == 'master') }}
needs
:
[
pdf-test
]
needs
:
[
pdf-test
]
...
...
tools/base_data.json
0 → 100644
View file @
b5b58d64
{
"accuracy"
:
1.0
,
"precision"
:
1.0
,
"recall"
:
1.0
,
"f1_score"
:
1.0
,
"pdf间的平均编辑距离"
:
133.10256410256412
,
"pdf间的平均bleu"
:
0.28838311595434046
,
"分段准确率"
:
0.07220216606498195
,
"行内公式准确率"
:
{
"accuracy"
:
0.004835727492533068
,
"precision"
:
0.008790072388831437
,
"recall"
:
0.010634970284641852
,
"f1_score"
:
0.009624911535739562
},
"行内公式编辑距离"
:
1.6176470588235294
,
"行内公式bleu"
:
0.17154724654721457
,
"行间公式准确率"
:
{
"accuracy"
:
0.08490566037735849
,
"precision"
:
0.1836734693877551
,
"recall"
:
0.13636363636363635
,
"f1_score"
:
0.1565217391304348
},
"行间公式编辑距离"
:
113.22222222222223
,
"行间公式bleu"
:
0.2531053359913409
,
"丢弃文本准确率"
:
{
"accuracy"
:
0.00035398230088495576
,
"precision"
:
0.0006389776357827476
,
"recall"
:
0.0007930214115781126
,
"f1_score"
:
0.0007077140835102619
},
"丢弃文本标签准确率"
:
{
"color_background_header_txt_block"
:
{
"precision"
:
0.0
,
"recall"
:
0.0
,
"f1-score"
:
0.0
,
"support"
:
41.0
},
"header"
:
{
"precision"
:
0.0
,
"recall"
:
0.0
,
"f1-score"
:
0.0
,
"support"
:
4.0
},
"footnote"
:
{
"precision"
:
1.0
,
"recall"
:
0.009708737864077669
,
"f1-score"
:
0.019230769230769232
,
"support"
:
103.0
},
"on-table"
:
{
"precision"
:
0.0
,
"recall"
:
0.0
,
"f1-score"
:
0.0
,
"support"
:
665.0
},
"rotate"
:
{
"precision"
:
0.0
,
"recall"
:
0.0
,
"f1-score"
:
0.0
,
"support"
:
63.0
},
"on-image"
:
{
"precision"
:
0.0
,
"recall"
:
0.0
,
"f1-score"
:
0.0
,
"support"
:
380.0
},
"micro avg"
:
{
"precision"
:
1.0
,
"recall"
:
0.0007961783439490446
,
"f1-score"
:
0.0015910898965791568
,
"support"
:
1256.0
}
},
"丢弃图片准确率"
:
{
"accuracy"
:
0.0
,
"precision"
:
0.0
,
"recall"
:
0.0
,
"f1_score"
:
0.0
},
"丢弃表格准确率"
:
{
"accuracy"
:
0.0
,
"precision"
:
0.0
,
"recall"
:
0.0
,
"f1_score"
:
0.0
}
}
\ No newline at end of file
tools/ocr_badcase.py
View file @
b5b58d64
...
@@ -413,7 +413,9 @@ def bbox_match_indicator_dropped_text_block(test_dropped_text_bboxs, standard_dr
...
@@ -413,7 +413,9 @@ def bbox_match_indicator_dropped_text_block(test_dropped_text_bboxs, standard_dr
# 计算和返回标签匹配指标
# 计算和返回标签匹配指标
text_block_tag_report
=
classification_report
(
y_true
=
standard_tag
,
y_pred
=
test_tag
,
labels
=
list
(
set
(
standard_tag
)
-
{
'None'
}),
output_dict
=
True
,
zero_division
=
0
)
text_block_tag_report
=
classification_report
(
y_true
=
standard_tag
,
y_pred
=
test_tag
,
labels
=
list
(
set
(
standard_tag
)
-
{
'None'
}),
output_dict
=
True
,
zero_division
=
0
)
del
text_block_tag_report
[
"macro avg"
]
del
text_block_tag_report
[
"weighted avg"
]
return
text_block_report
,
text_block_tag_report
return
text_block_report
,
text_block_tag_report
def
handle_multi_deletion
(
test_page
,
test_page_tag
,
test_page_bbox
,
standard_page_tag
,
standard_page_bbox
):
def
handle_multi_deletion
(
test_page
,
test_page_tag
,
test_page_bbox
,
standard_page_tag
,
standard_page_bbox
):
...
@@ -500,6 +502,142 @@ def merge_json_data(json_test_df, json_standard_df):
...
@@ -500,6 +502,142 @@ def merge_json_data(json_test_df, json_standard_df):
return
inner_merge
,
standard_exist
,
test_exist
return
inner_merge
,
standard_exist
,
test_exist
def
consolidate_data
(
test_data
,
standard_data
,
key_path
):
"""
Consolidates data from test and standard datasets based on the provided key path.
:param test_data: Dictionary containing the test dataset.
:param standard_data: Dictionary containing the standard dataset.
:param key_path: List of keys leading to the desired data within the dictionaries.
:return: List containing all items from both test and standard data at the specified key path.
"""
# Initialize an empty list to hold the consolidated data
overall_data_standard
=
[]
overall_data_test
=
[]
# Helper function to recursively navigate through the dictionaries based on the key path
def
extract_data
(
source_data
,
keys
):
for
key
in
keys
[:
-
1
]:
source_data
=
source_data
.
get
(
key
,
{})
return
source_data
.
get
(
keys
[
-
1
],
[])
for
data
in
extract_data
(
standard_data
,
key_path
):
# 假设每个 single_table_tags 已经是一个列表,直接将它的元素添加到总列表中
overall_data_standard
.
extend
(
data
)
for
data
in
extract_data
(
test_data
,
key_path
):
overall_data_test
.
extend
(
data
)
# Extract and extend the overall data list with items from both test and standard datasets
return
overall_data_standard
,
overall_data_test
def
overall_calculate_metrics
(
inner_merge
,
json_test
,
json_standard
,
standard_exist
,
test_exist
):
process_data_standard
=
process_equations_and_blocks
(
json_standard
,
is_standard
=
True
)
process_data_test
=
process_equations_and_blocks
(
json_test
,
is_standard
=
False
)
overall_report
=
{}
overall_report
[
'accuracy'
]
=
metrics
.
accuracy_score
(
standard_exist
,
test_exist
)
overall_report
[
'precision'
]
=
metrics
.
precision_score
(
standard_exist
,
test_exist
)
overall_report
[
'recall'
]
=
metrics
.
recall_score
(
standard_exist
,
test_exist
)
overall_report
[
'f1_score'
]
=
metrics
.
f1_score
(
standard_exist
,
test_exist
)
overall_report
test_para_text
=
np
.
asarray
(
process_data_test
[
'para_texts'
],
dtype
=
object
)[
inner_merge
[
'pass_label'
]
==
'yes'
]
standard_para_text
=
np
.
asarray
(
process_data_standard
[
'para_texts'
],
dtype
=
object
)[
inner_merge
[
'pass_label'
]
==
'yes'
]
ids_yes
=
inner_merge
[
'id'
][
inner_merge
[
'pass_label'
]
==
'yes'
].
tolist
()
pdf_dis
=
{}
pdf_bleu
=
{}
# 对pass_label为'yes'的数据计算编辑距离和BLEU得分
for
idx
,(
a
,
b
,
id
)
in
enumerate
(
zip
(
test_para_text
,
standard_para_text
,
ids_yes
)):
a1
=
''
.
join
(
a
)
b1
=
''
.
join
(
b
)
pdf_dis
[
id
]
=
Levenshtein_Distance
(
a
,
b
)
pdf_bleu
[
id
]
=
sentence_bleu
([
a1
],
b1
)
overall_report
[
'pdf间的平均编辑距离'
]
=
np
.
mean
(
list
(
pdf_dis
.
values
()))
overall_report
[
'pdf间的平均bleu'
]
=
np
.
mean
(
list
(
pdf_bleu
.
values
()))
# Consolidate equations bboxs inline
overall_equations_bboxs_inline_standard
,
overall_equations_bboxs_inline_test
=
consolidate_data
(
process_data_test
,
process_data_standard
,
[
"equations_bboxs"
,
"inline"
])
# # Consolidate equations texts inline
overall_equations_texts_inline_standard
,
overall_equations_texts_inline_test
=
consolidate_data
(
process_data_test
,
process_data_standard
,
[
"equations_texts"
,
"inline"
])
# Consolidate equations bboxs interline
overall_equations_bboxs_interline_standard
,
overall_equations_bboxs_interline_test
=
consolidate_data
(
process_data_test
,
process_data_standard
,
[
"equations_bboxs"
,
"interline"
])
# Consolidate equations texts interline
overall_equations_texts_interline_standard
,
overall_equations_texts_interline_test
=
consolidate_data
(
process_data_test
,
process_data_standard
,
[
"equations_texts"
,
"interline"
])
overall_dropped_bboxs_text_standard
,
overall_dropped_bboxs_text_test
=
consolidate_data
(
process_data_test
,
process_data_standard
,
[
"dropped_bboxs"
,
"text"
])
overall_dropped_tags_text_standard
,
overall_dropped_tags_text_test
=
consolidate_data
(
process_data_test
,
process_data_standard
,
[
"dropped_tags"
,
"text"
])
overall_dropped_bboxs_image_standard
,
overall_dropped_bboxs_image_test
=
consolidate_data
(
process_data_test
,
process_data_standard
,
[
"dropped_bboxs"
,
"image"
])
overall_dropped_bboxs_table_standard
,
overall_dropped_bboxs_table_test
=
consolidate_data
(
process_data_test
,
process_data_standard
,[
"dropped_bboxs"
,
"table"
])
para_nums_test
=
process_data_test
[
'para_nums'
]
para_nums_standard
=
process_data_standard
[
'para_nums'
]
overall_para_nums_standard
=
[
item
for
sublist
in
para_nums_standard
for
item
in
(
sublist
if
isinstance
(
sublist
,
list
)
else
[
sublist
])]
overall_para_nums_test
=
[
item
for
sublist
in
para_nums_test
for
item
in
(
sublist
if
isinstance
(
sublist
,
list
)
else
[
sublist
])]
test_para_num
=
np
.
array
(
overall_para_nums_test
)
standard_para_num
=
np
.
array
(
overall_para_nums_standard
)
acc_para
=
np
.
mean
(
test_para_num
==
standard_para_num
)
overall_report
[
'分段准确率'
]
=
acc_para
# 行内公式准确率和编辑距离、bleu
overall_report
[
'行内公式准确率'
]
=
bbox_match_indicator_general
(
overall_equations_bboxs_inline_test
,
overall_equations_bboxs_inline_standard
)
overall_report
[
'行内公式编辑距离'
],
overall_report
[
'行内公式bleu'
]
=
equations_indicator
(
overall_equations_bboxs_inline_test
,
overall_equations_bboxs_inline_standard
,
overall_equations_texts_inline_test
,
overall_equations_texts_inline_standard
)
# 行间公式准确率和编辑距离、bleu
overall_report
[
'行间公式准确率'
]
=
bbox_match_indicator_general
(
overall_equations_bboxs_interline_test
,
overall_equations_bboxs_interline_standard
)
overall_report
[
'行间公式编辑距离'
],
overall_report
[
'行间公式bleu'
]
=
equations_indicator
(
overall_equations_bboxs_interline_test
,
overall_equations_bboxs_interline_standard
,
overall_equations_texts_interline_test
,
overall_equations_texts_interline_standard
)
# 丢弃文本准确率,丢弃文本标签准确率
overall_report
[
'丢弃文本准确率'
],
overall_report
[
'丢弃文本标签准确率'
]
=
bbox_match_indicator_dropped_text_block
(
overall_dropped_bboxs_text_test
,
overall_dropped_bboxs_text_standard
,
overall_dropped_tags_text_standard
,
overall_dropped_tags_text_test
)
# 丢弃图片准确率
overall_report
[
'丢弃图片准确率'
]
=
bbox_match_indicator_general
(
overall_dropped_bboxs_image_test
,
overall_dropped_bboxs_image_standard
)
# 丢弃表格准确率
overall_report
[
'丢弃表格准确率'
]
=
bbox_match_indicator_general
(
overall_dropped_bboxs_table_test
,
overall_dropped_bboxs_table_standard
)
return
overall_report
def
calculate_metrics
(
inner_merge
,
json_test
,
json_standard
,
json_standard_origin
):
def
calculate_metrics
(
inner_merge
,
json_test
,
json_standard
,
json_standard_origin
):
...
@@ -602,21 +740,27 @@ def calculate_metrics(inner_merge, json_test, json_standard, json_standard_origi
...
@@ -602,21 +740,27 @@ def calculate_metrics(inner_merge, json_test, json_standard, json_standard_origi
return
result_dict
return
result_dict
def
save_results
(
result_dict
,
output_path
):
def
save_results
(
result_dict
,
overall_report_dict
,
badcase_path
,
overall_path
,):
"""
"""
将结果字典保存为JSON文件至指定路径。
将结果字典保存为JSON文件至指定路径。
参数:
参数:
- result_dict: 包含计算结果的字典。
- result_dict: 包含计算结果的字典。
- o
utput
_path: 结果文件的保存路径,包括文件名。
- o
verall
_path: 结果文件的保存路径,包括文件名。
"""
"""
# 打开指定的文件以写入
# 打开指定的文件以写入
with
open
(
output
_path
,
'w'
,
encoding
=
'utf-8'
)
as
f
:
with
open
(
badcase
_path
,
'w'
,
encoding
=
'utf-8'
)
as
f
:
# 将结果字典转换为JSON格式并写入文件
# 将结果字典转换为JSON格式并写入文件
json
.
dump
(
result_dict
,
f
,
ensure_ascii
=
False
,
indent
=
4
)
json
.
dump
(
result_dict
,
f
,
ensure_ascii
=
False
,
indent
=
4
)
print
(
f
"计算结果已经保存到文件:
{
output_path
}
"
)
print
(
f
"计算结果已经保存到文件:
{
badcase_path
}
"
)
with
open
(
overall_path
,
'w'
,
encoding
=
'utf-8'
)
as
f
:
# 将结果字典转换为JSON格式并写入文件
json
.
dump
(
overall_report_dict
,
f
,
ensure_ascii
=
False
,
indent
=
4
)
print
(
f
"计算结果已经保存到文件:
{
overall_path
}
"
)
def
upload_to_s3
(
file_path
,
bucket_name
,
s3_file_name
,
AWS_ACCESS_KEY
,
AWS_SECRET_KEY
,
END_POINT_URL
):
def
upload_to_s3
(
file_path
,
bucket_name
,
s3_file_name
,
AWS_ACCESS_KEY
,
AWS_SECRET_KEY
,
END_POINT_URL
):
"""
"""
...
@@ -634,7 +778,7 @@ def upload_to_s3(file_path, bucket_name, s3_file_name,AWS_ACCESS_KEY,AWS_SECRET_
...
@@ -634,7 +778,7 @@ def upload_to_s3(file_path, bucket_name, s3_file_name,AWS_ACCESS_KEY,AWS_SECRET_
except
ClientError
as
e
:
except
ClientError
as
e
:
print
(
f
"上传文件时发生错误:
{
e
}
"
)
print
(
f
"上传文件时发生错误:
{
e
}
"
)
def
generate_
output_
filename
(
ba
se
_path
):
def
generate_filename
(
ba
dcase_path
,
overall
_path
):
"""
"""
生成带有当前时间戳的输出文件名。
生成带有当前时间戳的输出文件名。
...
@@ -647,13 +791,24 @@ def generate_output_filename(base_path):
...
@@ -647,13 +791,24 @@ def generate_output_filename(base_path):
# 获取当前时间并格式化为字符串
# 获取当前时间并格式化为字符串
current_time
=
datetime
.
now
().
strftime
(
'%Y-%m-%d_%H-%M-%S'
)
current_time
=
datetime
.
now
().
strftime
(
'%Y-%m-%d_%H-%M-%S'
)
# 构建并返回完整的输出文件名
# 构建并返回完整的输出文件名
return
f
"
{
base_path
}
_
{
current_time
}
.json"
return
f
"
{
ba
dca
se_path
}
_
{
current_time
}
.json"
,
f
"
{
overall_path
}
_
{
current_time
}
.json"
def
compare_edit_distance
(
json_file
,
overall_report
):
with
open
(
json_file
,
'r'
,
encoding
=
'utf-8'
)
as
f
:
json_data
=
json
.
load
(
f
)
json_edit_distance
=
json_data
[
'pdf间的平均编辑距离'
]
if
overall_report
[
'pdf间的平均编辑距离'
]
>
json_edit_distance
:
return
0
else
:
return
1
def
main
(
standard_file
,
test_file
,
zip_file
,
ba
se_output
_path
,
s3_bucket_name
=
None
,
s3_file_name
=
None
,
AWS_ACCESS_KEY
=
None
,
AWS_SECRET_KEY
=
None
,
END_POINT_URL
=
None
):
def
main
(
standard_file
,
test_file
,
zip_file
,
ba
dcase_path
,
overall_path
,
base_data
_path
,
s3_bucket_name
=
None
,
s3_file_name
=
None
,
AWS_ACCESS_KEY
=
None
,
AWS_SECRET_KEY
=
None
,
END_POINT_URL
=
None
):
"""
"""
主函数,执行整个评估流程。
主函数,执行整个评估流程。
...
@@ -661,7 +816,8 @@ def main(standard_file, test_file, zip_file, base_output_path, s3_bucket_name=No
...
@@ -661,7 +816,8 @@ def main(standard_file, test_file, zip_file, base_output_path, s3_bucket_name=No
- standard_file: 标准文件的路径。
- standard_file: 标准文件的路径。
- test_file: 测试文件的路径。
- test_file: 测试文件的路径。
- zip_file: 压缩包的路径的路径。
- zip_file: 压缩包的路径的路径。
- base_output_path: 结果文件的基础路径和文件名前缀。
- badcase_path: badcase文件的基础路径和文件名前缀。
- overall_path: overall文件的基础路径和文件名前缀。
- s3_bucket_name: S3桶名称(可选)。
- s3_bucket_name: S3桶名称(可选)。
- s3_file_name: S3上的文件名(可选)。
- s3_file_name: S3上的文件名(可选)。
- AWS_ACCESS_KEY, AWS_SECRET_KEY, END_POINT_URL: AWS访问凭证和端点URL(可选)。
- AWS_ACCESS_KEY, AWS_SECRET_KEY, END_POINT_URL: AWS访问凭证和端点URL(可选)。
...
@@ -675,21 +831,29 @@ def main(standard_file, test_file, zip_file, base_output_path, s3_bucket_name=No
...
@@ -675,21 +831,29 @@ def main(standard_file, test_file, zip_file, base_output_path, s3_bucket_name=No
# 合并JSON数据
# 合并JSON数据
inner_merge
,
standard_exist
,
test_exist
=
merge_json_data
(
json_test_origin
,
json_standard_origin
)
inner_merge
,
standard_exist
,
test_exist
=
merge_json_data
(
json_test_origin
,
json_standard_origin
)
#计算总体指标
overall_report_dict
=
overall_calculate_metrics
(
inner_merge
,
inner_merge
[
'test_mid_json'
],
inner_merge
[
'standard_mid_json'
],
standard_exist
,
test_exist
)
# 计算指标
# 计算指标
result_dict
=
calculate_metrics
(
inner_merge
,
inner_merge
[
'test_mid_json'
],
inner_merge
[
'standard_mid_json'
],
json_standard_origin
)
result_dict
=
calculate_metrics
(
inner_merge
,
inner_merge
[
'test_mid_json'
],
inner_merge
[
'standard_mid_json'
],
json_standard_origin
)
# 生成带时间戳的输出文件名
# 生成带时间戳的输出文件名
output
_file
=
generate_
output_
filename
(
ba
se_output
_path
)
badcase_file
,
overall
_file
=
generate_filename
(
ba
dcase_path
,
overall
_path
)
# 保存结果到JSON文件
# 保存结果到JSON文件
save_results
(
result_dict
,
output_file
)
save_results
(
result_dict
,
overall_report_dict
,
badcase_file
,
overall_file
)
result
=
compare_edit_distance
(
base_data_path
,
overall_report_dict
)
print
(
result
)
assert
result
==
1
if
__name__
==
"__main__"
:
if
__name__
==
"__main__"
:
parser
=
argparse
.
ArgumentParser
(
description
=
"主函数,执行整个评估流程。"
)
parser
=
argparse
.
ArgumentParser
(
description
=
"主函数,执行整个评估流程。"
)
parser
.
add_argument
(
'standard_file'
,
type
=
str
,
help
=
'标准文件的路径。'
)
parser
.
add_argument
(
'standard_file'
,
type
=
str
,
help
=
'标准文件的路径。'
)
parser
.
add_argument
(
'test_file'
,
type
=
str
,
help
=
'测试文件的路径。'
)
parser
.
add_argument
(
'test_file'
,
type
=
str
,
help
=
'测试文件的路径。'
)
parser
.
add_argument
(
'zip_file'
,
type
=
str
,
help
=
'压缩包的路径。'
)
parser
.
add_argument
(
'zip_file'
,
type
=
str
,
help
=
'压缩包的路径。'
)
parser
.
add_argument
(
'base_output_path'
,
type
=
str
,
help
=
'结果文件的基础路径和文件名前缀。'
)
parser
.
add_argument
(
'badcase_path'
,
type
=
str
,
help
=
'badcase文件的基础路径和文件名前缀。'
)
parser
.
add_argument
(
'overall_path'
,
type
=
str
,
help
=
'overall文件的基础路径和文件名前缀。'
)
parser
.
add_argument
(
'base_data_path'
,
type
=
str
,
help
=
'基准文件的基础路径和文件名前缀。'
)
parser
.
add_argument
(
'--s3_bucket_name'
,
type
=
str
,
help
=
'S3桶名称。'
,
default
=
None
)
parser
.
add_argument
(
'--s3_bucket_name'
,
type
=
str
,
help
=
'S3桶名称。'
,
default
=
None
)
parser
.
add_argument
(
'--s3_file_name'
,
type
=
str
,
help
=
'S3上的文件名。'
,
default
=
None
)
parser
.
add_argument
(
'--s3_file_name'
,
type
=
str
,
help
=
'S3上的文件名。'
,
default
=
None
)
parser
.
add_argument
(
'--AWS_ACCESS_KEY'
,
type
=
str
,
help
=
'AWS访问密钥。'
,
default
=
None
)
parser
.
add_argument
(
'--AWS_ACCESS_KEY'
,
type
=
str
,
help
=
'AWS访问密钥。'
,
default
=
None
)
...
@@ -698,5 +862,5 @@ if __name__ == "__main__":
...
@@ -698,5 +862,5 @@ if __name__ == "__main__":
args
=
parser
.
parse_args
()
args
=
parser
.
parse_args
()
main
(
args
.
standard_file
,
args
.
test_file
,
args
.
zip_file
,
args
.
ba
se_output
_path
,
args
.
s3_bucket_name
,
args
.
s3_file_name
,
args
.
AWS_ACCESS_KEY
,
args
.
AWS_SECRET_KEY
,
args
.
END_POINT_URL
)
main
(
args
.
standard_file
,
args
.
test_file
,
args
.
zip_file
,
args
.
ba
dcase_path
,
args
.
overall_path
,
args
.
base_data
_path
,
args
.
s3_bucket_name
,
args
.
s3_file_name
,
args
.
AWS_ACCESS_KEY
,
args
.
AWS_SECRET_KEY
,
args
.
END_POINT_URL
)
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment