kie_metric.py 2.32 KB
Newer Older
LDOUBLEV's avatar
add kie  
LDOUBLEV committed
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
# Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

from __future__ import absolute_import
from __future__ import division
from __future__ import print_function

import numpy as np
LDOUBLEV's avatar
debug  
LDOUBLEV committed
20
import paddle
LDOUBLEV's avatar
add kie  
LDOUBLEV committed
21
22
23
24
25
26
27
28

__all__ = ['KIEMetric']


class KIEMetric(object):
    def __init__(self, main_indicator='hmean', **kwargs):
        self.main_indicator = main_indicator
        self.reset()
LDOUBLEV's avatar
debug  
LDOUBLEV committed
29
30
        self.node = []
        self.gt = []
LDOUBLEV's avatar
add kie  
LDOUBLEV committed
31
32
33
34
35

    def __call__(self, preds, batch, **kwargs):
        nodes, _ = preds
        gts, tag = batch[4].squeeze(0), batch[5].tolist()[0]
        gts = gts[:tag[0], :1].reshape([-1])
LDOUBLEV's avatar
debug  
LDOUBLEV committed
36
37
38
39
        self.node.append(nodes.numpy())
        self.gt.append(gts)
        # result = self.compute_f1_score(nodes, gts)
        # self.results.append(result)
LDOUBLEV's avatar
add kie  
LDOUBLEV committed
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54

    def compute_f1_score(self, preds, gts):
        ignores = [0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 25]
        C = preds.shape[1]
        classes = np.array(sorted(set(range(C)) - set(ignores)))
        hist = np.bincount(
            (gts * C).astype('int64') + preds.argmax(1), minlength=C
            **2).reshape([C, C]).astype('float32')
        diag = np.diag(hist)
        recalls = diag / hist.sum(1).clip(min=1)
        precisions = diag / hist.sum(0).clip(min=1)
        f1 = 2 * recalls * precisions / (recalls + precisions).clip(min=1e-8)
        return f1[classes]

    def combine_results(self, results):
LDOUBLEV's avatar
debug  
LDOUBLEV committed
55
56
57
58
        node = np.concatenate(self.node, 0)
        gts = np.concatenate(self.gt, 0)
        results = self.compute_f1_score(node, gts)
        data = {'hmean': results.mean()}
LDOUBLEV's avatar
add kie  
LDOUBLEV committed
59
60
61
        return data

    def get_metric(self):
LDOUBLEV's avatar
debug  
LDOUBLEV committed
62

LDOUBLEV's avatar
add kie  
LDOUBLEV committed
63
64
65
66
67
68
        metircs = self.combine_results(self.results)
        self.reset()
        return metircs

    def reset(self):
        self.results = []  # clear results
LDOUBLEV's avatar
debug  
LDOUBLEV committed
69
70
        self.node = []
        self.gt = []