rocm_onnxruntime_performance.py 3.18 KB
Newer Older
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
# Copyright (c) Microsoft Corporation.
# Licensed under the MIT license.

"""Module of the onnxruntime E2E model benchmarks.

Including:
  bert-large-uncased ngpu=1
  bert-large-uncased ngpu=8
  distilbert-base-uncased ngpu=1
  distilbert-base-uncased ngpu=8
  gpt2 ngpu=1
  gpt2 ngpu=8
  facebook/bart-large ngpu=1
  facebook/bart-large ngpu=8
  roberta-large ngpu=1
  roberta-large ngpu=8
"""

from superbench.common.utils import logger
from superbench.benchmarks import BenchmarkRegistry, Platform
from superbench.benchmarks.docker_benchmarks.docker_base import RocmDockerBenchmark


class RocmOnnxRuntimeModelBenchmark(RocmDockerBenchmark):
    """The onnxruntime E2E model benchmark class."""
    def __init__(self, name, parameters=''):
        """Constructor.

        Args:
            name (str): benchmark name.
            parameters (str): benchmark parameters.
        """
        super().__init__(name, parameters)

        # Image uri of the current docker-benchmark.
        self._image_uri = 'superbench/benchmark:rocm4.3.1-onnxruntime1.9.0'

        # Image digest of the current docker-benchmark.
        self._digest = 'f5e6c832e3cdcbba9820c619bb30fa47ca7117aa7f2c15944d17e6983d37ab9a'

        # Container name of the current docker-benchmark.
        self._container_name = 'rocm-onnxruntime-model-benchmarks'

        # Entrypoint option of the current docker-benchmark.
        self._entrypoint = '/stage/onnxruntime-training-examples/huggingface/azureml/run_benchmark.sh'

        # CMD option of the current docker-benchmark.
        self._cmd = None

    def _process_raw_result(self, cmd_idx, raw_output):
        """Function to parse raw results and save the summarized results.

          self._result.add_raw_data() and self._result.add_result() need to be called to save the results.

        Args:
            cmd_idx (int): the index of command corresponding with the raw_output.
            raw_output (str): raw output string of the micro-benchmark.

        Return:
            True if the raw output string is valid and result can be extracted.
        """
        self._result.add_raw_data('raw_output', raw_output)

        content = raw_output.splitlines(False)
        try:
            name_prefix = '__superbench__ begin '
            value_prefix = '    "samples_per_second": '
            model_name = None
            for line in content:
                if name_prefix in line:
                    model_name = line[len(name_prefix):]
                    for char in ['-', ' ', '=', '/']:
                        model_name = model_name.replace(char, '_')
                elif value_prefix in line and model_name is not None:
                    throughput = float(line[len(value_prefix):])
                    self._result.add_result(model_name, throughput)
                    model_name = None
        except BaseException as e:
            logger.error(
                'The result format is invalid - round: {}, benchmark: {}, message: {}.'.format(
                    self._curr_run_index, self._name, str(e)
                )
            )
            return False

        return True


BenchmarkRegistry.register_benchmark('onnxruntime-ort-models', RocmOnnxRuntimeModelBenchmark, platform=Platform.ROCM)