gemm.py 6.42 KB
Newer Older
PanZezhong's avatar
PanZezhong committed
1
import torch
PanZezhongQY's avatar
PanZezhongQY committed
2
import ctypes
PanZezhong's avatar
PanZezhong committed
3
4
from ctypes import POINTER, Structure, c_int32, c_size_t, c_uint64, c_void_p, c_float
from libinfiniop import (
PanZezhongQY's avatar
PanZezhongQY committed
5
6
    infiniopHandle_t,
    infiniopTensorDescriptor_t,
PanZezhong's avatar
PanZezhong committed
7
8
9
    open_lib,
    to_tensor,
    get_test_devices,
PanZezhongQY's avatar
PanZezhongQY committed
10
    check_error,
PanZezhong's avatar
PanZezhong committed
11
12
13
14
15
16
17
    rearrange_if_needed,
    create_workspace,
    test_operator,
    get_args,
    debug,
    get_tolerance,
    profile_operation,
PanZezhongQY's avatar
PanZezhongQY committed
18
19
)

PanZezhong's avatar
PanZezhong committed
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
# ==============================================================================
#  Configuration (Internal Use Only)
# ==============================================================================
# These are not meant to be imported from other modules
_TEST_CASES = [
    # alpha, beta, a_shape, b_shape, c_shape, a_stride, b_stride, c_stride
    (1.0, 0.0, (1, 2048), (2048, 2048), (1, 2048), None, None, None),
    (1.0, 0.0, (2, 4, 2048), (2, 2048, 2048), (2, 4, 2048), None, None, None),
    (1.0, 0.0, (1, 2048), (2048, 2048), (1, 2048), (4096, 1), (4096, 1), (4096, 1)),
    (1.0, 1.0, (6, 2048), (2048, 2560), (6, 2560), (2048, 1), (1, 2048), (2560, 1)),
    (1.0 / 8.0, 0.0, (4, 8 * 6, 64), (4, 64, 6), (4, 8 * 6, 6), None, None, None),
]

# Data types used for testing
_TENSOR_DTYPES = [torch.float16, torch.float32]

# Tolerance map for different data types
_TOLERANCE_MAP = {
    torch.float16: {"atol": 0, "rtol": 1e-2},
    torch.float32: {"atol": 0, "rtol": 1e-3},
}

DEBUG = False
PanZezhongQY's avatar
PanZezhongQY committed
43
44
45
46
PROFILE = False
NUM_PRERUN = 10
NUM_ITERATIONS = 1000

47

PanZezhong's avatar
PanZezhong committed
48
49
50
51
# ==============================================================================
#  Definitions
# ==============================================================================
class GemmDescriptor(Structure):
PanZezhongQY's avatar
PanZezhongQY committed
52
53
54
    _fields_ = [("device", c_int32)]


PanZezhong's avatar
PanZezhong committed
55
infiniopGemmDescriptor_t = POINTER(GemmDescriptor)
PanZezhongQY's avatar
PanZezhongQY committed
56
57


PanZezhong's avatar
PanZezhong committed
58
# PyTorch implementation for matrix multiplication
59
def gemm(d, _c, beta, _a, _b, alpha):
60
61
62
63
64
65
66
67
    try:
        if _c.ndim == 2:
            torch.addmm(_c, _a, _b, beta=beta, alpha=alpha, out=d)
        elif _c.ndim == 3:
            torch.baddbmm(_c, _a, _b, beta=beta, alpha=alpha, out=d)
        else:
            raise
    except Exception:
68
69
        torch.matmul(_a, _b, out=d)
        d.mul_(alpha).add_(_c, alpha=beta)
PanZezhongQY's avatar
PanZezhongQY committed
70
71


PanZezhong's avatar
PanZezhong committed
72
73
# The argument list should be (lib, handle, torch_device, <param list>, dtype)
# The <param list> should keep the same order as the one specified in _TEST_CASES
PanZezhongQY's avatar
PanZezhongQY committed
74
75
76
77
78
79
80
81
82
83
84
85
86
def test(
    lib,
    handle,
    torch_device,
    alpha,
    beta,
    a_shape,
    b_shape,
    c_shape,
    a_stride=None,
    b_stride=None,
    c_stride=None,
    dtype=torch.float16,
87
    sync=None
PanZezhongQY's avatar
PanZezhongQY committed
88
89
):
    print(
PanZezhong's avatar
PanZezhong committed
90
91
92
        f"Testing Gemm on {torch_device} with alpha:{alpha}, beta:{beta},"
        f" a_shape:{a_shape}, b_shape:{b_shape}, c_shape:{c_shape},"
        f" a_stride:{a_stride}, b_stride:{b_stride}, c_stride:{c_stride}, dtype:{dtype}"
PanZezhongQY's avatar
PanZezhongQY committed
93
94
    )

PanZezhong's avatar
PanZezhong committed
95
    # Initialize tensors
PanZezhongQY's avatar
PanZezhongQY committed
96
97
    a = torch.rand(a_shape, dtype=dtype).to(torch_device)
    b = torch.rand(b_shape, dtype=dtype).to(torch_device)
PanZezhong's avatar
PanZezhong committed
98
    c = torch.ones(c_shape, dtype=dtype).to(torch_device)
99
    ans = torch.zeros(c_shape, dtype=dtype).to(torch_device)
PanZezhongQY's avatar
PanZezhongQY committed
100

PanZezhong's avatar
PanZezhong committed
101
    # Compute the PyTorch reference result
102
    gemm(ans, c, beta, a, b, alpha)
PanZezhongQY's avatar
PanZezhongQY committed
103

PanZezhong's avatar
PanZezhong committed
104
105
106
107
108
    a, b, c = [
        rearrange_if_needed(tensor, stride)
        for tensor, stride in zip([a, b, c], [a_stride, b_stride, c_stride])
    ]
    a_tensor, b_tensor, c_tensor = [to_tensor(tensor, lib) for tensor in [a, b, c]]
PanZezhongQY's avatar
PanZezhongQY committed
109

110
111
112
    if sync is not None:
        sync()

PanZezhong's avatar
PanZezhong committed
113
    descriptor = infiniopGemmDescriptor_t()
PanZezhongQY's avatar
PanZezhongQY committed
114
    check_error(
PanZezhong's avatar
PanZezhong committed
115
        lib.infiniopCreateGemmDescriptor(
PanZezhongQY's avatar
PanZezhongQY committed
116
117
            handle,
            ctypes.byref(descriptor),
PanZezhong's avatar
PanZezhong committed
118
            c_tensor.descriptor,
PanZezhongQY's avatar
PanZezhongQY committed
119
120
121
122
123
124
            a_tensor.descriptor,
            b_tensor.descriptor,
        )
    )

    # Invalidate the shape and strides in the descriptor to prevent them from being directly used by the kernel
PanZezhong's avatar
PanZezhong committed
125
126
    for tensor in [a_tensor, b_tensor, c_tensor]:
        tensor.destroyDesc(lib)
PanZezhongQY's avatar
PanZezhongQY committed
127

PanZezhong's avatar
PanZezhong committed
128
129
    # Get workspace size and create workspace
    workspace_size = c_uint64(0)
PanZezhongQY's avatar
PanZezhongQY committed
130
    check_error(
PanZezhong's avatar
PanZezhong committed
131
        lib.infiniopGetGemmWorkspaceSize(descriptor, ctypes.byref(workspace_size))
PanZezhongQY's avatar
PanZezhongQY committed
132
    )
PanZezhong's avatar
PanZezhong committed
133
    workspace = create_workspace(workspace_size.value, a.device)
PanZezhongQY's avatar
PanZezhongQY committed
134

PanZezhong's avatar
PanZezhong committed
135
136
    # Execute infiniop gemm operator
    def lib_gemm():
PanZezhongQY's avatar
PanZezhongQY committed
137
        check_error(
PanZezhong's avatar
PanZezhong committed
138
            lib.infiniopGemm(
PanZezhongQY's avatar
PanZezhongQY committed
139
                descriptor,
PanZezhong's avatar
PanZezhong committed
140
141
142
                workspace.data_ptr() if workspace is not None else None,
                workspace_size.value,
                c_tensor.data,
PanZezhongQY's avatar
PanZezhongQY committed
143
144
                a_tensor.data,
                b_tensor.data,
PanZezhong's avatar
PanZezhong committed
145
146
                alpha,
                beta,
PanZezhongQY's avatar
PanZezhongQY committed
147
148
149
150
                None,
            )
        )

PanZezhong's avatar
PanZezhong committed
151
    lib_gemm()
PanZezhongQY's avatar
PanZezhongQY committed
152

PanZezhong's avatar
PanZezhong committed
153
154
155
156
157
    # Validate results
    atol, rtol = get_tolerance(_TOLERANCE_MAP, dtype)
    if DEBUG:
        debug(c, ans, atol=atol, rtol=rtol)
    assert torch.allclose(c, ans, atol=atol, rtol=rtol)
PanZezhongQY's avatar
PanZezhongQY committed
158

PanZezhong's avatar
PanZezhong committed
159
160
    # Profiling workflow
    if PROFILE:
161
        # fmt: off
162
        profile_operation("PyTorch", lambda: gemm(ans, c, beta, a, b, alpha), torch_device, NUM_PRERUN, NUM_ITERATIONS)
PanZezhong's avatar
PanZezhong committed
163
        profile_operation("    lib", lambda: lib_gemm(), torch_device, NUM_PRERUN, NUM_ITERATIONS)
164
        # fmt: on
PanZezhong's avatar
PanZezhong committed
165
    check_error(lib.infiniopDestroyGemmDescriptor(descriptor))
PanZezhongQY's avatar
PanZezhongQY committed
166
167


PanZezhong's avatar
PanZezhong committed
168
169
170
# ==============================================================================
#  Main Execution
# ==============================================================================
PanZezhongQY's avatar
PanZezhongQY committed
171
172
173
174
if __name__ == "__main__":
    args = get_args()
    lib = open_lib()

PanZezhong's avatar
PanZezhong committed
175
176
    lib.infiniopCreateGemmDescriptor.restype = c_int32
    lib.infiniopCreateGemmDescriptor.argtypes = [
PanZezhongQY's avatar
PanZezhongQY committed
177
        infiniopHandle_t,
PanZezhong's avatar
PanZezhong committed
178
        POINTER(infiniopGemmDescriptor_t),
PanZezhongQY's avatar
PanZezhongQY committed
179
180
181
182
183
        infiniopTensorDescriptor_t,
        infiniopTensorDescriptor_t,
        infiniopTensorDescriptor_t,
    ]

PanZezhong's avatar
PanZezhong committed
184
185
186
187
    lib.infiniopGetGemmWorkspaceSize.restype = c_int32
    lib.infiniopGetGemmWorkspaceSize.argtypes = [
        infiniopGemmDescriptor_t,
        POINTER(c_size_t),
PanZezhongQY's avatar
PanZezhongQY committed
188
189
    ]

PanZezhong's avatar
PanZezhong committed
190
191
192
    lib.infiniopGemm.restype = c_int32
    lib.infiniopGemm.argtypes = [
        infiniopGemmDescriptor_t,
PanZezhongQY's avatar
PanZezhongQY committed
193
194
195
196
197
        c_void_p,
        c_uint64,
        c_void_p,
        c_void_p,
        c_void_p,
PanZezhong's avatar
PanZezhong committed
198
199
        c_float,
        c_float,
PanZezhongQY's avatar
PanZezhongQY committed
200
201
202
        c_void_p,
    ]

PanZezhong's avatar
PanZezhong committed
203
204
205
    lib.infiniopDestroyGemmDescriptor.restype = c_int32
    lib.infiniopDestroyGemmDescriptor.argtypes = [
        infiniopGemmDescriptor_t,
PanZezhongQY's avatar
PanZezhongQY committed
206
207
    ]

PanZezhong's avatar
PanZezhong committed
208
209
210
211
212
213
214
215
216
217
    # Configure testing options
    DEBUG = args.debug
    PROFILE = args.profile
    NUM_PRERUN = args.num_prerun
    NUM_ITERATIONS = args.num_iterations

    # Execute tests
    for device in get_test_devices(args):
        test_operator(lib, device, test, _TEST_CASES, _TENSOR_DTYPES)

PanZezhongQY's avatar
PanZezhongQY committed
218
    print("\033[92mTest passed!\033[0m")