"vscode:/vscode.git/clone" did not exist on "df56f1ee5eb5f464b581402cb00c4fe7dd5de774"
adagrad_test.py 893 Bytes
Newer Older
aiss's avatar
aiss committed
1
2
3
4
# Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0

# DeepSpeed Team
aiss's avatar
aiss committed
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33

import torch
from deepspeed.ops.adagrad import DeepSpeedCPUAdagrad
import time

NUM_ITERS = 100


def _test_perf(param, optimizer_func):
    optimizer = optimizer_func(param)
    avg = 0
    for i in range(NUM_ITERS):
        for i, p in enumerate(param):
            p.grad = torch.ones_like(p) * 2
        start = time.time()
        optimizer.step()
        stop = time.time()
        avg += (stop - start)

    return avg / NUM_ITERS


def _main():
    device = 'cpu'
    model_size = 1 * 1024**3
    group_size = [model_size, 274432]
    param = [torch.nn.Parameter(torch.ones(size, device=device)) for size in group_size]
    torch_time = _test_perf(param, torch.optim.Adagrad)
    ds_time = _test_perf(param, DeepSpeedCPUAdagrad)
aiss's avatar
aiss committed
34
    print(f"Step time: {torch_time=} {ds_time=}")
aiss's avatar
aiss committed
35
36
37


_main()