"llama/llama.cpp/include/llama.h" did not exist on "c5bcf32823ec9870ab4f65943cf611c42a88f1e9"
adam_test1.py 793 Bytes
Newer Older
aiss's avatar
aiss committed
1
2
'''Copyright The Microsoft DeepSpeed Team'''

aiss's avatar
aiss committed
3
4
5
import torch
from deepspeed.ops.adam import DeepSpeedCPUAdam
import time
aiss's avatar
aiss committed
6
from deepspeed.accelerator import get_accelerator
aiss's avatar
aiss committed
7
8
9
10

device = 'cpu'
model_size = 1 * 1024**3
param = torch.nn.Parameter(torch.ones(model_size, device=device))
aiss's avatar
aiss committed
11
12
13
14
param_fp16 = torch.nn.Parameter(
    torch.ones(model_size,
               dtype=torch.half,
               device=get_accelerator().device_name(0)))
aiss's avatar
aiss committed
15
16
17
18
19
20
21
22
23
24
25
26

optimizer = DeepSpeedCPUAdam([param])
#torch.set_num_threads(128)
param.grad = torch.ones(model_size, device=device)
avg = 0
for i in range(100):
    start = time.time()
    optimizer.step(fp16_param_groups=[param_fp16])
    stop = time.time()
    avg += (stop - start)
    param.grad = torch.ones(model_size, device=device) * 2
print("Elapsed Time is ", avg / 100)