Commit d9b87898 authored by justheuristic's avatar justheuristic
Browse files

debug

parent 5d658171
...@@ -545,6 +545,7 @@ def test_linear8bitlt_no_fp16_weights(threshold, memory_efficient_backward): ...@@ -545,6 +545,7 @@ def test_linear8bitlt_no_fp16_weights(threshold, memory_efficient_backward):
.to(torch.float16) .to(torch.float16)
.to("cuda") .to("cuda")
) )
w1, w2 = mlp.fc1.weight.clone(), mlp.fc2.weight.clone()
for i in range(100): for i in range(100):
b1 = torch.randn(16, 8, 32, device="cuda").half() b1 = torch.randn(16, 8, 32, device="cuda").half()
...@@ -567,8 +568,15 @@ def test_linear8bitlt_no_fp16_weights(threshold, memory_efficient_backward): ...@@ -567,8 +568,15 @@ def test_linear8bitlt_no_fp16_weights(threshold, memory_efficient_backward):
assert o1.requires_grad assert o1.requires_grad
grad_proj = torch.randn_like(o1) grad_proj = torch.randn_like(o1)
mlp.zero_grad()
(o1 * grad_proj).sum().backward() (o1 * grad_proj).sum().backward()
grad_ref = grad_proj.flatten(2) @ w2 @ w1
assert torch.allclose(b1.grad, grad_ref)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment