Commit 5d658171 authored by justheuristic's avatar justheuristic
Browse files

debug

parent 4da2227f
...@@ -370,8 +370,6 @@ class MatMul8bitLt(torch.autograd.Function): ...@@ -370,8 +370,6 @@ class MatMul8bitLt(torch.autograd.Function):
if state.threshold > 0.0 and subA is not None: if state.threshold > 0.0 and subA is not None:
grad_B[:, idx] += torch.matmul(grad_output.t(), subA) grad_B[:, idx] += torch.matmul(grad_output.t(), subA)
raise NotImplementedError("!!")
if req_gradA: if req_gradA:
if state.CBt is not None: if state.CBt is not None:
C32grad, Sgrad = F.transform(Cgrad, "col32") C32grad, Sgrad = F.transform(Cgrad, "col32")
......
...@@ -237,7 +237,9 @@ class Linear8bitLt(nn.Linear): ...@@ -237,7 +237,9 @@ class Linear8bitLt(nn.Linear):
if threshold > 0.0 and not has_fp16_weights: if threshold > 0.0 and not has_fp16_weights:
self.state.use_pool = True self.state.use_pool = True
self.weight = Int8Params(self.weight.data, has_fp16_weights=has_fp16_weights) self.weight = Int8Params(
self.weight.data, has_fp16_weights=has_fp16_weights, requires_grad=has_fp16_weights
)
def init_8bit_state(self): def init_8bit_state(self):
self.state.CB = self.weight.CB self.state.CB = self.weight.CB
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment