Unverified Commit ceae1504 authored by Brian Vaughan's avatar Brian Vaughan Committed by GitHub
Browse files

fix a type in code comment (#1063)

was pointing to wrong class
parent 5b28fd3f
...@@ -275,7 +275,7 @@ class Linear4bit(nn.Linear): ...@@ -275,7 +275,7 @@ class Linear4bit(nn.Linear):
compute datatypes such as FP4 and NF4. compute datatypes such as FP4 and NF4.
In order to quantize a linear layer one should first load the original fp16 / bf16 weights into In order to quantize a linear layer one should first load the original fp16 / bf16 weights into
the Linear8bitLt module, then call `quantized_module.to("cuda")` to quantize the fp16 / bf16 weights. the Linear4bit module, then call `quantized_module.to("cuda")` to quantize the fp16 / bf16 weights.
Example: Example:
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment