"...text-generation-inference.git" did not exist on "9d8f21cace66e8593bc559174de0eed3ecdab6a2"
Unverified Commit ceae1504 authored by Brian Vaughan's avatar Brian Vaughan Committed by GitHub
Browse files

fix a type in code comment (#1063)

was pointing to wrong class
parent 5b28fd3f
......@@ -275,7 +275,7 @@ class Linear4bit(nn.Linear):
compute datatypes such as FP4 and NF4.
In order to quantize a linear layer one should first load the original fp16 / bf16 weights into
the Linear8bitLt module, then call `quantized_module.to("cuda")` to quantize the fp16 / bf16 weights.
the Linear4bit module, then call `quantized_module.to("cuda")` to quantize the fp16 / bf16 weights.
Example:
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment