Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
bitsandbytes
Commits
ceae1504
Unverified
Commit
ceae1504
authored
Feb 14, 2024
by
Brian Vaughan
Committed by
GitHub
Feb 14, 2024
Browse files
fix a type in code comment (#1063)
was pointing to wrong class
parent
5b28fd3f
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
1 addition
and
1 deletion
+1
-1
bitsandbytes/nn/modules.py
bitsandbytes/nn/modules.py
+1
-1
No files found.
bitsandbytes/nn/modules.py
View file @
ceae1504
...
@@ -275,7 +275,7 @@ class Linear4bit(nn.Linear):
...
@@ -275,7 +275,7 @@ class Linear4bit(nn.Linear):
compute datatypes such as FP4 and NF4.
compute datatypes such as FP4 and NF4.
In order to quantize a linear layer one should first load the original fp16 / bf16 weights into
In order to quantize a linear layer one should first load the original fp16 / bf16 weights into
the Linear
8
bit
Lt
module, then call `quantized_module.to("cuda")` to quantize the fp16 / bf16 weights.
the Linear
4
bit module, then call `quantized_module.to("cuda")` to quantize the fp16 / bf16 weights.
Example:
Example:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment