Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
21decb77
Unverified
Commit
21decb77
authored
Apr 05, 2022
by
Suraj Patil
Committed by
GitHub
Apr 05, 2022
Browse files
handle torch_dtype in low cpu mem usage (#16580)
parent
8bf6d28c
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
2 additions
and
1 deletion
+2
-1
src/transformers/modeling_utils.py
src/transformers/modeling_utils.py
+2
-1
No files found.
src/transformers/modeling_utils.py
View file @
21decb77
...
...
@@ -2165,7 +2165,8 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin, GenerationMixin, PushToHubMix
for
k
in
loaded_state_dict_keys
:
submodule
,
param_name
=
find_submodule_and_param_name
(
model
,
k
)
if
submodule
is
not
None
:
new_val
=
state_dict
[
k
]
param_dtype
=
getattr
(
submodule
,
param_name
).
dtype
new_val
=
state_dict
[
k
].
to
(
param_dtype
)
if
isinstance
(
getattr
(
submodule
,
param_name
),
torch
.
nn
.
Parameter
):
new_val
=
torch
.
nn
.
Parameter
(
new_val
)
setattr
(
submodule
,
param_name
,
new_val
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment