Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
693cb828
"example/Android/LLMAssistant/vscode:/vscode.git/clone" did not exist on "aefd9f11273430b5491a31f3966fd5149ec86ef0"
Unverified
Commit
693cb828
authored
Jul 16, 2024
by
Zach Mueller
Committed by
GitHub
Jul 16, 2024
Browse files
Fix bad test about slower init (#32002)
Bronked main
parent
25e5e3fa
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
3 additions
and
17 deletions
+3
-17
tests/utils/test_modeling_utils.py
tests/utils/test_modeling_utils.py
+3
-17
No files found.
tests/utils/test_modeling_utils.py
View file @
693cb828
...
...
@@ -20,7 +20,6 @@ import os.path
import
sys
import
tempfile
import
threading
import
time
import
unittest
import
unittest.mock
as
mock
import
uuid
...
...
@@ -895,28 +894,22 @@ class ModelUtilsTest(TestCasePlus):
@
require_usr_bin_time
@
require_accelerate
@
mark
.
accelerate_tests
def
test_from_pretrained_low_cpu_mem_usage_
slower
(
self
):
def
test_from_pretrained_low_cpu_mem_usage_
equal
(
self
):
# Before this would test that `from_pretrained(..., low_cpu_mem_usage=True)` uses less cpu memory than default
# Now though the
memory is the same, we simply test that loading with `low_cpu_mem_usage` winds up being *slower*
#
(mostly from extra logic needed)
# Now though the
se should be around the same.
#
TODO: Look for good bounds to check that their timings are near the same
mname
=
"hf-internal-testing/tiny-random-bert"
preamble
=
"from transformers import AutoModel"
one_liner_str
=
f
'
{
preamble
}
; AutoModel.from_pretrained("
{
mname
}
", low_cpu_mem_usage=False)'
start_time
=
time
.
time
()
# Save this output as `max_rss_normal` if testing memory results
max_rss_normal
=
self
.
python_one_liner_max_rss
(
one_liner_str
)
end_time
=
time
.
time
()
elapsed_time_normal
=
end_time
-
start_time
# print(f"{max_rss_normal=}")
one_liner_str
=
f
'
{
preamble
}
; AutoModel.from_pretrained("
{
mname
}
", low_cpu_mem_usage=True)'
start_time
=
time
.
time
()
# Save this output as `max_rss_low_mem` if testing memory results
max_rss_low_mem
=
self
.
python_one_liner_max_rss
(
one_liner_str
)
end_time
=
time
.
time
()
elapsed_time_low_mem
=
end_time
-
start_time
# Should be within 2MBs of each other (overhead)
self
.
assertAlmostEqual
(
...
...
@@ -926,13 +919,6 @@ class ModelUtilsTest(TestCasePlus):
msg
=
"using `low_cpu_mem_usage` should incur the same memory usage in both cases."
,
)
self
.
assertGreater
(
elapsed_time_low_mem
,
elapsed_time_normal
,
"using `low_cpu_mem_usage` should be slower due to extra logic, "
f
"but got elapsed_time_normal=
{
elapsed_time_normal
}
and elapsed_time_low_mem=
{
elapsed_time_low_mem
}
"
,
)
# if you want to compare things manually, let's first look at the size of the model in bytes
# model = BertModel.from_pretrained(mname, low_cpu_mem_usage=False)
# total_numel = sum(dict((p.data_ptr(), p.numel()) for p in model.parameters()).values())
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment