Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
8bbb53e2
"...git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "0f968ddaa391ceefb177a633d7122da132b38efe"
Unverified
Commit
8bbb53e2
authored
Oct 01, 2021
by
Suraj Patil
Committed by
GitHub
Sep 30, 2021
Browse files
skip gptj slow generate tests for now (#13809)
parent
41436d3d
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
4 additions
and
2 deletions
+4
-2
tests/test_modeling_gptj.py
tests/test_modeling_gptj.py
+4
-2
No files found.
tests/test_modeling_gptj.py
View file @
8bbb53e2
...
@@ -396,8 +396,9 @@ class GPTJModelTest(unittest.TestCase):
...
@@ -396,8 +396,9 @@ class GPTJModelTest(unittest.TestCase):
config_and_inputs
=
self
.
model_tester
.
prepare_config_and_inputs
()
config_and_inputs
=
self
.
model_tester
.
prepare_config_and_inputs
()
self
.
model_tester
.
create_and_check_forward_and_backwards
(
*
config_and_inputs
,
gradient_checkpointing
=
True
)
self
.
model_tester
.
create_and_check_forward_and_backwards
(
*
config_and_inputs
,
gradient_checkpointing
=
True
)
@
slow
@
too
slow
def
test_batch_generation
(
self
):
def
test_batch_generation
(
self
):
# Marked as @tooslow due to GPU OOM
model
=
GPTJForCausalLM
.
from_pretrained
(
"EleutherAI/gpt-j-6B"
,
revision
=
"float16"
,
torch_dtype
=
torch
.
float16
)
model
=
GPTJForCausalLM
.
from_pretrained
(
"EleutherAI/gpt-j-6B"
,
revision
=
"float16"
,
torch_dtype
=
torch
.
float16
)
model
.
to
(
torch_device
)
model
.
to
(
torch_device
)
tokenizer
=
AutoTokenizer
.
from_pretrained
(
"EleutherAI/gpt-j-6B"
,
revision
=
"float16"
)
tokenizer
=
AutoTokenizer
.
from_pretrained
(
"EleutherAI/gpt-j-6B"
,
revision
=
"float16"
)
...
@@ -464,8 +465,9 @@ class GPTJModelTest(unittest.TestCase):
...
@@ -464,8 +465,9 @@ class GPTJModelTest(unittest.TestCase):
@
require_torch
@
require_torch
class
GPTJModelLanguageGenerationTest
(
unittest
.
TestCase
):
class
GPTJModelLanguageGenerationTest
(
unittest
.
TestCase
):
@
slow
@
too
slow
def
test_lm_generate_gptj
(
self
):
def
test_lm_generate_gptj
(
self
):
# Marked as @tooslow due to GPU OOM
for
checkpointing
in
[
True
,
False
]:
for
checkpointing
in
[
True
,
False
]:
model
=
GPTJForCausalLM
.
from_pretrained
(
model
=
GPTJForCausalLM
.
from_pretrained
(
"EleutherAI/gpt-j-6B"
,
revision
=
"float16"
,
torch_dtype
=
torch
.
float16
"EleutherAI/gpt-j-6B"
,
revision
=
"float16"
,
torch_dtype
=
torch
.
float16
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment