Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
c037b2e3
"docs/source/en/model_doc/prophetnet.md" did not exist on "b9a768b3ffa80c4c19d024f9f42d5917e7d8109e"
Unverified
Commit
c037b2e3
authored
Oct 04, 2023
by
Arthur
Committed by
GitHub
Oct 04, 2023
Browse files
skip flaky hub tests (#26594)
skip flaky
parent
ca7912d1
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
2 additions
and
0 deletions
+2
-0
tests/test_modeling_utils.py
tests/test_modeling_utils.py
+2
-0
No files found.
tests/test_modeling_utils.py
View file @
c037b2e3
...
@@ -1096,6 +1096,7 @@ class ModelPushToHubTester(unittest.TestCase):
...
@@ -1096,6 +1096,7 @@ class ModelPushToHubTester(unittest.TestCase):
except
HTTPError
:
except
HTTPError
:
pass
pass
@
unittest
.
skip
(
"This test is flaky"
)
def
test_push_to_hub
(
self
):
def
test_push_to_hub
(
self
):
config
=
BertConfig
(
config
=
BertConfig
(
vocab_size
=
99
,
hidden_size
=
32
,
num_hidden_layers
=
5
,
num_attention_heads
=
4
,
intermediate_size
=
37
vocab_size
=
99
,
hidden_size
=
32
,
num_hidden_layers
=
5
,
num_attention_heads
=
4
,
intermediate_size
=
37
...
@@ -1118,6 +1119,7 @@ class ModelPushToHubTester(unittest.TestCase):
...
@@ -1118,6 +1119,7 @@ class ModelPushToHubTester(unittest.TestCase):
for
p1
,
p2
in
zip
(
model
.
parameters
(),
new_model
.
parameters
()):
for
p1
,
p2
in
zip
(
model
.
parameters
(),
new_model
.
parameters
()):
self
.
assertTrue
(
torch
.
equal
(
p1
,
p2
))
self
.
assertTrue
(
torch
.
equal
(
p1
,
p2
))
@
unittest
.
skip
(
"This test is flaky"
)
def
test_push_to_hub_in_organization
(
self
):
def
test_push_to_hub_in_organization
(
self
):
config
=
BertConfig
(
config
=
BertConfig
(
vocab_size
=
99
,
hidden_size
=
32
,
num_hidden_layers
=
5
,
num_attention_heads
=
4
,
intermediate_size
=
37
vocab_size
=
99
,
hidden_size
=
32
,
num_hidden_layers
=
5
,
num_attention_heads
=
4
,
intermediate_size
=
37
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment