Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
f2f32940
"git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "1e616c0af33f128569c7b81a7759acb095b29675"
Commit
f2f32940
authored
Nov 26, 2019
by
Lysandre
Committed by
Lysandre Debut
Nov 26, 2019
Browse files
Fix input embeddings
parent
bdfe21ab
Changes
3
Show whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
8 additions
and
3 deletions
+8
-3
transformers/tests/modeling_albert_test.py
transformers/tests/modeling_albert_test.py
+2
-0
transformers/tests/modeling_tf_albert_test.py
transformers/tests/modeling_tf_albert_test.py
+2
-0
transformers/tests/modeling_tf_common_test.py
transformers/tests/modeling_tf_common_test.py
+4
-3
No files found.
transformers/tests/modeling_albert_test.py
View file @
f2f32940
...
@@ -49,6 +49,7 @@ class AlbertModelTest(CommonTestCases.CommonModelTester):
...
@@ -49,6 +49,7 @@ class AlbertModelTest(CommonTestCases.CommonModelTester):
use_token_type_ids
=
True
,
use_token_type_ids
=
True
,
use_labels
=
True
,
use_labels
=
True
,
vocab_size
=
99
,
vocab_size
=
99
,
embedding_size
=
16
,
hidden_size
=
36
,
hidden_size
=
36
,
num_hidden_layers
=
6
,
num_hidden_layers
=
6
,
num_hidden_groups
=
6
,
num_hidden_groups
=
6
,
...
@@ -73,6 +74,7 @@ class AlbertModelTest(CommonTestCases.CommonModelTester):
...
@@ -73,6 +74,7 @@ class AlbertModelTest(CommonTestCases.CommonModelTester):
self
.
use_token_type_ids
=
use_token_type_ids
self
.
use_token_type_ids
=
use_token_type_ids
self
.
use_labels
=
use_labels
self
.
use_labels
=
use_labels
self
.
vocab_size
=
vocab_size
self
.
vocab_size
=
vocab_size
self
.
embedding_size
=
embedding_size
self
.
hidden_size
=
hidden_size
self
.
hidden_size
=
hidden_size
self
.
num_hidden_layers
=
num_hidden_layers
self
.
num_hidden_layers
=
num_hidden_layers
self
.
num_attention_heads
=
num_attention_heads
self
.
num_attention_heads
=
num_attention_heads
...
...
transformers/tests/modeling_tf_albert_test.py
View file @
f2f32940
...
@@ -54,6 +54,7 @@ class TFAlbertModelTest(TFCommonTestCases.TFCommonModelTester):
...
@@ -54,6 +54,7 @@ class TFAlbertModelTest(TFCommonTestCases.TFCommonModelTester):
use_token_type_ids
=
True
,
use_token_type_ids
=
True
,
use_labels
=
True
,
use_labels
=
True
,
vocab_size
=
99
,
vocab_size
=
99
,
embedding_size
=
16
,
hidden_size
=
32
,
hidden_size
=
32
,
num_hidden_layers
=
5
,
num_hidden_layers
=
5
,
num_attention_heads
=
4
,
num_attention_heads
=
4
,
...
@@ -77,6 +78,7 @@ class TFAlbertModelTest(TFCommonTestCases.TFCommonModelTester):
...
@@ -77,6 +78,7 @@ class TFAlbertModelTest(TFCommonTestCases.TFCommonModelTester):
self
.
use_token_type_ids
=
use_token_type_ids
self
.
use_token_type_ids
=
use_token_type_ids
self
.
use_labels
=
use_labels
self
.
use_labels
=
use_labels
self
.
vocab_size
=
vocab_size
self
.
vocab_size
=
vocab_size
self
.
embedding_size
=
embedding_size
self
.
hidden_size
=
hidden_size
self
.
hidden_size
=
hidden_size
self
.
num_hidden_layers
=
num_hidden_layers
self
.
num_hidden_layers
=
num_hidden_layers
self
.
num_attention_heads
=
num_attention_heads
self
.
num_attention_heads
=
num_attention_heads
...
...
transformers/tests/modeling_tf_common_test.py
View file @
f2f32940
...
@@ -426,9 +426,10 @@ class TFCommonTestCases:
...
@@ -426,9 +426,10 @@ class TFCommonTestCases:
try
:
try
:
x
=
wte
([
input_ids
],
mode
=
"embedding"
)
x
=
wte
([
input_ids
],
mode
=
"embedding"
)
except
:
except
:
if
hasattr
(
self
.
model_tester
,
"embedding_size"
):
x
=
tf
.
ones
(
input_ids
.
shape
+
[
model
.
config
.
embedding_size
],
dtype
=
tf
.
dtypes
.
float32
)
else
:
x
=
tf
.
ones
(
input_ids
.
shape
+
[
self
.
model_tester
.
hidden_size
],
dtype
=
tf
.
dtypes
.
float32
)
x
=
tf
.
ones
(
input_ids
.
shape
+
[
self
.
model_tester
.
hidden_size
],
dtype
=
tf
.
dtypes
.
float32
)
# ^^ In our TF models, the input_embeddings can take slightly different forms,
# so we try two of them and fall back to just synthetically creating a dummy tensor of ones.
inputs_dict
[
"inputs_embeds"
]
=
x
inputs_dict
[
"inputs_embeds"
]
=
x
outputs
=
model
(
inputs_dict
)
outputs
=
model
(
inputs_dict
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment