Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
ModelZoo
ResNet50_tensorflow
Commits
b1eddf4f
Commit
b1eddf4f
authored
May 21, 2020
by
A. Unique TensorFlower
Browse files
Makes token embedding projection consistent between Albert and BERT
PiperOrigin-RevId: 312751112
parent
99b5438a
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
11 additions
and
9 deletions
+11
-9
official/nlp/bert/tf2_encoder_checkpoint_converter.py
official/nlp/bert/tf2_encoder_checkpoint_converter.py
+2
-1
official/nlp/modeling/networks/transformer_encoder.py
official/nlp/modeling/networks/transformer_encoder.py
+9
-8
No files found.
official/nlp/bert/tf2_encoder_checkpoint_converter.py
View file @
b1eddf4f
...
@@ -64,7 +64,8 @@ def _create_bert_model(cfg):
...
@@ -64,7 +64,8 @@ def _create_bert_model(cfg):
sequence_length
=
cfg
.
max_position_embeddings
,
sequence_length
=
cfg
.
max_position_embeddings
,
type_vocab_size
=
cfg
.
type_vocab_size
,
type_vocab_size
=
cfg
.
type_vocab_size
,
initializer
=
tf
.
keras
.
initializers
.
TruncatedNormal
(
initializer
=
tf
.
keras
.
initializers
.
TruncatedNormal
(
stddev
=
cfg
.
initializer_range
))
stddev
=
cfg
.
initializer_range
),
embedding_width
=
cfg
.
embedding_size
)
return
bert_encoder
return
bert_encoder
...
...
official/nlp/modeling/networks/transformer_encoder.py
View file @
b1eddf4f
...
@@ -146,6 +146,15 @@ class TransformerEncoder(tf.keras.Model):
...
@@ -146,6 +146,15 @@ class TransformerEncoder(tf.keras.Model):
embeddings
=
tf
.
keras
.
layers
.
Add
()(
embeddings
=
tf
.
keras
.
layers
.
Add
()(
[
word_embeddings
,
position_embeddings
,
type_embeddings
])
[
word_embeddings
,
position_embeddings
,
type_embeddings
])
embeddings
=
(
tf
.
keras
.
layers
.
LayerNormalization
(
name
=
'embeddings/layer_norm'
,
axis
=-
1
,
epsilon
=
1e-12
,
dtype
=
tf
.
float32
)(
embeddings
))
embeddings
=
(
tf
.
keras
.
layers
.
Dropout
(
rate
=
dropout_rate
)(
embeddings
))
# We project the 'embedding' output to 'hidden_size' if it is not already
# We project the 'embedding' output to 'hidden_size' if it is not already
# 'hidden_size'.
# 'hidden_size'.
if
embedding_width
!=
hidden_size
:
if
embedding_width
!=
hidden_size
:
...
@@ -156,14 +165,6 @@ class TransformerEncoder(tf.keras.Model):
...
@@ -156,14 +165,6 @@ class TransformerEncoder(tf.keras.Model):
kernel_initializer
=
initializer
,
kernel_initializer
=
initializer
,
name
=
'embedding_projection'
)
name
=
'embedding_projection'
)
embeddings
=
self
.
_embedding_projection
(
embeddings
)
embeddings
=
self
.
_embedding_projection
(
embeddings
)
embeddings
=
(
tf
.
keras
.
layers
.
LayerNormalization
(
name
=
'embeddings/layer_norm'
,
axis
=-
1
,
epsilon
=
1e-12
,
dtype
=
tf
.
float32
)(
embeddings
))
embeddings
=
(
tf
.
keras
.
layers
.
Dropout
(
rate
=
dropout_rate
)(
embeddings
))
self
.
_transformer_layers
=
[]
self
.
_transformer_layers
=
[]
data
=
embeddings
data
=
embeddings
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment