Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
ModelZoo
ResNet50_tensorflow
Commits
088281ed
Commit
088281ed
authored
Sep 23, 2020
by
Hongkun Yu
Committed by
A. Unique TensorFlower
Sep 23, 2020
Browse files
Fix doc string.
PiperOrigin-RevId: 333314981
parent
47c77112
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
6 additions
and
1 deletion
+6
-1
official/nlp/projects/bigbird/attention.py
official/nlp/projects/bigbird/attention.py
+6
-1
No files found.
official/nlp/projects/bigbird/attention.py
View file @
088281ed
...
...
@@ -390,7 +390,12 @@ class BigBirdMasks(tf.keras.layers.Layer):
@
tf
.
keras
.
utils
.
register_keras_serializable
(
package
=
"Text"
)
class
BigBirdAttention
(
tf
.
keras
.
layers
.
MultiHeadAttention
):
"""Attention layer with cache used for auto-agressive decoding.
"""BigBird, a sparse attention mechanism.
This layer follows the paper "Big Bird: Transformers for Longer Sequences"
(https://arxiv.org/abs/2007.14062).
It reduces this quadratic dependency of attention
computation to linear.
Arguments are the same as `MultiHeadAttention` layer.
"""
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment