Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
01068abd
"git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "4b1970bb4cf4d7657ac1fc4e584600121c22e039"
Unverified
Commit
01068abd
authored
Mar 31, 2021
by
Patrick von Platen
Committed by
GitHub
Mar 31, 2021
Browse files
add blog to docs (#10997)
parent
cd56f3fe
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
2 additions
and
0 deletions
+2
-0
docs/source/model_doc/bigbird.rst
docs/source/model_doc/bigbird.rst
+2
-0
No files found.
docs/source/model_doc/bigbird.rst
View file @
01068abd
...
@@ -41,6 +41,8 @@ propose novel applications to genomics data.*
...
@@ -41,6 +41,8 @@ propose novel applications to genomics data.*
Tips:
Tips:
- For an in-detail explanation on how BigBird's attention works, see `this blog post
<https://huggingface.co/blog/big-bird>`__.
- BigBird comes with 2 implementations: **original_full** & **block_sparse**. For the sequence length < 1024, using
- BigBird comes with 2 implementations: **original_full** & **block_sparse**. For the sequence length < 1024, using
**original_full** is advised as there is no benefit in using **block_sparse** attention.
**original_full** is advised as there is no benefit in using **block_sparse** attention.
- The code currently uses window size of 3 blocks and 2 global blocks.
- The code currently uses window size of 3 blocks and 2 global blocks.
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment