Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
fa6113f9
"...sklearn/git@developer.sourcefind.cn:OpenDAS/nni.git" did not exist on "ae72aec87dbce3a3f328c9a70a82132f825634d6"
Unverified
Commit
fa6113f9
authored
May 18, 2020
by
Soham Chatterjee
Committed by
GitHub
May 18, 2020
Browse files
Fixed spelling of training (#4416)
parent
757baee8
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
1 addition
and
1 deletion
+1
-1
docs/source/model_doc/albert.rst
docs/source/model_doc/albert.rst
+1
-1
No files found.
docs/source/model_doc/albert.rst
View file @
fa6113f9
...
...
@@ -6,7 +6,7 @@ Overview
The ALBERT model was proposed in `ALBERT: A Lite BERT for Self-supervised Learning of Language Representations <https://arxiv.org/abs/1909.11942>`_
by Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut. It presents
two parameter-reduction techniques to lower memory consumption and increase the trainig speed of BERT:
two parameter-reduction techniques to lower memory consumption and increase the traini
n
g speed of BERT:
- Splitting the embedding matrix into two smaller matrices
- Using repeating layers split among groups
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment