Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
25b0f203
"git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "73efa694e6cc9141ce732ea8215b25b3083298dc"
Unverified
Commit
25b0f203
authored
Nov 18, 2023
by
Omar Sanseviero
Committed by
GitHub
Nov 18, 2023
Browse files
Fix broken distilbert url (#27579)
parent
d1a00f9d
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
1 addition
and
1 deletion
+1
-1
docs/source/en/model_doc/distilbert.md
docs/source/en/model_doc/distilbert.md
+1
-1
No files found.
docs/source/en/model_doc/distilbert.md
View file @
25b0f203
...
@@ -32,7 +32,7 @@ rendered properly in your Markdown viewer.
...
@@ -32,7 +32,7 @@ rendered properly in your Markdown viewer.
The DistilBERT model was proposed in the blog post
[
Smaller, faster, cheaper, lighter: Introducing DistilBERT, a
The DistilBERT model was proposed in the blog post
[
Smaller, faster, cheaper, lighter: Introducing DistilBERT, a
distilled version of BERT
](
https://medium.com/huggingface/distilbert-8cf3380435b5
)
, and the paper
[
DistilBERT, a
distilled version of BERT
](
https://medium.com/huggingface/distilbert-8cf3380435b5
)
, and the paper
[
DistilBERT, a
distilled version of BERT: smaller, faster, cheaper and lighter
](
https://arxiv.org/
paper
s/1910.01108
)
. DistilBERT is a
distilled version of BERT: smaller, faster, cheaper and lighter
](
https://arxiv.org/
ab
s/1910.01108
)
. DistilBERT is a
small, fast, cheap and light Transformer model trained by distilling BERT base. It has 40% less parameters than
small, fast, cheap and light Transformer model trained by distilling BERT base. It has 40% less parameters than
*bert-base-uncased*
, runs 60% faster while preserving over 95% of BERT's performances as measured on the GLUE language
*bert-base-uncased*
, runs 60% faster while preserving over 95% of BERT's performances as measured on the GLUE language
understanding benchmark.
understanding benchmark.
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment