Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
700331b5
Unverified
Commit
700331b5
authored
Oct 11, 2019
by
Thomas Wolf
Committed by
GitHub
Oct 11, 2019
Browse files
Merge pull request #1492 from stefan-it/bert-german-dbmdz-models
Add new BERT models for German (cased and uncased)
parents
573dde9b
5f25a5f3
Changes
5
Show whitespace changes
Inline
Side-by-side
Showing
5 changed files
with
20 additions
and
0 deletions
+20
-0
docs/source/pretrained_models.rst
docs/source/pretrained_models.rst
+8
-0
docs/source/serialization.rst
docs/source/serialization.rst
+2
-0
transformers/configuration_bert.py
transformers/configuration_bert.py
+2
-0
transformers/modeling_bert.py
transformers/modeling_bert.py
+2
-0
transformers/tokenization_bert.py
transformers/tokenization_bert.py
+6
-0
No files found.
docs/source/pretrained_models.rst
View file @
700331b5
...
@@ -53,6 +53,14 @@ Here is the full list of the currently provided pretrained models together with
...
@@ -53,6 +53,14 @@ Here is the full list of the currently provided pretrained models together with
| | ``bert-base-cased-finetuned-mrpc`` | | 12-layer, 768-hidden, 12-heads, 110M parameters. |
| | ``bert-base-cased-finetuned-mrpc`` | | 12-layer, 768-hidden, 12-heads, 110M parameters. |
| | | | The ``bert-base-cased`` model fine-tuned on MRPC |
| | | | The ``bert-base-cased`` model fine-tuned on MRPC |
| | | (see `details of fine-tuning in the example section <https://huggingface.co/transformers/examples.html>`__) |
| | | (see `details of fine-tuning in the example section <https://huggingface.co/transformers/examples.html>`__) |
| +------------------------------------------------------------+---------------------------------------------------------------------------------------------------------------------------------------+
| | ``bert-base-german-dbmdz-cased`` | | 12-layer, 768-hidden, 12-heads, 110M parameters. |
| | | | Trained on cased German text by DBMDZ |
| | | (see `details on dbmdz repository <https://github.com/dbmdz/german-bert>`__). |
| +------------------------------------------------------------+---------------------------------------------------------------------------------------------------------------------------------------+
| | ``bert-base-german-dbmdz-uncased`` | | 12-layer, 768-hidden, 12-heads, 110M parameters. |
| | | | Trained on uncased German text by DBMDZ |
| | | (see `details on dbmdz repository <https://github.com/dbmdz/german-bert>`__). |
+-------------------+------------------------------------------------------------+---------------------------------------------------------------------------------------------------------------------------------------+
+-------------------+------------------------------------------------------------+---------------------------------------------------------------------------------------------------------------------------------------+
| GPT | ``openai-gpt`` | | 12-layer, 768-hidden, 12-heads, 110M parameters. |
| GPT | ``openai-gpt`` | | 12-layer, 768-hidden, 12-heads, 110M parameters. |
| | | | OpenAI GPT English model |
| | | | OpenAI GPT English model |
...
...
docs/source/serialization.rst
View file @
700331b5
...
@@ -33,6 +33,8 @@ where
...
@@ -33,6 +33,8 @@ where
*
``
bert
-
large
-
uncased
-
whole
-
word
-
masking
``:
24
-
layer
,
1024
-
hidden
,
16
-
heads
,
340
M
parameters
-
Trained
with
Whole
Word
Masking
(
mask
all
of
the
the
tokens
corresponding
to
a
word
at
once
)
*
``
bert
-
large
-
uncased
-
whole
-
word
-
masking
``:
24
-
layer
,
1024
-
hidden
,
16
-
heads
,
340
M
parameters
-
Trained
with
Whole
Word
Masking
(
mask
all
of
the
the
tokens
corresponding
to
a
word
at
once
)
*
``
bert
-
large
-
cased
-
whole
-
word
-
masking
``:
24
-
layer
,
1024
-
hidden
,
16
-
heads
,
340
M
parameters
-
Trained
with
Whole
Word
Masking
(
mask
all
of
the
the
tokens
corresponding
to
a
word
at
once
)
*
``
bert
-
large
-
cased
-
whole
-
word
-
masking
``:
24
-
layer
,
1024
-
hidden
,
16
-
heads
,
340
M
parameters
-
Trained
with
Whole
Word
Masking
(
mask
all
of
the
the
tokens
corresponding
to
a
word
at
once
)
*
``
bert
-
large
-
uncased
-
whole
-
word
-
masking
-
finetuned
-
squad
``:
The
``
bert
-
large
-
uncased
-
whole
-
word
-
masking
``
model
finetuned
on
SQuAD
(
using
the
``
run_bert_squad
.
py
``
examples
).
Results
:
*
exact_match
:
86.91579943235573
,
f1
:
93.1532499015869
*
*
``
bert
-
large
-
uncased
-
whole
-
word
-
masking
-
finetuned
-
squad
``:
The
``
bert
-
large
-
uncased
-
whole
-
word
-
masking
``
model
finetuned
on
SQuAD
(
using
the
``
run_bert_squad
.
py
``
examples
).
Results
:
*
exact_match
:
86.91579943235573
,
f1
:
93.1532499015869
*
*
``
bert
-
base
-
german
-
dbmdz
-
cased
``:
Trained
on
German
data
only
,
12
-
layer
,
768
-
hidden
,
12
-
heads
,
110
M
parameters
`
Performance
Evaluation
<
https
://
github
.
com
/
dbmdz
/
german
-
bert
>`
__
*
``
bert
-
base
-
german
-
dbmdz
-
uncased
``:
Trained
on
(
uncased
)
German
data
only
,
12
-
layer
,
768
-
hidden
,
12
-
heads
,
110
M
parameters
`
Performance
Evaluation
<
https
://
github
.
com
/
dbmdz
/
german
-
bert
>`
__
*
``
openai
-
gpt
``:
OpenAI
GPT
English
model
,
12
-
layer
,
768
-
hidden
,
12
-
heads
,
110
M
parameters
*
``
openai
-
gpt
``:
OpenAI
GPT
English
model
,
12
-
layer
,
768
-
hidden
,
12
-
heads
,
110
M
parameters
*
``
gpt2
``:
OpenAI
GPT
-
2
English
model
,
12
-
layer
,
768
-
hidden
,
12
-
heads
,
117
M
parameters
*
``
gpt2
``:
OpenAI
GPT
-
2
English
model
,
12
-
layer
,
768
-
hidden
,
12
-
heads
,
117
M
parameters
*
``
gpt2
-
medium
``:
OpenAI
GPT
-
2
English
model
,
24
-
layer
,
1024
-
hidden
,
16
-
heads
,
345
M
parameters
*
``
gpt2
-
medium
``:
OpenAI
GPT
-
2
English
model
,
24
-
layer
,
1024
-
hidden
,
16
-
heads
,
345
M
parameters
...
...
transformers/configuration_bert.py
View file @
700331b5
...
@@ -40,6 +40,8 @@ BERT_PRETRAINED_CONFIG_ARCHIVE_MAP = {
...
@@ -40,6 +40,8 @@ BERT_PRETRAINED_CONFIG_ARCHIVE_MAP = {
'bert-large-uncased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-uncased-whole-word-masking-finetuned-squad-config.json"
,
'bert-large-uncased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-uncased-whole-word-masking-finetuned-squad-config.json"
,
'bert-large-cased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-cased-whole-word-masking-finetuned-squad-config.json"
,
'bert-large-cased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-cased-whole-word-masking-finetuned-squad-config.json"
,
'bert-base-cased-finetuned-mrpc'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-cased-finetuned-mrpc-config.json"
,
'bert-base-cased-finetuned-mrpc'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-cased-finetuned-mrpc-config.json"
,
'bert-base-german-dbmdz-cased'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-german-dbmdz-cased-config.json"
,
'bert-base-german-dbmdz-uncased'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-german-dbmdz-uncased-config.json"
,
}
}
...
...
transformers/modeling_bert.py
View file @
700331b5
...
@@ -48,6 +48,8 @@ BERT_PRETRAINED_MODEL_ARCHIVE_MAP = {
...
@@ -48,6 +48,8 @@ BERT_PRETRAINED_MODEL_ARCHIVE_MAP = {
'bert-large-uncased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-uncased-whole-word-masking-finetuned-squad-pytorch_model.bin"
,
'bert-large-uncased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-uncased-whole-word-masking-finetuned-squad-pytorch_model.bin"
,
'bert-large-cased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-cased-whole-word-masking-finetuned-squad-pytorch_model.bin"
,
'bert-large-cased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-cased-whole-word-masking-finetuned-squad-pytorch_model.bin"
,
'bert-base-cased-finetuned-mrpc'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-cased-finetuned-mrpc-pytorch_model.bin"
,
'bert-base-cased-finetuned-mrpc'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-cased-finetuned-mrpc-pytorch_model.bin"
,
'bert-base-german-dbmdz-cased'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-german-dbmdz-cased-pytorch_model.bin"
,
'bert-base-german-dbmdz-uncased'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-german-dbmdz-uncased-pytorch_model.bin"
,
}
}
def
load_tf_weights_in_bert
(
model
,
config
,
tf_checkpoint_path
):
def
load_tf_weights_in_bert
(
model
,
config
,
tf_checkpoint_path
):
...
...
transformers/tokenization_bert.py
View file @
700331b5
...
@@ -44,6 +44,8 @@ PRETRAINED_VOCAB_FILES_MAP = {
...
@@ -44,6 +44,8 @@ PRETRAINED_VOCAB_FILES_MAP = {
'bert-large-uncased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-uncased-whole-word-masking-finetuned-squad-vocab.txt"
,
'bert-large-uncased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-uncased-whole-word-masking-finetuned-squad-vocab.txt"
,
'bert-large-cased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-cased-whole-word-masking-finetuned-squad-vocab.txt"
,
'bert-large-cased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-cased-whole-word-masking-finetuned-squad-vocab.txt"
,
'bert-base-cased-finetuned-mrpc'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-cased-finetuned-mrpc-vocab.txt"
,
'bert-base-cased-finetuned-mrpc'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-cased-finetuned-mrpc-vocab.txt"
,
'bert-base-german-dbmdz-cased'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-german-dbmdz-cased-vocab.txt"
,
'bert-base-german-dbmdz-uncased'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-german-dbmdz-uncased-vocab.txt"
,
}
}
}
}
...
@@ -61,6 +63,8 @@ PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES = {
...
@@ -61,6 +63,8 @@ PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES = {
'bert-large-uncased-whole-word-masking-finetuned-squad'
:
512
,
'bert-large-uncased-whole-word-masking-finetuned-squad'
:
512
,
'bert-large-cased-whole-word-masking-finetuned-squad'
:
512
,
'bert-large-cased-whole-word-masking-finetuned-squad'
:
512
,
'bert-base-cased-finetuned-mrpc'
:
512
,
'bert-base-cased-finetuned-mrpc'
:
512
,
'bert-base-german-dbmdz-cased'
:
512
,
'bert-base-german-dbmdz-uncased'
:
512
,
}
}
PRETRAINED_INIT_CONFIGURATION
=
{
PRETRAINED_INIT_CONFIGURATION
=
{
...
@@ -77,6 +81,8 @@ PRETRAINED_INIT_CONFIGURATION = {
...
@@ -77,6 +81,8 @@ PRETRAINED_INIT_CONFIGURATION = {
'bert-large-uncased-whole-word-masking-finetuned-squad'
:
{
'do_lower_case'
:
True
},
'bert-large-uncased-whole-word-masking-finetuned-squad'
:
{
'do_lower_case'
:
True
},
'bert-large-cased-whole-word-masking-finetuned-squad'
:
{
'do_lower_case'
:
False
},
'bert-large-cased-whole-word-masking-finetuned-squad'
:
{
'do_lower_case'
:
False
},
'bert-base-cased-finetuned-mrpc'
:
{
'do_lower_case'
:
False
},
'bert-base-cased-finetuned-mrpc'
:
{
'do_lower_case'
:
False
},
'bert-base-german-dbmdz-cased'
:
{
'do_lower_case'
:
False
},
'bert-base-german-dbmdz-uncased'
:
{
'do_lower_case'
:
True
},
}
}
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment