Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
5f25a5f3
Commit
5f25a5f3
authored
Oct 11, 2019
by
Stefan Schweter
Browse files
model: add support for new German BERT models (cased and uncased) from @dbmdz
parent
f382a8de
Changes
5
Show whitespace changes
Inline
Side-by-side
Showing
5 changed files
with
20 additions
and
0 deletions
+20
-0
docs/source/pretrained_models.rst
docs/source/pretrained_models.rst
+8
-0
docs/source/serialization.rst
docs/source/serialization.rst
+2
-0
transformers/configuration_bert.py
transformers/configuration_bert.py
+2
-0
transformers/modeling_bert.py
transformers/modeling_bert.py
+2
-0
transformers/tokenization_bert.py
transformers/tokenization_bert.py
+6
-0
No files found.
docs/source/pretrained_models.rst
View file @
5f25a5f3
...
...
@@ -53,6 +53,14 @@ Here is the full list of the currently provided pretrained models together with
| | ``bert-base-cased-finetuned-mrpc`` | | 12-layer, 768-hidden, 12-heads, 110M parameters. |
| | | | The ``bert-base-cased`` model fine-tuned on MRPC |
| | | (see `details of fine-tuning in the example section <https://huggingface.co/transformers/examples.html>`__) |
| +------------------------------------------------------------+---------------------------------------------------------------------------------------------------------------------------------------+
| | ``bert-base-german-dbmdz-cased`` | | 12-layer, 768-hidden, 12-heads, 110M parameters. |
| | | | Trained on cased German text by DBMDZ |
| | | (see `details on dbmdz repository <https://github.com/dbmdz/german-bert>`__). |
| +------------------------------------------------------------+---------------------------------------------------------------------------------------------------------------------------------------+
| | ``bert-base-german-dbmdz-uncased`` | | 12-layer, 768-hidden, 12-heads, 110M parameters. |
| | | | Trained on uncased German text by DBMDZ |
| | | (see `details on dbmdz repository <https://github.com/dbmdz/german-bert>`__). |
+-------------------+------------------------------------------------------------+---------------------------------------------------------------------------------------------------------------------------------------+
| GPT | ``openai-gpt`` | | 12-layer, 768-hidden, 12-heads, 110M parameters. |
| | | | OpenAI GPT English model |
...
...
docs/source/serialization.rst
View file @
5f25a5f3
...
...
@@ -33,6 +33,8 @@ where
*
``
bert
-
large
-
uncased
-
whole
-
word
-
masking
``:
24
-
layer
,
1024
-
hidden
,
16
-
heads
,
340
M
parameters
-
Trained
with
Whole
Word
Masking
(
mask
all
of
the
the
tokens
corresponding
to
a
word
at
once
)
*
``
bert
-
large
-
cased
-
whole
-
word
-
masking
``:
24
-
layer
,
1024
-
hidden
,
16
-
heads
,
340
M
parameters
-
Trained
with
Whole
Word
Masking
(
mask
all
of
the
the
tokens
corresponding
to
a
word
at
once
)
*
``
bert
-
large
-
uncased
-
whole
-
word
-
masking
-
finetuned
-
squad
``:
The
``
bert
-
large
-
uncased
-
whole
-
word
-
masking
``
model
finetuned
on
SQuAD
(
using
the
``
run_bert_squad
.
py
``
examples
).
Results
:
*
exact_match
:
86.91579943235573
,
f1
:
93.1532499015869
*
*
``
bert
-
base
-
german
-
dbmdz
-
cased
``:
Trained
on
German
data
only
,
12
-
layer
,
768
-
hidden
,
12
-
heads
,
110
M
parameters
`
Performance
Evaluation
<
https
://
github
.
com
/
dbmdz
/
german
-
bert
>`
__
*
``
bert
-
base
-
german
-
dbmdz
-
uncased
``:
Trained
on
(
uncased
)
German
data
only
,
12
-
layer
,
768
-
hidden
,
12
-
heads
,
110
M
parameters
`
Performance
Evaluation
<
https
://
github
.
com
/
dbmdz
/
german
-
bert
>`
__
*
``
openai
-
gpt
``:
OpenAI
GPT
English
model
,
12
-
layer
,
768
-
hidden
,
12
-
heads
,
110
M
parameters
*
``
gpt2
``:
OpenAI
GPT
-
2
English
model
,
12
-
layer
,
768
-
hidden
,
12
-
heads
,
117
M
parameters
*
``
gpt2
-
medium
``:
OpenAI
GPT
-
2
English
model
,
24
-
layer
,
1024
-
hidden
,
16
-
heads
,
345
M
parameters
...
...
transformers/configuration_bert.py
View file @
5f25a5f3
...
...
@@ -40,6 +40,8 @@ BERT_PRETRAINED_CONFIG_ARCHIVE_MAP = {
'bert-large-uncased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-uncased-whole-word-masking-finetuned-squad-config.json"
,
'bert-large-cased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-cased-whole-word-masking-finetuned-squad-config.json"
,
'bert-base-cased-finetuned-mrpc'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-cased-finetuned-mrpc-config.json"
,
'bert-base-german-dbmdz-cased'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-german-dbmdz-cased-config.json"
,
'bert-base-german-dbmdz-uncased'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-german-dbmdz-uncased-config.json"
,
}
...
...
transformers/modeling_bert.py
View file @
5f25a5f3
...
...
@@ -48,6 +48,8 @@ BERT_PRETRAINED_MODEL_ARCHIVE_MAP = {
'bert-large-uncased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-uncased-whole-word-masking-finetuned-squad-pytorch_model.bin"
,
'bert-large-cased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-cased-whole-word-masking-finetuned-squad-pytorch_model.bin"
,
'bert-base-cased-finetuned-mrpc'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-cased-finetuned-mrpc-pytorch_model.bin"
,
'bert-base-german-dbmdz-cased'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-german-dbmdz-cased-pytorch_model.bin"
,
'bert-base-german-dbmdz-uncased'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-german-dbmdz-uncased-pytorch_model.bin"
,
}
def
load_tf_weights_in_bert
(
model
,
config
,
tf_checkpoint_path
):
...
...
transformers/tokenization_bert.py
View file @
5f25a5f3
...
...
@@ -44,6 +44,8 @@ PRETRAINED_VOCAB_FILES_MAP = {
'bert-large-uncased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-uncased-whole-word-masking-finetuned-squad-vocab.txt"
,
'bert-large-cased-whole-word-masking-finetuned-squad'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-cased-whole-word-masking-finetuned-squad-vocab.txt"
,
'bert-base-cased-finetuned-mrpc'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-cased-finetuned-mrpc-vocab.txt"
,
'bert-base-german-dbmdz-cased'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-german-dbmdz-cased-vocab.txt"
,
'bert-base-german-dbmdz-uncased'
:
"https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-german-dbmdz-uncased-vocab.txt"
,
}
}
...
...
@@ -61,6 +63,8 @@ PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES = {
'bert-large-uncased-whole-word-masking-finetuned-squad'
:
512
,
'bert-large-cased-whole-word-masking-finetuned-squad'
:
512
,
'bert-base-cased-finetuned-mrpc'
:
512
,
'bert-base-german-dbmdz-cased'
:
512
,
'bert-base-german-dbmdz-uncased'
:
512
,
}
PRETRAINED_INIT_CONFIGURATION
=
{
...
...
@@ -77,6 +81,8 @@ PRETRAINED_INIT_CONFIGURATION = {
'bert-large-uncased-whole-word-masking-finetuned-squad'
:
{
'do_lower_case'
:
True
},
'bert-large-cased-whole-word-masking-finetuned-squad'
:
{
'do_lower_case'
:
False
},
'bert-base-cased-finetuned-mrpc'
:
{
'do_lower_case'
:
False
},
'bert-base-german-dbmdz-cased'
:
{
'do_lower_case'
:
False
},
'bert-base-german-dbmdz-uncased'
:
{
'do_lower_case'
:
True
},
}
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment