Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
66f89332
Unverified
Commit
66f89332
authored
Jun 15, 2022
by
Stas Bekman
Committed by
GitHub
Jun 15, 2022
Browse files
normalize keys_to_ignore (#17722)
parent
c3c62b5d
Changes
23
Show whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
11 additions
and
11 deletions
+11
-11
src/transformers/models/speech_to_text/modeling_speech_to_text.py
...sformers/models/speech_to_text/modeling_speech_to_text.py
+2
-2
src/transformers/models/t5/modeling_t5.py
src/transformers/models/t5/modeling_t5.py
+8
-8
src/transformers/models/transfo_xl/modeling_transfo_xl.py
src/transformers/models/transfo_xl/modeling_transfo_xl.py
+1
-1
No files found.
src/transformers/models/speech_to_text/modeling_speech_to_text.py
View file @
66f89332
...
@@ -1252,8 +1252,8 @@ class Speech2TextModel(Speech2TextPreTrainedModel):
...
@@ -1252,8 +1252,8 @@ class Speech2TextModel(Speech2TextPreTrainedModel):
class
Speech2TextForConditionalGeneration
(
Speech2TextPreTrainedModel
):
class
Speech2TextForConditionalGeneration
(
Speech2TextPreTrainedModel
):
base_model_prefix
=
"model"
base_model_prefix
=
"model"
_keys_to_ignore_on_load_missing
=
[
_keys_to_ignore_on_load_missing
=
[
r
"encoder
\
.version"
,
r
"encoder.version"
,
r
"decoder
\
.version"
,
r
"decoder.version"
,
r
"model.encoder.embed_positions.weights"
,
r
"model.encoder.embed_positions.weights"
,
r
"model.decoder.embed_positions.weights"
,
r
"model.decoder.embed_positions.weights"
,
]
]
...
...
src/transformers/models/t5/modeling_t5.py
View file @
66f89332
...
@@ -1266,11 +1266,11 @@ num_heads)`.
...
@@ -1266,11 +1266,11 @@ num_heads)`.
)
)
class
T5Model
(
T5PreTrainedModel
):
class
T5Model
(
T5PreTrainedModel
):
_keys_to_ignore_on_load_missing
=
[
_keys_to_ignore_on_load_missing
=
[
r
"encoder
\
.embed_tokens
\
.weight"
,
r
"encoder.embed_tokens.weight"
,
r
"decoder
\
.embed_tokens
\
.weight"
,
r
"decoder.embed_tokens.weight"
,
]
]
_keys_to_ignore_on_load_unexpected
=
[
_keys_to_ignore_on_load_unexpected
=
[
r
"decoder
\
.block
\
.0
\
.layer
\
.1
\
.EncDecAttention
\
.relative_attention_bias
\
.weight"
,
r
"decoder.block.0.layer.1.EncDecAttention.relative_attention_bias.weight"
,
]
]
def
__init__
(
self
,
config
:
T5Config
):
def
__init__
(
self
,
config
:
T5Config
):
...
@@ -1455,12 +1455,12 @@ class T5Model(T5PreTrainedModel):
...
@@ -1455,12 +1455,12 @@ class T5Model(T5PreTrainedModel):
@
add_start_docstrings
(
"""T5 Model with a `language modeling` head on top."""
,
T5_START_DOCSTRING
)
@
add_start_docstrings
(
"""T5 Model with a `language modeling` head on top."""
,
T5_START_DOCSTRING
)
class
T5ForConditionalGeneration
(
T5PreTrainedModel
):
class
T5ForConditionalGeneration
(
T5PreTrainedModel
):
_keys_to_ignore_on_load_missing
=
[
_keys_to_ignore_on_load_missing
=
[
r
"encoder
\
.embed_tokens
\
.weight"
,
r
"encoder.embed_tokens.weight"
,
r
"decoder
\
.embed_tokens
\
.weight"
,
r
"decoder.embed_tokens.weight"
,
r
"lm_head
\
.weight"
,
r
"lm_head.weight"
,
]
]
_keys_to_ignore_on_load_unexpected
=
[
_keys_to_ignore_on_load_unexpected
=
[
r
"decoder
\
.block
\
.0
\
.layer
\
.1
\
.EncDecAttention
\
.relative_attention_bias
\
.weight"
,
r
"decoder.block.0.layer.1.EncDecAttention.relative_attention_bias.weight"
,
]
]
def
__init__
(
self
,
config
:
T5Config
):
def
__init__
(
self
,
config
:
T5Config
):
...
@@ -1749,7 +1749,7 @@ class T5ForConditionalGeneration(T5PreTrainedModel):
...
@@ -1749,7 +1749,7 @@ class T5ForConditionalGeneration(T5PreTrainedModel):
)
)
class
T5EncoderModel
(
T5PreTrainedModel
):
class
T5EncoderModel
(
T5PreTrainedModel
):
authorized_missing_keys
=
[
authorized_missing_keys
=
[
r
"encoder
\
.embed_tokens
\
.weight"
,
r
"encoder.embed_tokens.weight"
,
]
]
def
__init__
(
self
,
config
:
T5Config
):
def
__init__
(
self
,
config
:
T5Config
):
...
...
src/transformers/models/transfo_xl/modeling_transfo_xl.py
View file @
66f89332
...
@@ -1198,7 +1198,7 @@ class TransfoXLLMHeadModel(TransfoXLPreTrainedModel):
...
@@ -1198,7 +1198,7 @@ class TransfoXLLMHeadModel(TransfoXLPreTrainedModel):
TRANSFO_XL_START_DOCSTRING
,
TRANSFO_XL_START_DOCSTRING
,
)
)
class
TransfoXLForSequenceClassification
(
TransfoXLPreTrainedModel
):
class
TransfoXLForSequenceClassification
(
TransfoXLPreTrainedModel
):
_keys_to_ignore_on_load_missing
=
[
r
"h\.\d+\.attn\.masked_bias"
,
r
"lm_head
\
.weight"
]
_keys_to_ignore_on_load_missing
=
[
r
"h\.\d+\.attn\.masked_bias"
,
r
"lm_head.weight"
]
def
__init__
(
self
,
config
):
def
__init__
(
self
,
config
):
super
().
__init__
(
config
)
super
().
__init__
(
config
)
...
...
Prev
1
2
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment