Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
36b60ce9
"...git@developer.sourcefind.cn:chenpangpang/ComfyUI.git" did not exist on "ca2ae98470fdebd951cdb750998b82ecb532c901"
Unverified
Commit
36b60ce9
authored
Nov 28, 2020
by
Patrick von Platen
Committed by
GitHub
Nov 28, 2020
Browse files
fix mt5 config (#8832)
parent
18c32eeb
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
4 additions
and
0 deletions
+4
-0
src/transformers/models/mt5/configuration_mt5.py
src/transformers/models/mt5/configuration_mt5.py
+4
-0
No files found.
src/transformers/models/mt5/configuration_mt5.py
View file @
36b60ce9
...
@@ -60,6 +60,8 @@ class MT5Config(PretrainedConfig):
...
@@ -60,6 +60,8 @@ class MT5Config(PretrainedConfig):
testing).
testing).
feed_forward_proj (:obj:`string`, `optional`, defaults to :obj:`"gated-gelu"`):
feed_forward_proj (:obj:`string`, `optional`, defaults to :obj:`"gated-gelu"`):
Type of feed forward layer to be used. Should be one of :obj:`"relu"` or :obj:`"gated-gelu"`.
Type of feed forward layer to be used. Should be one of :obj:`"relu"` or :obj:`"gated-gelu"`.
use_cache (:obj:`bool`, `optional`, defaults to :obj:`True`):
Whether or not the model should return the last key/values attentions (not used by all models).
"""
"""
model_type
=
"mt5"
model_type
=
"mt5"
keys_to_ignore_at_inference
=
[
"past_key_values"
]
keys_to_ignore_at_inference
=
[
"past_key_values"
]
...
@@ -79,6 +81,7 @@ class MT5Config(PretrainedConfig):
...
@@ -79,6 +81,7 @@ class MT5Config(PretrainedConfig):
initializer_factor
=
1.0
,
initializer_factor
=
1.0
,
feed_forward_proj
=
"gated-gelu"
,
feed_forward_proj
=
"gated-gelu"
,
is_encoder_decoder
=
True
,
is_encoder_decoder
=
True
,
use_cache
=
True
,
tokenizer_class
=
"T5Tokenizer"
,
tokenizer_class
=
"T5Tokenizer"
,
tie_word_embeddings
=
False
,
tie_word_embeddings
=
False
,
pad_token_id
=
0
,
pad_token_id
=
0
,
...
@@ -109,6 +112,7 @@ class MT5Config(PretrainedConfig):
...
@@ -109,6 +112,7 @@ class MT5Config(PretrainedConfig):
self
.
layer_norm_epsilon
=
layer_norm_epsilon
self
.
layer_norm_epsilon
=
layer_norm_epsilon
self
.
initializer_factor
=
initializer_factor
self
.
initializer_factor
=
initializer_factor
self
.
feed_forward_proj
=
feed_forward_proj
self
.
feed_forward_proj
=
feed_forward_proj
self
.
use_cache
=
use_cache
@
property
@
property
def
hidden_size
(
self
):
def
hidden_size
(
self
):
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment