Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
d9e848c1
"docs/vscode:/vscode.git/clone" did not exist on "ae320fa53f74cc4dfa0e4fc3c95b6129a86b0512"
Unverified
Commit
d9e848c1
authored
Jan 05, 2021
by
Stas Bekman
Committed by
GitHub
Jan 05, 2021
Browse files
add experimental warning (#9412)
parent
29acabd8
Changes
2
Show whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
4 additions
and
0 deletions
+4
-0
src/transformers/models/gpt2/modeling_gpt2.py
src/transformers/models/gpt2/modeling_gpt2.py
+2
-0
src/transformers/models/t5/modeling_t5.py
src/transformers/models/t5/modeling_t5.py
+2
-0
No files found.
src/transformers/models/gpt2/modeling_gpt2.py
View file @
d9e848c1
...
@@ -480,6 +480,8 @@ GPT2_INPUTS_DOCSTRING = r"""
...
@@ -480,6 +480,8 @@ GPT2_INPUTS_DOCSTRING = r"""
Whether or not to return a :class:`~transformers.file_utils.ModelOutput` instead of a plain tuple.
Whether or not to return a :class:`~transformers.file_utils.ModelOutput` instead of a plain tuple.
"""
"""
PARALLELIZE_DOCSTRING
=
r
"""
PARALLELIZE_DOCSTRING
=
r
"""
This is an experimental feature and is a subject to change at a moment's notice.
Uses a device map to distribute attention modules of the model across several devices. If no device map is given,
Uses a device map to distribute attention modules of the model across several devices. If no device map is given,
it will evenly distribute blocks across all devices.
it will evenly distribute blocks across all devices.
...
...
src/transformers/models/t5/modeling_t5.py
View file @
d9e848c1
...
@@ -179,6 +179,8 @@ def load_tf_weights_in_t5(model, config, tf_checkpoint_path):
...
@@ -179,6 +179,8 @@ def load_tf_weights_in_t5(model, config, tf_checkpoint_path):
# - PreTrainedModel for the models (it-self a sub-class of torch.nn.Module)
# - PreTrainedModel for the models (it-self a sub-class of torch.nn.Module)
####################################################
####################################################
PARALLELIZE_DOCSTRING
=
r
"""
PARALLELIZE_DOCSTRING
=
r
"""
This is an experimental feature and is a subject to change at a moment's notice.
Uses a device map to distribute attention modules of the model across several devices. If no device map is given,
Uses a device map to distribute attention modules of the model across several devices. If no device map is given,
it will evenly distribute blocks across all devices.
it will evenly distribute blocks across all devices.
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment