Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
0f443436
"git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "f2c1df93f5bb13a57de21e355836b7aa7c820d63"
Unverified
Commit
0f443436
authored
Jan 25, 2021
by
Lysandre Debut
Committed by
GitHub
Jan 25, 2021
Browse files
Actual fix (#9787)
parent
fac7cfb1
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
15 additions
and
0 deletions
+15
-0
src/transformers/models/gpt2/modeling_gpt2.py
src/transformers/models/gpt2/modeling_gpt2.py
+11
-0
src/transformers/models/t5/modeling_t5.py
src/transformers/models/t5/modeling_t5.py
+4
-0
No files found.
src/transformers/models/gpt2/modeling_gpt2.py
View file @
0f443436
...
@@ -541,6 +541,7 @@ class GPT2Model(GPT2PreTrainedModel):
...
@@ -541,6 +541,7 @@ class GPT2Model(GPT2PreTrainedModel):
self
.
ln_f
=
nn
.
LayerNorm
(
config
.
n_embd
,
eps
=
config
.
layer_norm_epsilon
)
self
.
ln_f
=
nn
.
LayerNorm
(
config
.
n_embd
,
eps
=
config
.
layer_norm_epsilon
)
self
.
init_weights
()
self
.
init_weights
()
# Model parallel
# Model parallel
self
.
model_parallel
=
False
self
.
model_parallel
=
False
self
.
device_map
=
None
self
.
device_map
=
None
...
@@ -805,7 +806,9 @@ class GPT2LMHeadModel(GPT2PreTrainedModel):
...
@@ -805,7 +806,9 @@ class GPT2LMHeadModel(GPT2PreTrainedModel):
self
.
init_weights
()
self
.
init_weights
()
# Model parallel
self
.
model_parallel
=
False
self
.
model_parallel
=
False
self
.
device_map
=
None
@
add_start_docstrings
(
PARALLELIZE_DOCSTRING
)
@
add_start_docstrings
(
PARALLELIZE_DOCSTRING
)
def
parallelize
(
self
,
device_map
=
None
):
def
parallelize
(
self
,
device_map
=
None
):
...
@@ -971,6 +974,10 @@ class GPT2DoubleHeadsModel(GPT2PreTrainedModel):
...
@@ -971,6 +974,10 @@ class GPT2DoubleHeadsModel(GPT2PreTrainedModel):
self
.
init_weights
()
self
.
init_weights
()
# Model parallel
self
.
model_parallel
=
False
self
.
device_map
=
None
def
get_output_embeddings
(
self
):
def
get_output_embeddings
(
self
):
return
self
.
lm_head
return
self
.
lm_head
...
@@ -1153,6 +1160,10 @@ class GPT2ForSequenceClassification(GPT2PreTrainedModel):
...
@@ -1153,6 +1160,10 @@ class GPT2ForSequenceClassification(GPT2PreTrainedModel):
self
.
init_weights
()
self
.
init_weights
()
# Model parallel
self
.
model_parallel
=
False
self
.
device_map
=
None
@
add_start_docstrings_to_model_forward
(
GPT2_INPUTS_DOCSTRING
)
@
add_start_docstrings_to_model_forward
(
GPT2_INPUTS_DOCSTRING
)
@
add_code_sample_docstrings
(
@
add_code_sample_docstrings
(
tokenizer_class
=
_TOKENIZER_FOR_DOC
,
tokenizer_class
=
_TOKENIZER_FOR_DOC
,
...
...
src/transformers/models/t5/modeling_t5.py
View file @
0f443436
...
@@ -1651,6 +1651,10 @@ class T5EncoderModel(T5PreTrainedModel):
...
@@ -1651,6 +1651,10 @@ class T5EncoderModel(T5PreTrainedModel):
self
.
init_weights
()
self
.
init_weights
()
# Model parallel
self
.
model_parallel
=
False
self
.
device_map
=
None
@
add_start_docstrings
(
PARALLELIZE_DOCSTRING
)
@
add_start_docstrings
(
PARALLELIZE_DOCSTRING
)
def
parallelize
(
self
,
device_map
=
None
):
def
parallelize
(
self
,
device_map
=
None
):
self
.
device_map
=
(
self
.
device_map
=
(
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment