Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
c7f3abc2
Unverified
Commit
c7f3abc2
authored
Feb 27, 2023
by
Stas Bekman
Committed by
GitHub
Feb 27, 2023
Browse files
introduce `logger.warning_once` and use it for grad checkpointing code (#21804)
* logger.warning_once * style
parent
f95f60c8
Changes
58
Hide whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
20 additions
and
20 deletions
+20
-20
src/transformers/models/altclip/modeling_altclip.py
src/transformers/models/altclip/modeling_altclip.py
+1
-1
src/transformers/models/bart/modeling_bart.py
src/transformers/models/bart/modeling_bart.py
+1
-1
src/transformers/models/bert/modeling_bert.py
src/transformers/models/bert/modeling_bert.py
+1
-1
src/transformers/models/bert_generation/modeling_bert_generation.py
...ormers/models/bert_generation/modeling_bert_generation.py
+1
-1
src/transformers/models/big_bird/modeling_big_bird.py
src/transformers/models/big_bird/modeling_big_bird.py
+1
-1
src/transformers/models/bigbird_pegasus/modeling_bigbird_pegasus.py
...ormers/models/bigbird_pegasus/modeling_bigbird_pegasus.py
+1
-1
src/transformers/models/biogpt/modeling_biogpt.py
src/transformers/models/biogpt/modeling_biogpt.py
+1
-1
src/transformers/models/blenderbot/modeling_blenderbot.py
src/transformers/models/blenderbot/modeling_blenderbot.py
+1
-1
src/transformers/models/blenderbot_small/modeling_blenderbot_small.py
...mers/models/blenderbot_small/modeling_blenderbot_small.py
+1
-1
src/transformers/models/bloom/modeling_bloom.py
src/transformers/models/bloom/modeling_bloom.py
+1
-1
src/transformers/models/bridgetower/modeling_bridgetower.py
src/transformers/models/bridgetower/modeling_bridgetower.py
+1
-1
src/transformers/models/camembert/modeling_camembert.py
src/transformers/models/camembert/modeling_camembert.py
+1
-1
src/transformers/models/chinese_clip/modeling_chinese_clip.py
...transformers/models/chinese_clip/modeling_chinese_clip.py
+1
-1
src/transformers/models/clap/modeling_clap.py
src/transformers/models/clap/modeling_clap.py
+1
-1
src/transformers/models/codegen/modeling_codegen.py
src/transformers/models/codegen/modeling_codegen.py
+1
-1
src/transformers/models/data2vec/modeling_data2vec_text.py
src/transformers/models/data2vec/modeling_data2vec_text.py
+1
-1
src/transformers/models/decision_transformer/modeling_decision_transformer.py
...els/decision_transformer/modeling_decision_transformer.py
+1
-1
src/transformers/models/electra/modeling_electra.py
src/transformers/models/electra/modeling_electra.py
+1
-1
src/transformers/models/ernie/modeling_ernie.py
src/transformers/models/ernie/modeling_ernie.py
+1
-1
src/transformers/models/esm/modeling_esm.py
src/transformers/models/esm/modeling_esm.py
+1
-1
No files found.
src/transformers/models/altclip/modeling_altclip.py
View file @
c7f3abc2
...
@@ -638,7 +638,7 @@ class AltRobertaEncoder(nn.Module):
...
@@ -638,7 +638,7 @@ class AltRobertaEncoder(nn.Module):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/bart/modeling_bart.py
View file @
c7f3abc2
...
@@ -1085,7 +1085,7 @@ class BartDecoder(BartPretrainedModel):
...
@@ -1085,7 +1085,7 @@ class BartDecoder(BartPretrainedModel):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/bert/modeling_bert.py
View file @
c7f3abc2
...
@@ -585,7 +585,7 @@ class BertEncoder(nn.Module):
...
@@ -585,7 +585,7 @@ class BertEncoder(nn.Module):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/bert_generation/modeling_bert_generation.py
View file @
c7f3abc2
...
@@ -395,7 +395,7 @@ class BertEncoder(nn.Module):
...
@@ -395,7 +395,7 @@ class BertEncoder(nn.Module):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/big_bird/modeling_big_bird.py
View file @
c7f3abc2
...
@@ -1606,7 +1606,7 @@ class BigBirdEncoder(nn.Module):
...
@@ -1606,7 +1606,7 @@ class BigBirdEncoder(nn.Module):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/bigbird_pegasus/modeling_bigbird_pegasus.py
View file @
c7f3abc2
...
@@ -2265,7 +2265,7 @@ class BigBirdPegasusDecoder(BigBirdPegasusPreTrainedModel):
...
@@ -2265,7 +2265,7 @@ class BigBirdPegasusDecoder(BigBirdPegasusPreTrainedModel):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/biogpt/modeling_biogpt.py
View file @
c7f3abc2
...
@@ -557,7 +557,7 @@ class BioGptModel(BioGptPreTrainedModel):
...
@@ -557,7 +557,7 @@ class BioGptModel(BioGptPreTrainedModel):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/blenderbot/modeling_blenderbot.py
View file @
c7f3abc2
...
@@ -1016,7 +1016,7 @@ class BlenderbotDecoder(BlenderbotPreTrainedModel):
...
@@ -1016,7 +1016,7 @@ class BlenderbotDecoder(BlenderbotPreTrainedModel):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/blenderbot_small/modeling_blenderbot_small.py
View file @
c7f3abc2
...
@@ -1012,7 +1012,7 @@ class BlenderbotSmallDecoder(BlenderbotSmallPreTrainedModel):
...
@@ -1012,7 +1012,7 @@ class BlenderbotSmallDecoder(BlenderbotSmallPreTrainedModel):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/bloom/modeling_bloom.py
View file @
c7f3abc2
...
@@ -757,7 +757,7 @@ class BloomModel(BloomPreTrainedModel):
...
@@ -757,7 +757,7 @@ class BloomModel(BloomPreTrainedModel):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/bridgetower/modeling_bridgetower.py
View file @
c7f3abc2
...
@@ -769,7 +769,7 @@ class BridgeTowerTextEncoder(nn.Module):
...
@@ -769,7 +769,7 @@ class BridgeTowerTextEncoder(nn.Module):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/camembert/modeling_camembert.py
View file @
c7f3abc2
...
@@ -516,7 +516,7 @@ class CamembertEncoder(nn.Module):
...
@@ -516,7 +516,7 @@ class CamembertEncoder(nn.Module):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/chinese_clip/modeling_chinese_clip.py
View file @
c7f3abc2
...
@@ -901,7 +901,7 @@ class ChineseCLIPTextEncoder(nn.Module):
...
@@ -901,7 +901,7 @@ class ChineseCLIPTextEncoder(nn.Module):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/clap/modeling_clap.py
View file @
c7f3abc2
...
@@ -1588,7 +1588,7 @@ class ClapTextEncoder(nn.Module):
...
@@ -1588,7 +1588,7 @@ class ClapTextEncoder(nn.Module):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/codegen/modeling_codegen.py
View file @
c7f3abc2
...
@@ -548,7 +548,7 @@ class CodeGenModel(CodeGenPreTrainedModel):
...
@@ -548,7 +548,7 @@ class CodeGenModel(CodeGenPreTrainedModel):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with `config.gradient_checkpointing=True`. Setting "
"`use_cache=True` is incompatible with `config.gradient_checkpointing=True`. Setting "
"`use_cache=False`..."
"`use_cache=False`..."
)
)
...
...
src/transformers/models/data2vec/modeling_data2vec_text.py
View file @
c7f3abc2
...
@@ -502,7 +502,7 @@ class Data2VecTextEncoder(nn.Module):
...
@@ -502,7 +502,7 @@ class Data2VecTextEncoder(nn.Module):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/decision_transformer/modeling_decision_transformer.py
View file @
c7f3abc2
...
@@ -609,7 +609,7 @@ class DecisionTransformerGPT2Model(DecisionTransformerGPT2PreTrainedModel):
...
@@ -609,7 +609,7 @@ class DecisionTransformerGPT2Model(DecisionTransformerGPT2PreTrainedModel):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/electra/modeling_electra.py
View file @
c7f3abc2
...
@@ -563,7 +563,7 @@ class ElectraEncoder(nn.Module):
...
@@ -563,7 +563,7 @@ class ElectraEncoder(nn.Module):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/ernie/modeling_ernie.py
View file @
c7f3abc2
...
@@ -498,7 +498,7 @@ class ErnieEncoder(nn.Module):
...
@@ -498,7 +498,7 @@ class ErnieEncoder(nn.Module):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
)
)
use_cache
=
False
use_cache
=
False
...
...
src/transformers/models/esm/modeling_esm.py
View file @
c7f3abc2
...
@@ -597,7 +597,7 @@ class EsmEncoder(nn.Module):
...
@@ -597,7 +597,7 @@ class EsmEncoder(nn.Module):
if
self
.
gradient_checkpointing
and
self
.
training
:
if
self
.
gradient_checkpointing
and
self
.
training
:
if
use_cache
:
if
use_cache
:
logger
.
warning
(
logger
.
warning
_once
(
"`use_cache=True` is incompatible with `config.gradient_checkpointing=True`. Setting "
"`use_cache=True` is incompatible with `config.gradient_checkpointing=True`. Setting "
"`use_cache=False`..."
"`use_cache=False`..."
)
)
...
...
Prev
1
2
3
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment