Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
40ea9ab2
Unverified
Commit
40ea9ab2
authored
Oct 12, 2023
by
Tom Aarsen
Committed by
GitHub
Oct 12, 2023
Browse files
Add many missing spaces in adjacent strings (#26751)
Add missing spaces in adjacent strings
parent
3bc65505
Changes
154
Show whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
42 additions
and
42 deletions
+42
-42
src/transformers/models/pop2piano/feature_extraction_pop2piano.py
...sformers/models/pop2piano/feature_extraction_pop2piano.py
+2
-2
src/transformers/models/sew/configuration_sew.py
src/transformers/models/sew/configuration_sew.py
+3
-3
src/transformers/models/sew/modeling_sew.py
src/transformers/models/sew/modeling_sew.py
+2
-2
src/transformers/models/sew_d/configuration_sew_d.py
src/transformers/models/sew_d/configuration_sew_d.py
+3
-3
src/transformers/models/sew_d/modeling_sew_d.py
src/transformers/models/sew_d/modeling_sew_d.py
+2
-2
src/transformers/models/speech_to_text_2/tokenization_speech_to_text_2.py
.../models/speech_to_text_2/tokenization_speech_to_text_2.py
+1
-1
src/transformers/models/switch_transformers/modeling_switch_transformers.py
...odels/switch_transformers/modeling_switch_transformers.py
+3
-3
src/transformers/models/t5/configuration_t5.py
src/transformers/models/t5/configuration_t5.py
+1
-1
src/transformers/models/t5/modeling_t5.py
src/transformers/models/t5/modeling_t5.py
+1
-1
src/transformers/models/t5/modeling_tf_t5.py
src/transformers/models/t5/modeling_tf_t5.py
+1
-1
src/transformers/models/tapas/modeling_tf_tapas.py
src/transformers/models/tapas/modeling_tf_tapas.py
+2
-2
src/transformers/models/transfo_xl/modeling_transfo_xl.py
src/transformers/models/transfo_xl/modeling_transfo_xl.py
+1
-1
src/transformers/models/umt5/configuration_umt5.py
src/transformers/models/umt5/configuration_umt5.py
+1
-1
src/transformers/models/umt5/modeling_umt5.py
src/transformers/models/umt5/modeling_umt5.py
+1
-1
src/transformers/models/unispeech/modeling_unispeech.py
src/transformers/models/unispeech/modeling_unispeech.py
+3
-3
src/transformers/models/unispeech_sat/modeling_unispeech_sat.py
...ansformers/models/unispeech_sat/modeling_unispeech_sat.py
+5
-5
src/transformers/models/vision_encoder_decoder/modeling_tf_vision_encoder_decoder.py
...ion_encoder_decoder/modeling_tf_vision_encoder_decoder.py
+1
-1
src/transformers/models/vit/modeling_vit.py
src/transformers/models/vit/modeling_vit.py
+1
-1
src/transformers/models/wav2vec2/modeling_tf_wav2vec2.py
src/transformers/models/wav2vec2/modeling_tf_wav2vec2.py
+2
-2
src/transformers/models/wav2vec2/modeling_wav2vec2.py
src/transformers/models/wav2vec2/modeling_wav2vec2.py
+6
-6
No files found.
src/transformers/models/pop2piano/feature_extraction_pop2piano.py
View file @
40ea9ab2
...
...
@@ -407,9 +407,9 @@ class Pop2PianoFeatureExtractor(SequenceFeatureExtractor):
)
else
:
warnings
.
warn
(
f
"The sampling_rate of the provided audio is different from the target sampling_rate"
f
"The sampling_rate of the provided audio is different from the target sampling_rate
"
f
"of the Feature Extractor,
{
self
.
sampling_rate
}
vs
{
single_sampling_rate
}
. "
f
"In these cases it is recommended to use `resample=True` in the `__call__` method to"
f
"In these cases it is recommended to use `resample=True` in the `__call__` method to
"
f
"get the optimal behaviour."
)
...
...
src/transformers/models/sew/configuration_sew.py
View file @
40ea9ab2
...
...
@@ -229,9 +229,9 @@ class SEWConfig(PretrainedConfig):
or
(
len
(
self
.
conv_dim
)
!=
self
.
num_feat_extract_layers
)
):
raise
ValueError
(
"Configuration for convolutional layers is incorrect."
"It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,"
f
"but is `len(config.conv_dim) =
{
len
(
self
.
conv_dim
)
}
`, `len(config.conv_stride)"
"Configuration for convolutional layers is incorrect.
"
"It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,
"
f
"but is `len(config.conv_dim) =
{
len
(
self
.
conv_dim
)
}
`, `len(config.conv_stride)
"
f
"=
{
len
(
self
.
conv_stride
)
}
`, `len(config.conv_kernel) =
{
len
(
self
.
conv_kernel
)
}
`."
)
...
...
src/transformers/models/sew/modeling_sew.py
View file @
40ea9ab2
...
...
@@ -1018,7 +1018,7 @@ class SEWForCTC(SEWPreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
@@ -1151,7 +1151,7 @@ class SEWForSequenceClassification(SEWPreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
src/transformers/models/sew_d/configuration_sew_d.py
View file @
40ea9ab2
...
...
@@ -256,9 +256,9 @@ class SEWDConfig(PretrainedConfig):
or
(
len
(
self
.
conv_dim
)
!=
self
.
num_feat_extract_layers
)
):
raise
ValueError
(
"Configuration for convolutional layers is incorrect."
"It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,"
f
"but is `len(config.conv_dim) =
{
len
(
self
.
conv_dim
)
}
`, `len(config.conv_stride)"
"Configuration for convolutional layers is incorrect.
"
"It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,
"
f
"but is `len(config.conv_dim) =
{
len
(
self
.
conv_dim
)
}
`, `len(config.conv_stride)
"
f
"=
{
len
(
self
.
conv_stride
)
}
`, `len(config.conv_kernel) =
{
len
(
self
.
conv_kernel
)
}
`."
)
...
...
src/transformers/models/sew_d/modeling_sew_d.py
View file @
40ea9ab2
...
...
@@ -1558,7 +1558,7 @@ class SEWDForCTC(SEWDPreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
@@ -1691,7 +1691,7 @@ class SEWDForSequenceClassification(SEWDPreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
src/transformers/models/speech_to_text_2/tokenization_speech_to_text_2.py
View file @
40ea9ab2
...
...
@@ -199,7 +199,7 @@ class Speech2Text2Tokenizer(PreTrainedTokenizer):
if
self
.
bpe_ranks
is
None
:
raise
ValueError
(
"This tokenizer was instantiated without a `merges.txt` file, so"
" that it can only be used for decoding, not for encoding."
" that it can only be used for decoding, not for encoding.
"
"Make sure to provide `merges.txt` file at instantiation to enable "
"encoding."
)
...
...
src/transformers/models/switch_transformers/modeling_switch_transformers.py
View file @
40ea9ab2
...
...
@@ -1774,13 +1774,13 @@ class SwitchTransformersForConditionalGeneration(SwitchTransformersPreTrainedMod
if
reordered_layer_past_states
[
0
].
shape
!=
layer_past_states
[
0
].
shape
:
raise
ValueError
(
"expected reordered_layer_past_states to have the same shape than layer_past_states"
"expected reordered_layer_past_states to have the same shape than layer_past_states
,
"
f
"but got
{
reordered_layer_past_states
[
0
].
shape
}
and
{
layer_past_states
[
0
].
shape
}
"
)
if
len
(
reordered_layer_past_states
)
!=
len
(
layer_past_states
):
raise
ValueError
(
"expected layer_past_states to have the same length as reordered_layer_past_states"
f
"got
{
len
(
layer_past_states
)
}
and
{
len
(
reordered_layer_past_states
)
}
"
"expected layer_past_states to have the same length as reordered_layer_past_states
,
"
f
"
but
got
{
len
(
layer_past_states
)
}
and
{
len
(
reordered_layer_past_states
)
}
"
)
reordered_decoder_past
=
reordered_decoder_past
+
(
reordered_layer_past_states
,)
...
...
src/transformers/models/t5/configuration_t5.py
View file @
40ea9ab2
...
...
@@ -127,7 +127,7 @@ class T5Config(PretrainedConfig):
if
len
(
act_info
)
>
1
and
act_info
[
0
]
!=
"gated"
or
len
(
act_info
)
>
2
:
raise
ValueError
(
f
"`feed_forward_proj`:
{
feed_forward_proj
}
is not a valid activation function of the dense layer."
f
"`feed_forward_proj`:
{
feed_forward_proj
}
is not a valid activation function of the dense layer.
"
"Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. "
"'gated-gelu' or 'relu'"
)
...
...
src/transformers/models/t5/modeling_t5.py
View file @
40ea9ab2
...
...
@@ -883,7 +883,7 @@ class T5PreTrainedModel(PreTrainedModel):
if
decoder_start_token_id
is
None
:
raise
ValueError
(
"self.model.config.decoder_start_token_id has to be defined. In T5 it is usually set to the pad_token_id."
"self.model.config.decoder_start_token_id has to be defined. In T5 it is usually set to the pad_token_id.
"
"See T5 docs for more information."
)
...
...
src/transformers/models/t5/modeling_tf_t5.py
View file @
40ea9ab2
...
...
@@ -555,7 +555,7 @@ class TFT5Block(tf.keras.layers.Layer):
if
len
(
past_key_value
)
!=
expected_num_past_key_values
:
raise
ValueError
(
f
"There should be
{
expected_num_past_key_values
}
past states. "
f
"
{
'2 (past / key) for cross attention'
if
expected_num_past_key_values
==
4
else
''
}
."
f
"
{
'2 (past / key) for cross attention'
if
expected_num_past_key_values
==
4
else
''
}
.
"
f
"Got
{
len
(
past_key_value
)
}
past key / value states"
)
...
...
src/transformers/models/tapas/modeling_tf_tapas.py
View file @
40ea9ab2
...
...
@@ -66,8 +66,8 @@ if is_tensorflow_probability_available():
n
=
tfp
.
distributions
.
Normal
(
loc
=
0.0
,
scale
=
1.0
)
except
ImportError
:
logger
.
error
(
"TAPAS models are not usable since `tensorflow_probability` can't be loaded."
"It seems you have `tensorflow_probability` installed with the wrong tensorflow version."
"TAPAS models are not usable since `tensorflow_probability` can't be loaded.
"
"It seems you have `tensorflow_probability` installed with the wrong tensorflow version.
"
"Please try to reinstall it following the instructions here: https://github.com/tensorflow/probability."
)
...
...
src/transformers/models/transfo_xl/modeling_transfo_xl.py
View file @
40ea9ab2
...
...
@@ -1012,7 +1012,7 @@ class TransfoXLLMHeadModel(TransfoXLPreTrainedModel):
if
not
self
.
trainer_compatible
:
warnings
.
warn
(
"The output of TransfoXL will be updated in v5 to support a single loss as first argument. In order"
"The output of TransfoXL will be updated in v5 to support a single loss as first argument. In order
"
"to use that updated output, please specify `trainer_compatible=True` as your configuration"
" attribute."
,
DeprecationWarning
,
...
...
src/transformers/models/umt5/configuration_umt5.py
View file @
40ea9ab2
...
...
@@ -134,7 +134,7 @@ class UMT5Config(PretrainedConfig):
if
len
(
act_info
)
>
1
and
act_info
[
0
]
!=
"gated"
or
len
(
act_info
)
>
2
:
raise
ValueError
(
f
"`feed_forward_proj`:
{
feed_forward_proj
}
is not a valid activation function of the dense layer."
f
"`feed_forward_proj`:
{
feed_forward_proj
}
is not a valid activation function of the dense layer.
"
"Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. "
"'gated-gelu' or 'relu'"
)
...
...
src/transformers/models/umt5/modeling_umt5.py
View file @
40ea9ab2
...
...
@@ -566,7 +566,7 @@ class UMT5PreTrainedModel(PreTrainedModel):
if
decoder_start_token_id
is
None
:
raise
ValueError
(
"self.model.config.decoder_start_token_id has to be defined. In UMT5 it is usually set to the pad_token_id."
"self.model.config.decoder_start_token_id has to be defined. In UMT5 it is usually set to the pad_token_id.
"
"See UMT5 docs for more information."
)
...
...
src/transformers/models/unispeech/modeling_unispeech.py
View file @
40ea9ab2
...
...
@@ -1256,7 +1256,7 @@ class UniSpeechForPreTraining(UniSpeechPreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
@@ -1427,7 +1427,7 @@ class UniSpeechForCTC(UniSpeechPreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
@@ -1560,7 +1560,7 @@ class UniSpeechForSequenceClassification(UniSpeechPreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
src/transformers/models/unispeech_sat/modeling_unispeech_sat.py
View file @
40ea9ab2
...
...
@@ -1276,7 +1276,7 @@ class UniSpeechSatForPreTraining(UniSpeechSatPreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
@@ -1434,7 +1434,7 @@ class UniSpeechSatForCTC(UniSpeechSatPreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
@@ -1567,7 +1567,7 @@ class UniSpeechSatForSequenceClassification(UniSpeechSatPreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
@@ -1690,7 +1690,7 @@ class UniSpeechSatForAudioFrameClassification(UniSpeechSatPreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
@@ -1857,7 +1857,7 @@ class UniSpeechSatForXVector(UniSpeechSatPreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
src/transformers/models/vision_encoder_decoder/modeling_tf_vision_encoder_decoder.py
View file @
40ea9ab2
...
...
@@ -709,6 +709,6 @@ class TFVisionEncoderDecoderModel(TFPreTrainedModel, TFCausalLanguageModelingLos
def
resize_token_embeddings
(
self
,
*
args
,
**
kwargs
):
raise
NotImplementedError
(
"Resizing the embedding layers via the TFVisionEncoderDecoderModel directly is not supported."
"Resizing the embedding layers via the TFVisionEncoderDecoderModel directly is not supported.
"
"Please use the respective methods of the wrapped objects (model.decoder.resize_token_embeddings(...))"
)
src/transformers/models/vit/modeling_vit.py
View file @
40ea9ab2
...
...
@@ -698,7 +698,7 @@ class ViTForMaskedImageModeling(ViTPreTrainedModel):
if
bool_masked_pos
is
not
None
and
(
self
.
config
.
patch_size
!=
self
.
config
.
encoder_stride
):
raise
ValueError
(
"When `bool_masked_pos` is provided, `patch_size` must be equal to `encoder_stride` to ensure that "
"the reconstructed image has the same dimensions as the input."
"the reconstructed image has the same dimensions as the input.
"
f
"Got `patch_size` =
{
self
.
config
.
patch_size
}
and `encoder_stride` =
{
self
.
config
.
encoder_stride
}
."
)
...
...
src/transformers/models/wav2vec2/modeling_tf_wav2vec2.py
View file @
40ea9ab2
...
...
@@ -1438,7 +1438,7 @@ class TFWav2Vec2ForCTC(TFWav2Vec2PreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
@@ -1593,7 +1593,7 @@ class TFWav2Vec2ForSequenceClassification(TFWav2Vec2PreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
src/transformers/models/wav2vec2/modeling_wav2vec2.py
View file @
40ea9ab2
...
...
@@ -1480,7 +1480,7 @@ class Wav2Vec2Model(Wav2Vec2PreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
@@ -1627,7 +1627,7 @@ class Wav2Vec2ForPreTraining(Wav2Vec2PreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
@@ -1923,7 +1923,7 @@ class Wav2Vec2ForCTC(Wav2Vec2PreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
@@ -2055,7 +2055,7 @@ class Wav2Vec2ForSequenceClassification(Wav2Vec2PreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
@@ -2176,7 +2176,7 @@ class Wav2Vec2ForAudioFrameClassification(Wav2Vec2PreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
@@ -2340,7 +2340,7 @@ class Wav2Vec2ForXVector(Wav2Vec2PreTrainedModel):
not be updated during training.
"""
warnings
.
warn
(
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5."
"The method `freeze_feature_extractor` is deprecated and will be removed in Transformers v5.
"
"Please use the equivalent `freeze_feature_encoder` method instead."
,
FutureWarning
,
)
...
...
Prev
1
…
3
4
5
6
7
8
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment