Unverified Commit 9ffbed26 authored by Yih-Dar's avatar Yih-Dar Committed by GitHub
Browse files

Cleanup some config attributes (#20554)



* Remove is_encoder_decoder from some vision models

* cleanup more

* cleanup more
Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
parent e1782653
...@@ -129,7 +129,6 @@ class BeitConfig(PretrainedConfig): ...@@ -129,7 +129,6 @@ class BeitConfig(PretrainedConfig):
attention_probs_dropout_prob=0.0, attention_probs_dropout_prob=0.0,
initializer_range=0.02, initializer_range=0.02,
layer_norm_eps=1e-12, layer_norm_eps=1e-12,
is_encoder_decoder=False,
image_size=224, image_size=224,
patch_size=16, patch_size=16,
num_channels=3, num_channels=3,
......
...@@ -119,7 +119,6 @@ class BigBirdConfig(PretrainedConfig): ...@@ -119,7 +119,6 @@ class BigBirdConfig(PretrainedConfig):
initializer_range=0.02, initializer_range=0.02,
layer_norm_eps=1e-12, layer_norm_eps=1e-12,
use_cache=True, use_cache=True,
is_encoder_decoder=False,
pad_token_id=0, pad_token_id=0,
bos_token_id=1, bos_token_id=1,
eos_token_id=2, eos_token_id=2,
...@@ -153,7 +152,6 @@ class BigBirdConfig(PretrainedConfig): ...@@ -153,7 +152,6 @@ class BigBirdConfig(PretrainedConfig):
self.type_vocab_size = type_vocab_size self.type_vocab_size = type_vocab_size
self.layer_norm_eps = layer_norm_eps self.layer_norm_eps = layer_norm_eps
self.use_cache = use_cache self.use_cache = use_cache
self.is_encoder_decoder = is_encoder_decoder
self.rescale_embeddings = rescale_embeddings self.rescale_embeddings = rescale_embeddings
self.attention_type = attention_type self.attention_type = attention_type
......
...@@ -105,7 +105,6 @@ class CanineConfig(PretrainedConfig): ...@@ -105,7 +105,6 @@ class CanineConfig(PretrainedConfig):
initializer_range=0.02, initializer_range=0.02,
layer_norm_eps=1e-12, layer_norm_eps=1e-12,
use_cache=True, use_cache=True,
is_encoder_decoder=False,
pad_token_id=0, pad_token_id=0,
bos_token_id=0xE000, bos_token_id=0xE000,
eos_token_id=0xE001, eos_token_id=0xE001,
......
...@@ -102,7 +102,6 @@ class ConvBertConfig(PretrainedConfig): ...@@ -102,7 +102,6 @@ class ConvBertConfig(PretrainedConfig):
self, self,
vocab_size=30522, vocab_size=30522,
hidden_size=768, hidden_size=768,
is_encoder_decoder=False,
num_hidden_layers=12, num_hidden_layers=12,
num_attention_heads=12, num_attention_heads=12,
intermediate_size=3072, intermediate_size=3072,
...@@ -125,7 +124,6 @@ class ConvBertConfig(PretrainedConfig): ...@@ -125,7 +124,6 @@ class ConvBertConfig(PretrainedConfig):
): ):
super().__init__( super().__init__(
pad_token_id=pad_token_id, pad_token_id=pad_token_id,
is_encoder_decoder=is_encoder_decoder,
bos_token_id=bos_token_id, bos_token_id=bos_token_id,
eos_token_id=eos_token_id, eos_token_id=eos_token_id,
**kwargs, **kwargs,
......
...@@ -90,7 +90,6 @@ class ConvNextConfig(PretrainedConfig): ...@@ -90,7 +90,6 @@ class ConvNextConfig(PretrainedConfig):
hidden_act="gelu", hidden_act="gelu",
initializer_range=0.02, initializer_range=0.02,
layer_norm_eps=1e-12, layer_norm_eps=1e-12,
is_encoder_decoder=False,
layer_scale_init_value=1e-6, layer_scale_init_value=1e-6,
drop_path_rate=0.0, drop_path_rate=0.0,
image_size=224, image_size=224,
......
...@@ -128,7 +128,6 @@ class Data2VecVisionConfig(PretrainedConfig): ...@@ -128,7 +128,6 @@ class Data2VecVisionConfig(PretrainedConfig):
attention_probs_dropout_prob=0.0, attention_probs_dropout_prob=0.0,
initializer_range=0.02, initializer_range=0.02,
layer_norm_eps=1e-12, layer_norm_eps=1e-12,
is_encoder_decoder=False,
image_size=224, image_size=224,
patch_size=16, patch_size=16,
num_channels=3, num_channels=3,
......
...@@ -104,7 +104,6 @@ class DeiTConfig(PretrainedConfig): ...@@ -104,7 +104,6 @@ class DeiTConfig(PretrainedConfig):
attention_probs_dropout_prob=0.0, attention_probs_dropout_prob=0.0,
initializer_range=0.02, initializer_range=0.02,
layer_norm_eps=1e-12, layer_norm_eps=1e-12,
is_encoder_decoder=False,
image_size=224, image_size=224,
patch_size=16, patch_size=16,
num_channels=3, num_channels=3,
......
...@@ -109,7 +109,6 @@ class GLPNConfig(PretrainedConfig): ...@@ -109,7 +109,6 @@ class GLPNConfig(PretrainedConfig):
initializer_range=0.02, initializer_range=0.02,
drop_path_rate=0.1, drop_path_rate=0.1,
layer_norm_eps=1e-6, layer_norm_eps=1e-6,
is_encoder_decoder=False,
decoder_hidden_size=64, decoder_hidden_size=64,
max_depth=10, max_depth=10,
head_in_index=-1, head_in_index=-1,
......
...@@ -136,7 +136,6 @@ class PerceiverConfig(PretrainedConfig): ...@@ -136,7 +136,6 @@ class PerceiverConfig(PretrainedConfig):
position_embedding_init_scale=0.02, position_embedding_init_scale=0.02,
initializer_range=0.02, initializer_range=0.02,
layer_norm_eps=1e-12, layer_norm_eps=1e-12,
is_encoder_decoder=False,
use_query_residual=True, use_query_residual=True,
vocab_size=262, vocab_size=262,
max_position_embeddings=2048, max_position_embeddings=2048,
......
...@@ -100,7 +100,6 @@ class QDQBertConfig(PretrainedConfig): ...@@ -100,7 +100,6 @@ class QDQBertConfig(PretrainedConfig):
initializer_range=0.02, initializer_range=0.02,
layer_norm_eps=1e-12, layer_norm_eps=1e-12,
use_cache=True, use_cache=True,
is_encoder_decoder=False,
pad_token_id=1, pad_token_id=1,
bos_token_id=0, bos_token_id=0,
eos_token_id=2, eos_token_id=2,
......
...@@ -111,7 +111,6 @@ class RemBertConfig(PretrainedConfig): ...@@ -111,7 +111,6 @@ class RemBertConfig(PretrainedConfig):
initializer_range=0.02, initializer_range=0.02,
layer_norm_eps=1e-12, layer_norm_eps=1e-12,
use_cache=True, use_cache=True,
is_encoder_decoder=False,
pad_token_id=0, pad_token_id=0,
bos_token_id=312, bos_token_id=312,
eos_token_id=313, eos_token_id=313,
......
...@@ -121,7 +121,6 @@ class SegformerConfig(PretrainedConfig): ...@@ -121,7 +121,6 @@ class SegformerConfig(PretrainedConfig):
drop_path_rate=0.1, drop_path_rate=0.1,
layer_norm_eps=1e-6, layer_norm_eps=1e-6,
decoder_hidden_size=256, decoder_hidden_size=256,
is_encoder_decoder=False,
semantic_loss_ignore_index=255, semantic_loss_ignore_index=255,
**kwargs **kwargs
): ):
......
...@@ -136,7 +136,6 @@ class TrajectoryTransformerConfig(PretrainedConfig): ...@@ -136,7 +136,6 @@ class TrajectoryTransformerConfig(PretrainedConfig):
layer_norm_eps=1e-12, layer_norm_eps=1e-12,
kaiming_initializer_range=1, kaiming_initializer_range=1,
use_cache=True, use_cache=True,
is_encoder_decoder=False,
pad_token_id=1, pad_token_id=1,
bos_token_id=50256, bos_token_id=50256,
eos_token_id=50256, eos_token_id=50256,
......
...@@ -113,7 +113,6 @@ class ViltConfig(PretrainedConfig): ...@@ -113,7 +113,6 @@ class ViltConfig(PretrainedConfig):
attention_probs_dropout_prob=0.0, attention_probs_dropout_prob=0.0,
initializer_range=0.02, initializer_range=0.02,
layer_norm_eps=1e-12, layer_norm_eps=1e-12,
is_encoder_decoder=False,
image_size=384, image_size=384,
patch_size=32, patch_size=32,
num_channels=3, num_channels=3,
......
...@@ -101,7 +101,6 @@ class ViTConfig(PretrainedConfig): ...@@ -101,7 +101,6 @@ class ViTConfig(PretrainedConfig):
attention_probs_dropout_prob=0.0, attention_probs_dropout_prob=0.0,
initializer_range=0.02, initializer_range=0.02,
layer_norm_eps=1e-12, layer_norm_eps=1e-12,
is_encoder_decoder=False,
image_size=224, image_size=224,
patch_size=16, patch_size=16,
num_channels=3, num_channels=3,
......
...@@ -106,7 +106,6 @@ class ViTMAEConfig(PretrainedConfig): ...@@ -106,7 +106,6 @@ class ViTMAEConfig(PretrainedConfig):
attention_probs_dropout_prob=0.0, attention_probs_dropout_prob=0.0,
initializer_range=0.02, initializer_range=0.02,
layer_norm_eps=1e-12, layer_norm_eps=1e-12,
is_encoder_decoder=False,
image_size=224, image_size=224,
patch_size=16, patch_size=16,
num_channels=3, num_channels=3,
......
...@@ -163,7 +163,6 @@ class {{cookiecutter.camelcase_modelname}}Config(PretrainedConfig): ...@@ -163,7 +163,6 @@ class {{cookiecutter.camelcase_modelname}}Config(PretrainedConfig):
initializer_range=0.02, initializer_range=0.02,
layer_norm_eps=1e-12, layer_norm_eps=1e-12,
use_cache=True, use_cache=True,
is_encoder_decoder=False,
{% else -%} {% else -%}
vocab_size=50265, vocab_size=50265,
max_position_embeddings=1024, max_position_embeddings=1024,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment