Unverified Commit 58022e41 authored by Vijeth Moudgalya's avatar Vijeth Moudgalya Committed by GitHub
Browse files

#23388 Issue: Update RoBERTa configuration (#23863)

parent 6fc0454b
...@@ -46,7 +46,7 @@ class RobertaConfig(PretrainedConfig): ...@@ -46,7 +46,7 @@ class RobertaConfig(PretrainedConfig):
Args: Args:
vocab_size (`int`, *optional*, defaults to 30522): vocab_size (`int`, *optional*, defaults to 50265):
Vocabulary size of the RoBERTa model. Defines the number of different tokens that can be represented by the Vocabulary size of the RoBERTa model. Defines the number of different tokens that can be represented by the
`inputs_ids` passed when calling [`RobertaModel`] or [`TFRobertaModel`]. `inputs_ids` passed when calling [`RobertaModel`] or [`TFRobertaModel`].
hidden_size (`int`, *optional*, defaults to 768): hidden_size (`int`, *optional*, defaults to 768):
...@@ -105,7 +105,7 @@ class RobertaConfig(PretrainedConfig): ...@@ -105,7 +105,7 @@ class RobertaConfig(PretrainedConfig):
def __init__( def __init__(
self, self,
vocab_size=30522, vocab_size=50265,
hidden_size=768, hidden_size=768,
num_hidden_layers=12, num_hidden_layers=12,
num_attention_heads=12, num_attention_heads=12,
......
...@@ -45,7 +45,7 @@ class RobertaPreLayerNormConfig(PretrainedConfig): ...@@ -45,7 +45,7 @@ class RobertaPreLayerNormConfig(PretrainedConfig):
Args: Args:
vocab_size (`int`, *optional*, defaults to 30522): vocab_size (`int`, *optional*, defaults to 50265):
Vocabulary size of the RoBERTa-PreLayerNorm model. Defines the number of different tokens that can be Vocabulary size of the RoBERTa-PreLayerNorm model. Defines the number of different tokens that can be
represented by the `inputs_ids` passed when calling [`RobertaPreLayerNormModel`] or represented by the `inputs_ids` passed when calling [`RobertaPreLayerNormModel`] or
[`TFRobertaPreLayerNormModel`]. [`TFRobertaPreLayerNormModel`].
...@@ -106,7 +106,7 @@ class RobertaPreLayerNormConfig(PretrainedConfig): ...@@ -106,7 +106,7 @@ class RobertaPreLayerNormConfig(PretrainedConfig):
def __init__( def __init__(
self, self,
vocab_size=30522, vocab_size=50265,
hidden_size=768, hidden_size=768,
num_hidden_layers=12, num_hidden_layers=12,
num_attention_heads=12, num_attention_heads=12,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment