Unverified Commit 58022e41 authored by Vijeth Moudgalya's avatar Vijeth Moudgalya Committed by GitHub
Browse files

#23388 Issue: Update RoBERTa configuration (#23863)

parent 6fc0454b
......@@ -46,7 +46,7 @@ class RobertaConfig(PretrainedConfig):
Args:
vocab_size (`int`, *optional*, defaults to 30522):
vocab_size (`int`, *optional*, defaults to 50265):
Vocabulary size of the RoBERTa model. Defines the number of different tokens that can be represented by the
`inputs_ids` passed when calling [`RobertaModel`] or [`TFRobertaModel`].
hidden_size (`int`, *optional*, defaults to 768):
......@@ -105,7 +105,7 @@ class RobertaConfig(PretrainedConfig):
def __init__(
self,
vocab_size=30522,
vocab_size=50265,
hidden_size=768,
num_hidden_layers=12,
num_attention_heads=12,
......
......@@ -45,7 +45,7 @@ class RobertaPreLayerNormConfig(PretrainedConfig):
Args:
vocab_size (`int`, *optional*, defaults to 30522):
vocab_size (`int`, *optional*, defaults to 50265):
Vocabulary size of the RoBERTa-PreLayerNorm model. Defines the number of different tokens that can be
represented by the `inputs_ids` passed when calling [`RobertaPreLayerNormModel`] or
[`TFRobertaPreLayerNormModel`].
......@@ -106,7 +106,7 @@ class RobertaPreLayerNormConfig(PretrainedConfig):
def __init__(
self,
vocab_size=30522,
vocab_size=50265,
hidden_size=768,
num_hidden_layers=12,
num_attention_heads=12,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment