Unverified Commit 51042ae8 authored by louietouie's avatar louietouie Committed by GitHub
Browse files

[docstring] Fix docstring for LukeConfig (#26858)



* Deleted LukeConfig and ran check_docstrings.py

* Filled docstring information

---------
Co-authored-by: default avatarlouie <louisparizeau@Chicken.local>
parent db611aab
......@@ -38,7 +38,7 @@ class LukeConfig(PretrainedConfig):
Args:
vocab_size (`int`, *optional*, defaults to 30522):
vocab_size (`int`, *optional*, defaults to 50267):
Vocabulary size of the LUKE model. Defines the number of different tokens that can be represented by the
`inputs_ids` passed when calling [`LukeModel`].
entity_vocab_size (`int`, *optional*, defaults to 500000):
......@@ -70,12 +70,18 @@ class LukeConfig(PretrainedConfig):
The standard deviation of the truncated_normal_initializer for initializing all weight matrices.
layer_norm_eps (`float`, *optional*, defaults to 1e-12):
The epsilon used by the layer normalization layers.
use_entity_aware_attention (`bool`, defaults to `True`):
use_entity_aware_attention (`bool`, *optional*, defaults to `True`):
Whether or not the model should use the entity-aware self-attention mechanism proposed in [LUKE: Deep
Contextualized Entity Representations with Entity-aware Self-attention (Yamada et
al.)](https://arxiv.org/abs/2010.01057).
classifier_dropout (`float`, *optional*):
The dropout ratio for the classification head.
pad_token_id (`int`, *optional*, defaults to 1):
Padding token id.
bos_token_id (`int`, *optional*, defaults to 0):
Beginning of stream token id.
eos_token_id (`int`, *optional*, defaults to 2):
End of stream token id.
Examples:
......
......@@ -356,7 +356,6 @@ OBJECTS_TO_IGNORE = [
"LongformerConfig",
"LongformerModel",
"LongformerTokenizerFast",
"LukeConfig",
"LukeModel",
"LukeTokenizer",
"LxmertTokenizerFast",
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment