Unverified Commit 51042ae8 authored by louietouie's avatar louietouie Committed by GitHub
Browse files

[docstring] Fix docstring for LukeConfig (#26858)



* Deleted LukeConfig and ran check_docstrings.py

* Filled docstring information

---------
Co-authored-by: default avatarlouie <louisparizeau@Chicken.local>
parent db611aab
...@@ -38,7 +38,7 @@ class LukeConfig(PretrainedConfig): ...@@ -38,7 +38,7 @@ class LukeConfig(PretrainedConfig):
Args: Args:
vocab_size (`int`, *optional*, defaults to 30522): vocab_size (`int`, *optional*, defaults to 50267):
Vocabulary size of the LUKE model. Defines the number of different tokens that can be represented by the Vocabulary size of the LUKE model. Defines the number of different tokens that can be represented by the
`inputs_ids` passed when calling [`LukeModel`]. `inputs_ids` passed when calling [`LukeModel`].
entity_vocab_size (`int`, *optional*, defaults to 500000): entity_vocab_size (`int`, *optional*, defaults to 500000):
...@@ -70,12 +70,18 @@ class LukeConfig(PretrainedConfig): ...@@ -70,12 +70,18 @@ class LukeConfig(PretrainedConfig):
The standard deviation of the truncated_normal_initializer for initializing all weight matrices. The standard deviation of the truncated_normal_initializer for initializing all weight matrices.
layer_norm_eps (`float`, *optional*, defaults to 1e-12): layer_norm_eps (`float`, *optional*, defaults to 1e-12):
The epsilon used by the layer normalization layers. The epsilon used by the layer normalization layers.
use_entity_aware_attention (`bool`, defaults to `True`): use_entity_aware_attention (`bool`, *optional*, defaults to `True`):
Whether or not the model should use the entity-aware self-attention mechanism proposed in [LUKE: Deep Whether or not the model should use the entity-aware self-attention mechanism proposed in [LUKE: Deep
Contextualized Entity Representations with Entity-aware Self-attention (Yamada et Contextualized Entity Representations with Entity-aware Self-attention (Yamada et
al.)](https://arxiv.org/abs/2010.01057). al.)](https://arxiv.org/abs/2010.01057).
classifier_dropout (`float`, *optional*): classifier_dropout (`float`, *optional*):
The dropout ratio for the classification head. The dropout ratio for the classification head.
pad_token_id (`int`, *optional*, defaults to 1):
Padding token id.
bos_token_id (`int`, *optional*, defaults to 0):
Beginning of stream token id.
eos_token_id (`int`, *optional*, defaults to 2):
End of stream token id.
Examples: Examples:
......
...@@ -356,7 +356,6 @@ OBJECTS_TO_IGNORE = [ ...@@ -356,7 +356,6 @@ OBJECTS_TO_IGNORE = [
"LongformerConfig", "LongformerConfig",
"LongformerModel", "LongformerModel",
"LongformerTokenizerFast", "LongformerTokenizerFast",
"LukeConfig",
"LukeModel", "LukeModel",
"LukeTokenizer", "LukeTokenizer",
"LxmertTokenizerFast", "LxmertTokenizerFast",
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment