Unverified Commit 4d40109c authored by Susnato Dhar's avatar Susnato Dhar Committed by GitHub
Browse files

Fix typo in `configuration_gpt2.py` (#25676)

Update configuration_gpt2.py
parent 3c2383b1
......@@ -60,7 +60,7 @@ class GPT2Config(PretrainedConfig):
Number of attention heads for each attention layer in the Transformer encoder.
n_inner (`int`, *optional*, defaults to None):
Dimensionality of the inner feed-forward layers. `None` will set it to 4 times n_embd
activation_function (`str`, *optional*, defaults to `"gelu"`):
activation_function (`str`, *optional*, defaults to `"gelu_new"`):
Activation function, to be selected in the list `["relu", "silu", "gelu", "tanh", "gelu_new"]`.
resid_pdrop (`float`, *optional*, defaults to 0.1):
The dropout probability for all fully connected layers in the embeddings, encoder, and pooler.
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment