"...git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "893122f6662acb339c7e1a014834b2670ae00e0b"
Unverified Commit 4d40109c authored by Susnato Dhar's avatar Susnato Dhar Committed by GitHub
Browse files

Fix typo in `configuration_gpt2.py` (#25676)

Update configuration_gpt2.py
parent 3c2383b1
...@@ -60,7 +60,7 @@ class GPT2Config(PretrainedConfig): ...@@ -60,7 +60,7 @@ class GPT2Config(PretrainedConfig):
Number of attention heads for each attention layer in the Transformer encoder. Number of attention heads for each attention layer in the Transformer encoder.
n_inner (`int`, *optional*, defaults to None): n_inner (`int`, *optional*, defaults to None):
Dimensionality of the inner feed-forward layers. `None` will set it to 4 times n_embd Dimensionality of the inner feed-forward layers. `None` will set it to 4 times n_embd
activation_function (`str`, *optional*, defaults to `"gelu"`): activation_function (`str`, *optional*, defaults to `"gelu_new"`):
Activation function, to be selected in the list `["relu", "silu", "gelu", "tanh", "gelu_new"]`. Activation function, to be selected in the list `["relu", "silu", "gelu", "tanh", "gelu_new"]`.
resid_pdrop (`float`, *optional*, defaults to 0.1): resid_pdrop (`float`, *optional*, defaults to 0.1):
The dropout probability for all fully connected layers in the embeddings, encoder, and pooler. The dropout probability for all fully connected layers in the embeddings, encoder, and pooler.
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment