Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
14bb196c
Unverified
Commit
14bb196c
authored
Oct 31, 2023
by
Seungwoo, Jeong
Committed by
GitHub
Oct 31, 2023
Browse files
[doctring] Fix docstring for BlipTextConfig, BlipVisionConfig (#27173)
Update configuration_blip.py edit docstrings
parent
9234caef
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
6 additions
and
6 deletions
+6
-6
src/transformers/models/blip/configuration_blip.py
src/transformers/models/blip/configuration_blip.py
+6
-6
No files found.
src/transformers/models/blip/configuration_blip.py
View file @
14bb196c
...
...
@@ -55,7 +55,7 @@ class BlipTextConfig(PretrainedConfig):
Args:
vocab_size (`int`, *optional*, defaults to 3052
2
):
vocab_size (`int`, *optional*, defaults to 3052
4
):
Vocabulary size of the `Blip` text model. Defines the number of different tokens that can be represented by
the `inputs_ids` passed when calling [`BlipModel`].
hidden_size (`int`, *optional*, defaults to 768):
...
...
@@ -68,7 +68,7 @@ class BlipTextConfig(PretrainedConfig):
Number of hidden layers in the Transformer encoder.
num_attention_heads (`int`, *optional*, defaults to 8):
Number of attention heads for each attention layer in the Transformer encoder.
max_position_embeddings (`int`, *optional*, defaults to
77
):
max_position_embeddings (`int`, *optional*, defaults to
512
):
The maximum sequence length that this model might ever be used with. Typically set this to something large
just in case (e.g., 512 or 1024 or 2048).
hidden_act (`str` or `function`, *optional*, defaults to `"gelu"`):
...
...
@@ -90,7 +90,7 @@ class BlipTextConfig(PretrainedConfig):
The id of the `padding` token.
sep_token_id (`int`, *optional*, defaults to 102):
The id of the `separator` token.
is_decoder (`bool`, *optional*, defaults to `
Fals
e`):
is_decoder (`bool`, *optional*, defaults to `
Tru
e`):
Whether the model is used as a decoder.
use_cache (`bool`, *optional*, defaults to `True`):
Whether or not the model should return the last key/values attentions (not used by all models).
...
...
@@ -197,9 +197,9 @@ class BlipVisionConfig(PretrainedConfig):
Number of hidden layers in the Transformer encoder.
num_attention_heads (`int`, *optional*, defaults to 12):
Number of attention heads for each attention layer in the Transformer encoder.
image_size (`int`, *optional*, defaults to
22
4):
image_size (`int`, *optional*, defaults to
38
4):
The size (resolution) of each image.
patch_size (`int`, *optional*, defaults to
32
):
patch_size (`int`, *optional*, defaults to
16
):
The size (resolution) of each patch.
hidden_act (`str` or `function`, *optional*, defaults to `"gelu"`):
The non-linear activation function (function or string) in the encoder and pooler. If string, `"gelu"`,
...
...
@@ -208,7 +208,7 @@ class BlipVisionConfig(PretrainedConfig):
The epsilon used by the layer normalization layers.
attention_dropout (`float`, *optional*, defaults to 0.0):
The dropout ratio for the attention probabilities.
initializer_range (`float`, *optional*, defaults to
0.02
):
initializer_range (`float`, *optional*, defaults to
1e-10
):
The standard deviation of the truncated_normal_initializer for initializing all weight matrices.
Example:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment