Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
9acce7de
Unverified
Commit
9acce7de
authored
Mar 13, 2024
by
Younes Belkada
Committed by
GitHub
Mar 13, 2024
Browse files
Core: Fix copies on main (#29624)
fix fix copies
parent
be3fd8a2
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
1 addition
and
1 deletion
+1
-1
src/transformers/models/gptj/modeling_gptj.py
src/transformers/models/gptj/modeling_gptj.py
+1
-1
No files found.
src/transformers/models/gptj/modeling_gptj.py
View file @
9acce7de
...
@@ -454,7 +454,7 @@ class GPTJFlashAttention2(GPTJAttention):
...
@@ -454,7 +454,7 @@ class GPTJFlashAttention2(GPTJAttention):
attention_mask (`torch.Tensor`):
attention_mask (`torch.Tensor`):
The padding mask - corresponds to a tensor of size `(batch_size, seq_len)` where 0 stands for the
The padding mask - corresponds to a tensor of size `(batch_size, seq_len)` where 0 stands for the
position of padding tokens and 1 for the position of non-padding tokens.
position of padding tokens and 1 for the position of non-padding tokens.
dropout (`
int`, *optional*
):
dropout (`
float`
):
Attention dropout
Attention dropout
softmax_scale (`float`, *optional*):
softmax_scale (`float`, *optional*):
The scaling of QK^T before applying softmax. Default to 1 / sqrt(head_dim)
The scaling of QK^T before applying softmax. Default to 1 / sqrt(head_dim)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment