Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
63e91f5f
"src/vscode:/vscode.git/clone" did not exist on "36a2b2f3fe06dfdbdd92f286938b6235815e73ff"
Unverified
Commit
63e91f5f
authored
Nov 20, 2020
by
Sylvain Gugger
Committed by
GitHub
Nov 20, 2020
Browse files
Document adam betas TrainingArguments (#8688)
parent
94caaa93
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
4 additions
and
0 deletions
+4
-0
src/transformers/training_args.py
src/transformers/training_args.py
+4
-0
No files found.
src/transformers/training_args.py
View file @
63e91f5f
...
@@ -89,6 +89,10 @@ class TrainingArguments:
...
@@ -89,6 +89,10 @@ class TrainingArguments:
The initial learning rate for Adam.
The initial learning rate for Adam.
weight_decay (:obj:`float`, `optional`, defaults to 0):
weight_decay (:obj:`float`, `optional`, defaults to 0):
The weight decay to apply (if not zero).
The weight decay to apply (if not zero).
adam_beta1 (:obj:`float`, `optional`, defaults to 0.9):
The beta1 for the Adam optimizer.
adam_beta2 (:obj:`float`, `optional`, defaults to 0.999):
The beta2 for the Adam optimizer.
adam_epsilon (:obj:`float`, `optional`, defaults to 1e-8):
adam_epsilon (:obj:`float`, `optional`, defaults to 1e-8):
Epsilon for the Adam optimizer.
Epsilon for the Adam optimizer.
max_grad_norm (:obj:`float`, `optional`, defaults to 1.0):
max_grad_norm (:obj:`float`, `optional`, defaults to 1.0):
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment