Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
ModelZoo
ResNet50_tensorflow
Commits
56761bcb
Commit
56761bcb
authored
May 31, 2022
by
Jaehong Kim
Committed by
A. Unique TensorFlower
May 31, 2022
Browse files
Add adamW experimental optimzier.
PiperOrigin-RevId: 452047178
parent
9985d123
Changes
3
Show whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
29 additions
and
0 deletions
+29
-0
official/modeling/optimization/configs/optimization_config.py
...cial/modeling/optimization/configs/optimization_config.py
+2
-0
official/modeling/optimization/configs/optimizer_config.py
official/modeling/optimization/configs/optimizer_config.py
+26
-0
official/modeling/optimization/optimizer_factory.py
official/modeling/optimization/optimizer_factory.py
+1
-0
No files found.
official/modeling/optimization/configs/optimization_config.py
View file @
56761bcb
...
@@ -51,6 +51,8 @@ class OptimizerConfig(oneof.OneOfConfig):
...
@@ -51,6 +51,8 @@ class OptimizerConfig(oneof.OneOfConfig):
adam_experimental
:
opt_cfg
.
AdamExperimentalConfig
=
(
adam_experimental
:
opt_cfg
.
AdamExperimentalConfig
=
(
opt_cfg
.
AdamExperimentalConfig
())
opt_cfg
.
AdamExperimentalConfig
())
adamw
:
opt_cfg
.
AdamWeightDecayConfig
=
opt_cfg
.
AdamWeightDecayConfig
()
adamw
:
opt_cfg
.
AdamWeightDecayConfig
=
opt_cfg
.
AdamWeightDecayConfig
()
adamw_experimental
:
opt_cfg
.
AdamWeightDecayExperimentalConfig
=
(
opt_cfg
.
AdamWeightDecayExperimentalConfig
())
lamb
:
opt_cfg
.
LAMBConfig
=
opt_cfg
.
LAMBConfig
()
lamb
:
opt_cfg
.
LAMBConfig
=
opt_cfg
.
LAMBConfig
()
rmsprop
:
opt_cfg
.
RMSPropConfig
=
opt_cfg
.
RMSPropConfig
()
rmsprop
:
opt_cfg
.
RMSPropConfig
=
opt_cfg
.
RMSPropConfig
()
lars
:
opt_cfg
.
LARSConfig
=
opt_cfg
.
LARSConfig
()
lars
:
opt_cfg
.
LARSConfig
=
opt_cfg
.
LARSConfig
()
...
...
official/modeling/optimization/configs/optimizer_config.py
View file @
56761bcb
...
@@ -190,6 +190,32 @@ class AdamWeightDecayConfig(BaseOptimizerConfig):
...
@@ -190,6 +190,32 @@ class AdamWeightDecayConfig(BaseOptimizerConfig):
gradient_clip_norm
:
float
=
1.0
gradient_clip_norm
:
float
=
1.0
@
dataclasses
.
dataclass
class
AdamWeightDecayExperimentalConfig
(
BaseOptimizerConfig
):
"""Configuration for Adam optimizer with weight decay.
Attributes:
name: name of the optimizer.
beta_1: decay rate for 1st order moments.
beta_2: decay rate for 2st order moments.
epsilon: epsilon value used for numerical stability in the optimizer.
amsgrad: boolean. Whether to apply AMSGrad variant of this algorithm from
the paper "On the Convergence of Adam and beyond".
weight_decay: float. Weight decay rate. Default to 0.
global_clipnorm: A positive float. Clips the gradients to this maximum
L2-norm. Default to 1.0.
jit_compile: if True, jit compile will be used.
"""
name
:
str
=
"AdamWeightDecayExperimental"
beta_1
:
float
=
0.9
beta_2
:
float
=
0.999
epsilon
:
float
=
1e-07
amsgrad
:
bool
=
False
weight_decay
:
float
=
0.0
global_clipnorm
:
float
=
1.0
jit_compile
:
bool
=
False
@
dataclasses
.
dataclass
@
dataclasses
.
dataclass
class
LAMBConfig
(
BaseOptimizerConfig
):
class
LAMBConfig
(
BaseOptimizerConfig
):
"""Configuration for LAMB optimizer.
"""Configuration for LAMB optimizer.
...
...
official/modeling/optimization/optimizer_factory.py
View file @
56761bcb
...
@@ -33,6 +33,7 @@ OPTIMIZERS_CLS = {
...
@@ -33,6 +33,7 @@ OPTIMIZERS_CLS = {
'adam'
:
tf
.
keras
.
optimizers
.
Adam
,
'adam'
:
tf
.
keras
.
optimizers
.
Adam
,
# TODO(chenmoneygithub): experimental.Adam
# TODO(chenmoneygithub): experimental.Adam
'adamw'
:
legacy_adamw
.
AdamWeightDecay
,
'adamw'
:
legacy_adamw
.
AdamWeightDecay
,
'adamw_experimental'
:
tf
.
keras
.
optimizers
.
experimental
.
AdamW
,
'lamb'
:
tfa_optimizers
.
LAMB
,
'lamb'
:
tfa_optimizers
.
LAMB
,
'rmsprop'
:
tf
.
keras
.
optimizers
.
RMSprop
,
'rmsprop'
:
tf
.
keras
.
optimizers
.
RMSprop
,
'lars'
:
lars_optimizer
.
LARS
,
'lars'
:
lars_optimizer
.
LARS
,
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment