Commit 1dc59163 authored by Zachary Nado's avatar Zachary Nado Committed by A. Unique TensorFlower
Browse files

adding `beta_1` parameter to `create_optimizer` in `optimization.py` to allow...

adding `beta_1` parameter to `create_optimizer` in `optimization.py` to allow for a tunable `beta_1`. Defaults to the original `0.9` to ensure existing compatibility.

PiperOrigin-RevId: 365634937
parent 9369b1a3
...@@ -69,7 +69,8 @@ def create_optimizer(init_lr, ...@@ -69,7 +69,8 @@ def create_optimizer(init_lr,
num_train_steps, num_train_steps,
num_warmup_steps, num_warmup_steps,
end_lr=0.0, end_lr=0.0,
optimizer_type='adamw'): optimizer_type='adamw',
beta_1=0.9):
"""Creates an optimizer with learning rate schedule.""" """Creates an optimizer with learning rate schedule."""
# Implements linear decay of the learning rate. # Implements linear decay of the learning rate.
lr_schedule = tf.keras.optimizers.schedules.PolynomialDecay( lr_schedule = tf.keras.optimizers.schedules.PolynomialDecay(
...@@ -87,7 +88,7 @@ def create_optimizer(init_lr, ...@@ -87,7 +88,7 @@ def create_optimizer(init_lr,
optimizer = AdamWeightDecay( optimizer = AdamWeightDecay(
learning_rate=lr_schedule, learning_rate=lr_schedule,
weight_decay_rate=0.01, weight_decay_rate=0.01,
beta_1=0.9, beta_1=beta_1,
beta_2=0.999, beta_2=0.999,
epsilon=1e-6, epsilon=1e-6,
exclude_from_weight_decay=['LayerNorm', 'layer_norm', 'bias']) exclude_from_weight_decay=['LayerNorm', 'layer_norm', 'bias'])
...@@ -96,7 +97,7 @@ def create_optimizer(init_lr, ...@@ -96,7 +97,7 @@ def create_optimizer(init_lr,
optimizer = tfa_optimizers.LAMB( optimizer = tfa_optimizers.LAMB(
learning_rate=lr_schedule, learning_rate=lr_schedule,
weight_decay_rate=0.01, weight_decay_rate=0.01,
beta_1=0.9, beta_1=beta_1,
beta_2=0.999, beta_2=0.999,
epsilon=1e-6, epsilon=1e-6,
exclude_from_weight_decay=['LayerNorm', 'layer_norm', 'bias']) exclude_from_weight_decay=['LayerNorm', 'layer_norm', 'bias'])
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment