|
|
@ -1549,7 +1549,7 @@ class AdamOptimizer(Optimizer):
|
|
|
|
avg_cost = fluid.layers.mean(cost)
|
|
|
|
avg_cost = fluid.layers.mean(cost)
|
|
|
|
|
|
|
|
|
|
|
|
# define beta decay variable
|
|
|
|
# define beta decay variable
|
|
|
|
def get_decayed_betas(beta1_init, beta2_init, decay_steps, decay_rate)
|
|
|
|
def get_decayed_betas(beta1_init, beta2_init, decay_steps, decay_rate):
|
|
|
|
global_step = lr_scheduler._decay_step_counter()
|
|
|
|
global_step = lr_scheduler._decay_step_counter()
|
|
|
|
|
|
|
|
|
|
|
|
beta1 = fluid.layers.create_global_var(
|
|
|
|
beta1 = fluid.layers.create_global_var(
|
|
|
@ -1578,7 +1578,7 @@ class AdamOptimizer(Optimizer):
|
|
|
|
beta1, beta2 = get_decayed_betas(0.9, 0.99, 1e5, 0.9)
|
|
|
|
beta1, beta2 = get_decayed_betas(0.9, 0.99, 1e5, 0.9)
|
|
|
|
adam_optimizer = fluid.optimizer.AdamOptimizer(
|
|
|
|
adam_optimizer = fluid.optimizer.AdamOptimizer(
|
|
|
|
learning_rate=0.01,
|
|
|
|
learning_rate=0.01,
|
|
|
|
beta1=beta1
|
|
|
|
beta1=beta1,
|
|
|
|
beta2=beta2)
|
|
|
|
beta2=beta2)
|
|
|
|
adam_optimizer.minimize(avg_cost)
|
|
|
|
adam_optimizer.minimize(avg_cost)
|
|
|
|
|
|
|
|
|
|
|
|