|
|
@ -1084,17 +1084,11 @@ class Fleet(object):
|
|
|
|
loss_name=loss.name, share_vars_from=None)
|
|
|
|
loss_name=loss.name, share_vars_from=None)
|
|
|
|
loss.block.program._graph = compiled_program
|
|
|
|
loss.block.program._graph = compiled_program
|
|
|
|
return self.user_defined_optimizer.minimize(
|
|
|
|
return self.user_defined_optimizer.minimize(
|
|
|
|
loss,
|
|
|
|
loss, startup_program, parameter_list, no_grad_set=no_grad_set)
|
|
|
|
startup_program=startup_program,
|
|
|
|
|
|
|
|
parameter_list=parameter_list,
|
|
|
|
|
|
|
|
no_grad_set=no_grad_set)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if meta_optimizer:
|
|
|
|
if meta_optimizer:
|
|
|
|
optimize_ops, params_grads = meta_optimizer.minimize(
|
|
|
|
optimize_ops, params_grads = meta_optimizer.minimize(
|
|
|
|
loss,
|
|
|
|
loss, startup_program, parameter_list, no_grad_set=no_grad_set)
|
|
|
|
startup_program=startup_program,
|
|
|
|
|
|
|
|
parameter_list=parameter_list,
|
|
|
|
|
|
|
|
no_grad_set=no_grad_set)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
default_program = paddle.static.default_main_program()
|
|
|
|
default_program = paddle.static.default_main_program()
|
|
|
|
|
|
|
|
|
|
|
@ -1103,20 +1097,14 @@ class Fleet(object):
|
|
|
|
|
|
|
|
|
|
|
|
else:
|
|
|
|
else:
|
|
|
|
optimize_ops, params_grads = self.user_defined_optimizer.minimize(
|
|
|
|
optimize_ops, params_grads = self.user_defined_optimizer.minimize(
|
|
|
|
loss,
|
|
|
|
loss, startup_program, parameter_list, no_grad_set=no_grad_set)
|
|
|
|
startup_program=startup_program,
|
|
|
|
|
|
|
|
parameter_list=parameter_list,
|
|
|
|
|
|
|
|
no_grad_set=no_grad_set)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
context["program_optimize_ops"] = optimize_ops
|
|
|
|
context["program_optimize_ops"] = optimize_ops
|
|
|
|
context["program_params_grads"] = params_grads
|
|
|
|
context["program_params_grads"] = params_grads
|
|
|
|
|
|
|
|
|
|
|
|
if graph_optimizer:
|
|
|
|
if graph_optimizer:
|
|
|
|
optimize_ops, params_grads = graph_optimizer.minimize(
|
|
|
|
optimize_ops, params_grads = graph_optimizer.minimize(
|
|
|
|
loss,
|
|
|
|
loss, startup_program, parameter_list, no_grad_set=no_grad_set)
|
|
|
|
startup_program=startup_program,
|
|
|
|
|
|
|
|
parameter_list=parameter_list,
|
|
|
|
|
|
|
|
no_grad_set=no_grad_set)
|
|
|
|
|
|
|
|
# since we do not encourage users to use graph operations
|
|
|
|
# since we do not encourage users to use graph operations
|
|
|
|
# if a graph optimizer takes effect, mostly
|
|
|
|
# if a graph optimizer takes effect, mostly
|
|
|
|
# optimizers_ops and params_grads are None
|
|
|
|
# optimizers_ops and params_grads are None
|
|
|
|