You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
364 lines
13 KiB
364 lines
13 KiB
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License. */
|
|
|
|
#include "paddle/framework/backward.h"
|
|
|
|
#include <gtest/gtest.h>
|
|
#include "paddle/framework/op_registry.h"
|
|
#include "paddle/operators/net_op.h"
|
|
#include "paddle/operators/type_alias.h"
|
|
|
|
namespace paddle {
|
|
namespace framework {
|
|
|
|
class EmptyOp : public OperatorBase {
|
|
public:
|
|
void InferShape(const Scope &scope) const override {}
|
|
void Run(const Scope &scope,
|
|
const platform::DeviceContext &dev_ctx) const override {}
|
|
};
|
|
|
|
class RowWiseAddOpMaker : public OpProtoAndCheckerMaker {
|
|
public:
|
|
RowWiseAddOpMaker(OpProto *proto, OpAttrChecker *op_checker)
|
|
: OpProtoAndCheckerMaker(proto, op_checker) {
|
|
AddInput("X", "Input X of Add").IgnoreGradient();
|
|
AddInput("b", "Bias of Add").IgnoreGradient();
|
|
AddOutput("Out", "Out of Add").IgnoreGradient();
|
|
AddComment("Add Op");
|
|
}
|
|
};
|
|
|
|
class MulOpMaker : public OpProtoAndCheckerMaker {
|
|
public:
|
|
MulOpMaker(OpProto *proto, OpAttrChecker *op_checker)
|
|
: OpProtoAndCheckerMaker(proto, op_checker) {
|
|
AddInput("A", "A");
|
|
AddInput("B", "B");
|
|
AddOutput("Out", "Out");
|
|
AddComment("Mul");
|
|
}
|
|
};
|
|
|
|
class SigmoidOpMaker : public OpProtoAndCheckerMaker {
|
|
public:
|
|
SigmoidOpMaker(OpProto *proto, OpAttrChecker *op_checker)
|
|
: OpProtoAndCheckerMaker(proto, op_checker) {
|
|
AddInput("X", "X");
|
|
AddOutput("Y", "Y");
|
|
AddComment("Sigmoid");
|
|
}
|
|
};
|
|
|
|
class NoGradOpMaker : public OpProtoAndCheckerMaker {
|
|
public:
|
|
NoGradOpMaker(OpProto *proto, OpAttrChecker *op_checker)
|
|
: OpProtoAndCheckerMaker(proto, op_checker) {
|
|
AddInput("X", "X input");
|
|
AddOutput("Y", "Y output");
|
|
AddComment("NoGradOp, same input output. no Grad");
|
|
}
|
|
};
|
|
|
|
class FcOp : public ops::NetOp {
|
|
public:
|
|
void Init() override {
|
|
AddOp(OpRegistry::CreateOp("mul", {Input("X"), Input("W")},
|
|
{Output("mul_result")}, {}));
|
|
auto b_name = Input("b");
|
|
std::string before_act = "mul_result";
|
|
if (b_name != kEmptyVarName) {
|
|
AddOp(OpRegistry::CreateOp("rowwise_add", {Output("mul_result"), b_name},
|
|
{Output("add_result")}, {}));
|
|
before_act = "add_result";
|
|
} else {
|
|
auto out_varname = Output("add_result");
|
|
if (out_varname != kEmptyVarName) {
|
|
this->Rename(out_varname, kEmptyVarName);
|
|
}
|
|
}
|
|
|
|
AddOp(OpRegistry::CreateOp("sigmoid", {Output(before_act)}, {Output("Out")},
|
|
{}));
|
|
CompleteAddOp(false);
|
|
}
|
|
};
|
|
|
|
class FcOpMaker : public OpProtoAndCheckerMaker {
|
|
public:
|
|
FcOpMaker(OpProto *proto, OpAttrChecker *op_checker)
|
|
: OpProtoAndCheckerMaker(proto, op_checker) {
|
|
AddInput("X", "x");
|
|
AddInput("W", "w");
|
|
AddInput("b", "b");
|
|
AddOutput("mul_result", "").SetTemporary();
|
|
AddOutput("add_result", "").SetTemporary();
|
|
AddOutput("Out", "");
|
|
AddComment("");
|
|
}
|
|
};
|
|
|
|
class ManyOutputOpMaker : public OpProtoAndCheckerMaker {
|
|
public:
|
|
ManyOutputOpMaker(OpProto *proto, OpAttrChecker *op_checker)
|
|
: OpProtoAndCheckerMaker(proto, op_checker) {
|
|
AddInput("x", "x");
|
|
AddOutput("y", "y");
|
|
AddOutput("z", "z");
|
|
AddComment("");
|
|
}
|
|
};
|
|
|
|
class FillZeroOpMaker : public OpProtoAndCheckerMaker {
|
|
public:
|
|
FillZeroOpMaker(OpProto *proto, OpAttrChecker *op_checker)
|
|
: OpProtoAndCheckerMaker(proto, op_checker) {
|
|
AddInput("x", "x");
|
|
AddOutput("out", "out");
|
|
AddComment("");
|
|
}
|
|
};
|
|
|
|
class AddOpMaker : public OpProtoAndCheckerMaker {
|
|
public:
|
|
AddOpMaker(OpProto *proto, OpAttrChecker *op_checker)
|
|
: OpProtoAndCheckerMaker(proto, op_checker) {
|
|
AddInput("X", "x").SetMultiple();
|
|
AddOutput("Y", "y");
|
|
AddComment("");
|
|
}
|
|
};
|
|
} // namespace framework
|
|
} // namespace paddle
|
|
|
|
namespace f = paddle::framework;
|
|
using EnforceNotMet = paddle::platform::EnforceNotMet;
|
|
REGISTER_OP(rowwise_add, f::EmptyOp, f::RowWiseAddOpMaker);
|
|
REGISTER_GRADIENT_OP(rowwise_add, rowwise_add_grad, f::EmptyOp);
|
|
REGISTER_OP(mul, f::EmptyOp, f::MulOpMaker);
|
|
REGISTER_GRADIENT_OP(mul, mul_grad, f::EmptyOp);
|
|
REGISTER_OP(sigmoid, f::EmptyOp, f::SigmoidOpMaker);
|
|
REGISTER_GRADIENT_OP(sigmoid, sigmoid_grad, f::EmptyOp);
|
|
REGISTER_OP(nograd, f::EmptyOp, f::NoGradOpMaker);
|
|
REGISTER_OP(fill_zeros_like, f::EmptyOp, f::FillZeroOpMaker);
|
|
REGISTER_OP(add, f::EmptyOp, f::AddOpMaker);
|
|
REGISTER_GRADIENT_OP(add, add_grad, f::EmptyOp);
|
|
REGISTER_OP(fc, f::FcOp, f::FcOpMaker);
|
|
REGISTER_OP(many_output_op, f::EmptyOp, f::ManyOutputOpMaker);
|
|
REGISTER_GRADIENT_OP(many_output_op, many_output_op_grad, f::EmptyOp);
|
|
|
|
TEST(Backward, simple_op_grad) {
|
|
auto fwd = f::OpRegistry::CreateOp("rowwise_add", {"X", "b"}, {"Out"}, {});
|
|
ASSERT_NE(fwd, nullptr);
|
|
auto gop = f::OpRegistry::CreateGradOp(*fwd);
|
|
ASSERT_EQ(4UL, gop->inputs_.size());
|
|
ASSERT_EQ(f::kEmptyVarName, gop->inputs_[0]);
|
|
ASSERT_EQ("rowwise_add_grad", gop->type_);
|
|
ASSERT_EQ("X" + f::kGradVarSuffix, gop->outputs_[0]);
|
|
ASSERT_EQ("b" + f::kGradVarSuffix, gop->outputs_[1]);
|
|
|
|
ASSERT_EQ("X" + f::kGradVarSuffix, gop->Output("X" + f::kGradVarSuffix));
|
|
}
|
|
|
|
TEST(Backward, simple_op_not_need_grad) {
|
|
auto fwd = f::OpRegistry::CreateOp("rowwise_add", {"X", "b"}, {"Out"}, {});
|
|
ASSERT_NE(fwd, nullptr);
|
|
auto gop = f::Backward(*fwd, {"X"});
|
|
ASSERT_EQ(std::find(gop->outputs_.begin(), gop->outputs_.end(),
|
|
"X" + f::kGradVarSuffix),
|
|
gop->outputs_.end());
|
|
|
|
auto no_input_gop = f::Backward(*fwd, {"X", "b"});
|
|
ASSERT_NE(no_input_gop, nullptr);
|
|
ASSERT_TRUE(no_input_gop->IsNetOp());
|
|
ASSERT_EQ(0UL,
|
|
std::static_pointer_cast<ops::NetOp>(no_input_gop)->ops_.size());
|
|
}
|
|
|
|
TEST(Backward, net_fc_backward_normal) {
|
|
std::shared_ptr<f::OperatorBase> fwd = f::OpRegistry::CreateOp(
|
|
"fc", {"X", "w", "b"}, {"mul_result", "add_result", "out"}, {});
|
|
ASSERT_NE(fwd, nullptr);
|
|
std::shared_ptr<f::OperatorBase> gop = f::Backward(*fwd, {});
|
|
ASSERT_TRUE(gop->IsNetOp());
|
|
auto net = static_cast<ops::NetOp *>(gop.get());
|
|
|
|
ASSERT_NO_THROW(net->DebugString());
|
|
|
|
ASSERT_EQ(3UL, net->ops_.size());
|
|
|
|
f::OperatorBase &d_sigmoid = *net->ops_[0];
|
|
ASSERT_EQ("sigmoid_grad", d_sigmoid.type_);
|
|
|
|
f::OperatorBase &d_add = *net->ops_[1];
|
|
ASSERT_EQ("rowwise_add_grad", d_add.type_);
|
|
|
|
f::OperatorBase &d_mul = *net->ops_[2];
|
|
ASSERT_EQ("mul_grad", d_mul.type_);
|
|
}
|
|
|
|
TEST(Backward, net_fc_backward_not_have_b) {
|
|
std::shared_ptr<f::OperatorBase> fwd =
|
|
f::OpRegistry::CreateOp("fc", {"X", "w", f::kEmptyVarName},
|
|
{"mul_result", "add_result", "tmp"}, {});
|
|
ASSERT_NE(fwd, nullptr);
|
|
std::shared_ptr<f::OperatorBase> gop = f::Backward(*fwd, {});
|
|
ASSERT_TRUE(gop->IsNetOp());
|
|
auto net = static_cast<ops::NetOp *>(gop.get());
|
|
|
|
ASSERT_NO_THROW(net->DebugString());
|
|
|
|
ASSERT_EQ(2UL, net->ops_.size());
|
|
|
|
f::OperatorBase &d_sigmoid = *net->ops_[0];
|
|
ASSERT_EQ("sigmoid_grad", d_sigmoid.type_);
|
|
|
|
f::OperatorBase &d_mul = *net->ops_[1];
|
|
ASSERT_EQ("mul_grad", d_mul.type_);
|
|
}
|
|
|
|
TEST(Backward, net_input_of_network_not_need_grad) {
|
|
ops::NetOp net;
|
|
net.AddOp(f::OpRegistry::CreateOp("fc", {"X", "W1", "b1"},
|
|
{"mul_tmp_0", "add_tmp_0", "hidden0"}, {}));
|
|
net.AddOp(f::OpRegistry::CreateOp("fc", {"hidden0", "W2", "b2"},
|
|
{"mul_tmp_1", "add_tmp_1", "hidden1"}, {}));
|
|
net.CompleteAddOp();
|
|
auto bwd = Backward(net, {"X"}); // X@GRAD is not need.
|
|
ASSERT_TRUE(bwd->IsNetOp());
|
|
auto bwd_net = static_cast<ops::NetOp *>(bwd.get());
|
|
|
|
std::unordered_set<std::string> all_output = std::unordered_set<std::string>(
|
|
bwd_net->outputs_.begin(), bwd_net->outputs_.end());
|
|
all_output.erase(f::kEmptyVarName);
|
|
|
|
for (auto &out : {"W1", "b1", "hidden0", "W2", "b2"}) {
|
|
ASSERT_NE(all_output.find(out + f::kGradVarSuffix), all_output.end());
|
|
}
|
|
|
|
// Not Generated X
|
|
ASSERT_EQ(all_output.find("X" + f::kGradVarSuffix), all_output.end());
|
|
|
|
ASSERT_EQ(2UL, bwd_net->ops_.size());
|
|
ASSERT_TRUE(bwd_net->ops_[1]->IsNetOp());
|
|
auto first_fc_grad = static_cast<ops::NetOp *>(bwd_net->ops_[1].get());
|
|
ASSERT_EQ(3UL, first_fc_grad->ops_.size());
|
|
ASSERT_EQ(f::kEmptyVarName,
|
|
first_fc_grad->ops_[2]->Output("A" + f::kGradVarSuffix));
|
|
}
|
|
|
|
TEST(Backward, net_shared_weight) {
|
|
ops::NetOp net;
|
|
net.AddOp(f::OpRegistry::CreateOp("mul", {"X", "W"}, {"Out"}, {}));
|
|
net.AddOp(f::OpRegistry::CreateOp("mul", {"Out", "W"}, {"FinalOut"}, {}));
|
|
net.CompleteAddOp();
|
|
|
|
auto bwd = f::Backward(net, {});
|
|
ASSERT_TRUE(bwd->IsNetOp());
|
|
auto bwd_net = static_cast<ops::NetOp *>(bwd.get());
|
|
ASSERT_EQ(3UL, bwd_net->ops_.size());
|
|
ASSERT_EQ("add", bwd_net->ops_[2]->type_);
|
|
}
|
|
|
|
TEST(Backward, op_register_grad_not_for_network) {
|
|
auto fwd = f::OpRegistry::CreateOp(
|
|
"fc", {"X", "W", "b"}, {"mul_out", "add_out", "out1"},
|
|
{{"temporary_index", std::vector<int>{0, 1}}});
|
|
|
|
ASSERT_THROW(f::OpRegistry::CreateGradOp(*fwd), EnforceNotMet);
|
|
}
|
|
|
|
TEST(Backward, op_all_input_are_not_need) {
|
|
auto fwd = f::OpRegistry::CreateOp("rowwise_add", {"X", "b"}, {"Out"}, {});
|
|
auto backward = f::Backward(*fwd, {"X", "b"});
|
|
ASSERT_TRUE(backward->IsNetOp());
|
|
auto net = static_cast<ops::NetOp *>(backward.get());
|
|
ASSERT_TRUE(net->ops_.empty());
|
|
}
|
|
|
|
TEST(Backward, op_all_output_are_not_need) {
|
|
auto fwd = f::OpRegistry::CreateOp("rowwise_add", {"X", "b"}, {"Out"}, {});
|
|
auto backward = f::Backward(*fwd, {"Out"});
|
|
ASSERT_TRUE(backward->IsNetOp());
|
|
auto net = static_cast<ops::NetOp *>(backward.get());
|
|
ASSERT_TRUE(net->ops_.empty());
|
|
}
|
|
|
|
TEST(Backward, op_part_of_output_are_not_need) {
|
|
auto fwd = f::OpRegistry::CreateOp("many_output_op", {"X"}, {"Y", "Z"}, {});
|
|
auto backward = f::Backward(*fwd, {"Z"});
|
|
ASSERT_TRUE(backward->IsNetOp());
|
|
auto net = static_cast<ops::NetOp *>(backward.get());
|
|
ASSERT_EQ(net->ops_.size(), 2UL);
|
|
|
|
auto &fill_zero = *net->ops_[0];
|
|
ASSERT_EQ("fill_zeros_like", fill_zero.type_);
|
|
ASSERT_EQ(1UL, fill_zero.inputs_.size());
|
|
ASSERT_EQ("Z", fill_zero.inputs_[0]);
|
|
ASSERT_EQ(1UL, fill_zero.outputs_.size());
|
|
ASSERT_EQ("Z" + f::kZeroVarSuffix, fill_zero.outputs_[0]);
|
|
|
|
auto &d_many_out = *net->ops_[1];
|
|
ASSERT_EQ("many_output_op_grad", d_many_out.type_);
|
|
ASSERT_EQ(1UL + 2UL + 2UL, d_many_out.inputs_.size()); // I/O/OG
|
|
ASSERT_EQ("Z" + f::kZeroVarSuffix, d_many_out.Input("z" + f::kGradVarSuffix));
|
|
ASSERT_EQ("Y" + f::kGradVarSuffix, d_many_out.Input("y" + f::kGradVarSuffix));
|
|
ASSERT_EQ("X" + f::kGradVarSuffix,
|
|
d_many_out.Output("x" + f::kGradVarSuffix));
|
|
}
|
|
|
|
TEST(Backward, op_part_of_input_are_not_need) {
|
|
auto fwd = f::OpRegistry::CreateOp("mul", {"a", "b"}, {"out"}, {});
|
|
auto backward = f::Backward(*fwd, {"a"});
|
|
auto &grad_mul = *backward;
|
|
ASSERT_EQ(grad_mul.type_, "mul_grad");
|
|
ASSERT_EQ(grad_mul.inputs_.size(), 2UL + 1UL + 1UL);
|
|
ASSERT_EQ(grad_mul.outputs_.size(), 2UL);
|
|
ASSERT_EQ(grad_mul.Output("A" + f::kGradVarSuffix), f::kEmptyVarName);
|
|
ASSERT_EQ(grad_mul.Output("B" + f::kGradVarSuffix), "b" + f::kGradVarSuffix);
|
|
ASSERT_EQ(grad_mul.Input("Out" + f::kGradVarSuffix),
|
|
"out" + f::kGradVarSuffix);
|
|
ASSERT_EQ(grad_mul.Input("A"), "a");
|
|
ASSERT_EQ(grad_mul.Input("B"), "b");
|
|
ASSERT_EQ(grad_mul.Input("Out"), "out");
|
|
}
|
|
|
|
TEST(Backward, linear_net_intermediate_variable_has_no_grad) {
|
|
ops::NetOp net;
|
|
net.AddOp(f::OpRegistry::CreateOp("fc", {"x1", "w1", "b1"},
|
|
{"mul_out1", "add_out1", "out1"}, {}));
|
|
net.AddOp(f::OpRegistry::CreateOp("fc", {"out1", "w2", "b2"},
|
|
{"mul_out2", "tmp_out2", "out2"}, {}));
|
|
net.AddOp(f::OpRegistry::CreateOp("fc", {"out2", "w3", "b3"},
|
|
{"mul_out3", "tmp_out3", "out3"}, {}));
|
|
net.CompleteAddOp();
|
|
auto backward = f::Backward(net, {"mul_out2", "tmp_out2", "out2"});
|
|
ASSERT_TRUE(backward->IsNetOp());
|
|
auto bwd_net = static_cast<ops::NetOp *>(backward.get());
|
|
ASSERT_EQ(bwd_net->ops_.size(), 3UL);
|
|
auto &grad_fc = *bwd_net->ops_[0];
|
|
EXPECT_EQ(grad_fc.inputs_.size(),
|
|
3UL /* external input number */
|
|
+ 1UL /* external output number*/
|
|
+ 1UL /* number of gradient of external output*/
|
|
+ 2U /* internal variable number*/);
|
|
EXPECT_EQ(grad_fc.outputs_.size(), 2UL /* input number of mul*/
|
|
+ 2UL /* input number of rowwise_add */
|
|
+ 1UL /* input number of sigmod */);
|
|
EXPECT_EQ(bwd_net->ops_[1]->inputs_.size(), 0UL);
|
|
EXPECT_EQ(bwd_net->ops_[1]->outputs_.size(), 0UL);
|
|
EXPECT_EQ(bwd_net->ops_[2]->inputs_.size(), 0UL);
|
|
EXPECT_EQ(bwd_net->ops_[2]->outputs_.size(), 0UL);
|
|
}
|