|
|
|
@ -12,8 +12,11 @@
|
|
|
|
|
See the License for the specific language governing permissions and
|
|
|
|
|
limitations under the License. */
|
|
|
|
|
|
|
|
|
|
#include "paddle/framework/backward.h"
|
|
|
|
|
#include <gtest/gtest.h>
|
|
|
|
|
#include "paddle/framework/net.h"
|
|
|
|
|
#include "paddle/framework/op_registry.h"
|
|
|
|
|
|
|
|
|
|
namespace paddle {
|
|
|
|
|
namespace framework {
|
|
|
|
|
|
|
|
|
@ -24,10 +27,9 @@ class EmptyOp : public OperatorBase {
|
|
|
|
|
const platform::DeviceContext &dev_ctx) const override {}
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
class RowwiseAddOp : public EmptyOp {};
|
|
|
|
|
class RowwiseAddOpMaker : public OpProtoAndCheckerMaker {
|
|
|
|
|
class RowWiseAddOpMaker : public OpProtoAndCheckerMaker {
|
|
|
|
|
public:
|
|
|
|
|
RowwiseAddOpMaker(OpProto *proto, OpAttrChecker *op_checker)
|
|
|
|
|
RowWiseAddOpMaker(OpProto *proto, OpAttrChecker *op_checker)
|
|
|
|
|
: OpProtoAndCheckerMaker(proto, op_checker) {
|
|
|
|
|
AddInput("X", "Input X of Add").IgnoreGradient();
|
|
|
|
|
AddInput("b", "Bias of Add").IgnoreGradient();
|
|
|
|
@ -36,15 +38,143 @@ class RowwiseAddOpMaker : public OpProtoAndCheckerMaker {
|
|
|
|
|
}
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
class RowwiseAddGradOp : public EmptyOp {};
|
|
|
|
|
class MulOpMaker : public OpProtoAndCheckerMaker {
|
|
|
|
|
public:
|
|
|
|
|
MulOpMaker(OpProto *proto, OpAttrChecker *op_checker)
|
|
|
|
|
: OpProtoAndCheckerMaker(proto, op_checker) {
|
|
|
|
|
AddInput("A", "A");
|
|
|
|
|
AddInput("B", "B");
|
|
|
|
|
AddOutput("Out", "Out");
|
|
|
|
|
AddComment("Mul");
|
|
|
|
|
}
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
class SigmoidOpMaker : public OpProtoAndCheckerMaker {
|
|
|
|
|
public:
|
|
|
|
|
SigmoidOpMaker(OpProto *proto, OpAttrChecker *op_checker)
|
|
|
|
|
: OpProtoAndCheckerMaker(proto, op_checker) {
|
|
|
|
|
AddInput("X", "X");
|
|
|
|
|
AddOutput("Y", "Y");
|
|
|
|
|
AddComment("Sigmoid");
|
|
|
|
|
}
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
class FcOp : public NetOp {
|
|
|
|
|
public:
|
|
|
|
|
void Init() override {
|
|
|
|
|
AddOp(OpRegistry::CreateOp("mul", {Input("X"), Input("W")},
|
|
|
|
|
{Output("before_act")}, {}));
|
|
|
|
|
auto b_name = Input("b");
|
|
|
|
|
if (b_name != EMPTY_VAR_NAME()) {
|
|
|
|
|
AddOp(OpRegistry::CreateOp("rowwise_add", {Output("before_act"), b_name},
|
|
|
|
|
{Output("before_act")}, {}));
|
|
|
|
|
}
|
|
|
|
|
AddOp(OpRegistry::CreateOp("sigmoid", {Output("before_act")},
|
|
|
|
|
{Output("Out")}, {}));
|
|
|
|
|
CompleteAddOp(false);
|
|
|
|
|
}
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
class FcOpMaker : public OpProtoAndCheckerMaker {
|
|
|
|
|
public:
|
|
|
|
|
FcOpMaker(OpProto *proto, OpAttrChecker *op_checker)
|
|
|
|
|
: OpProtoAndCheckerMaker(proto, op_checker) {
|
|
|
|
|
AddInput("X", "x");
|
|
|
|
|
AddInput("W", "w");
|
|
|
|
|
AddInput("b", "b");
|
|
|
|
|
AddOutput("before_act", "before act").SetTemporary();
|
|
|
|
|
AddOutput("Out", "");
|
|
|
|
|
AddComment("");
|
|
|
|
|
}
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
class ManyOutputOpMaker : public OpProtoAndCheckerMaker {
|
|
|
|
|
public:
|
|
|
|
|
ManyOutputOpMaker(OpProto *proto, OpAttrChecker *op_checker)
|
|
|
|
|
: OpProtoAndCheckerMaker(proto, op_checker) {
|
|
|
|
|
AddInput("x", "x");
|
|
|
|
|
AddOutput("y", "y");
|
|
|
|
|
AddOutput("z", "z");
|
|
|
|
|
AddComment("");
|
|
|
|
|
}
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
class FillZeroOpMaker : public OpProtoAndCheckerMaker {
|
|
|
|
|
public:
|
|
|
|
|
FillZeroOpMaker(OpProto *proto, OpAttrChecker *op_checker)
|
|
|
|
|
: OpProtoAndCheckerMaker(proto, op_checker) {
|
|
|
|
|
AddInput("x", "x");
|
|
|
|
|
AddOutput("out", "out");
|
|
|
|
|
AddComment("");
|
|
|
|
|
}
|
|
|
|
|
};
|
|
|
|
|
} // namespace framework
|
|
|
|
|
} // namespace paddle
|
|
|
|
|
|
|
|
|
|
namespace f = paddle::framework;
|
|
|
|
|
REGISTER_OP(rowwise_add, f::RowwiseAddOp, f::RowwiseAddOpMaker);
|
|
|
|
|
REGISTER_GRADIENT_OP(rowwise_add, rowwise_add_grad, f::RowwiseAddGradOp);
|
|
|
|
|
using EnforceNotMet = paddle::platform::EnforceNotMet;
|
|
|
|
|
REGISTER_OP(rowwise_add, f::EmptyOp, f::RowWiseAddOpMaker);
|
|
|
|
|
REGISTER_GRADIENT_OP(rowwise_add, rowwise_add_grad, f::EmptyOp);
|
|
|
|
|
REGISTER_OP(mul, f::EmptyOp, f::MulOpMaker);
|
|
|
|
|
REGISTER_GRADIENT_OP(mul, mul_grad, f::EmptyOp);
|
|
|
|
|
REGISTER_OP(sigmoid, f::EmptyOp, f::SigmoidOpMaker);
|
|
|
|
|
REGISTER_GRADIENT_OP(sigmoid, sigmoid_grad, f::EmptyOp);
|
|
|
|
|
REGISTER_OP(fc, f::FcOp, f::FcOpMaker);
|
|
|
|
|
REGISTER_OP(many_output_op, f::EmptyOp, f::ManyOutputOpMaker);
|
|
|
|
|
REGISTER_GRADIENT_OP(many_output_op, many_output_op_grad, f::EmptyOp);
|
|
|
|
|
REGISTER_OP(fill_zeros_like, f::EmptyOp, f::FillZeroOpMaker);
|
|
|
|
|
|
|
|
|
|
TEST(Backward, simple_grad) {
|
|
|
|
|
auto fwd = f::OpRegistry::CreateOp("rowwise_add", {"X", "b"}, {"Out"}, {});
|
|
|
|
|
ASSERT_NE(fwd, nullptr);
|
|
|
|
|
auto gop = f::OpRegistry::CreateGradOp(*fwd);
|
|
|
|
|
ASSERT_EQ("Out" + f::OperatorBase::GRAD_VAR_SUFFIX(), gop->inputs_[0]);
|
|
|
|
|
ASSERT_EQ("rowwise_add_grad", gop->type_);
|
|
|
|
|
ASSERT_EQ("X" + f::OperatorBase::GRAD_VAR_SUFFIX(), gop->outputs_[0]);
|
|
|
|
|
ASSERT_EQ("b" + f::OperatorBase::GRAD_VAR_SUFFIX(), gop->outputs_[1]);
|
|
|
|
|
|
|
|
|
|
// LOG(INFO) << gop->Output("X" + "@GRAD");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
TEST(Backward, not_for_network) {
|
|
|
|
|
auto fwd =
|
|
|
|
|
f::OpRegistry::CreateOp("fc", {"X", "W", "b"}, {"Out", "tmp_out"},
|
|
|
|
|
{{"temporary_index", std::vector<int>{1}}});
|
|
|
|
|
ASSERT_THROW(f::OpRegistry::CreateGradOp(*fwd), EnforceNotMet);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
TEST(Backward, all_input_are_not_need) {
|
|
|
|
|
auto fwd = f::OpRegistry::CreateOp("rowwise_add", {"X", "b"}, {"Out"}, {});
|
|
|
|
|
auto backward = f::Backward(*fwd, {"X", "b"});
|
|
|
|
|
ASSERT_TRUE(backward->IsNetOp());
|
|
|
|
|
auto net = static_cast<f::NetOp *>(backward.get());
|
|
|
|
|
ASSERT_TRUE(net->ops_.empty());
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
TEST(Backward, all_output_are_not_need) {
|
|
|
|
|
auto fwd = f::OpRegistry::CreateOp("rowwise_add", {"X", "b"}, {"Out"}, {});
|
|
|
|
|
auto backward = f::Backward(*fwd, {"Out"});
|
|
|
|
|
ASSERT_TRUE(backward->IsNetOp());
|
|
|
|
|
auto net = static_cast<f::NetOp *>(backward.get());
|
|
|
|
|
ASSERT_TRUE(net->ops_.empty());
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
TEST(Backward, part_of_output_are_not_need) {
|
|
|
|
|
auto fwd = f::OpRegistry::CreateOp("many_output_op", {"X"}, {"Y", "Z"}, {});
|
|
|
|
|
auto backward = f::Backward(*fwd, {"Z"});
|
|
|
|
|
ASSERT_TRUE(backward->IsNetOp());
|
|
|
|
|
auto net = static_cast<f::NetOp *>(backward.get());
|
|
|
|
|
ASSERT_EQ(net->ops_.size(), 2);
|
|
|
|
|
|
|
|
|
|
auto &fill_zero = *net->ops_[0];
|
|
|
|
|
ASSERT_EQ("fill_zeros_like", fill_zero.type_);
|
|
|
|
|
ASSERT_EQ(1, fill_zero.inputs_.size());
|
|
|
|
|
ASSERT_EQ("Z", fill_zero.inputs_[0]);
|
|
|
|
|
ASSERT_EQ(1, fill_zero.outputs_.size());
|
|
|
|
|
ASSERT_EQ("Z@ZERO", fill_zero.outputs_[0]);
|
|
|
|
|
|
|
|
|
|
auto &d_many_out = *net->ops_[1];
|
|
|
|
|
ASSERT_EQ("many_output_op_grad", d_many_out.type_);
|
|
|
|
|
ASSERT_EQ(1 + 2 + 2, d_many_out.inputs_.size()); // I/O/OG
|
|
|
|
|
ASSERT_EQ("Z@ZERO", d_many_out.Input("z@GRAD"));
|
|
|
|
|
}
|