|
|
|
@ -119,7 +119,7 @@ TEST(test_tracer, test_track_backward_output) {
|
|
|
|
|
std::shared_ptr<imperative::VarBase> x_in(
|
|
|
|
|
new imperative::VarBase(true, "x_in"));
|
|
|
|
|
std::shared_ptr<imperative::VarBase> y_in(
|
|
|
|
|
new imperative::VarBase(false, "y_in"));
|
|
|
|
|
new imperative::VarBase(true, "y_in"));
|
|
|
|
|
x_in->SetOverridedStopGradient(false);
|
|
|
|
|
std::shared_ptr<imperative::VarBase> vout(
|
|
|
|
|
new imperative::VarBase(true, "vout"));
|
|
|
|
@ -146,7 +146,10 @@ TEST(test_tracer, test_track_backward_output) {
|
|
|
|
|
imperative::NameVarBaseMap outs = {out_pair};
|
|
|
|
|
framework::AttributeMap mul_attr_map;
|
|
|
|
|
mul_attr_map["use_mkldnn"] = false;
|
|
|
|
|
ASSERT_ANY_THROW(tracer.TraceOp("mul", ins, outs, mul_attr_map, place, true));
|
|
|
|
|
tracer.TraceOp("mul", ins, outs, mul_attr_map, place, true);
|
|
|
|
|
auto* engine = tracer.GetDefaultEngine();
|
|
|
|
|
ASSERT_NE(engine->GradVars().size(), 0UL);
|
|
|
|
|
ASSERT_NE(engine->GradOps().size(), 0UL); // trace_backward already ran.
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
TEST(test_tracer, test_track_backward_input) {
|
|
|
|
@ -157,7 +160,7 @@ TEST(test_tracer, test_track_backward_input) {
|
|
|
|
|
std::shared_ptr<imperative::VarBase> y_in(
|
|
|
|
|
new imperative::VarBase(true, "y_in"));
|
|
|
|
|
std::shared_ptr<imperative::VarBase> vout(
|
|
|
|
|
new imperative::VarBase(false, "vout"));
|
|
|
|
|
new imperative::VarBase(true, "vout"));
|
|
|
|
|
platform::CPUPlace place;
|
|
|
|
|
x_in->SetOverridedStopGradient(false);
|
|
|
|
|
std::vector<float> src_data(10, 2.0);
|
|
|
|
@ -182,7 +185,10 @@ TEST(test_tracer, test_track_backward_input) {
|
|
|
|
|
imperative::NameVarBaseMap outs = {out_pair};
|
|
|
|
|
framework::AttributeMap mul_attr_map;
|
|
|
|
|
mul_attr_map["use_mkldnn"] = false;
|
|
|
|
|
ASSERT_ANY_THROW(tracer.TraceOp("mul", ins, outs, mul_attr_map, place, true));
|
|
|
|
|
tracer.TraceOp("mul", ins, outs, mul_attr_map, place, true);
|
|
|
|
|
auto* engine = tracer.GetDefaultEngine();
|
|
|
|
|
ASSERT_NE(engine->GradVars().size(), 0UL);
|
|
|
|
|
ASSERT_NE(engine->GradOps().size(), 0UL); // trace_backward already ran.
|
|
|
|
|
}
|
|
|
|
|
#if defined(PADDLE_WITH_CUDA)
|
|
|
|
|
TEST(test_tracer, test_trace_op_with_multi_device_inputs) {
|
|
|
|
@ -296,6 +302,73 @@ TEST(test_tracer, test_expected_place) {
|
|
|
|
|
ASSERT_EQ(platform::is_gpu_place(tracer.ExpectedPlace()), true);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
TEST(test_tracer, test_var_without_grad_var) {
|
|
|
|
|
// Doing an mul
|
|
|
|
|
imperative::Tracer tracer;
|
|
|
|
|
std::shared_ptr<imperative::VarBase> x_in(
|
|
|
|
|
new imperative::VarBase(true, "x_in"));
|
|
|
|
|
x_in->ClearGradVarBase();
|
|
|
|
|
std::shared_ptr<imperative::VarBase> y_in(
|
|
|
|
|
new imperative::VarBase(true, "y_in"));
|
|
|
|
|
std::shared_ptr<imperative::VarBase> vout(
|
|
|
|
|
new imperative::VarBase(true, "vout"));
|
|
|
|
|
x_in->SetOverridedStopGradient(false);
|
|
|
|
|
y_in->SetOverridedStopGradient(false);
|
|
|
|
|
platform::CPUPlace place;
|
|
|
|
|
std::vector<float> src_data(10, 2.0);
|
|
|
|
|
std::vector<int64_t> dims1 = {2, 5};
|
|
|
|
|
std::vector<int64_t> dims2 = {5, 2};
|
|
|
|
|
|
|
|
|
|
auto* x_in_tensor = x_in->MutableVar()->GetMutable<framework::LoDTensor>();
|
|
|
|
|
auto* y_in_tensor = y_in->MutableVar()->GetMutable<framework::LoDTensor>();
|
|
|
|
|
x_in_tensor->Resize(framework::make_ddim(dims1));
|
|
|
|
|
auto* mutable_x = x_in_tensor->mutable_data<float>(place);
|
|
|
|
|
paddle::memory::Copy(place, mutable_x, place, src_data.data(),
|
|
|
|
|
sizeof(float) * src_data.size());
|
|
|
|
|
y_in_tensor->Resize(framework::make_ddim(dims2));
|
|
|
|
|
auto* mutable_y = y_in_tensor->mutable_data<float>(place);
|
|
|
|
|
paddle::memory::Copy(place, mutable_y, place, src_data.data(),
|
|
|
|
|
sizeof(float) * src_data.size());
|
|
|
|
|
|
|
|
|
|
var_pair x_pair = var_pair("X", vb_vector(1, x_in));
|
|
|
|
|
var_pair y_pair = var_pair("Y", vb_vector(1, y_in));
|
|
|
|
|
var_pair out_pair = var_pair("Out", vb_vector(1, vout));
|
|
|
|
|
imperative::NameVarBaseMap ins = {x_pair, y_pair};
|
|
|
|
|
imperative::NameVarBaseMap outs = {out_pair};
|
|
|
|
|
framework::AttributeMap mul_attr_map;
|
|
|
|
|
mul_attr_map["use_mkldnn"] = false;
|
|
|
|
|
tracer.TraceOp("mul", ins, outs, mul_attr_map, place, true);
|
|
|
|
|
|
|
|
|
|
const auto& out_tensor = vout->Var().Get<framework::LoDTensor>();
|
|
|
|
|
for (int i = 0; i < vout->Var().Get<framework::LoDTensor>().numel(); i++) {
|
|
|
|
|
ASSERT_EQ(out_tensor.data<float>()[i], 20.0);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
detail::BackwardStrategy back_st;
|
|
|
|
|
imperative::Engine* engine = tracer.GetDefaultEngine();
|
|
|
|
|
ASSERT_NE(engine->GradVars().size(), 0UL);
|
|
|
|
|
ASSERT_NE(engine->GradOps().size(), 0UL); // trace_backward already ran.
|
|
|
|
|
engine->Init(vout.get(), back_st);
|
|
|
|
|
engine->Execute();
|
|
|
|
|
|
|
|
|
|
// check the grad
|
|
|
|
|
framework::LoDTensor x_grad;
|
|
|
|
|
framework::TensorCopySync(x_in->GradVar().Get<framework::LoDTensor>(), place,
|
|
|
|
|
&x_grad);
|
|
|
|
|
|
|
|
|
|
for (int i = 0; i < x_grad.numel(); ++i) {
|
|
|
|
|
ASSERT_EQ(x_grad.data<float>()[i], 4.0);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
framework::LoDTensor y_grad;
|
|
|
|
|
framework::TensorCopySync(y_in->GradVar().Get<framework::LoDTensor>(), place,
|
|
|
|
|
&y_grad);
|
|
|
|
|
|
|
|
|
|
for (int i = 0; i < y_grad.numel(); ++i) {
|
|
|
|
|
ASSERT_EQ(y_grad.data<float>()[i], 4.0);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
} // namespace imperative
|
|
|
|
|
} // namespace paddle
|
|
|
|
|
|
|
|
|
|