cherry-pick from feature/anakin-engine: Add subgraph fuse support and anakin engine #16018
parent
084310f536
commit
b21770a2aa
@ -1,8 +1,12 @@
|
||||
cc_library(anakin_op_converter SRCS fc.cc conv2d.cc activation.cc pool2d.cc concat.cc split.cc DEPS anakin_engine framework_proto scope op_registry)
|
||||
cc_test(test_anakin_fc SRCS test_fc_op.cc DEPS ${FLUID_CORE_MODULES} ${GLOB_OPERATOR_DEPS} anakin_op_converter mul_op)
|
||||
cc_test(test_anakin_conv2d SRCS test_conv2d_op.cc DEPS ${FLUID_CORE_MODULES} ${GLOB_OPERATOR_DEPS} anakin_op_converter conv_op im2col vol2col depthwise_conv SERIAL)
|
||||
cc_test(test_anakin_activation SRCS test_activation_op.cc DEPS ${FLUID_CORE_MODULES} ${GLOB_OPERATOR_DEPS} activation_op anakin_op_converter
|
||||
SERIAL)
|
||||
cc_test(test_anakin_pool2d SRCS test_pool2d_op.cc DEPS ${FLUID_CORE_MODULES} ${GLOB_OPERATOR_DEPS} anakin_op_converter pool_op pooling)
|
||||
cc_test(test_anakin_concat SRCS test_concat_op.cc DEPS ${FLUID_CORE_MODULES} ${GLOB_OPERATOR_DEPS} anakin_op_converter concat_op concat_and_split)
|
||||
cc_test(test_anakin_split SRCS test_split_op.cc DEPS ${FLUID_CORE_MODULES} ${GLOB_OPERATOR_DEPS} anakin_op_converter split_op concat_and_split)
|
||||
cc_library(anakin_op_converter SRCS fc.cc conv2d.cc conv2d_fusion.cc
|
||||
elementwise.cc activation.cc pool2d.cc concat.cc split.cc relu.cc DEPS anakin_engine framework_proto scope op_registry)
|
||||
cc_test(test_anakin_fc SRCS test_fc_op.cc DEPS anakin_op_converter mul_op)
|
||||
cc_test(test_anakin_conv2d SRCS test_conv2d_op.cc DEPS anakin_op_converter conv_op im2col vol2col depthwise_conv SERIAL)
|
||||
cc_test(test_anakin_activation SRCS test_activation_op.cc DEPS activation_op anakin_op_converter SERIAL)
|
||||
cc_test(test_anakin_pool2d SRCS test_pool2d_op.cc DEPS anakin_op_converter pool_op pooling)
|
||||
cc_test(test_anakin_concat SRCS test_concat_op.cc DEPS anakin_op_converter concat_op concat_and_split)
|
||||
cc_test(test_anakin_split SRCS test_split_op.cc DEPS anakin_op_converter split_op concat_and_split)
|
||||
cc_test(test_anakin_elementwise SRCS test_elementwise_op.cc DEPS
|
||||
anakin_op_converter elementwise_add_op)
|
||||
|
||||
cc_test(test_anakin_relu SRCS test_relu_op.cc DEPS activation_op anakin_op_converter SERIAL)
|
||||
|
@ -0,0 +1,113 @@
|
||||
// Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
|
||||
//
|
||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||
// you may not use this file except in compliance with the License.
|
||||
// You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing, software
|
||||
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
// See the License for the specific language governing permissions and
|
||||
// limitations under the License.
|
||||
|
||||
#include "paddle/fluid/inference/anakin/convert/conv2d_fusion.h"
|
||||
#include <algorithm>
|
||||
#include <memory>
|
||||
#include <vector>
|
||||
|
||||
using anakin::graph::GraphGlobalMem;
|
||||
using anakin::AK_FLOAT;
|
||||
using anakin::saber::NV;
|
||||
using anakin::saber::Shape;
|
||||
using anakin::PTuple;
|
||||
|
||||
namespace paddle {
|
||||
namespace inference {
|
||||
namespace anakin {
|
||||
|
||||
void Conv2dFusionOpConverter::operator()(const framework::proto::OpDesc &op,
|
||||
const framework::Scope &scope,
|
||||
bool test_mode) {
|
||||
framework::OpDesc op_desc(op, nullptr);
|
||||
PADDLE_ENFORCE_EQ(op_desc.Input("Input").size(), 1UL);
|
||||
PADDLE_ENFORCE_EQ(op_desc.Input("Filter").size(), 1UL);
|
||||
PADDLE_ENFORCE_EQ(op_desc.Input("Bias").size(), 1UL);
|
||||
PADDLE_ENFORCE_EQ(op_desc.Output("Output").size(), 1UL);
|
||||
|
||||
auto input_name = op_desc.Input("Input").front();
|
||||
auto output_name = op_desc.Output("Output").front();
|
||||
auto op_name = op_desc.Type() + ":" + op_desc.Output("Output").front();
|
||||
engine_->AddOp(op_name, "Convolution", {input_name}, {output_name});
|
||||
|
||||
auto *filter_v = scope.FindVar(op_desc.Input("Filter").front());
|
||||
PADDLE_ENFORCE_NOT_NULL(filter_v);
|
||||
auto *filter_t = filter_v->GetMutable<framework::LoDTensor>();
|
||||
|
||||
auto *b_v = scope.FindVar(op_desc.Input("Bias").front());
|
||||
PADDLE_ENFORCE_NOT_NULL(b_v);
|
||||
auto *b_t = b_v->GetMutable<framework::LoDTensor>();
|
||||
|
||||
std::unique_ptr<framework::LoDTensor> weight_tensor(
|
||||
new framework::LoDTensor());
|
||||
weight_tensor->Resize(filter_t->dims());
|
||||
TensorCopySync((*filter_t), platform::CPUPlace(), weight_tensor.get());
|
||||
|
||||
PADDLE_ENFORCE_EQ(weight_tensor->dims().size(), 4UL);
|
||||
|
||||
// const int n_output = weight_tensor->dims()[0];
|
||||
// const int n_input = weight_tensor->dims()[1];
|
||||
const int filter_h = weight_tensor->dims()[2];
|
||||
const int filter_w = weight_tensor->dims()[3];
|
||||
// auto filter_num = n_input * filter_h * filter_w ;
|
||||
auto filter_num = weight_tensor->dims()[0];
|
||||
engine_->AddOpAttr<int>(op_name, "filter_num", filter_num);
|
||||
engine_->AddOpAttr<PTuple<int>>(op_name, "kernel_size", {filter_h, filter_w});
|
||||
auto strides = boost::get<std::vector<int>>(op_desc.GetAttr("strides"));
|
||||
engine_->AddOpAttr<PTuple<int>>(op_name, "strides", strides);
|
||||
auto paddings = boost::get<std::vector<int>>(op_desc.GetAttr("paddings"));
|
||||
engine_->AddOpAttr<PTuple<int>>(op_name, "padding", paddings);
|
||||
auto dilations = boost::get<std::vector<int>>(op_desc.GetAttr("dilations"));
|
||||
engine_->AddOpAttr<PTuple<int>>(op_name, "dilation_rate", dilations);
|
||||
const int groups = boost::get<int>(op_desc.GetAttr("groups"));
|
||||
engine_->AddOpAttr(op_name, "group", groups);
|
||||
engine_->AddOpAttr(op_name, "axis", 1);
|
||||
engine_->AddOpAttr(op_name, "bias_term", true);
|
||||
|
||||
auto weight_shape = framework::vectorize2int(filter_t->dims());
|
||||
Shape anakin_shape(weight_shape);
|
||||
auto *weight1 =
|
||||
GraphGlobalMem<NV>::Global().template new_block<AK_FLOAT>(anakin_shape);
|
||||
float *cpu_data = static_cast<float *>(weight1->h_tensor().mutable_data());
|
||||
std::copy_n(weight_tensor->data<float>(), weight_tensor->numel(), cpu_data);
|
||||
weight1->d_tensor().set_shape(anakin_shape);
|
||||
weight1->d_tensor().copy_from(weight1->h_tensor());
|
||||
engine_->AddOpAttr(op_name, "weight_1", *weight1);
|
||||
|
||||
auto bias_shape = framework::vectorize2int(b_t->dims());
|
||||
framework::LoDTensor bias_tensor;
|
||||
bias_tensor.Resize(b_t->dims());
|
||||
TensorCopySync((*b_t), platform::CPUPlace(), &bias_tensor);
|
||||
auto *bias_data = bias_tensor.data<float>();
|
||||
bias_shape.insert(bias_shape.begin(), 1);
|
||||
bias_shape.insert(bias_shape.begin(), 1);
|
||||
bias_shape.insert(bias_shape.begin(), 1);
|
||||
// bias_shape.push_back(1);
|
||||
// bias_shape.push_back(1);
|
||||
Shape anakin_bias_shape(bias_shape);
|
||||
|
||||
auto *weight2 = GraphGlobalMem<NV>::Global().template new_block<AK_FLOAT>(
|
||||
anakin_bias_shape);
|
||||
float *cpu_data2 = static_cast<float *>(weight2->h_tensor().mutable_data());
|
||||
std::copy_n(bias_data, bias_tensor.numel(), cpu_data2);
|
||||
weight2->d_tensor().set_shape(anakin_bias_shape);
|
||||
weight2->d_tensor().copy_from(weight2->h_tensor());
|
||||
engine_->AddOpAttr(op_name, "weight_2", *weight2);
|
||||
}
|
||||
|
||||
} // namespace anakin
|
||||
} // namespace inference
|
||||
} // namespace paddle
|
||||
|
||||
REGISTER_ANAKIN_OP_CONVERTER(conv2d_fusion, Conv2dFusionOpConverter);
|
@ -0,0 +1,35 @@
|
||||
// Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
|
||||
//
|
||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||
// you may not use this file except in compliance with the License.
|
||||
// You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing, software
|
||||
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
// See the License for the specific language governing permissions and
|
||||
// limitations under the License.
|
||||
|
||||
#pragma once
|
||||
|
||||
#include "paddle/fluid/inference/anakin/convert/op_converter.h"
|
||||
|
||||
namespace paddle {
|
||||
namespace inference {
|
||||
namespace anakin {
|
||||
|
||||
class Conv2dFusionOpConverter : public AnakinOpConverter {
|
||||
public:
|
||||
Conv2dFusionOpConverter() = default;
|
||||
|
||||
virtual void operator()(const framework::proto::OpDesc &op,
|
||||
const framework::Scope &scope,
|
||||
bool test_mode) override;
|
||||
virtual ~Conv2dFusionOpConverter() {}
|
||||
};
|
||||
|
||||
} // namespace anakin
|
||||
} // namespace inference
|
||||
} // namespace paddle
|
@ -0,0 +1,57 @@
|
||||
// Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
|
||||
//
|
||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||
// you may not use this file except in compliance with the License.
|
||||
// You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing, software
|
||||
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
// See the License for the specific language governing permissions and
|
||||
// limitations under the License.
|
||||
|
||||
#include "paddle/fluid/inference/anakin/convert/elementwise.h"
|
||||
#include <algorithm>
|
||||
#include <string>
|
||||
#include <vector>
|
||||
|
||||
using anakin::graph::GraphGlobalMem;
|
||||
using anakin::AK_FLOAT;
|
||||
using anakin::Precision;
|
||||
using anakin::saber::NV;
|
||||
using anakin::saber::X86;
|
||||
using anakin::saber::Shape;
|
||||
using anakin::PBlock;
|
||||
using anakin::PTuple;
|
||||
|
||||
namespace paddle {
|
||||
namespace inference {
|
||||
namespace anakin {
|
||||
|
||||
void ElementwiseAddOpConverter::operator()(const framework::proto::OpDesc &op,
|
||||
const framework::Scope &scope,
|
||||
bool test_mode) {
|
||||
framework::OpDesc op_desc(op, nullptr);
|
||||
PADDLE_ENFORCE_EQ(op_desc.Input("X").size(), 1);
|
||||
PADDLE_ENFORCE_EQ(op_desc.Input("Y").size(), 1); // Y is a weight
|
||||
PADDLE_ENFORCE_EQ(op_desc.Output("Out").size(), 1);
|
||||
|
||||
auto x_name = op_desc.Input("X").front();
|
||||
auto y_name = op_desc.Input("Y").front();
|
||||
auto out_name = op_desc.Output("Out").front();
|
||||
auto op_name = op_desc.Type() + ":" + op_desc.Output("Out").front();
|
||||
|
||||
engine_->AddOp(op_name, "Eltwise", {x_name, y_name}, {out_name});
|
||||
std::string elementwise_type = "Add";
|
||||
engine_->AddOpAttr<std::string>(op_name, "type", elementwise_type);
|
||||
std::vector<float> coeff = {1.0, 1.0};
|
||||
engine_->AddOpAttr<PTuple<float>>(op_name, "coeff", coeff);
|
||||
}
|
||||
|
||||
} // namespace anakin
|
||||
} // namespace inference
|
||||
} // namespace paddle
|
||||
|
||||
REGISTER_ANAKIN_OP_CONVERTER(elementwise_add, ElementwiseAddOpConverter);
|
@ -0,0 +1,37 @@
|
||||
// Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
|
||||
//
|
||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||
// you may not use this file except in compliance with the License.
|
||||
// You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing, software
|
||||
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
// See the License for the specific language governing permissions and
|
||||
// limitations under the License.
|
||||
|
||||
#pragma once
|
||||
|
||||
#include "paddle/fluid/inference/anakin/convert/op_converter.h"
|
||||
|
||||
namespace paddle {
|
||||
namespace inference {
|
||||
namespace anakin {
|
||||
|
||||
class ElementwiseAddOpConverter : public AnakinOpConverter {
|
||||
public:
|
||||
ElementwiseAddOpConverter() = default;
|
||||
|
||||
virtual void operator()(const framework::proto::OpDesc &op,
|
||||
const framework::Scope &scope,
|
||||
bool test_mode) override;
|
||||
virtual ~ElementwiseAddOpConverter() {}
|
||||
|
||||
private:
|
||||
};
|
||||
|
||||
} // namespace anakin
|
||||
} // namespace inference
|
||||
} // namespace paddle
|
@ -0,0 +1,47 @@
|
||||
// Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
|
||||
//
|
||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||
// you may not use this file except in compliance with the License.
|
||||
// You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing, software
|
||||
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
// See the License for the specific language governing permissions and
|
||||
// limitations under the License.
|
||||
|
||||
#include "paddle/fluid/inference/anakin/convert/relu.h"
|
||||
#include <algorithm>
|
||||
#include <map>
|
||||
|
||||
using anakin::graph::GraphGlobalMem;
|
||||
using anakin::AK_FLOAT;
|
||||
using anakin::saber::NV;
|
||||
using anakin::saber::Shape;
|
||||
|
||||
namespace paddle {
|
||||
namespace inference {
|
||||
namespace anakin {
|
||||
|
||||
void ReluOpConverter::operator()(const framework::proto::OpDesc &op,
|
||||
const framework::Scope &scope,
|
||||
bool test_mode) {
|
||||
framework::OpDesc op_desc(op, nullptr);
|
||||
PADDLE_ENFORCE_EQ(op_desc.Input("X").size(), 1);
|
||||
PADDLE_ENFORCE_EQ(op_desc.Output("Out").size(), 1);
|
||||
|
||||
auto op_name = op_desc.Type() + ":" + op_desc.Output("Out").front();
|
||||
auto input_name = op_desc.Input("X").front();
|
||||
auto output_name = op_desc.Output("Out").front();
|
||||
|
||||
engine_->AddOp(op_name, "ReLU", {input_name}, {output_name});
|
||||
engine_->AddOpAttr(op_name, "alpha", 0);
|
||||
}
|
||||
|
||||
} // namespace anakin
|
||||
} // namespace inference
|
||||
} // namespace paddle
|
||||
|
||||
REGISTER_ANAKIN_OP_CONVERTER(relu, ReluOpConverter);
|
@ -0,0 +1,37 @@
|
||||
// Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
|
||||
//
|
||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||
// you may not use this file except in compliance with the License.
|
||||
// You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing, software
|
||||
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
// See the License for the specific language governing permissions and
|
||||
// limitations under the License.
|
||||
|
||||
#pragma once
|
||||
|
||||
#include <map>
|
||||
#include <string>
|
||||
#include "paddle/fluid/inference/anakin/convert/op_converter.h"
|
||||
|
||||
namespace paddle {
|
||||
namespace inference {
|
||||
namespace anakin {
|
||||
|
||||
class ReluOpConverter : public AnakinOpConverter {
|
||||
public:
|
||||
ReluOpConverter() = default;
|
||||
|
||||
virtual void operator()(const framework::proto::OpDesc &op,
|
||||
const framework::Scope &scope,
|
||||
bool test_mode) override;
|
||||
virtual ~ReluOpConverter() {}
|
||||
};
|
||||
|
||||
} // namespace anakin
|
||||
} // namespace inference
|
||||
} // namespace paddle
|
@ -0,0 +1,50 @@
|
||||
/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License. */
|
||||
|
||||
#include <gtest/gtest.h>
|
||||
#include "paddle/fluid/inference/anakin/convert/op_converter.h"
|
||||
#include "paddle/fluid/inference/anakin/convert/ut_helper.h"
|
||||
|
||||
namespace paddle {
|
||||
namespace inference {
|
||||
namespace anakin {
|
||||
|
||||
TEST(elementwise_op, native) {
|
||||
std::unordered_set<std::string> parameters;
|
||||
framework::Scope scope;
|
||||
AnakinConvertValidation validator(parameters, scope);
|
||||
validator.DeclInputVar("elementwise_add_x", {1, 1, 2, 2});
|
||||
validator.DeclInputVar("elementwise_y", {1, 1, 2, 2});
|
||||
validator.DeclOutputVar("elementwise_out", {1, 1, 2, 2});
|
||||
|
||||
// Prepare Op description
|
||||
framework::OpDesc desc;
|
||||
desc.SetType("elementwise_add");
|
||||
desc.SetInput("X", {"elementwise_add_x"});
|
||||
desc.SetInput("Y", {"elementwise_y"});
|
||||
desc.SetOutput("Out", {"elementwise_out"});
|
||||
|
||||
int axis = -1;
|
||||
desc.SetAttr("axis", axis);
|
||||
|
||||
validator.SetOp(*desc.Proto());
|
||||
validator.Execute(1);
|
||||
}
|
||||
|
||||
} // namespace anakin
|
||||
} // namespace inference
|
||||
} // namespace paddle
|
||||
|
||||
USE_OP(elementwise_add);
|
||||
USE_ANAKIN_CONVERTER(elementwise_add);
|
@ -0,0 +1,50 @@
|
||||
/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License. */
|
||||
|
||||
#include <gtest/gtest.h>
|
||||
#include "paddle/fluid/inference/anakin/convert/op_converter.h"
|
||||
#include "paddle/fluid/inference/anakin/convert/relu.h"
|
||||
#include "paddle/fluid/inference/anakin/convert/ut_helper.h"
|
||||
|
||||
namespace paddle {
|
||||
namespace inference {
|
||||
namespace anakin {
|
||||
|
||||
static void test_activation_op(const std::string &op_type) {
|
||||
auto *converter = Registry<AnakinOpConverter>::Global().Lookup(op_type);
|
||||
PADDLE_ENFORCE(converter != nullptr);
|
||||
std::unordered_set<std::string> parameters;
|
||||
framework::Scope scope;
|
||||
AnakinConvertValidation validator(parameters, scope);
|
||||
validator.DeclInputVar("act-X", {10, 6, 1, 1});
|
||||
validator.DeclOutputVar("act-Out", {10, 6, 1, 1});
|
||||
framework::OpDesc desc;
|
||||
desc.SetType(op_type);
|
||||
desc.SetInput("X", {"act-X"});
|
||||
desc.SetOutput("Out", {"act-Out"});
|
||||
|
||||
LOG(INFO) << "set OP";
|
||||
validator.SetOp(*desc.Proto());
|
||||
LOG(INFO) << "execute";
|
||||
|
||||
validator.Execute(5);
|
||||
}
|
||||
|
||||
TEST(sigm_op, test) { test_activation_op("relu"); }
|
||||
} // namespace anakin
|
||||
} // namespace inference
|
||||
} // namespace paddle
|
||||
|
||||
USE_OP(relu);
|
||||
USE_ANAKIN_CONVERTER(relu);
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in new issue