Merge pull request #14903 from NHZlX/add_conv_elementwise_pass
Add conv + elementwiseAdd passfor_weibo
commit
a9fb34fad8
@ -0,0 +1,91 @@
|
|||||||
|
// Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
|
||||||
|
//
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
//
|
||||||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
|
#include <string>
|
||||||
|
|
||||||
|
#include "paddle/fluid/framework/ir/conv_elementwise_add_fuse_pass.h"
|
||||||
|
#include "paddle/fluid/framework/ir/graph_viz_pass.h"
|
||||||
|
|
||||||
|
namespace paddle {
|
||||||
|
namespace framework {
|
||||||
|
namespace ir {
|
||||||
|
|
||||||
|
#define GET_IR_NODE(node__) GET_IR_NODE_FROM_SUBGRAPH(node__, node__, pattern);
|
||||||
|
#define GET_NODES \
|
||||||
|
GET_IR_NODE(conv_op); \
|
||||||
|
GET_IR_NODE(conv_out); \
|
||||||
|
GET_IR_NODE(conv_filter); \
|
||||||
|
GET_IR_NODE(elementwise_add_op); \
|
||||||
|
GET_IR_NODE(elementwise_add_in_y); \
|
||||||
|
GET_IR_NODE(elementwise_add_out);
|
||||||
|
|
||||||
|
std::unique_ptr<ir::Graph> ConvElementwiseAddFusePass::ApplyImpl(
|
||||||
|
std::unique_ptr<ir::Graph> graph) const {
|
||||||
|
const std::string pattern_name = "conv_elementwise_add_fuse";
|
||||||
|
FusePassBase::Init(pattern_name, graph.get());
|
||||||
|
|
||||||
|
GraphPatternDetector gpd;
|
||||||
|
auto* x = gpd.mutable_pattern()
|
||||||
|
->NewNode("x")
|
||||||
|
->assert_is_op_input("conv2d", "Input")
|
||||||
|
->AsInput();
|
||||||
|
|
||||||
|
patterns::ConvElementwiseadd pattern(gpd.mutable_pattern(), pattern_name);
|
||||||
|
pattern(x);
|
||||||
|
|
||||||
|
auto handler = [&](const GraphPatternDetector::subgraph_t& subgraph,
|
||||||
|
Graph* g) {
|
||||||
|
GET_NODES;
|
||||||
|
|
||||||
|
auto base_op_desc = *conv_op->Op()->Proto();
|
||||||
|
std::string bias_name = elementwise_add_in_y->Name();
|
||||||
|
std::string output_name = elementwise_add_out->Name();
|
||||||
|
|
||||||
|
std::string act_type = "identity";
|
||||||
|
framework::OpDesc new_op_desc(base_op_desc, nullptr);
|
||||||
|
new_op_desc.SetType("conv2d_fusion");
|
||||||
|
new_op_desc.SetInput("Bias", {bias_name});
|
||||||
|
new_op_desc.SetInput("ResidualData", {});
|
||||||
|
new_op_desc.SetAttr("activation", act_type);
|
||||||
|
new_op_desc.SetOutput("Output", {output_name});
|
||||||
|
new_op_desc.SetAttr("is_test", true);
|
||||||
|
new_op_desc.SetAttr("use_cudnn", false);
|
||||||
|
new_op_desc.Flush();
|
||||||
|
|
||||||
|
// Create a new node for the fused op.
|
||||||
|
auto* new_conv_op = graph->CreateOpNode(&new_op_desc);
|
||||||
|
|
||||||
|
// Link inputs and outputs.
|
||||||
|
PADDLE_ENFORCE(subgraph.count(x));
|
||||||
|
auto* conv_in_node = subgraph.at(x);
|
||||||
|
|
||||||
|
IR_NODE_LINK_TO(conv_in_node, new_conv_op); // Input
|
||||||
|
IR_NODE_LINK_TO(conv_filter, new_conv_op); // Filter
|
||||||
|
IR_NODE_LINK_TO(elementwise_add_in_y, new_conv_op); // Bias
|
||||||
|
IR_NODE_LINK_TO(new_conv_op, elementwise_add_out); // Output
|
||||||
|
|
||||||
|
// Delete the unneeded nodes.
|
||||||
|
GraphSafeRemoveNodes(graph.get(), {conv_op, conv_out, elementwise_add_op});
|
||||||
|
};
|
||||||
|
|
||||||
|
gpd(graph.get(), handler);
|
||||||
|
return graph;
|
||||||
|
}
|
||||||
|
|
||||||
|
} // namespace ir
|
||||||
|
} // namespace framework
|
||||||
|
} // namespace paddle
|
||||||
|
|
||||||
|
REGISTER_PASS(conv_elementwise_add_fuse_pass,
|
||||||
|
paddle::framework::ir::ConvElementwiseAddFusePass);
|
@ -0,0 +1,33 @@
|
|||||||
|
// Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
|
||||||
|
//
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
//
|
||||||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
#include "paddle/fluid/framework/ir/fuse_pass_base.h"
|
||||||
|
#include "paddle/fluid/framework/ir/graph_pattern_detector.h"
|
||||||
|
|
||||||
|
namespace paddle {
|
||||||
|
namespace framework {
|
||||||
|
namespace ir {
|
||||||
|
|
||||||
|
class ConvElementwiseAddFusePass : public FusePassBase {
|
||||||
|
public:
|
||||||
|
virtual ~ConvElementwiseAddFusePass() {}
|
||||||
|
|
||||||
|
protected:
|
||||||
|
std::unique_ptr<ir::Graph> ApplyImpl(std::unique_ptr<ir::Graph> graph) const;
|
||||||
|
};
|
||||||
|
|
||||||
|
} // namespace ir
|
||||||
|
} // namespace framework
|
||||||
|
} // namespace paddle
|
@ -1,5 +1,5 @@
|
|||||||
op_library(tensorrt_engine_op DEPS tensorrt_engine tensorrt_converter)
|
op_library(tensorrt_engine_op DEPS tensorrt_engine tensorrt_converter)
|
||||||
file(APPEND ${pybind_file} "USE_CUDA_ONLY_OP(tensorrt_engine);\n")
|
file(APPEND ${pybind_file} "USE_NO_KERNEL_OP(tensorrt_engine);\n")
|
||||||
nv_test(test_tensorrt_engine_op SRCS tensorrt_engine_op_test.cc
|
nv_test(test_tensorrt_engine_op SRCS tensorrt_engine_op_test.cc
|
||||||
DEPS tensorrt_engine_op
|
DEPS tensorrt_engine_op
|
||||||
analysis)
|
analysis)
|
||||||
|
@ -1,24 +0,0 @@
|
|||||||
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
|
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
you may not use this file except in compliance with the License.
|
|
||||||
You may obtain a copy of the License at
|
|
||||||
|
|
||||||
http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
|
|
||||||
Unless required by applicable law or agreed to in writing, software
|
|
||||||
distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
See the License for the specific language governing permissions and
|
|
||||||
limitations under the License. */
|
|
||||||
|
|
||||||
#include "paddle/fluid/operators/tensorrt/tensorrt_engine_op.h"
|
|
||||||
|
|
||||||
namespace ops = paddle::operators;
|
|
||||||
|
|
||||||
REGISTER_OP_CUDA_KERNEL(
|
|
||||||
tensorrt_engine,
|
|
||||||
ops::TensorRTEngineKernel<paddle::platform::CUDADeviceContext, float>,
|
|
||||||
ops::TensorRTEngineKernel<paddle::platform::CUDADeviceContext, double>,
|
|
||||||
ops::TensorRTEngineKernel<paddle::platform::CUDADeviceContext, int>,
|
|
||||||
ops::TensorRTEngineKernel<paddle::platform::CUDADeviceContext, int64_t>);
|
|
File diff suppressed because it is too large
Load Diff
Loading…
Reference in new issue