|
|
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
|
|
|
|
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
you may not use this file except in compliance with the License.
|
|
|
|
You may obtain a copy of the License at
|
|
|
|
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
See the License for the specific language governing permissions and
|
|
|
|
limitations under the License. */
|
|
|
|
|
|
|
|
#include "paddle/fluid/operators/conv_op.h"
|
|
|
|
|
|
|
|
#include <memory>
|
|
|
|
#include <string>
|
|
|
|
#include <vector>
|
|
|
|
|
|
|
|
#ifdef PADDLE_WITH_CUDA
|
|
|
|
#include "paddle/fluid/operators/conv_cudnn_op_cache.h"
|
|
|
|
#include "paddle/fluid/platform/cudnn_helper.h"
|
|
|
|
#endif
|
|
|
|
#ifdef PADDLE_WITH_MKLDNN
|
|
|
|
#include "paddle/fluid/platform/mkldnn_helper.h"
|
|
|
|
#endif
|
|
|
|
#include "paddle/fluid/platform/cudnn_workspace_helper.h"
|
|
|
|
|
|
|
|
namespace paddle {
|
|
|
|
namespace operators {
|
|
|
|
|
|
|
|
std::vector<int64_t> ConvOp::ComputeOutputShape(
|
|
|
|
framework::InferShapeContext* ctx) const {
|
|
|
|
OP_INOUT_CHECK(ctx->HasInput("Input"), "Input", "Input", "Conv");
|
|
|
|
OP_INOUT_CHECK(ctx->HasInput("Filter"), "Input", "Filter", "Conv");
|
|
|
|
|
|
|
|
auto in_dims = ctx->GetInputDim("Input");
|
|
|
|
auto filter_dims = ctx->GetInputDim("Filter");
|
|
|
|
|
|
|
|
std::vector<int> strides = ctx->Attrs().Get<std::vector<int>>("strides");
|
|
|
|
std::vector<int> paddings = ctx->Attrs().Get<std::vector<int>>("paddings");
|
|
|
|
std::string padding_algorithm =
|
|
|
|
ctx->Attrs().Get<std::string>("padding_algorithm");
|
|
|
|
int groups = ctx->Attrs().Get<int>("groups");
|
|
|
|
std::vector<int> dilations = ctx->Attrs().Get<std::vector<int>>("dilations");
|
|
|
|
const std::string data_format = ctx->Attrs().Get<std::string>("data_format");
|
|
|
|
|
|
|
|
// MKL-DNN Kernels are using NCHW order of dims description
|
|
|
|
// so we ignore data_format consideration for MKL-DNN kernel
|
|
|
|
const bool channel_last = (this->IsMKLDNNType() == false) &&
|
|
|
|
(data_format == "NHWC" || data_format == "NDHWC");
|
|
|
|
|
|
|
|
PADDLE_ENFORCE_EQ(
|
|
|
|
in_dims.size() == 4 || in_dims.size() == 5, true,
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
"The input of Op(Conv) should be a 4-D or 5-D Tensor. But "
|
|
|
|
"received: input's dimension is %u, input's shape is [%s].",
|
|
|
|
in_dims.size(), in_dims));
|
|
|
|
|
|
|
|
PADDLE_ENFORCE_EQ(
|
|
|
|
in_dims.size(), filter_dims.size(),
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
"The input's dimension and filter's dimension of "
|
|
|
|
"Op(Conv) should be equal. But received: the input's shape is [%s], "
|
|
|
|
"the input's dimension is %d; the filter's shape is [%s], "
|
|
|
|
"the filter's dimension is %d.",
|
|
|
|
in_dims, in_dims.size(), filter_dims, filter_dims.size()));
|
|
|
|
|
|
|
|
int in_sub_stride_size = in_dims.size() - strides.size();
|
|
|
|
PADDLE_ENFORCE_EQ(
|
|
|
|
in_dims.size(), strides.size() + 2U,
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
"The difference of input's dimension and Attr(strides)'s "
|
|
|
|
"length must be euqal to 2 for Op(Conv). "
|
|
|
|
"But received: input's dimension is %d, input's shape is [%s]; "
|
|
|
|
"Attr(stride)'s length is %d, Attr(stride) is [%s]; "
|
|
|
|
"difference of input's dimention and Attr(strides)'s length = %u.",
|
|
|
|
in_dims.size(), in_dims, strides.size(),
|
|
|
|
framework::make_ddim(strides), in_sub_stride_size));
|
|
|
|
|
|
|
|
const auto input_channels =
|
|
|
|
channel_last ? in_dims[in_dims.size() - 1] : in_dims[1];
|
|
|
|
|
|
|
|
PADDLE_ENFORCE_EQ(
|
|
|
|
input_channels, filter_dims[1] * groups,
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
"The number of input's channels should be equal to filter's channels "
|
|
|
|
"* groups for Op(Conv). But received: the input's channels is %d, "
|
|
|
|
"the input's shape is [%s]; the filter's channels is %d, the "
|
|
|
|
"filter's shape is [%s]; the groups is %d, the data_format is %s. "
|
|
|
|
"The error may come from wrong data_format setting.",
|
|
|
|
input_channels, in_dims, filter_dims[1], filter_dims, groups,
|
|
|
|
data_format));
|
|
|
|
PADDLE_ENFORCE_EQ(
|
|
|
|
filter_dims[0] % groups, 0,
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
"The number of output's channels (filter's first dimension) of "
|
|
|
|
"Op(Conv) should be divided by groups. But received: "
|
|
|
|
"the output channels is %d, the filter's shape is [%s], "
|
|
|
|
"the groups is %d.",
|
|
|
|
filter_dims[0], filter_dims, groups));
|
|
|
|
|
|
|
|
framework::DDim in_data_dims;
|
|
|
|
if (channel_last) {
|
|
|
|
in_data_dims = framework::slice_ddim(in_dims, 1, in_dims.size() - 1);
|
|
|
|
} else {
|
|
|
|
in_data_dims = framework::slice_ddim(in_dims, 2, in_dims.size());
|
|
|
|
}
|
|
|
|
|
|
|
|
framework::DDim filter_data_dims =
|
|
|
|
framework::slice_ddim(filter_dims, 2, filter_dims.size());
|
|
|
|
|
|
|
|
std::vector<int> ksize = framework::vectorize<int>(filter_data_dims);
|
|
|
|
UpdatePaddingAndDilation(&paddings, &dilations, padding_algorithm,
|
|
|
|
in_data_dims, strides, ksize);
|
|
|
|
|
|
|
|
std::vector<int64_t> output_shape({in_dims[0]});
|
|
|
|
if (!channel_last) {
|
|
|
|
output_shape.push_back(filter_dims[0]);
|
|
|
|
}
|
|
|
|
for (int i = 0; i < in_data_dims.size(); ++i) {
|
|
|
|
if ((!ctx->IsRuntime()) &&
|
|
|
|
(in_data_dims[i] <= 0 || filter_dims[i + 2] <= 0)) {
|
|
|
|
output_shape.push_back(-1);
|
|
|
|
} else {
|
|
|
|
output_shape.push_back(
|
|
|
|
ConvOutputSize(in_data_dims[i], filter_data_dims[i], dilations[i],
|
|
|
|
paddings[2 * i], paddings[2 * i + 1], strides[i]));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (channel_last) {
|
|
|
|
output_shape.push_back(filter_dims[0]);
|
|
|
|
}
|
|
|
|
|
|
|
|
return output_shape;
|
|
|
|
}
|
|
|
|
|
|
|
|
framework::OpKernelType ConvOp::GetExpectedKernelType(
|
|
|
|
const framework::ExecutionContext& ctx) const {
|
|
|
|
int customized_type_value =
|
|
|
|
framework::OpKernelType::kDefaultCustomizedTypeValue;
|
|
|
|
framework::LibraryType library{framework::LibraryType::kPlain};
|
|
|
|
// TODO(pzelazko-intel): enable MKLDNN layout when it's ready
|
|
|
|
auto input_data_type = OperatorWithKernel::IndicateVarDataType(ctx, "Input");
|
|
|
|
std::string data_format =
|
|
|
|
"AnyLayout"; // todo enable data layout when it's ready
|
|
|
|
framework::DataLayout layout = framework::StringToDataLayout(data_format);
|
|
|
|
|
|
|
|
#ifdef PADDLE_WITH_CUDA
|
|
|
|
if (platform::CanCUDNNBeUsed(ctx)) {
|
|
|
|
library = framework::LibraryType::kCUDNN;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
#ifdef PADDLE_WITH_MKLDNN
|
|
|
|
if (library == framework::LibraryType::kPlain &&
|
|
|
|
platform::CanMKLDNNBeUsed(ctx)) {
|
|
|
|
library = framework::LibraryType::kMKLDNN;
|
|
|
|
layout = framework::DataLayout::kMKLDNN;
|
|
|
|
customized_type_value =
|
|
|
|
(input_data_type == framework::DataTypeTrait<int8_t>::DataType() ||
|
|
|
|
input_data_type == framework::DataTypeTrait<uint8_t>::DataType())
|
|
|
|
? kConvMKLDNNINT8
|
|
|
|
: kConvMKLDNNFP32;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
if (input_data_type != framework::proto::VarType::INT8 &&
|
|
|
|
input_data_type != framework::proto::VarType::UINT8) {
|
|
|
|
auto filter_data_type = ctx.Input<Tensor>("Filter")->type();
|
|
|
|
PADDLE_ENFORCE_EQ(input_data_type, filter_data_type,
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
"input and filter data type should be consistent"));
|
|
|
|
}
|
|
|
|
if (input_data_type == framework::proto::VarType::FP16) {
|
|
|
|
PADDLE_ENFORCE_EQ(library, framework::LibraryType::kCUDNN,
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
"float16 can only be used when CUDNN is used"));
|
|
|
|
}
|
|
|
|
|
|
|
|
auto type = framework::OpKernelType(input_data_type, ctx.GetPlace(), layout,
|
|
|
|
library, customized_type_value);
|
|
|
|
return type;
|
|
|
|
}
|
|
|
|
|
|
|
|
framework::OpKernelType ConvOp::GetKernelTypeForVar(
|
|
|
|
const std::string& var_name, const Tensor& tensor,
|
|
|
|
const framework::OpKernelType& expected_kernel_type) const {
|
|
|
|
#ifdef PADDLE_WITH_MKLDNN
|
|
|
|
// Only input require reshaping, weights and
|
|
|
|
// bias are having shape in NCHW order
|
|
|
|
if ((var_name == "Input") &&
|
|
|
|
(expected_kernel_type.data_layout_ == framework::DataLayout::kMKLDNN) &&
|
|
|
|
(tensor.layout() != framework::DataLayout::kMKLDNN)) {
|
|
|
|
auto attrs = Attrs();
|
|
|
|
auto ar = paddle::framework::AttrReader(attrs);
|
|
|
|
const std::string data_format = ar.Get<std::string>("data_format");
|
|
|
|
auto dl = framework::StringToDataLayout(data_format);
|
|
|
|
// Some models may have intentionally set "AnyLayout" for pool
|
|
|
|
// op. Treat this as NCHW (default data_format value)
|
|
|
|
if (dl != framework::DataLayout::kAnyLayout) {
|
|
|
|
return framework::OpKernelType(expected_kernel_type.data_type_,
|
|
|
|
tensor.place(), dl);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
return framework::OpKernelType(expected_kernel_type.data_type_,
|
|
|
|
tensor.place(), tensor.layout());
|
|
|
|
}
|
|
|
|
|
|
|
|
void Conv2DOpMaker::Make() {
|
|
|
|
AddAttr<bool>("is_test",
|
|
|
|
"(bool, default false) Set to true for inference only, false "
|
|
|
|
"for training. Some layers may run faster when this is true.")
|
|
|
|
.SetDefault(false);
|
|
|
|
AddInput("Input",
|
|
|
|
"(Tensor) The input tensor of convolution operator. "
|
|
|
|
"The format of input tensor is NCHW or NHWC, where N is batch size, "
|
|
|
|
"C is the "
|
|
|
|
"number of channels, H is the height of the feature, "
|
|
|
|
"and W is the width of the feature.");
|
|
|
|
AddInput("Filter",
|
|
|
|
"(Tensor) The filter tensor of convolution operator. "
|
|
|
|
"The format of the filter tensor is MCHW, where M is the number of "
|
|
|
|
"output image channels, C is the number of input image channels, "
|
|
|
|
"H is the height of the filter, and W is the width of the filter. "
|
|
|
|
"If the groups attribute is greater than 1, C equals the number of "
|
|
|
|
"input image channels divided by the groups.");
|
|
|
|
AddInput("Bias",
|
|
|
|
"(Tensor) Bias to be added to each output of filter application."
|
|
|
|
"The format of output tensor is X (one-dimensional) of size equal"
|
|
|
|
"to the number of output channels. Only used with MKL-DNN.")
|
|
|
|
.AsDispensable();
|
|
|
|
AddInput("ResidualData",
|
|
|
|
"(Tensor) Tensor with residual data "
|
|
|
|
"to which convolution output will be added."
|
|
|
|
"Used with fuse_residual_connection fusion.")
|
|
|
|
.AsDispensable();
|
|
|
|
AddOutput("Output",
|
|
|
|
"(Tensor) The output tensor of convolution operator. "
|
|
|
|
"It has same data fromat and data type as the Input.");
|
|
|
|
AddAttr<std::vector<int>>("strides",
|
|
|
|
"(vector<int> default:{1, 1}), the "
|
|
|
|
"strides(h_stride, w_stride) of "
|
|
|
|
"convolution operator.")
|
|
|
|
.SetDefault({1, 1});
|
|
|
|
AddAttr<std::vector<int>>("paddings",
|
|
|
|
"(vector<int> default:{0, 0}), the "
|
|
|
|
"paddings(pad_height_top, pad_height_bottom, "
|
|
|
|
"pad_width_left, pad_wifth_right) of "
|
|
|
|
"convolution operator.")
|
|
|
|
.SetDefault({0, 0});
|
|
|
|
AddAttr<std::string>(
|
|
|
|
"padding_algorithm",
|
|
|
|
"(string, default \"EXPLICIT\") An optional string from: \"EXPLICIT\","
|
|
|
|
"\"SAME\",\"VALID\". Set to \"EXPLICIT\" for explicit padding. "
|
|
|
|
"Set to \"SAME\" or \"VALID\" for algorithm of padding. ")
|
|
|
|
.SetDefault("EXPLICIT");
|
|
|
|
AddAttr<int>(
|
|
|
|
"groups",
|
|
|
|
"(int default:1), the groups number of the convolution operator. "
|
|
|
|
"According to grouped convolution in Alex Krizhevsky's Deep CNN paper: "
|
|
|
|
"when group=2, the first half of the filters is only connected to the "
|
|
|
|
"first half of the input channels, while the second half of the filters "
|
|
|
|
"is only connected to the second half of the input channels.")
|
|
|
|
.SetDefault(1);
|
|
|
|
AddAttr<std::vector<int>>("dilations",
|
|
|
|
"(vector<int> default:{1, 1}), the "
|
|
|
|
"dilations(h_dilation, w_dilation) of "
|
|
|
|
"convolution operator.")
|
|
|
|
.SetDefault({1, 1});
|
|
|
|
AddAttr<bool>(
|
|
|
|
"use_cudnn",
|
|
|
|
"(bool, default false) Only used in cudnn kernel, need install cudnn")
|
|
|
|
.SetDefault(false);
|
|
|
|
AddAttr<bool>("fuse_relu_before_depthwise_conv",
|
|
|
|
"(bool, default false) Only used in cuda depthwise kernel")
|
|
|
|
.SetDefault(false);
|
|
|
|
AddAttr<bool>("use_mkldnn",
|
|
|
|
"(bool, default false) Only used in mkldnn kernel")
|
|
|
|
.SetDefault(false);
|
|
|
|
AddAttr<bool>("use_quantizer",
|
|
|
|
"(bool, default false) "
|
|
|
|
"Set to true for operators that should be quantized and use "
|
|
|
|
"int8 kernel. "
|
|
|
|
"Only used on CPU.")
|
|
|
|
.SetDefault(false);
|
|
|
|
AddAttr<bool>("fuse_relu", "(bool, default false) Only used in mkldnn kernel")
|
|
|
|
.SetDefault(false);
|
|
|
|
AddAttr<bool>("fuse_brelu",
|
|
|
|
"(bool, default false) Only used in mkldnn kernel")
|
|
|
|
.SetDefault(false);
|
|
|
|
AddAttr<float>("fuse_brelu_threshold",
|
|
|
|
"(float, default false 6.0) Only used in mkldnn kernel")
|
|
|
|
.SetDefault(6.0f);
|
|
|
|
AddAttr<std::string>("fuse_activation",
|
|
|
|
"(string, default \"\") Only used in mkldnn kernel")
|
|
|
|
.SetDefault("");
|
|
|
|
AddAttr<float>("fuse_alpha",
|
|
|
|
"(float, default 0.0) Only used in mkldnn kernel")
|
|
|
|
.SetDefault(0.0f);
|
|
|
|
AddAttr<float>("fuse_beta", "(float, default 0.0) Only used in mkldnn kernel")
|
|
|
|
.SetDefault(0.0f);
|
|
|
|
AddAttr<bool>("fuse_residual_connection",
|
|
|
|
"(bool, default false) Only used in mkldnn kernel. Used "
|
|
|
|
"whenever convolution output is as an input to residual "
|
|
|
|
"connection.")
|
|
|
|
.SetDefault(false);
|
|
|
|
AddAttr<float>("Scale_in",
|
|
|
|
"Scale_in to be used for int8 input data."
|
|
|
|
"Only used with MKL-DNN INT8.")
|
|
|
|
.SetDefault(1.0f);
|
|
|
|
AddAttr<float>("Scale_out",
|
|
|
|
"Scale_out to be used for int8 output data."
|
|
|
|
"Only used with MKL-DNN INT8.")
|
|
|
|
.SetDefault(1.0f);
|
|
|
|
AddAttr<float>("Scale_in_eltwise",
|
|
|
|
"Scale_in_eltwise to be used for int8 eltwise input data."
|
|
|
|
"Only used with MKL-DNN INT8.")
|
|
|
|
.SetDefault(1.0f);
|
|
|
|
AddAttr<std::vector<float>>("Scale_weights",
|
|
|
|
"Scale_weights to be used for int8 weights data."
|
|
|
|
"Only used with MKL-DNN INT8.")
|
|
|
|
.SetDefault({1.0f});
|
|
|
|
AddAttr<bool>("force_fp32_output",
|
|
|
|
"(bool, default false) Force INT8 kernel output FP32, only "
|
|
|
|
"used in MKL-DNN INT8")
|
|
|
|
.SetDefault(false);
|
|
|
|
AddAttr<std::string>(
|
|
|
|
"data_format",
|
|
|
|
"(string, default NCHW) Only used in "
|
|
|
|
"An optional string from: \"NHWC\", \"NCHW\". "
|
|
|
|
"Defaults to \"NHWC\". Specify the data format of the output data, "
|
|
|
|
"the input will be transformed automatically. ")
|
|
|
|
.SetDefault("NCHW");
|
|
|
|
// TODO(dzhwinter): need to registered layout transform function
|
|
|
|
AddAttr<int>("workspace_size_MB",
|
|
|
|
"Only used in cudnn kernel. Need set use_cudnn to true."
|
|
|
|
"workspace size for cudnn, in MB, "
|
|
|
|
"workspace is a section of GPU memory which will be "
|
|
|
|
"allocated/freed each time the operator runs, larger "
|
|
|
|
"workspace size can increase performance but also requires "
|
|
|
|
"better hardware. This size should be chosen carefully.")
|
|
|
|
.SetDefault(platform::GetDefaultConvWorkspaceSizeLimitMB());
|
|
|
|
AddAttr<bool>("exhaustive_search",
|
|
|
|
"(bool, default false) cuDNN has many algorithm to calculation "
|
|
|
|
"convolution, whether enable exhaustive search "
|
|
|
|
"for cuDNN convolution or not, default is False.")
|
|
|
|
.SetDefault(false);
|
|
|
|
|
|
|
|
AddComment(R"DOC(
|
|
|
|
Convolution Operator.
|
|
|
|
|
|
|
|
The convolution operation calculates the output based on the input, filter
|
|
|
|
and strides, paddings, dilations, groups parameters. The size of each dimension of the
|
|
|
|
parameters is checked in the infer-shape.
|
|
|
|
Input(Input) and Output(Output) are in NCHW or NHWC format. Where N is batch
|
|
|
|
size, C is the number of channels, H is the height of the feature, and W is
|
|
|
|
the width of the feature.
|
|
|
|
Filters(Input) is MCHW format format. Where M is the number of output image channels, C is
|
|
|
|
the number of input image channels, H is the height of the filter, and W
|
|
|
|
is the width of the filter.
|
|
|
|
Parameters(strides, paddings, dilations) are two elements. These two elements represent
|
|
|
|
height and width, respectively.
|
|
|
|
The input(X) size and output(Out) size may be different.
|
|
|
|
|
|
|
|
Example:
|
|
|
|
Input:
|
|
|
|
Input shape: $(N, C_{in}, H_{in}, W_{in})$
|
|
|
|
Filter shape: $(C_{out}, C_{in}, H_f, W_f)$
|
|
|
|
Output:
|
|
|
|
Output shape: $(N, C_{out}, H_{out}, W_{out})$
|
|
|
|
Where
|
|
|
|
$$
|
|
|
|
H_{out}= \frac{(H_{in} + pad_height_top + pad_height_bottom - (dilations[0] * (H_f - 1) + 1))}{strides[0]}+ 1 \\
|
|
|
|
W_{out}= \frac{(W_{in} + pad_width_left + pad_width_right - (dilations[1] * (W_f - 1) + 1))}{strides[1]}+ 1
|
|
|
|
$$
|
|
|
|
)DOC");
|
|
|
|
Apply();
|
|
|
|
}
|
|
|
|
|
|
|
|
void Conv3DOpMaker::Make() {
|
|
|
|
AddAttr<bool>("is_test",
|
|
|
|
"(bool, default false) Set to true for inference only, false "
|
|
|
|
"for training. Some layers may run faster when this is true.")
|
|
|
|
.SetDefault(false);
|
|
|
|
AddInput(
|
|
|
|
"Input",
|
|
|
|
"(Tensor) The input tensor of convolution operator. "
|
|
|
|
"The format of input tensor is NCDHW or NDHWC. Where N is batch size, C "
|
|
|
|
"is the "
|
|
|
|
"number of channels, D is the depth of the feature, H is the height of "
|
|
|
|
"the feature, "
|
|
|
|
"and W is the width of the feature.");
|
|
|
|
AddInput("Filter",
|
|
|
|
"(Tensor) The filter tensor of convolution operator. "
|
|
|
|
"The format of the filter tensor is MCDHW, where M is the number of "
|
|
|
|
"output image channels, C is the number of input image channels, "
|
|
|
|
"D is the depth of the filter, H is the height of the filter, and W "
|
|
|
|
"is the width of the filter."
|
|
|
|
"If the groups attribute is greater than 1, C equals the number of "
|
|
|
|
"input image channels divided by the groups.");
|
|
|
|
AddInput("ResidualData",
|
|
|
|
"(Tensor) Tensor with residual data "
|
|
|
|
"to which convolution output will be added."
|
|
|
|
"Used with fuse_residual_connection fusion.")
|
|
|
|
.AsDispensable();
|
|
|
|
AddOutput("Output",
|
|
|
|
"(Tensor) The output tensor of convolution operator."
|
|
|
|
"It has same data fromat and data type as the Input.");
|
|
|
|
AddAttr<std::vector<int>>("strides",
|
|
|
|
"(vector<int>, default:{1, 1, 1}), the "
|
|
|
|
"strides(d_stride, h_stride, w_stride) of "
|
|
|
|
"convolution operator.")
|
|
|
|
.SetDefault({1, 1, 1});
|
|
|
|
AddAttr<std::vector<int>>(
|
|
|
|
"paddings",
|
|
|
|
"(vector<int>, default:{0, 0, 0}), the "
|
|
|
|
"paddings(pad_depth_front, pad_depth_back, pad_height_top, "
|
|
|
|
"pad_height_bottom, pad_width_left, pad_width_right) of convolution "
|
|
|
|
"operator.")
|
|
|
|
.SetDefault({0, 0, 0});
|
|
|
|
AddAttr<std::string>(
|
|
|
|
"padding_algorithm",
|
|
|
|
"(string, default \"EXPLICIT\") An optional string from: \"EXPLICIT\","
|
|
|
|
"\"SAME\",\"VALID\". Set to \"EXPLICIT\" for explicit padding. "
|
|
|
|
"Set to \"SAME\" or \"VALID\" for algorithm of padding. ")
|
|
|
|
.SetDefault("EXPLICIT");
|
|
|
|
AddAttr<int>(
|
|
|
|
"groups",
|
|
|
|
"(int default:1), the groups number of the convolution operator. "
|
|
|
|
"According to grouped convolution in Alex Krizhevsky's Deep CNN paper: "
|
|
|
|
"when group=2, the first half of the filters is only connected to the "
|
|
|
|
"first half of the input channels, while the second half of the filters "
|
|
|
|
"is only connected to the second half of the input channels.")
|
|
|
|
.SetDefault(1);
|
|
|
|
AddAttr<std::vector<int>>("dilations",
|
|
|
|
"(vector<int> default:{1, 1, 1}), the "
|
|
|
|
"dilations(d_dilation, h_dilation, w_dilation) of "
|
|
|
|
"convolution operator.")
|
|
|
|
.SetDefault({1, 1, 1});
|
|
|
|
AddAttr<bool>(
|
|
|
|
"use_cudnn",
|
|
|
|
"(bool, default false) Only used in cudnn kernel, need install cudnn")
|
|
|
|
.SetDefault(false);
|
|
|
|
AddAttr<bool>("use_mkldnn",
|
|
|
|
"(bool, default false) Only used in mkldnn kernel")
|
|
|
|
.SetDefault(false);
|
|
|
|
AddAttr<bool>("fuse_relu", "(bool, default false) Only used in mkldnn kernel")
|
|
|
|
.SetDefault(false);
|
|
|
|
AddAttr<std::string>("fuse_activation",
|
|
|
|
"(string, default \"\") Only used in mkldnn kernel")
|
|
|
|
.SetDefault("");
|
|
|
|
AddAttr<float>("fuse_alpha",
|
|
|
|
"(float, default 0.0) Only used in mkldnn kernel")
|
|
|
|
.SetDefault(0.0f);
|
|
|
|
AddAttr<float>("fuse_beta", "(float, default 0.0) Only used in mkldnn kernel")
|
|
|
|
.SetDefault(0.0f);
|
|
|
|
AddAttr<bool>("fuse_residual_connection",
|
|
|
|
"(bool, default false) Only used in mkldnn kernel. Used "
|
|
|
|
"whenever convolution output is as an input to residual "
|
|
|
|
"connection.")
|
|
|
|
.SetDefault(false);
|
|
|
|
AddAttr<std::string>(
|
|
|
|
"data_format",
|
|
|
|
"(string, default NCDHW) Only used in "
|
|
|
|
"An optional string from: \"NDHWC\", \"NCDHW\". "
|
|
|
|
"Defaults to \"NDHWC\". Specify the data format of the output data, "
|
|
|
|
"the input will be transformed automatically. ")
|
|
|
|
.SetDefault("NCDHW");
|
|
|
|
AddAttr<bool>("force_fp32_output",
|
|
|
|
"(bool, default false) Only used in mkldnn INT8 kernel")
|
|
|
|
.SetDefault(false);
|
|
|
|
// TODO(dzhwinter): need to registered layout transform function
|
|
|
|
AddAttr<int>("workspace_size_MB",
|
|
|
|
"Only used in cudnn kernel. workspace size for cudnn, in MB, "
|
|
|
|
"workspace is a section of GPU memory which will be "
|
|
|
|
"allocated/freed each time the operator runs, larger "
|
|
|
|
"workspace size can increase performance but also requires "
|
|
|
|
"better hardware. This size should be chosen carefully.")
|
|
|
|
.SetDefault(platform::GetDefaultConvWorkspaceSizeLimitMB());
|
|
|
|
AddAttr<bool>("exhaustive_search",
|
|
|
|
"(bool, default false) cuDNN has many algorithm to calculation "
|
|
|
|
"convolution, whether enable exhaustive search "
|
|
|
|
"for cuDNN convolution or not, default is False.")
|
|
|
|
.SetDefault(false);
|
|
|
|
AddComment(R"DOC(
|
|
|
|
Convolution3D Operator.
|
|
|
|
|
|
|
|
The convolution operation calculates the output based on the input, filter
|
|
|
|
and strides, paddings, dilations, groups parameters. The size of each dimension of the
|
|
|
|
parameters is checked in the infer-shape.
|
|
|
|
Input(Input) and output(Output) are in NCDHW or NDHWC format, where N is batch
|
|
|
|
size, C is the number of channels,D is the depth of the feature, H is the height of
|
|
|
|
the feature, and W is the width of the feature.
|
|
|
|
Filters(Input) is MCDHW format, where M is the number of output image channels,
|
|
|
|
C is the number of input image channels, D is the depth of the filter,
|
|
|
|
H is the height of the filter, and W is the width of the filter.
|
|
|
|
Parameters(strides, paddings, dilations) are three elements. These three elements
|
|
|
|
represent depth, height and width, respectively.
|
|
|
|
The input(X) size and output(Out) size may be different.
|
|
|
|
|
|
|
|
Example:
|
|
|
|
Input:
|
|
|
|
Input shape: $(N, C_{in}, D_{in}, H_{in}, W_{in})$
|
|
|
|
Filter shape: $(C_{out}, C_{in}, D_f, H_f, W_f)$
|
|
|
|
Output:
|
|
|
|
Output shape: $(N, C_{out}, D_{out}, H_{out}, W_{out})$
|
|
|
|
Where
|
|
|
|
$$
|
|
|
|
D_{out}= \frac{(D_{in} + pad_depth_front + pad_depth_back - (dilations[0] * (D_f - 1) + 1))}{ strides[0]}+ 1 \\
|
|
|
|
H_{out}= \frac{(H_{in} + pad_height_top + pad_height_bottom - (dilations[1] * (H_f - 1) + 1))}{ strides[1]}+ 1 \\
|
|
|
|
W_{out}= \frac{(W_{in} + pad_width_left + pad_width_right - (dilations[2] * (W_f - 1) + 1))}{ strides[2]}+ 1
|
|
|
|
$$
|
|
|
|
)DOC");
|
|
|
|
Apply();
|
|
|
|
}
|
|
|
|
|
|
|
|
void ConvOpGrad::InferShape(framework::InferShapeContext* ctx) const {
|
|
|
|
auto in_dims = ctx->GetInputDim("Input");
|
|
|
|
auto filter_dims = ctx->GetInputDim("Filter");
|
|
|
|
if (ctx->HasOutput(framework::GradVarName("Input"))) {
|
|
|
|
ctx->SetOutputDim(framework::GradVarName("Input"), in_dims);
|
|
|
|
}
|
|
|
|
if (ctx->HasOutput(framework::GradVarName("Filter"))) {
|
|
|
|
ctx->SetOutputDim(framework::GradVarName("Filter"), filter_dims);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
framework::OpKernelType ConvOpGrad::GetExpectedKernelType(
|
|
|
|
const framework::ExecutionContext& ctx) const {
|
|
|
|
int customized_type_value =
|
|
|
|
framework::OpKernelType::kDefaultCustomizedTypeValue;
|
|
|
|
framework::LibraryType library_{framework::LibraryType::kPlain};
|
|
|
|
// TODO(pzelazko-intel): enable MKLDNN layout when it's ready
|
|
|
|
std::string data_format = "AnyLayout";
|
|
|
|
framework::DataLayout layout_ = framework::StringToDataLayout(data_format);
|
|
|
|
|
|
|
|
#ifdef PADDLE_WITH_CUDA
|
|
|
|
if (platform::CanCUDNNBeUsed(ctx)) {
|
|
|
|
library_ = framework::LibraryType::kCUDNN;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
#ifdef PADDLE_WITH_MKLDNN
|
|
|
|
if (library_ == framework::LibraryType::kPlain &&
|
|
|
|
platform::CanMKLDNNBeUsed(ctx)) {
|
|
|
|
const std::string data_format = ctx.Attr<std::string>("data_format");
|
|
|
|
library_ = framework::LibraryType::kMKLDNN;
|
|
|
|
layout_ = framework::DataLayout::kMKLDNN;
|
|
|
|
customized_type_value = kConvMKLDNNFP32;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
auto type = framework::OpKernelType(
|
|
|
|
OperatorWithKernel::IndicateVarDataType(ctx, "Input"), ctx.GetPlace(),
|
|
|
|
layout_, library_, customized_type_value);
|
|
|
|
return type;
|
|
|
|
}
|
|
|
|
|
|
|
|
framework::OpKernelType ConvOpGrad::GetKernelTypeForVar(
|
|
|
|
const std::string& var_name, const Tensor& tensor,
|
|
|
|
const framework::OpKernelType& expected_kernel_type) const {
|
|
|
|
#ifdef PADDLE_WITH_MKLDNN
|
|
|
|
// Only input require reshaping, weights and
|
|
|
|
// bias are having shape in NCHW order
|
|
|
|
if (((var_name == "Input") ||
|
|
|
|
(var_name == framework::GradVarName("Output"))) &&
|
|
|
|
(expected_kernel_type.data_layout_ == framework::DataLayout::kMKLDNN) &&
|
|
|
|
(tensor.layout() != framework::DataLayout::kMKLDNN)) {
|
|
|
|
auto attrs = Attrs();
|
|
|
|
auto ar = paddle::framework::AttrReader(attrs);
|
|
|
|
const std::string data_format = ar.Get<std::string>("data_format");
|
|
|
|
auto dl = framework::StringToDataLayout(data_format);
|
|
|
|
// Some models may have intentionally set "AnyLayout" for pool
|
|
|
|
// op. Treat this as NCHW (default data_format value)
|
|
|
|
if (dl != framework::DataLayout::kAnyLayout) {
|
|
|
|
return framework::OpKernelType(expected_kernel_type.data_type_,
|
|
|
|
tensor.place(), dl);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
return framework::OpKernelType(expected_kernel_type.data_type_,
|
|
|
|
tensor.place(), tensor.layout());
|
|
|
|
}
|
|
|
|
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
template <typename T>
|
|
|
|
class Conv2DGradMaker : public framework::SingleGradOpMaker<T> {
|
|
|
|
public:
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
using framework::SingleGradOpMaker<T>::SingleGradOpMaker;
|
|
|
|
|
|
|
|
void Apply(GradOpPtr<T> op) const override {
|
|
|
|
op->SetType(this->ForwardOpType() + "_grad");
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
op->SetInput("Input", this->Input("Input"));
|
|
|
|
op->SetInput("Filter", this->Input("Filter"));
|
|
|
|
op->SetInput("Bias", this->Input("Bias"));
|
|
|
|
op->SetInput(framework::GradVarName("Output"), this->OutputGrad("Output"));
|
|
|
|
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
op->SetOutput(framework::GradVarName("Input"), this->InputGrad("Input"));
|
|
|
|
op->SetOutput(framework::GradVarName("Filter"), this->InputGrad("Filter"));
|
|
|
|
op->SetOutput(framework::GradVarName("Bias"), this->InputGrad("Bias"));
|
|
|
|
op->SetAttrMap(this->Attrs());
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
template <typename T>
|
|
|
|
class Conv3DGradMaker : public framework::SingleGradOpMaker<T> {
|
|
|
|
public:
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
using framework::SingleGradOpMaker<T>::SingleGradOpMaker;
|
|
|
|
|
|
|
|
void Apply(GradOpPtr<T> op) const override {
|
|
|
|
op->SetType(this->ForwardOpType() + "_grad");
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
op->SetInput("Input", this->Input("Input"));
|
|
|
|
op->SetInput("Filter", this->Input("Filter"));
|
|
|
|
op->SetInput(framework::GradVarName("Output"), this->OutputGrad("Output"));
|
|
|
|
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
op->SetOutput(framework::GradVarName("Input"), this->InputGrad("Input"));
|
|
|
|
op->SetOutput(framework::GradVarName("Filter"), this->InputGrad("Filter"));
|
|
|
|
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
if (this->HasInput("ResidualData")) {
|
|
|
|
op->SetInput("ResidualData", this->Input("ResidualData"));
|
|
|
|
}
|
|
|
|
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
op->SetAttrMap(this->Attrs());
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Inputs: I, W, dO, ddI, ddW
|
|
|
|
* Outputs: ddO, dW, dI
|
|
|
|
*/
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
template <typename T>
|
|
|
|
class Conv2DDoubleGradMaker : public framework::SingleGradOpMaker<T> {
|
|
|
|
public:
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
using framework::SingleGradOpMaker<T>::SingleGradOpMaker;
|
|
|
|
|
|
|
|
void Apply(GradOpPtr<T> op) const override {
|
|
|
|
op->SetType(this->ForwardOpType() + "_grad");
|
|
|
|
// I, W, dO, ddI, ddW
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
op->SetInput("Input", this->Input("Input"));
|
|
|
|
op->SetInput("Filter", this->Input("Filter"));
|
|
|
|
op->SetInput("DOutput", this->Input(framework::GradVarName("Output")));
|
|
|
|
op->SetInput("DDInput", this->OutputGrad(framework::GradVarName("Input")));
|
|
|
|
op->SetInput("DDFilter",
|
|
|
|
this->OutputGrad(framework::GradVarName("Filter")));
|
|
|
|
|
|
|
|
// ddO, dI, dW
|
|
|
|
// Unlike grad op, double grad op does not use name@GRAD@GRAD
|
|
|
|
// as key of ops' inputs and outputs.
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
auto ddx = this->OutputGrad(framework::GradVarName("Input"));
|
|
|
|
auto ddw = this->OutputGrad(framework::GradVarName("Filter"));
|
|
|
|
|
|
|
|
op->SetOutput("DDOutput",
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
ddx.empty()
|
|
|
|
? this->EmptyInputGrad()
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
: this->InputGrad(framework::GradVarName("Output")));
|
|
|
|
op->SetOutput("DFilter", ddx.empty() ? this->EmptyInputGrad()
|
|
|
|
: this->InputGrad("Filter"));
|
|
|
|
op->SetOutput("DInput", ddw.empty() ? this->EmptyInputGrad()
|
|
|
|
: this->InputGrad("Input"));
|
|
|
|
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
op->SetAttrMap(this->Attrs());
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Inputs: I, W, dO, ddI, ddW
|
|
|
|
* Outputs: ddO, dW, dI
|
|
|
|
*/
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
template <typename T>
|
|
|
|
class Conv3DDoubleGradMaker : public framework::SingleGradOpMaker<T> {
|
|
|
|
public:
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
using framework::SingleGradOpMaker<T>::SingleGradOpMaker;
|
|
|
|
|
|
|
|
void Apply(GradOpPtr<T> op) const override {
|
|
|
|
op->SetType(this->ForwardOpType() + "_grad");
|
|
|
|
// I, W, dO, ddI, ddW
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
op->SetInput("Input", this->Input("Input"));
|
|
|
|
op->SetInput("Filter", this->Input("Filter"));
|
|
|
|
op->SetInput("DOutput", this->Input(framework::GradVarName("Output")));
|
|
|
|
op->SetInput("DDInput", this->OutputGrad(framework::GradVarName("Input")));
|
|
|
|
op->SetInput("DDFilter",
|
|
|
|
this->OutputGrad(framework::GradVarName("Filter")));
|
|
|
|
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
auto ddx = this->OutputGrad(framework::GradVarName("Input"));
|
|
|
|
auto ddw = this->OutputGrad(framework::GradVarName("Filter"));
|
|
|
|
|
|
|
|
op->SetOutput("DDOutput",
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
ddx.empty()
|
|
|
|
? this->EmptyInputGrad()
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
: this->InputGrad(framework::GradVarName("Output")));
|
|
|
|
op->SetOutput("DFilter", ddx.empty() ? this->EmptyInputGrad()
|
|
|
|
: this->InputGrad("Filter"));
|
|
|
|
op->SetOutput("DInput", ddw.empty() ? this->EmptyInputGrad()
|
|
|
|
: this->InputGrad("Input"));
|
|
|
|
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
op->SetAttrMap(this->Attrs());
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
void ConvOpDoubleGrad::InferShape(framework::InferShapeContext* ctx) const {
|
|
|
|
auto x_dims = ctx->GetInputDim("Input");
|
|
|
|
auto w_dims = ctx->GetInputDim("Filter");
|
|
|
|
auto do_dims = ctx->GetInputDim("DOutput");
|
|
|
|
|
|
|
|
if (ctx->HasOutput("DDOutput") &&
|
|
|
|
(ctx->HasInput("DDInput") || (ctx->HasInput("DDFilter")))) {
|
|
|
|
ctx->SetOutputDim("DDOutput", do_dims);
|
|
|
|
}
|
|
|
|
if (ctx->HasOutput("DFilter") && ctx->HasInput("DDInput")) {
|
|
|
|
ctx->SetOutputDim("DFilter", w_dims);
|
|
|
|
}
|
|
|
|
if (ctx->HasOutput("DInput") && ctx->HasInput("DDFilter")) {
|
|
|
|
ctx->SetOutputDim("DInput", x_dims);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
framework::OpKernelType ConvOpDoubleGrad::GetExpectedKernelType(
|
|
|
|
const framework::ExecutionContext& ctx) const {
|
|
|
|
int customized_type_value =
|
|
|
|
framework::OpKernelType::kDefaultCustomizedTypeValue;
|
|
|
|
framework::LibraryType library_{framework::LibraryType::kPlain};
|
|
|
|
std::string data_format = "AnyLayout";
|
|
|
|
framework::DataLayout layout_ = framework::StringToDataLayout(data_format);
|
|
|
|
|
|
|
|
#ifdef PADDLE_WITH_CUDA
|
|
|
|
if (platform::CanCUDNNBeUsed(ctx)) {
|
|
|
|
library_ = framework::LibraryType::kCUDNN;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
auto type = framework::OpKernelType(
|
|
|
|
OperatorWithKernel::IndicateVarDataType(ctx, "Input"), ctx.GetPlace(),
|
|
|
|
layout_, library_, customized_type_value);
|
|
|
|
return type;
|
|
|
|
}
|
|
|
|
|
|
|
|
} // namespace operators
|
|
|
|
} // namespace paddle
|
|
|
|
|
|
|
|
namespace ops = paddle::operators;
|
|
|
|
REGISTER_OPERATOR(conv2d, ops::ConvOp, ops::Conv2DOpMaker,
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
ops::ConvOpInferVarType,
|
|
|
|
ops::Conv2DGradMaker<paddle::framework::OpDesc>,
|
|
|
|
ops::Conv2DGradMaker<paddle::imperative::OpBase>);
|
|
|
|
REGISTER_OPERATOR(conv2d_grad, ops::ConvOpGrad,
|
|
|
|
ops::Conv2DDoubleGradMaker<paddle::framework::OpDesc>,
|
|
|
|
ops::Conv2DDoubleGradMaker<paddle::imperative::OpBase>);
|
|
|
|
REGISTER_OPERATOR(conv2d_grad_grad, ops::ConvOpDoubleGrad);
|
|
|
|
|
|
|
|
// depthwise convolution op
|
|
|
|
REGISTER_OPERATOR(depthwise_conv2d, ops::ConvOp, ops::Conv2DOpMaker,
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
ops::ConvOpInferVarType,
|
|
|
|
ops::Conv2DGradMaker<paddle::framework::OpDesc>,
|
|
|
|
ops::Conv2DGradMaker<paddle::imperative::OpBase>);
|
|
|
|
REGISTER_OPERATOR(depthwise_conv2d_grad, ops::ConvOpGrad);
|
|
|
|
|
|
|
|
REGISTER_OPERATOR(conv3d, ops::ConvOp, ops::Conv3DOpMaker,
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
ops::ConvOpInferVarType,
|
|
|
|
ops::Conv3DGradMaker<paddle::framework::OpDesc>,
|
|
|
|
ops::Conv3DGradMaker<paddle::imperative::OpBase>);
|
|
|
|
REGISTER_OPERATOR(conv3d_grad, ops::ConvOpGrad,
|
|
|
|
ops::Conv3DDoubleGradMaker<paddle::framework::OpDesc>,
|
|
|
|
ops::Conv3DDoubleGradMaker<paddle::imperative::OpBase>);
|
|
|
|
REGISTER_OPERATOR(conv3d_grad_grad, ops::ConvOpDoubleGrad);
|
|
|
|
|
|
|
|
// depthwise conv kernel
|
|
|
|
// TODO(xingzhaolong): neon kernel for mobile
|
|
|
|
REGISTER_OP_CPU_KERNEL(
|
|
|
|
depthwise_conv2d,
|
|
|
|
ops::GemmConvKernel<paddle::platform::CPUDeviceContext, float>,
|
|
|
|
ops::GemmConvKernel<paddle::platform::CPUDeviceContext, double>);
|
|
|
|
|
|
|
|
REGISTER_OP_CPU_KERNEL(
|
|
|
|
depthwise_conv2d_grad,
|
|
|
|
ops::GemmConvGradKernel<paddle::platform::CPUDeviceContext, float>,
|
|
|
|
ops::GemmConvGradKernel<paddle::platform::CPUDeviceContext, double>);
|
|
|
|
|
|
|
|
REGISTER_OP_CPU_KERNEL(
|
|
|
|
conv2d, ops::GemmConvKernel<paddle::platform::CPUDeviceContext, float>,
|
|
|
|
ops::GemmConvKernel<paddle::platform::CPUDeviceContext, double>);
|
|
|
|
REGISTER_OP_CPU_KERNEL(
|
|
|
|
conv2d_grad,
|
|
|
|
ops::GemmConvGradKernel<paddle::platform::CPUDeviceContext, float>,
|
|
|
|
ops::GemmConvGradKernel<paddle::platform::CPUDeviceContext, double>);
|
|
|
|
REGISTER_OP_CPU_KERNEL(
|
|
|
|
conv2d_grad_grad,
|
|
|
|
ops::GemmConvDoubleGradKernel<paddle::platform::CPUDeviceContext, float>,
|
|
|
|
ops::GemmConvDoubleGradKernel<paddle::platform::CPUDeviceContext, double>);
|
|
|
|
|
|
|
|
REGISTER_OP_CPU_KERNEL(
|
|
|
|
conv3d, ops::GemmConvKernel<paddle::platform::CPUDeviceContext, float>,
|
|
|
|
ops::GemmConvKernel<paddle::platform::CPUDeviceContext, double>);
|
|
|
|
REGISTER_OP_CPU_KERNEL(
|
|
|
|
conv3d_grad,
|
|
|
|
ops::GemmConvGradKernel<paddle::platform::CPUDeviceContext, float>,
|
|
|
|
ops::GemmConvGradKernel<paddle::platform::CPUDeviceContext, double>);
|
|
|
|
REGISTER_OP_CPU_KERNEL(
|
|
|
|
conv3d_grad_grad,
|
|
|
|
ops::GemmConvDoubleGradKernel<paddle::platform::CPUDeviceContext, float>,
|
|
|
|
ops::GemmConvDoubleGradKernel<paddle::platform::CPUDeviceContext, double>);
|