|
|
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
|
|
|
|
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
you may not use this file except in compliance with the License.
|
|
|
|
You may obtain a copy of the License at
|
|
|
|
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
See the License for the specific language governing permissions and
|
|
|
|
limitations under the License. */
|
|
|
|
|
|
|
|
#include "paddle/fluid/operators/l1_norm_op.h"
|
|
|
|
#include <memory>
|
|
|
|
|
|
|
|
namespace paddle {
|
|
|
|
namespace operators {
|
|
|
|
|
|
|
|
using framework::Tensor;
|
|
|
|
|
|
|
|
class L1NormOp : public framework::OperatorWithKernel {
|
|
|
|
public:
|
|
|
|
using framework::OperatorWithKernel::OperatorWithKernel;
|
|
|
|
|
|
|
|
void InferShape(framework::InferShapeContext* ctx) const override {
|
|
|
|
PADDLE_ENFORCE(ctx->HasInput("X"), "Input(X) should be not null.");
|
|
|
|
PADDLE_ENFORCE(ctx->HasOutput("Out"), "Output(Out) should be not null.");
|
|
|
|
|
|
|
|
ctx->SetOutputDim("Out", {1});
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
class L1NormGradOp : public framework::OperatorWithKernel {
|
|
|
|
public:
|
|
|
|
using framework::OperatorWithKernel::OperatorWithKernel;
|
|
|
|
|
|
|
|
void InferShape(framework::InferShapeContext* ctx) const override {
|
|
|
|
PADDLE_ENFORCE(ctx->HasInput("X"), "Input(X) should be not null.");
|
|
|
|
PADDLE_ENFORCE(ctx->HasInput(framework::GradVarName("Out")),
|
|
|
|
"Input(Out@GRAD) should be not null.");
|
|
|
|
PADDLE_ENFORCE(ctx->HasOutput(framework::GradVarName("X")),
|
|
|
|
"Output(X@GRAD) should be not null.");
|
|
|
|
|
|
|
|
ctx->SetOutputDim(framework::GradVarName("X"), ctx->GetInputDim("X"));
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
class L1NormOpMaker : public framework::OpProtoAndCheckerMaker {
|
|
|
|
public:
|
|
|
|
void Make() override {
|
|
|
|
AddInput("X", "(Tensor) The input of l1_norm op.");
|
|
|
|
AddOutput("Out", "(Scalar) The output of l1_norm op.");
|
|
|
|
AddComment(R"DOC(
|
|
|
|
L1 Norm Operator.
|
|
|
|
|
|
|
|
Computes the L1 norm of a tensor.
|
|
|
|
|
|
|
|
$$Out = \sum{|X|}$$
|
|
|
|
|
|
|
|
)DOC");
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
template <typename T>
|
|
|
|
class L1NormGradMaker : public framework::SingleGradOpMaker<T> {
|
|
|
|
public:
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
using framework::SingleGradOpMaker<T>::SingleGradOpMaker;
|
|
|
|
|
|
|
|
protected:
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
std::unique_ptr<T> Apply() const override {
|
|
|
|
std::unique_ptr<T> op(new T());
|
|
|
|
op->SetType("l1_norm_grad");
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
op->SetInput("X", this->Input("X"));
|
|
|
|
op->SetInput(framework::GradVarName("Out"), this->OutputGrad("Out"));
|
|
|
|
op->SetOutput(framework::GradVarName("X"), this->InputGrad("X"));
|
|
|
|
op->SetAttrMap(this->Attrs());
|
|
|
|
return op;
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
} // namespace operators
|
|
|
|
} // namespace paddle
|
|
|
|
|
|
|
|
namespace ops = paddle::operators;
|
|
|
|
REGISTER_OPERATOR(l1_norm, ops::L1NormOp, ops::L1NormOpMaker,
|
GradMaker for dygraph (#19706)
* refactor dygraph,test=develop
* fix failed unittest,test=develop
* polish code,test=develop
* check windows ci error,test=develop
try to fix windows ci error by np.allclose,test=develop
* polish vlog and profiler, test=develop
* try to fix preceding ops order,test=develop
* test transformer in windows ci, test=develop
* use python c-api to speed up tracer.trace,test=develop
* test=develop, fix docker with paddle nccl problem
* test=develop, add ut for debug string and gradient_accumulator
* test=develop, add tests for layer/gradient_accumulator/prepared_op
* test=develop, fix complie error for test_prepared_op
* test=develop, add more ut for dygraph
* test=develop, create API.spec for dygraph api change
* optimize grad maker; test=develop
* optimize grad maker
* test
* grad make optim; test=develop
* fix unittest bugs; test=develop
* add dygraph grad op maker and split_op
* grad op maker refactor; test=develop
* add dygraph grad maker; test=develop
* fix op deformable_conv_v1_op bug; test=develop
* fix deformable_conv prroi pool bugs;
* fix new op grad op maker bug; test=develop
* fix split by ref bug; test=develop
* fix dygraph auto prune bug; test=develop
* fix test_trace bug; test=develop
* fix fused emb seq pool bug; test=develop
* remove useless code in op_desc file; test=develop
* remove useless code, StrVarBaseNode; test=develop
* fix review issues; test=develop
* fix rank_loss grad maker; test=develop
* remove flag in VarBase; test=develop
* fix distributed_notify_op compile bug ; test=develop
* fix reshape op double grad; test=develop
* fix expand as op; test=develop
* add impertive type_defs.h for demo_train; test=develop
* fix inference lib cmake; test=develop
* fix inference lib; test=develop
* fix infernce_lib; test=develop
* fix inference cmake; test=develop
* fix inference lib; test=develop
* fix inference lib; test=develop
* remove condition dygraph grad maker, modify local name; test=develop
* fix split grad maker bug; test=develop
* fix pyramid_op bug; test=develop
* change travis time out limit; test=develop
* restore travis; test=develop
* change timeout limit; test=develop
5 years ago
|
|
|
ops::L1NormGradMaker<paddle::framework::OpDesc>,
|
|
|
|
ops::L1NormGradMaker<paddle::imperative::OpBase>);
|
|
|
|
REGISTER_OPERATOR(l1_norm_grad, ops::L1NormGradOp);
|
|
|
|
REGISTER_OP_CPU_KERNEL(
|
|
|
|
l1_norm, ops::L1NormKernel<paddle::platform::CPUDeviceContext, float>);
|
|
|
|
REGISTER_OP_CPU_KERNEL(
|
|
|
|
l1_norm_grad,
|
|
|
|
ops::L1NormGradKernel<paddle::platform::CPUDeviceContext, float>);
|