You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
76 lines
4.1 KiB
76 lines
4.1 KiB
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License. */
|
|
|
|
#include "paddle/fluid/operators/activation_op.h"
|
|
#include "paddle/fluid/platform/float16.h"
|
|
|
|
namespace ops = paddle::operators;
|
|
namespace plat = paddle::platform;
|
|
|
|
#define REGISTER_ACTIVATION_CUDA_KERNEL(act_type, op_name, functor, \
|
|
grad_functor) \
|
|
REGISTER_OP_CUDA_KERNEL( \
|
|
act_type, \
|
|
ops::ActivationKernel<plat::CUDADeviceContext, ops::functor<float>>, \
|
|
ops::ActivationKernel<plat::CUDADeviceContext, ops::functor<double>>, \
|
|
ops::ActivationKernel<plat::CUDADeviceContext, \
|
|
ops::functor<plat::float16>>); \
|
|
REGISTER_OP_CUDA_KERNEL( \
|
|
act_type##_grad, ops::ActivationGradKernel<plat::CUDADeviceContext, \
|
|
ops::grad_functor<float>>, \
|
|
ops::ActivationGradKernel<plat::CUDADeviceContext, \
|
|
ops::grad_functor<double>>, \
|
|
ops::ActivationGradKernel<plat::CUDADeviceContext, \
|
|
ops::grad_functor<plat::float16>>);
|
|
|
|
FOR_EACH_ACTIVATION_OP(REGISTER_ACTIVATION_CUDA_KERNEL);
|
|
|
|
/* ======================== leaky relu register ============================ */
|
|
REGISTER_ACTIVATION_CUDA_KERNEL(leaky_relu, LeakyRelu, LeakyReluFunctor,
|
|
LeakyReluGradFunctor);
|
|
|
|
REGISTER_OP_CUDA_KERNEL(
|
|
leaky_relu_grad_grad,
|
|
ops::ActivationDoubleGradKernel<plat::CUDADeviceContext,
|
|
ops::LeakyReluGradGradFunctor<float>>,
|
|
ops::ActivationDoubleGradKernel<plat::CUDADeviceContext,
|
|
ops::LeakyReluGradGradFunctor<double>>,
|
|
ops::ActivationDoubleGradKernel<
|
|
plat::CUDADeviceContext, ops::LeakyReluGradGradFunctor<plat::float16>>);
|
|
/* ========================================================================== */
|
|
|
|
/* =========================== relu register ============================ */
|
|
REGISTER_ACTIVATION_CUDA_KERNEL(relu, Relu, ReluFunctor, ReluGradFunctor);
|
|
|
|
REGISTER_OP_CUDA_KERNEL(
|
|
relu_grad_grad,
|
|
ops::ActivationDoubleGradKernel<paddle::platform::CUDADeviceContext,
|
|
ops::ReluGradGradFunctor<float>>,
|
|
ops::ActivationDoubleGradKernel<paddle::platform::CUDADeviceContext,
|
|
ops::ReluGradGradFunctor<double>>,
|
|
ops::ActivationDoubleGradKernel<plat::CUDADeviceContext,
|
|
ops::ReluGradGradFunctor<plat::float16>>);
|
|
/* ========================================================================== */
|
|
|
|
/* =========================== square register ============================ */
|
|
REGISTER_ACTIVATION_CUDA_KERNEL(square, Square, SquareFunctor,
|
|
SquareGradFunctor);
|
|
|
|
REGISTER_OP_CUDA_KERNEL(
|
|
square_grad_grad,
|
|
ops::SquareDoubleGradKernel<paddle::platform::CUDADeviceContext,
|
|
ops::SquareGradGradFunctor<float>>,
|
|
ops::SquareDoubleGradKernel<paddle::platform::CUDADeviceContext,
|
|
ops::SquareGradGradFunctor<double>>,
|
|
ops::SquareDoubleGradKernel<plat::CUDADeviceContext,
|
|
ops::SquareGradGradFunctor<plat::float16>>);
|
|
/* ========================================================================== */
|