From 22f03c3981ae930e608b3e53dcaf32c85408be55 Mon Sep 17 00:00:00 2001
From: dongzhihong <dzhwinter@gmail.com>
Date: Tue, 8 Aug 2017 23:35:48 +0800
Subject: [PATCH] "fix clang format"

---
 paddle/operators/add_op.cc           | 1 +
 paddle/operators/add_op_test.cc      | 1 -
 paddle/operators/cross_entropy_op.cc | 5 +++--
 paddle/operators/mul_op.cc           | 1 +
 paddle/operators/mul_op.cu           | 1 +
 paddle/operators/mul_op.h            | 1 +
 paddle/operators/sigmoid_op.cu       | 1 +
 paddle/operators/softmax_op.cc       | 1 +
 paddle/operators/softmax_op.cu       | 1 +
 9 files changed, 10 insertions(+), 3 deletions(-)

diff --git a/paddle/operators/add_op.cc b/paddle/operators/add_op.cc
index 64f856ac6f..086245ef62 100644
--- a/paddle/operators/add_op.cc
+++ b/paddle/operators/add_op.cc
@@ -57,5 +57,6 @@ class AddOpGrad : public framework::OperatorWithKernel {
 namespace ops = paddle::operators;
 REGISTER_OP(add_two, ops::AddOp, ops::AddOpMaker);
 REGISTER_GRADIENT_OP(add_two, add_two_grad, ops::AddOpGrad);
+
 REGISTER_OP_CPU_KERNEL(add_two,
                        ops::AddKernel<paddle::platform::CPUPlace, float>);
diff --git a/paddle/operators/add_op_test.cc b/paddle/operators/add_op_test.cc
index 4f33f46bb2..bf529defb2 100644
--- a/paddle/operators/add_op_test.cc
+++ b/paddle/operators/add_op_test.cc
@@ -17,7 +17,6 @@ limitations under the License. */
 #include "paddle/framework/op_registry.h"
 
 USE_OP(add_two);
-// USE_OP(add_two_grad);
 
 TEST(AddOp, GetOpProto) {
   auto& protos = paddle::framework::OpRegistry::protos();
diff --git a/paddle/operators/cross_entropy_op.cc b/paddle/operators/cross_entropy_op.cc
index 7c03fbd9e9..c813d54e17 100644
--- a/paddle/operators/cross_entropy_op.cc
+++ b/paddle/operators/cross_entropy_op.cc
@@ -70,8 +70,9 @@ OnehotCrossEntropy Operator.
 namespace ops = paddle::operators;
 REGISTER_OP(onehot_cross_entropy, ops::OnehotCrossEntropyOp,
             ops::OnehotCrossEntropyOpMaker);
-REGISTER_OP_CPU_KERNEL(onehot_cross_entropy,
-                       ops::OnehotCrossEntropyOpKernel<paddle::platform::CPUPlace, float>);
+REGISTER_OP_CPU_KERNEL(
+    onehot_cross_entropy,
+    ops::OnehotCrossEntropyOpKernel<paddle::platform::CPUPlace, float>);
 REGISTER_GRADIENT_OP(onehot_cross_entropy, onehot_cross_entropy_grad,
                      ops::OnehotCrossEntropyGradientOp);
 REGISTER_OP_CPU_KERNEL(
diff --git a/paddle/operators/mul_op.cc b/paddle/operators/mul_op.cc
index 06bc6172db..db81fd555d 100644
--- a/paddle/operators/mul_op.cc
+++ b/paddle/operators/mul_op.cc
@@ -67,4 +67,5 @@ class MulOpGrad : public framework::OperatorWithKernel {
 namespace ops = paddle::operators;
 REGISTER_OP(mul, ops::MulOp, ops::MulOpMaker);
 REGISTER_GRADIENT_OP(mul, mul_grad, ops::MulOpGrad);
+
 REGISTER_OP_CPU_KERNEL(mul, ops::MulKernel<paddle::platform::CPUPlace, float>);
diff --git a/paddle/operators/mul_op.cu b/paddle/operators/mul_op.cu
index 346a7e505d..43debbc21a 100644
--- a/paddle/operators/mul_op.cu
+++ b/paddle/operators/mul_op.cu
@@ -16,4 +16,5 @@
 #include "paddle/operators/mul_op.h"
 
 namespace ops = paddle::operators;
+
 REGISTER_OP_GPU_KERNEL(mul, ops::MulKernel<paddle::platform::GPUPlace, float>);
diff --git a/paddle/operators/mul_op.h b/paddle/operators/mul_op.h
index 67c3958149..ab12631c03 100644
--- a/paddle/operators/mul_op.h
+++ b/paddle/operators/mul_op.h
@@ -45,5 +45,6 @@ class MulKernel : public framework::OpKernel {
     Z.device(place) = X.contract(Y, dim_pair);
   }
 };
+
 }  // namespace operators
 }  // namespace paddle
diff --git a/paddle/operators/sigmoid_op.cu b/paddle/operators/sigmoid_op.cu
index 9518c3091a..1a50dfe14a 100644
--- a/paddle/operators/sigmoid_op.cu
+++ b/paddle/operators/sigmoid_op.cu
@@ -16,6 +16,7 @@
 #include "paddle/operators/sigmoid_op.h"
 
 namespace ops = paddle::operators;
+
 REGISTER_OP_GPU_KERNEL(sigmoid,
                        ops::SigmoidKernel<paddle::platform::GPUPlace, float>);
 REGISTER_OP_GPU_KERNEL(
diff --git a/paddle/operators/softmax_op.cc b/paddle/operators/softmax_op.cc
index 1143fb66c6..3dd4e86918 100644
--- a/paddle/operators/softmax_op.cc
+++ b/paddle/operators/softmax_op.cc
@@ -63,6 +63,7 @@ class SoftmaxOpGrad : public framework::OperatorWithKernel {
 }  // namespace paddle
 
 namespace ops = paddle::operators;
+
 REGISTER_OP(softmax, ops::SoftmaxOp, ops::SoftmaxOpMaker);
 REGISTER_OP_CPU_KERNEL(softmax,
                        ops::SoftmaxKernel<paddle::platform::CPUPlace, float>);
diff --git a/paddle/operators/softmax_op.cu b/paddle/operators/softmax_op.cu
index 92d2214273..2e99a89699 100644
--- a/paddle/operators/softmax_op.cu
+++ b/paddle/operators/softmax_op.cu
@@ -16,6 +16,7 @@
 #include "paddle/operators/softmax_op.h"
 
 namespace ops = paddle::operators;
+
 REGISTER_OP_GPU_KERNEL(softmax,
                        ops::SoftmaxKernel<paddle::platform::GPUPlace, float>);
 REGISTER_OP_GPU_KERNEL(