From 920aaa230424ed7c126febed46eccd8d5351544c Mon Sep 17 00:00:00 2001 From: lz Date: Wed, 16 Sep 2020 17:53:22 +0800 Subject: [PATCH] fix codex checking --- mindspore/lite/include/train_session.h | 1 - mindspore/lite/nnacl/fp32_grad/pooling_grad.h | 1 - mindspore/lite/src/ops/apply_momentum.cc | 4 ---- .../runtime/kernel/arm/fp32_grad/bn_grad.cc | 1 - .../arm/fp32_grad/convolution_grad_input.h | 3 --- .../runtime/kernel/arm/fp32_grad/depend.cc | 20 +++++++++---------- .../kernel/arm/fp32_grad/softmax_grad.cc | 5 ----- mindspore/lite/src/train/train_session.cc | 4 ---- 8 files changed, 9 insertions(+), 30 deletions(-) diff --git a/mindspore/lite/include/train_session.h b/mindspore/lite/include/train_session.h index 1a641b9119..7d09bbd783 100644 --- a/mindspore/lite/include/train_session.h +++ b/mindspore/lite/include/train_session.h @@ -18,7 +18,6 @@ #include #include #include -// #include "include/lite_session.h" #include "src/lite_session.h" namespace mindspore { diff --git a/mindspore/lite/nnacl/fp32_grad/pooling_grad.h b/mindspore/lite/nnacl/fp32_grad/pooling_grad.h index 4d27e21d7b..80fd98ccbd 100644 --- a/mindspore/lite/nnacl/fp32_grad/pooling_grad.h +++ b/mindspore/lite/nnacl/fp32_grad/pooling_grad.h @@ -23,7 +23,6 @@ extern "C" { #endif void AvgPoolingGrad(const float *input_ptr, float *output_ptr, PoolingParameter *pooling_param); -// void MaxPoolingGrad(const float *dy, const int *indices_ptr, float *output_ptr, PoolingParameter *pooling_param); void MaxPoolingGrad(const float *input_ptr, const float *dx_ptr, const float *dy_ptr, float *output_ptr, PoolingParameter *pooling_param); #ifdef __cplusplus diff --git a/mindspore/lite/src/ops/apply_momentum.cc b/mindspore/lite/src/ops/apply_momentum.cc index fa86ad0242..12a061522c 100644 --- a/mindspore/lite/src/ops/apply_momentum.cc +++ b/mindspore/lite/src/ops/apply_momentum.cc @@ -65,10 +65,6 @@ int ApplyMomentum::InferShape(std::vector inputs, std::vector
  • ElementsNum() != inputs[1]->ElementsNum() || inputs[0]->ElementsNum() != inputs[3]->ElementsNum() || inputs[2]->ElementsNum() != 1 || inputs[4]->ElementsNum() != 1) { diff --git a/mindspore/lite/src/runtime/kernel/arm/fp32_grad/bn_grad.cc b/mindspore/lite/src/runtime/kernel/arm/fp32_grad/bn_grad.cc index fcbfd02bd7..0cc3b67134 100644 --- a/mindspore/lite/src/runtime/kernel/arm/fp32_grad/bn_grad.cc +++ b/mindspore/lite/src/runtime/kernel/arm/fp32_grad/bn_grad.cc @@ -58,7 +58,6 @@ int BNGradCPUKernel::Run() { auto *output_dx = out_tensors_.at(0); auto *output_scale = out_tensors_.at(1); auto *output_bias = out_tensors_.at(2); - // Tensor *bias = input[5]; int batch = input_x->Batch(); int channels = input_x->Channel(); int spatial = input_x->Height() * input_x->Width(); diff --git a/mindspore/lite/src/runtime/kernel/arm/fp32_grad/convolution_grad_input.h b/mindspore/lite/src/runtime/kernel/arm/fp32_grad/convolution_grad_input.h index 0090608f32..df0a3f84cd 100644 --- a/mindspore/lite/src/runtime/kernel/arm/fp32_grad/convolution_grad_input.h +++ b/mindspore/lite/src/runtime/kernel/arm/fp32_grad/convolution_grad_input.h @@ -40,9 +40,6 @@ class ConvolutionGradInputCPUKernel : public LiteKernel { private: float *workspace; }; - -// OpParameter *PopulateConvolutionGradInputParameter(const lite::Primitive *primitive); - } // namespace mindspore::kernel #endif // MINDSPORE_LITE_SRC_RUNTIME_KERNEL_ARM_FP32_GRAD_CONVOLUTION_GRAD_INPUT_H diff --git a/mindspore/lite/src/runtime/kernel/arm/fp32_grad/depend.cc b/mindspore/lite/src/runtime/kernel/arm/fp32_grad/depend.cc index 811c43c2d7..7700c84143 100644 --- a/mindspore/lite/src/runtime/kernel/arm/fp32_grad/depend.cc +++ b/mindspore/lite/src/runtime/kernel/arm/fp32_grad/depend.cc @@ -33,17 +33,15 @@ int DependCPUKernel::Init() { return RET_OK; } int DependCPUKernel::ReSize() { return 0; } int DependCPUKernel::Run() { -#if 0 - auto ret = Prepare(); - if (ret != RET_OK) { - MS_LOG(ERROR) << "Prepare failed."; - return RET_ERROR; - } - auto in = reinterpret_cast(in_tensors_.at(0)->MutableData()); - auto out = reinterpret_cast(out_tensors_.at(0)->MutableData()); - - memcpy(out, in, in_tensors_.at(0)->Size()); -#endif +// auto ret = Prepare(); +// if (ret != RET_OK) { +// MS_LOG(ERROR) << "Prepare failed."; +// return RET_ERROR; +// } +// auto in = reinterpret_cast(in_tensors_.at(0)->MutableData()); +// auto out = reinterpret_cast(out_tensors_.at(0)->MutableData()); +// +// memcpy(out, in, in_tensors_.at(0)->Size()); return RET_OK; } diff --git a/mindspore/lite/src/runtime/kernel/arm/fp32_grad/softmax_grad.cc b/mindspore/lite/src/runtime/kernel/arm/fp32_grad/softmax_grad.cc index 51f4b93272..e29fc89d25 100644 --- a/mindspore/lite/src/runtime/kernel/arm/fp32_grad/softmax_grad.cc +++ b/mindspore/lite/src/runtime/kernel/arm/fp32_grad/softmax_grad.cc @@ -22,11 +22,9 @@ #include "src/kernel_registry.h" #include "include/errorcode.h" -// using mindspore::kernel::KERNEL_ARCH::kCPU; using mindspore::lite::KernelRegistrar; using mindspore::lite::RET_ERROR; using mindspore::lite::RET_OK; -// using mindspore::schema::PrimitiveType_SoftMaxGrad; namespace mindspore::kernel { int SoftmaxGradCPUKernel::Init() { @@ -71,7 +69,6 @@ int SoftmaxGradCPUKernel::Init() { int SoftmaxGradCPUKernel::ReSize() { return RET_OK; } int SoftmaxGradCPUKernel::Run() { - // auto input_ptr = reinterpret_cast(in_tensors_.at(kInputIndex)->MutableData()); auto input_ptr = reinterpret_cast(in_tensors_.at(kInputIndex)->MutableData()); auto yt_ptr = reinterpret_cast(in_tensors_.at(1)->MutableData()); auto output_ptr = reinterpret_cast(out_tensors_.at(kOutputIndex)->MutableData()); @@ -85,7 +82,6 @@ kernel::LiteKernel *CpuSoftmaxGradFp32KernelCreator(const std::vectorExportBuf(buf, len); return nullptr; } @@ -79,9 +78,6 @@ int TrainSession::RunGraph(const session::KernelCallBack &before, const session: } MS_EXCEPTION_IF_NULL(this->context_); - // TODO(Emir) - // SetMaxWokerNum(context_->thread_num_); - // context_->running_ = true; lite::Executor executor; if (before == nullptr && after == nullptr) { return executor.Run(this->inputs_, this->outputs_, infference_kernels, this->context_->allocator.get());