2. Make SwitchOrderLayer support for softmax activation 3. Fix bugsAdaptive_data_structure_for_SwitchOrderLayer
parent
475dd708b9
commit
1cdf149b6f
@ -1,89 +0,0 @@
|
||||
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License. */
|
||||
|
||||
#include "PixelSoftmaxLayer.h"
|
||||
#include "paddle/utils/Stat.h"
|
||||
|
||||
namespace paddle {
|
||||
|
||||
REGISTER_LAYER(pixel_softmax, PixelSoftmaxLayer);
|
||||
|
||||
bool PixelSoftmaxLayer::init(const LayerMap& layerMap,
|
||||
const ParameterMap& parameterMap) {
|
||||
/* Initialize the basic parent class */
|
||||
Layer::init(layerMap, parameterMap);
|
||||
auto& img_conf = config_.inputs(0).image_conf();
|
||||
inH_ =
|
||||
img_conf.has_img_size_y() ? img_conf.img_size_y() : img_conf.img_size();
|
||||
inW_ = img_conf.img_size();
|
||||
inC_ = img_conf.channels();
|
||||
createFunction(forward_, "NCHW2NHWC", FuncConfig());
|
||||
createFunction(backward_, "NHWC2NCHW", FuncConfig());
|
||||
inDims_ = TensorShape({0, inH_, inW_, inC_});
|
||||
outDims_ = TensorShape({0, inC_, inH_, inW_});
|
||||
return true;
|
||||
}
|
||||
|
||||
void PixelSoftmaxLayer::forward(PassType passType) {
|
||||
Layer::forward(passType);
|
||||
MatrixPtr input = inputLayers_[0]->getOutputValue();
|
||||
size_t batchSize = input->getHeight();
|
||||
// cout<<"useGpu:"<<useGpu(deviceId_)<<endl;
|
||||
Matrix::resizeOrCreate(
|
||||
tmpInput_, batchSize * inH_ * inW_, inC_, false, useGpu_);
|
||||
Matrix::resizeOrCreate(
|
||||
tmpOutput_, batchSize * inH_ * inW_, inC_, false, useGpu_);
|
||||
tmpOutput_->zeroMem();
|
||||
resetOutput(batchSize, inH_ * inW_ * inC_);
|
||||
inDims_.setDim(0, batchSize);
|
||||
outDims_.setDim(0, batchSize);
|
||||
|
||||
// switch NCHW to NHWC
|
||||
BufferArgs inputs;
|
||||
BufferArgs outputs;
|
||||
inputs.addArg(*getInputValue(0), inDims_);
|
||||
outputs.addArg(*tmpInput_, outDims_);
|
||||
forward_[0]->calc(inputs, outputs);
|
||||
// softmax forward and save softmax result into tmpMatrix_
|
||||
tmpInput_->softmax(*tmpOutput_);
|
||||
|
||||
// switch NHWC to NCHW
|
||||
BufferArgs inputs_1;
|
||||
BufferArgs outputs_1;
|
||||
inputs_1.addArg(*tmpOutput_, outDims_);
|
||||
outputs_1.addArg(*getOutputValue(), inDims_);
|
||||
backward_[0]->calc(inputs_1, outputs_1);
|
||||
}
|
||||
|
||||
void PixelSoftmaxLayer::backward(const UpdateCallback& callback) {
|
||||
(void)callback;
|
||||
REGISTER_TIMER_INFO("PixelSoftmaxBackward", getName().c_str());
|
||||
|
||||
// switch NCHW to NHWC
|
||||
BufferArgs inputs;
|
||||
BufferArgs outputs;
|
||||
inputs.addArg(*getOutputGrad(), inDims_);
|
||||
outputs.addArg(*tmpInput_, outDims_);
|
||||
forward_[0]->calc(inputs, outputs);
|
||||
// softmax backward and save grad result into tmpOutput_
|
||||
tmpInput_->softmaxBackward(*tmpOutput_);
|
||||
|
||||
// switch NHWC to NCHW
|
||||
BufferArgs inputs_1;
|
||||
BufferArgs outputs_1;
|
||||
inputs_1.addArg(*tmpInput_, outDims_);
|
||||
outputs_1.addArg(*getInputGrad(0), inDims_);
|
||||
backward_[0]->calc(inputs_1, outputs_1);
|
||||
}
|
||||
} // namespace paddle
|
@ -0,0 +1,112 @@
|
||||
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License. */
|
||||
|
||||
#include "SwitchOrderLayer.h"
|
||||
#include "paddle/utils/Stat.h"
|
||||
|
||||
namespace paddle {
|
||||
|
||||
REGISTER_LAYER(switch_order, SwitchOrderLayer);
|
||||
|
||||
bool SwitchOrderLayer::init(const LayerMap& layerMap,
|
||||
const ParameterMap& parameterMap) {
|
||||
/* Initialize the basic parent class */
|
||||
Layer::init(layerMap, parameterMap);
|
||||
auto& img_conf = config_.inputs(0).image_conf();
|
||||
size_t inH =
|
||||
img_conf.has_img_size_y() ? img_conf.img_size_y() : img_conf.img_size();
|
||||
size_t inW = img_conf.img_size();
|
||||
size_t inC = img_conf.channels();
|
||||
inDims_ = TensorShape({0, inC, inH, inW});
|
||||
outDims_ = TensorShape(4);
|
||||
|
||||
auto& reshape_conf = config_.reshape_conf();
|
||||
for (size_t i = 0; i < reshape_conf.heightaxis_size(); i++) {
|
||||
LOG(INFO) << "reshape height axis: " << reshape_conf.heightaxis(i);
|
||||
heightAxis_.push_back(reshape_conf.heightaxis(i));
|
||||
}
|
||||
for (size_t i = 0; i < reshape_conf.widthaxis_size(); i++) {
|
||||
LOG(INFO) << "reshape width axis: " << reshape_conf.widthaxis(i);
|
||||
widthAxis_.push_back(reshape_conf.widthaxis(i));
|
||||
}
|
||||
createFunction(nchw2nhwc_, "NCHW2NHWC", FuncConfig());
|
||||
createFunction(nhwc2nchw_, "NHWC2NCHW", FuncConfig());
|
||||
return true;
|
||||
}
|
||||
|
||||
void SwitchOrderLayer::setOutDims() {
|
||||
outDims_.setDim(0, inDims_[0]);
|
||||
outDims_.setDim(1, inDims_[2]);
|
||||
outDims_.setDim(2, inDims_[3]);
|
||||
outDims_.setDim(3, inDims_[1]);
|
||||
reshapeHeight_ = 1;
|
||||
for (size_t i = 0; i < heightAxis_.size(); i++) {
|
||||
reshapeHeight_ *= outDims_[heightAxis_[i]];
|
||||
}
|
||||
output_.setFrameHeight(reshapeHeight_);
|
||||
reshapeWidth_ = 1;
|
||||
for (size_t i = 0; i < widthAxis_.size(); i++) {
|
||||
reshapeWidth_ *= outDims_[widthAxis_[i]];
|
||||
}
|
||||
output_.setFrameWidth(reshapeWidth_);
|
||||
LOG(INFO) << "outDims: " << outDims_[0] << "; " << outDims_[1] << ";"
|
||||
<< outDims_[2] << ";" << outDims_[3];
|
||||
}
|
||||
|
||||
void SwitchOrderLayer::setInDims() {
|
||||
MatrixPtr input = inputLayers_[0]->getOutputValue();
|
||||
size_t batchSize = input->getHeight();
|
||||
inDims_.setDim(0, batchSize);
|
||||
|
||||
int h = inputLayers_[0]->getOutput().getFrameHeight();
|
||||
if (h != 0) inDims_.setDim(2, h);
|
||||
int w = inputLayers_[0]->getOutput().getFrameWidth();
|
||||
if (w != 0) inDims_.setDim(3, w);
|
||||
int totalCount = input->getElementCnt();
|
||||
int channels = totalCount / (inDims_[0] * inDims_[2] * inDims_[3]);
|
||||
if (channels != 0) inDims_.setDim(1, channels);
|
||||
LOG(INFO) << "inDims: " << inDims_[0] << "; " << inDims_[1] << ";"
|
||||
<< inDims_[2] << ";" << inDims_[3];
|
||||
}
|
||||
|
||||
void SwitchOrderLayer::forward(PassType passType) {
|
||||
Layer::forward(passType);
|
||||
setInDims();
|
||||
setOutDims();
|
||||
resetOutput(outDims_[0], outDims_[1] * outDims_[2] * outDims_[3]);
|
||||
if (heightAxis_.size() > 0) {
|
||||
getOutputValue()->reshape(reshapeHeight_, reshapeWidth_);
|
||||
}
|
||||
|
||||
// switch NCHW to NHWC
|
||||
BufferArgs inputs;
|
||||
BufferArgs outputs;
|
||||
inputs.addArg(*getInputValue(0), inDims_);
|
||||
outputs.addArg(*getOutputValue(), outDims_);
|
||||
nchw2nhwc_[0]->calc(inputs, outputs);
|
||||
// forwardActivation();
|
||||
}
|
||||
|
||||
void SwitchOrderLayer::backward(const UpdateCallback& callback) {
|
||||
(void)callback;
|
||||
// backwardActivation();
|
||||
|
||||
// switch NHWC to NCHW
|
||||
BufferArgs inputs;
|
||||
BufferArgs outputs;
|
||||
inputs.addArg(*getOutputGrad(), outDims_);
|
||||
outputs.addArg(*getInputGrad(0), inDims_, ADD_TO);
|
||||
nhwc2nchw_[0]->calc(inputs, outputs);
|
||||
}
|
||||
} // namespace paddle
|
Loading…
Reference in new issue