Merge pull request #409 from luotao1/conv

Support rectangle input for CNN
avx_docs
Tao Luo 9 years ago committed by GitHub
commit 7bb627d3f9

@ -60,14 +60,12 @@ bool BatchNormBaseLayer::init(const LayerMap& layerMap,
void BatchNormBaseLayer::calFeatureMapSize() { void BatchNormBaseLayer::calFeatureMapSize() {
const ImageConfig& conf = config_.inputs(0).image_conf(); const ImageConfig& conf = config_.inputs(0).image_conf();
if (inputLayers_[0]->getOutput().getFrameHeight() == 0 && imageH_ = inputLayers_[0]->getOutput().getFrameHeight();
inputLayers_[0]->getOutput().getFrameWidth() == 0) { imageW_ = inputLayers_[0]->getOutput().getFrameWidth();
imgSize_ = conf.img_size(); if (imageH_ == 0 && imageW_ == 0) {
imageH_ = imgSize_; imageH_ = conf.has_img_size_y() ? conf.img_size_y() : conf.img_size();
imageW_ = imgSize_; imageW_ = conf.img_size();
} else { } else {
imageH_ = inputLayers_[0]->getOutput().getFrameHeight();
imageW_ = inputLayers_[0]->getOutput().getFrameWidth();
getOutput().setFrameHeight(imageH_); getOutput().setFrameHeight(imageH_);
getOutput().setFrameWidth(imageW_); getOutput().setFrameWidth(imageW_);
} }

@ -77,9 +77,8 @@ protected:
MatrixPtr savedMean_; MatrixPtr savedMean_;
MatrixPtr savedInvVar_; MatrixPtr savedInvVar_;
/// Height or width of input image feature, now height is equal to width. /// Height or width of input image feature.
/// imgSize is 1 if the input is fully-connected layer. /// Both of them are 1 if the input is fully-connected layer.
int imgSize_;
int imageH_; int imageH_;
int imageW_; int imageW_;
/// Height * Width. /// Height * Width.

@ -26,15 +26,15 @@ size_t BilinearInterpLayer::getSize() {
const BilinearInterpConfig& conf = config_.inputs(0).bilinear_interp_conf(); const BilinearInterpConfig& conf = config_.inputs(0).bilinear_interp_conf();
if (inImgH_ == 0) { if (inImgH_ == 0) {
inImgH_ = conf.img_size_y(); inImgH_ = conf.image_conf().img_size_y();
} }
if (inImgW_ == 0) { if (inImgW_ == 0) {
inImgW_ = conf.img_size_x(); inImgW_ = conf.image_conf().img_size();
} }
outImgH_ = conf.out_size_y(); outImgH_ = conf.out_size_y();
outImgW_ = conf.out_size_x(); outImgW_ = conf.out_size_x();
numChannels_ = conf.num_channels(); numChannels_ = conf.image_conf().channels();
CHECK(outImgH_ > 0 && outImgW_ > 0); CHECK(outImgH_ > 0 && outImgW_ > 0);
CHECK(inImgH_ > 0 && inImgW_ > 0); CHECK(inImgH_ > 0 && inImgW_ > 0);

@ -38,11 +38,12 @@ bool ConvBaseLayer::init(const LayerMap& layerMap,
filterSizeY_.push_back(conf.filter_size_y()); filterSizeY_.push_back(conf.filter_size_y());
filterPixels_.push_back(filterSize_.back() * filterSizeY_.back()); filterPixels_.push_back(filterSize_.back() * filterSizeY_.back());
channels_.push_back(conf.channels()); channels_.push_back(conf.channels());
imgSizeH_.push_back(conf.img_size()); imgSizeH_.push_back(conf.has_img_size_y() ? conf.img_size_y()
: conf.img_size());
imgSizeW_.push_back(conf.img_size()); imgSizeW_.push_back(conf.img_size());
groups_.push_back(conf.groups()); groups_.push_back(conf.groups());
filterChannels_.push_back(conf.filter_channels()); filterChannels_.push_back(conf.filter_channels());
outputH_.push_back(conf.output_x()); outputH_.push_back(conf.has_output_y() ? conf.output_y() : conf.output_x());
outputW_.push_back(conf.output_x()); outputW_.push_back(conf.output_x());
} }
@ -91,16 +92,19 @@ size_t ConvBaseLayer::calOutputSize() {
for (size_t i = 0; i < inputLayers_.size(); i++) { for (size_t i = 0; i < inputLayers_.size(); i++) {
inH.push_back(inputLayers_[i]->getOutput().getFrameHeight()); inH.push_back(inputLayers_[i]->getOutput().getFrameHeight());
inW.push_back(inputLayers_[i]->getOutput().getFrameWidth()); inW.push_back(inputLayers_[i]->getOutput().getFrameWidth());
const ConvConfig& conf = config_.inputs(i).conv_conf();
if (isDeconv_) { if (isDeconv_) {
if (inH[i] == 0) inH[i] = config_.inputs(i).conv_conf().output_x(); if (inH[i] == 0)
if (inW[i] == 0) inW[i] = config_.inputs(i).conv_conf().output_x(); inH[i] = conf.has_output_y() ? conf.output_y() : conf.output_x();
if (inW[i] == 0) inW[i] = conf.output_x();
outH.push_back(imageSize( outH.push_back(imageSize(
inH[i], filterSizeY_[i], paddingY_[i], strideY_[i], caffeMode_)); inH[i], filterSizeY_[i], paddingY_[i], strideY_[i], caffeMode_));
outW.push_back(imageSize( outW.push_back(imageSize(
inW[i], filterSize_[i], padding_[i], stride_[i], caffeMode_)); inW[i], filterSize_[i], padding_[i], stride_[i], caffeMode_));
} else { } else {
if (inH[i] == 0) inH[i] = config_.inputs(i).conv_conf().img_size(); if (inH[i] == 0)
if (inW[i] == 0) inW[i] = config_.inputs(i).conv_conf().img_size(); inH[i] = conf.has_img_size_y() ? conf.img_size_y() : conf.img_size();
if (inW[i] == 0) inW[i] = conf.img_size();
outH.push_back(outputSize( outH.push_back(outputSize(
inH[i], filterSizeY_[i], paddingY_[i], strideY_[i], caffeMode_)); inH[i], filterSizeY_[i], paddingY_[i], strideY_[i], caffeMode_));
outW.push_back(outputSize( outW.push_back(outputSize(

@ -93,9 +93,9 @@ private:
bool caffeMode_; bool caffeMode_;
int inputOffset_, outputOffset_, weightOffset_; int inputOffset_, outputOffset_, weightOffset_;
int numFilters_; int numFilters_;
int padding_, stride_, filterSize_, channels_, imgSize_; int padding_, stride_, filterSize_, channels_, imgSize_, imgSizeY_;
int paddingY_, strideY_, filterSizeY_; int paddingY_, strideY_, filterSizeY_;
int imgPixels_, filterPixels_, filterChannels_, outputX_, outputs_; int imgPixels_, filterPixels_, filterChannels_, outputX_, outputY_, outputs_;
/// Following member variables are same with CudnnConvLayer. /// Following member variables are same with CudnnConvLayer.
/// There is no explanation here. /// There is no explanation here.
@ -144,7 +144,7 @@ void ConvOperator::allocConvWorkSpace(size_t maxWorkSpace) {
void ConvOperator::reshape(int batchSize) { void ConvOperator::reshape(int batchSize) {
imageH_ = ins_[0]->getFrameHeight(); imageH_ = ins_[0]->getFrameHeight();
imageW_ = ins_[0]->getFrameWidth(); imageW_ = ins_[0]->getFrameWidth();
if (imageH_ == 0) imageH_ = imgSize_; if (imageH_ == 0) imageH_ = imgSizeY_;
if (imageW_ == 0) imageW_ = imgSize_; if (imageW_ == 0) imageW_ = imgSize_;
outputH_ = outputSize(imageH_, filterSizeY_, paddingY_, strideY_, caffeMode_); outputH_ = outputSize(imageH_, filterSizeY_, paddingY_, strideY_, caffeMode_);
outputW_ = outputSize(imageW_, filterSize_, padding_, stride_, caffeMode_); outputW_ = outputSize(imageW_, filterSize_, padding_, stride_, caffeMode_);
@ -182,7 +182,10 @@ void ConvOperator::computeConvSizes() {
hl_create_tensor_descriptor(&inputDesc_); hl_create_tensor_descriptor(&inputDesc_);
int outputX = int outputX =
outputSize(imgSize_, filterSize_, padding_, stride_, caffeMode_); outputSize(imgSize_, filterSize_, padding_, stride_, caffeMode_);
int outputY =
outputSize(imgSizeY_, filterSizeY_, paddingY_, strideY_, caffeMode_);
CHECK_EQ(outputX, outputX_); CHECK_EQ(outputX, outputX_);
CHECK_EQ(outputY, outputY_);
hl_create_tensor_descriptor(&outputDesc_); hl_create_tensor_descriptor(&outputDesc_);
hl_create_convolution_descriptor(&convDesc_, hl_create_convolution_descriptor(&convDesc_,
inputDesc_, inputDesc_,
@ -236,10 +239,12 @@ void ConvOperator::getConvParams() {
filterPixels_ = filterSize_ * filterSizeY_; filterPixels_ = filterSize_ * filterSizeY_;
channels_ = conf.channels(); channels_ = conf.channels();
imgSize_ = conf.img_size(); imgSize_ = conf.img_size();
imgPixels_ = imgSize_ * imgSize_; imgSizeY_ = conf.has_img_size_y() ? conf.img_size_y() : conf.img_size();
imgPixels_ = imgSize_ * imgSizeY_;
CHECK_EQ(conf.groups(), 1U); CHECK_EQ(conf.groups(), 1U);
filterChannels_ = conf.filter_channels(); filterChannels_ = conf.filter_channels();
outputX_ = conf.output_x(); outputX_ = conf.output_x();
outputY_ = conf.has_output_y() ? conf.output_y() : conf.output_x();
outputs_ = outputX_ * outputX_; outputs_ = outputX_ * outputX_;
} }

@ -46,7 +46,7 @@ void ConvProjection::getConvParams() {
filterH_ = conf.filter_size_y(); filterH_ = conf.filter_size_y();
filterW_ = conf.filter_size(); filterW_ = conf.filter_size();
configImgH_ = conf.img_size(); configImgH_ = conf.has_img_size_y() ? conf.img_size_y() : conf.img_size();
configImgW_ = conf.img_size(); configImgW_ = conf.img_size();
channels_ = conf.channels(); channels_ = conf.channels();
@ -58,9 +58,11 @@ void ConvProjection::getConvParams() {
} }
void ConvProjection::initCudnn() { void ConvProjection::initCudnn() {
hl_create_filter_descriptor( hl_create_filter_descriptor(&filterDesc_,
&filterDesc_, channels_ / groups_, numFilters_ / groups_, channels_ / groups_,
filterH_, filterW_); numFilters_ / groups_,
filterH_,
filterW_);
hl_create_tensor_descriptor(&inputDesc_); hl_create_tensor_descriptor(&inputDesc_);
hl_create_tensor_descriptor(&outputDesc_); hl_create_tensor_descriptor(&outputDesc_);
hl_create_convolution_descriptor(&convDesc_, hl_create_convolution_descriptor(&convDesc_,

@ -49,8 +49,13 @@ void DataLayer::copyDataToOutput(Argument& output) {
output.ids->copyFrom(*data_.ids); output.ids->copyFrom(*data_.ids);
} }
} }
output.setFrameHeight(data_.getFrameHeight()); if (config_.height() && config_.width()) {
output.setFrameWidth(data_.getFrameWidth()); output.setFrameHeight(config_.height());
output.setFrameWidth(config_.width());
} else {
output.setFrameHeight(data_.getFrameHeight());
output.setFrameHeight(data_.getFrameHeight());
}
output.cpuSequenceDims = data_.cpuSequenceDims; output.cpuSequenceDims = data_.cpuSequenceDims;
output.sequenceStartPositions = data_.sequenceStartPositions; output.sequenceStartPositions = data_.sequenceStartPositions;
output.subSequenceStartPositions = data_.subSequenceStartPositions; output.subSequenceStartPositions = data_.subSequenceStartPositions;

@ -29,17 +29,19 @@ bool ExpandConvBaseLayer::init(const LayerMap &layerMap,
* meaning as in conv, we need to swap channels_ and numFilters here for * meaning as in conv, we need to swap channels_ and numFilters here for
* convTrans, and in other functions too. * convTrans, and in other functions too.
* */ * */
int channel;
int numFilters;
/* Initialize the projection */ /* Initialize the projection */
for (auto &inputConfig : config_.inputs()) { for (auto &inputConfig : config_.inputs()) {
const ConvConfig &conf = inputConfig.conv_conf(); const ConvConfig &conf = inputConfig.conv_conf();
numFilters = isDeconv_ ? conf.channels() : numFilters_; int numFilters = isDeconv_ ? conf.channels() : numFilters_;
subM_.push_back(numFilters / conf.groups()); subM_.push_back(numFilters / conf.groups());
subN_.push_back(conf.output_x() * conf.output_x()); subN_.push_back(conf.output_x() *
channel = isDeconv_ ? numFilters_ : conf.channels(); (conf.has_output_y() ? conf.output_y() : conf.output_x()));
subK_.push_back(channel * conf.filter_size() * conf.filter_size() / int channel = isDeconv_ ? numFilters_ : conf.channels();
conf.groups()); subK_.push_back(
channel * conf.filter_size() *
(conf.has_filter_size_y() ? conf.filter_size_y() : conf.filter_size()) /
conf.groups());
/* Consistent caffe mode for multiple input */ /* Consistent caffe mode for multiple input */
caffeMode_ = conf.caffe_mode(); caffeMode_ = conf.caffe_mode();
} }
@ -116,11 +118,11 @@ void ExpandConvBaseLayer::expandOneFrame(MatrixPtr image,
imgSizeH_[inIdx], imgSizeH_[inIdx],
imgSizeW_[inIdx], imgSizeW_[inIdx],
channel, channel,
filterSizeY_[inIdx],
filterSize_[inIdx], filterSize_[inIdx],
filterSize_[inIdx], strideY_[inIdx],
stride_[inIdx], stride_[inIdx],
stride_[inIdx], paddingY_[inIdx],
padding_[inIdx],
padding_[inIdx], padding_[inIdx],
outputH_[inIdx], outputH_[inIdx],
outputW_[inIdx]); outputW_[inIdx]);
@ -208,11 +210,11 @@ void ExpandConvBaseLayer::bpropActs(MatrixPtr out,
imgSizeH_[inpIdx], imgSizeH_[inpIdx],
imgSizeW_[inpIdx], imgSizeW_[inpIdx],
channel, channel,
filterSizeY_[inpIdx],
filterSize_[inpIdx], filterSize_[inpIdx],
filterSize_[inpIdx], strideY_[inpIdx],
stride_[inpIdx],
stride_[inpIdx], stride_[inpIdx],
padding_[inpIdx], paddingY_[inpIdx],
padding_[inpIdx], padding_[inpIdx],
outputH_[inpIdx], outputH_[inpIdx],
outputW_[inpIdx], outputW_[inpIdx],

@ -25,10 +25,10 @@ size_t MaxOutLayer::getSize() {
imgSizeH_ = inputLayers_[0]->getOutput().getFrameHeight(); imgSizeH_ = inputLayers_[0]->getOutput().getFrameHeight();
imgSizeW_ = inputLayers_[0]->getOutput().getFrameWidth(); imgSizeW_ = inputLayers_[0]->getOutput().getFrameWidth();
if (imgSizeH_ == 0) { if (imgSizeH_ == 0) {
imgSizeH_ = maxoutConf.img_size_y(); imgSizeH_ = maxoutConf.image_conf().img_size_y();
} }
if (imgSizeW_ == 0) { if (imgSizeW_ == 0) {
imgSizeW_ = maxoutConf.img_size_x(); imgSizeW_ = maxoutConf.image_conf().img_size();
} }
featLen_ = imgSizeH_ * imgSizeW_; featLen_ = imgSizeH_ * imgSizeW_;
@ -50,7 +50,7 @@ bool MaxOutLayer::init(const LayerMap& layerMap,
const MaxOutConfig& conf = config_.inputs(0).maxout_conf(); const MaxOutConfig& conf = config_.inputs(0).maxout_conf();
groups_ = conf.groups(); groups_ = conf.groups();
channels_ = conf.channels(); channels_ = conf.image_conf().channels();
CHECK_EQ(channels_ % groups_, 0UL); CHECK_EQ(channels_ % groups_, 0UL);
outputChannels_ = channels_ / groups_; outputChannels_ = channels_ / groups_;

@ -48,6 +48,9 @@ bool ResponseNormLayer::init(const LayerMap& layerMap,
outputX_ = conf.output_x(); outputX_ = conf.output_x();
imgSize_ = conf.img_size(); imgSize_ = conf.img_size();
denoms_ = NULL; denoms_ = NULL;
outputY_ = conf.has_output_y() ? conf.output_y() : conf.output_x();
imgSizeY_ = conf.has_img_size_y() ? conf.img_size_y() : conf.img_size();
return true; return true;
} }

@ -49,7 +49,7 @@ public:
*/ */
class ResponseNormLayer : public NormLayer { class ResponseNormLayer : public NormLayer {
protected: protected:
size_t channels_, size_, outputX_, imgSize_; size_t channels_, size_, outputX_, imgSize_, outputY_, imgSizeY_;
float scale_, pow_; float scale_, pow_;
MatrixPtr denoms_; MatrixPtr denoms_;

@ -23,7 +23,7 @@ size_t CMRProjectionNormLayer::getSize() {
imgSizeH_ = inputLayers_[0]->getOutput().getFrameHeight(); imgSizeH_ = inputLayers_[0]->getOutput().getFrameHeight();
imgSizeW_ = inputLayers_[0]->getOutput().getFrameWidth(); imgSizeW_ = inputLayers_[0]->getOutput().getFrameWidth();
if (imgSizeH_ == 0) { if (imgSizeH_ == 0) {
imgSizeH_ = imgSize_; imgSizeH_ = imgSizeY_;
} }
if (imgSizeW_ == 0) { if (imgSizeW_ == 0) {
imgSizeW_ = imgSize_; imgSizeW_ = imgSize_;

@ -56,14 +56,14 @@ ProjectionConfig SpatialPyramidPoolLayer::getConfig(size_t imgSizeW,
size_t SpatialPyramidPoolLayer::getSize() { size_t SpatialPyramidPoolLayer::getSize() {
CHECK_EQ(inputLayers_.size(), 1UL); CHECK_EQ(inputLayers_.size(), 1UL);
size_t layerSize = 0; size_t layerSize = 0;
const SppConfig& sppConf = config_.inputs(0).spp_conf(); const ImageConfig& conf = config_.inputs(0).spp_conf().image_conf();
imgSizeH_ = inputLayers_[0]->getOutput().getFrameHeight(); imgSizeH_ = inputLayers_[0]->getOutput().getFrameHeight();
imgSizeW_ = inputLayers_[0]->getOutput().getFrameWidth(); imgSizeW_ = inputLayers_[0]->getOutput().getFrameWidth();
if (imgSizeH_ == 0) { if (imgSizeH_ == 0) {
imgSizeH_ = sppConf.has_img_size_y() ? sppConf.img_size_y() : imgSizeW_; imgSizeH_ = conf.has_img_size_y() ? conf.img_size_y() : conf.img_size();
} }
if (imgSizeW_ == 0) { if (imgSizeW_ == 0) {
imgSizeW_ = sppConf.img_size(); imgSizeW_ = conf.img_size();
} }
size_t outputH = 1; size_t outputH = 1;
@ -82,9 +82,10 @@ bool SpatialPyramidPoolLayer::init(const LayerMap& layerMap,
pyramidHeight_ = sppConf.pyramid_height(); pyramidHeight_ = sppConf.pyramid_height();
poolType_ = sppConf.pool_type(); poolType_ = sppConf.pool_type();
channels_ = sppConf.channels(); const ImageConfig& imageConf = sppConf.image_conf();
imgSizeW_ = sppConf.img_size(); channels_ = imageConf.channels();
imgSizeH_ = sppConf.has_img_size_y() ? sppConf.img_size_y() : imgSizeW_; imgSizeW_ = imageConf.img_size();
imgSizeH_ = imageConf.has_img_size_y() ? imageConf.img_size_y() : imgSizeW_;
poolProjections_.reserve(pyramidHeight_); poolProjections_.reserve(pyramidHeight_);
projCol_.reserve(pyramidHeight_); projCol_.reserve(pyramidHeight_);
projOutput_.resize(pyramidHeight_); projOutput_.resize(pyramidHeight_);

@ -28,7 +28,6 @@ maxpool = img_pool_layer(input=conv,
stride_y=2, stride_y=2,
padding=1, padding=1,
padding_y=2, padding_y=2,
img_width=16,
pool_type=MaxPooling(), pool_type=MaxPooling(),
) )
avgpool = img_pool_layer(input=conv, avgpool = img_pool_layer(input=conv,
@ -39,7 +38,6 @@ avgpool = img_pool_layer(input=conv,
stride_y=2, stride_y=2,
padding=1, padding=1,
padding_y=2, padding_y=2,
img_width=16,
pool_type=AvgPooling(), pool_type=AvgPooling(),
) )

@ -202,16 +202,15 @@ void testProjectionConv(size_t groups) {
conf.set_input_size(IMAGE_SIZE * IMAGE_SIZE * CHANNELS); conf.set_input_size(IMAGE_SIZE * IMAGE_SIZE * CHANNELS);
conf.set_output_size(output_x * output_y * NUM_FILTERS); conf.set_output_size(output_x * output_y * NUM_FILTERS);
testProjectionGrad( testProjectionGrad(conf,
conf, INPUT_DATA,
INPUT_DATA, /* parameterSize */ NUM_FILTERS * CHANNELS * FILTER_SIZE *
/* parameterSize */ NUM_FILTERS * CHANNELS * FILTER_SIZE * FILTER_SIZE_Y FILTER_SIZE_Y / groups,
/ groups, /* batchSize */ 100,
/* batchSize */ 100, true,
true, false,
false, NUM_FILTERS,
NUM_FILTERS, true);
true);
} }
#ifndef PADDLE_ONLY_CPU #ifndef PADDLE_ONLY_CPU
@ -229,9 +228,10 @@ TEST(Layer, BilinearInterpLayer) {
LayerInputConfig* input = config.layerConfig.add_inputs(); LayerInputConfig* input = config.layerConfig.add_inputs();
BilinearInterpConfig* bilinear = input->mutable_bilinear_interp_conf(); BilinearInterpConfig* bilinear = input->mutable_bilinear_interp_conf();
bilinear->set_img_size_x(32); ImageConfig* image = bilinear->mutable_image_conf();
bilinear->set_img_size_y(32); image->set_img_size(32);
bilinear->set_num_channels(4); image->set_img_size_y(32);
image->set_channels(4);
for (auto useGpu : {false, true}) { for (auto useGpu : {false, true}) {
for (auto outSize : {32, 64}) { for (auto outSize : {32, 64}) {
@ -354,7 +354,7 @@ void testConvLayer(const string& type, bool trans, bool useGpu) {
config.layerConfig.set_partial_sum(1); config.layerConfig.set_partial_sum(1);
config.layerConfig.set_shared_biases(true); config.layerConfig.set_shared_biases(true);
config.inputDefs.push_back({INPUT_DATA, "layer_0", 768, 288}); config.inputDefs.push_back({INPUT_DATA, "layer_0", 384, 288});
LayerInputConfig* input = config.layerConfig.add_inputs(); LayerInputConfig* input = config.layerConfig.add_inputs();
ConvConfig* conv = input->mutable_conv_conf(); ConvConfig* conv = input->mutable_conv_conf();
conv->set_filter_size(2); conv->set_filter_size(2);
@ -367,12 +367,18 @@ void testConvLayer(const string& type, bool trans, bool useGpu) {
conv->set_groups(1); conv->set_groups(1);
conv->set_filter_channels(conv->channels() / conv->groups()); conv->set_filter_channels(conv->channels() / conv->groups());
conv->set_img_size(16); conv->set_img_size(16);
conv->set_img_size_y(8);
conv->set_output_x(outputSize(conv->img_size(), conv->set_output_x(outputSize(conv->img_size(),
conv->filter_size(), conv->filter_size(),
conv->padding(), conv->padding(),
conv->stride(), conv->stride(),
/* caffeMode */ true)); /* caffeMode */ true));
config.layerConfig.set_size(conv->output_x() * conv->output_x() * conv->set_output_y(outputSize(conv->img_size_y(),
conv->filter_size_y(),
conv->padding_y(),
conv->stride_y(),
/* caffeMode */ true));
config.layerConfig.set_size(conv->output_x() * conv->output_y() *
config.layerConfig.num_filters()); config.layerConfig.num_filters());
testLayerGrad(config, "conv", 100, trans, useGpu); testLayerGrad(config, "conv", 100, trans, useGpu);
@ -472,10 +478,11 @@ TEST(Layer, maxoutLayer) {
config.inputDefs.push_back({INPUT_DATA, "layer_0", 4096, 0}); config.inputDefs.push_back({INPUT_DATA, "layer_0", 4096, 0});
LayerInputConfig* input = config.layerConfig.add_inputs(); LayerInputConfig* input = config.layerConfig.add_inputs();
MaxOutConfig* maxout = input->mutable_maxout_conf(); MaxOutConfig* maxout = input->mutable_maxout_conf();
ImageConfig* image = maxout->mutable_image_conf();
maxout->set_img_size_x(32); image->set_img_size(32);
maxout->set_img_size_y(32); image->set_img_size_y(32);
maxout->set_channels(4); image->set_channels(4);
maxout->set_groups(2); maxout->set_groups(2);
for (auto useGpu : {false, true}) { for (auto useGpu : {false, true}) {
@ -987,7 +994,7 @@ void testNormLayer(const string& normType, bool trans, bool useGpu) {
config.layerConfig.set_type("norm"); config.layerConfig.set_type("norm");
config.layerConfig.set_active_type("relu"); config.layerConfig.set_active_type("relu");
config.inputDefs.push_back({INPUT_DATA, "layer_0", 3136, 0}); config.inputDefs.push_back({INPUT_DATA, "layer_0", 1568, 0});
LayerInputConfig* input = config.layerConfig.add_inputs(); LayerInputConfig* input = config.layerConfig.add_inputs();
NormConfig* norm = input->mutable_norm_conf(); NormConfig* norm = input->mutable_norm_conf();
norm->set_norm_type(normType); norm->set_norm_type(normType);
@ -997,7 +1004,9 @@ void testNormLayer(const string& normType, bool trans, bool useGpu) {
norm->set_pow(0.75); norm->set_pow(0.75);
norm->set_blocked(0); norm->set_blocked(0);
norm->set_img_size(14); norm->set_img_size(14);
norm->set_img_size_y(7);
norm->set_output_x(norm->img_size()); norm->set_output_x(norm->img_size());
norm->set_output_y(norm->img_size_y());
if (norm->norm_type() == "cmrnorm" || if (norm->norm_type() == "cmrnorm" ||
norm->norm_type() == "cmrnorm-projection") { norm->norm_type() == "cmrnorm-projection") {
norm->set_scale(norm->scale() / norm->size()); norm->set_scale(norm->scale() / norm->size());
@ -1005,7 +1014,7 @@ void testNormLayer(const string& normType, bool trans, bool useGpu) {
norm->set_scale(norm->scale() / (norm->size() * norm->size())); norm->set_scale(norm->scale() / (norm->size() * norm->size()));
} }
config.layerConfig.set_size(norm->output_x() * norm->output_x() * config.layerConfig.set_size(norm->output_x() * norm->output_y() *
norm->channels()); norm->channels());
config.biasSize = 0; config.biasSize = 0;
@ -1106,11 +1115,12 @@ void testSppLayer(const string& poolType,
SppConfig* sppConfig = input->mutable_spp_conf(); SppConfig* sppConfig = input->mutable_spp_conf();
sppConfig->set_pool_type(poolType); sppConfig->set_pool_type(poolType);
sppConfig->set_pyramid_height(pyramidHeight); sppConfig->set_pyramid_height(pyramidHeight);
sppConfig->set_channels(16); ImageConfig* imageConfig = sppConfig->mutable_image_conf();
sppConfig->set_img_size(10); imageConfig->set_channels(16);
sppConfig->set_img_size_y(20); imageConfig->set_img_size(10);
imageConfig->set_img_size_y(20);
int outputSize = (std::pow(4, sppConfig->pyramid_height()) - 1) / (4 - 1); int outputSize = (std::pow(4, sppConfig->pyramid_height()) - 1) / (4 - 1);
config.layerConfig.set_size(outputSize * sppConfig->channels()); config.layerConfig.set_size(outputSize * imageConfig->channels());
testLayerGrad(config, "spp", 100, trans, useGpu); testLayerGrad(config, "spp", 100, trans, useGpu);
} }
@ -1420,13 +1430,15 @@ void testBatchNormLayer(const string& type, bool trans, bool useGpu) {
TestConfig config; TestConfig config;
const int CHANNELS = 10; const int CHANNELS = 10;
const int IMG_SIZE = 16; const int IMG_SIZE = 16;
const int IMG_SIZE_Y = 8;
size_t size = CHANNELS * IMG_SIZE * IMG_SIZE_Y;
config.layerConfig.set_type(type); config.layerConfig.set_type(type);
config.layerConfig.set_size(CHANNELS * IMG_SIZE * IMG_SIZE); config.layerConfig.set_size(size);
config.layerConfig.set_active_type("sigmoid"); config.layerConfig.set_active_type("sigmoid");
config.biasSize = CHANNELS; config.biasSize = CHANNELS;
config.inputDefs.push_back({INPUT_DATA, config.inputDefs.push_back({INPUT_DATA,
"layer_0", "layer_0",
/* dim= */ IMG_SIZE * IMG_SIZE * CHANNELS, /* dim= */ size,
/* paraSize= */ CHANNELS}); /* paraSize= */ CHANNELS});
config.inputDefs.push_back({INPUT_DATA, "layer_1_running_mean", 1, CHANNELS}); config.inputDefs.push_back({INPUT_DATA, "layer_1_running_mean", 1, CHANNELS});
@ -1441,6 +1453,7 @@ void testBatchNormLayer(const string& type, bool trans, bool useGpu) {
ImageConfig* img_conf = input->mutable_image_conf(); ImageConfig* img_conf = input->mutable_image_conf();
img_conf->set_channels(CHANNELS); img_conf->set_channels(CHANNELS);
img_conf->set_img_size(IMG_SIZE); img_conf->set_img_size(IMG_SIZE);
img_conf->set_img_size_y(IMG_SIZE_Y);
testLayerGrad(config, testLayerGrad(config,
"batch_norm", "batch_norm",
@ -1467,6 +1480,7 @@ TEST(Operator, conv) {
const int FILTER_SIZE_Y = 3; const int FILTER_SIZE_Y = 3;
const int CHANNELS = 3; const int CHANNELS = 3;
const int IMAGE_SIZE = 16; const int IMAGE_SIZE = 16;
const int IMAGE_SIZE_Y = 8;
OperatorConfig& operatorConf = *config.layerConfig.add_operator_confs(); OperatorConfig& operatorConf = *config.layerConfig.add_operator_confs();
operatorConf.set_type("conv"); operatorConf.set_type("conv");
ConvConfig* conv = operatorConf.mutable_conv_conf(); ConvConfig* conv = operatorConf.mutable_conv_conf();
@ -1481,19 +1495,22 @@ TEST(Operator, conv) {
conv->set_groups(1); conv->set_groups(1);
conv->set_filter_channels(conv->channels() / conv->groups()); conv->set_filter_channels(conv->channels() / conv->groups());
conv->set_img_size(IMAGE_SIZE); conv->set_img_size(IMAGE_SIZE);
int output_x = outputSize(conv->img_size(), conv->set_img_size_y(IMAGE_SIZE_Y);
conv->filter_size(), conv->set_output_x(outputSize(conv->img_size(),
conv->padding(), conv->filter_size(),
conv->stride(), conv->padding(),
/* caffeMode */ true); conv->stride(),
conv->set_output_x(output_x); /* caffeMode */ true));
config.layerConfig.set_size(output_x * output_x * conv->set_output_y(outputSize(conv->img_size_y(),
config.layerConfig.num_filters()); conv->filter_size_y(),
config.layerConfig.set_size(conv->output_x() * conv->output_x() * conv->padding_y(),
conv->stride_y(),
/* caffeMode */ true));
config.layerConfig.set_size(conv->output_x() * conv->output_y() *
NUM_FILTERS); NUM_FILTERS);
config.inputDefs.push_back( config.inputDefs.push_back(
{INPUT_DATA, "layer_0", IMAGE_SIZE * IMAGE_SIZE * CHANNELS, 0}); {INPUT_DATA, "layer_0", IMAGE_SIZE * IMAGE_SIZE_Y * CHANNELS, 0});
config.inputDefs.push_back( config.inputDefs.push_back(
{INPUT_DATA, {INPUT_DATA,
"layer_1", "layer_1",

@ -225,6 +225,8 @@ void Argument::resizeAndCopyFrom(const Argument& src,
} }
resizeAndCopy(udp, src.udp, useGpu, stream); resizeAndCopy(udp, src.udp, useGpu, stream);
resizeAndCopy(strs, src.strs, useGpu, stream); resizeAndCopy(strs, src.strs, useGpu, stream);
frameWidth = src.frameWidth;
frameHeight = src.frameHeight;
} }
int32_t Argument::resizeAndCopyFrom(const Argument& src, int32_t Argument::resizeAndCopyFrom(const Argument& src,

@ -59,7 +59,6 @@ pool = img_pool_layer(input=fc2,
padding_y=2, padding_y=2,
stride=2, stride=2,
stride_y=3, stride_y=3,
img_width=3,
pool_type=CudnnAvgPooling()) pool_type=CudnnAvgPooling())
concat = concat_layer(input=[fc3, fc4]) concat = concat_layer(input=[fc3, fc4])

@ -77,6 +77,12 @@ message ConvConfig {
required uint32 filter_size_y = 10; required uint32 filter_size_y = 10;
required uint32 padding_y = 11; required uint32 padding_y = 11;
required uint32 stride_y = 12; required uint32 stride_y = 12;
// if not set, use output_x
optional uint32 output_y = 13;
// if not set, use img_size
optional uint32 img_size_y = 14;
} }
message PoolConfig { message PoolConfig {
@ -122,11 +128,9 @@ message PoolConfig {
} }
message SppConfig { message SppConfig {
required string pool_type = 1; required ImageConfig image_conf = 1;
required uint32 pyramid_height = 2; required string pool_type = 2;
required uint32 channels = 3; required uint32 pyramid_height = 3;
required uint32 img_size = 4;
optional uint32 img_size_y = 5;
} }
message NormConfig { message NormConfig {
@ -156,6 +160,12 @@ message NormConfig {
// fixed window: shared a fixed window for each value // fixed window: shared a fixed window for each value
// sliding window: have a different window for each value // sliding window: have a different window for each value
optional bool blocked = 8; optional bool blocked = 8;
// if not set, use output_x
optional uint32 output_y = 9;
// if not set, use img_size
optional uint32 img_size_y = 10;
} }
message BlockExpandConfig { message BlockExpandConfig {
@ -180,12 +190,8 @@ message BlockExpandConfig {
} }
message MaxOutConfig { message MaxOutConfig {
required uint32 channels = 1; required ImageConfig image_conf = 1;
required uint32 groups = 2; required uint32 groups = 2;
// The size of input feature map.
required uint32 img_size_x = 3;
required uint32 img_size_y = 4;
} }
message ProjectionConfig { message ProjectionConfig {
@ -226,12 +232,10 @@ message OperatorConfig {
message BilinearInterpConfig { message BilinearInterpConfig {
// The size of input feature map. // The size of input feature map.
optional uint32 img_size_x = 1; required ImageConfig image_conf = 1;
optional uint32 img_size_y = 2;
// The size of output feature map. // The size of output feature map.
required uint32 out_size_x = 3; required uint32 out_size_x = 2;
required uint32 out_size_y = 4; required uint32 out_size_y = 3;
required uint32 num_channels = 5;
} }
message ImageConfig { message ImageConfig {
@ -241,6 +245,7 @@ message ImageConfig {
// The size of input feature map. // The size of input feature map.
required uint32 img_size = 8; required uint32 img_size = 8;
required uint32 img_size_y = 9;
} }
message LayerInputConfig { message LayerInputConfig {
@ -413,7 +418,10 @@ sinclude(`ModelConfigLayer.proto.m4')
// string type is used for flexibility: different types can be converted // string type is used for flexibility: different types can be converted
// to string and reinterpreted in the user's own layer implementation. // to string and reinterpreted in the user's own layer implementation.
optional string user_arg = 49; optional string user_arg = 49;
// to indicate rectangle image data
optional uint64 height = 50;
optional uint64 width = 51;
} }
message EvaluatorConfig { message EvaluatorConfig {

File diff suppressed because it is too large Load Diff

@ -766,7 +766,7 @@ def mixed_layer(size=0,
@layer_support() @layer_support()
def data_layer(name, size, layer_attr=None): def data_layer(name, size, height=None, width=None, layer_attr=None):
""" """
Define DataLayer For NeuralNetwork. Define DataLayer For NeuralNetwork.
@ -781,6 +781,10 @@ def data_layer(name, size, layer_attr=None):
:type name: basestring :type name: basestring
:param size: Size of this data layer. :param size: Size of this data layer.
:type size: int :type size: int
:param height: Height of this data layer, used for image
:type size: int|None
:param width: Width of this data layer, used for image
:type size: int|None
:param layer_attr: Extra Layer Attribute. :param layer_attr: Extra Layer Attribute.
:type layer_attr: ExtraLayerAttribute. :type layer_attr: ExtraLayerAttribute.
:return: LayerOutput object. :return: LayerOutput object.
@ -790,6 +794,8 @@ def data_layer(name, size, layer_attr=None):
type=LayerType.DATA, type=LayerType.DATA,
name=name, name=name,
size=size, size=size,
height=height,
width=width,
**ExtraLayerAttribute.to_kwargs(layer_attr)) **ExtraLayerAttribute.to_kwargs(layer_attr))
return LayerOutput(name, LayerType.DATA, size=size) return LayerOutput(name, LayerType.DATA, size=size)
@ -1483,7 +1489,7 @@ def bilinear_interp_layer(input,
bilinear_interp=BilinearInterp( bilinear_interp=BilinearInterp(
out_size_x=out_size_x, out_size_x=out_size_x,
out_size_y=out_size_y, out_size_y=out_size_y,
num_channels=num_channels)), channels=num_channels)),
type=LayerType.BILINEAR_INTERP_LAYER, type=LayerType.BILINEAR_INTERP_LAYER,
**ExtraLayerAttribute.to_kwargs(layer_attr)) **ExtraLayerAttribute.to_kwargs(layer_attr))
return LayerOutput( return LayerOutput(
@ -1872,7 +1878,7 @@ def img_conv_layer(input,
param_attr.attr["initial_std"] = init_w param_attr.attr["initial_std"] = init_w
param_attr.attr["initial_strategy"] = 0 param_attr.attr["initial_strategy"] = 0
param_attr.attr["initial_smart"] = False param_attr.attr["initial_smart"] = False
if layer_type: if layer_type:
if trans: if trans:
assert layer_type in ["exconvt"] assert layer_type in ["exconvt"]
@ -1923,8 +1929,7 @@ def img_pool_layer(input,
layer_attr=None, layer_attr=None,
pool_size_y=None, pool_size_y=None,
stride_y=None, stride_y=None,
padding_y=None, padding_y=None):
img_width=None):
""" """
Image pooling Layer. Image pooling Layer.
@ -1955,9 +1960,6 @@ def img_pool_layer(input,
:type stride_y: int|None :type stride_y: int|None
:param layer_attr: Extra Layer attribute. :param layer_attr: Extra Layer attribute.
:type layer_attr: ExtraLayerAttribute :type layer_attr: ExtraLayerAttribute
:param img_width: the width of input feature map. If it is None, the input feature
map should be square.
:type img_width: int|None
:return: LayerOutput object. :return: LayerOutput object.
:rtype: LayerOutput :rtype: LayerOutput
""" """
@ -1993,8 +1995,7 @@ def img_pool_layer(input,
padding=padding, padding=padding,
size_y=pool_size_y, size_y=pool_size_y,
stride_y=stride_y, stride_y=stride_y,
padding_y=padding_y, padding_y=padding_y))
img_width=img_width))
], ],
**ExtraLayerAttribute.to_kwargs(layer_attr)) **ExtraLayerAttribute.to_kwargs(layer_attr))
return LayerOutput( return LayerOutput(
@ -2012,7 +2013,6 @@ def spp_layer(input,
num_channels=None, num_channels=None,
pool_type=None, pool_type=None,
pyramid_height=None, pyramid_height=None,
img_width=None,
layer_attr=None): layer_attr=None):
""" """
Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition.
@ -2029,9 +2029,6 @@ def spp_layer(input,
:type scale: BasePoolingType :type scale: BasePoolingType
:param pyramid_height: pyramid height. :param pyramid_height: pyramid height.
:type pyramid_height: int :type pyramid_height: int
:param img_width: the width of input feature map. If it is None, the input feature
map should be square.
:type img_width: int|None
:param layer_attr: Extra Layer Attribute. :param layer_attr: Extra Layer Attribute.
:type layer_attr: ExtraLayerAttribute :type layer_attr: ExtraLayerAttribute
:return: LayerOutput object. :return: LayerOutput object.
@ -2058,8 +2055,7 @@ def spp_layer(input,
spp=SpatialPyramidPool( spp=SpatialPyramidPool(
pool_type=type_name, pool_type=type_name,
channels=num_channels, channels=num_channels,
pyramid_height=pyramid_height, pyramid_height=pyramid_height)),
img_width=img_width)),
**ExtraLayerAttribute.to_kwargs(layer_attr)) **ExtraLayerAttribute.to_kwargs(layer_attr))
return LayerOutput( return LayerOutput(
name, name,

@ -26,11 +26,15 @@ layers {
filter_size_y: 32 filter_size_y: 32
padding_y: 1 padding_y: 1
stride_y: 1 stride_y: 1
output_y: 227
img_size_y: 256
} }
} }
bias_parameter_name: "___conv_0__.wbias" bias_parameter_name: "___conv_0__.wbias"
num_filters: 64 num_filters: 64
shared_biases: true shared_biases: true
height: 227
width: 227
} }
layers { layers {
name: "__batch_norm_0__" name: "__batch_norm_0__"
@ -43,6 +47,7 @@ layers {
image_conf { image_conf {
channels: 64 channels: 64
img_size: 227 img_size: 227
img_size_y: 227
} }
} }
inputs { inputs {
@ -55,6 +60,8 @@ layers {
} }
bias_parameter_name: "___batch_norm_0__.wbias" bias_parameter_name: "___batch_norm_0__.wbias"
moving_average_fraction: 0.9 moving_average_fraction: 0.9
height: 227
width: 227
} }
layers { layers {
name: "__crmnorm_0__" name: "__crmnorm_0__"
@ -72,8 +79,12 @@ layers {
output_x: 227 output_x: 227
img_size: 227 img_size: 227
blocked: false blocked: false
output_y: 227
img_size_y: 227
} }
} }
height: 227
width: 227
} }
layers { layers {
name: "__pool_0__" name: "__pool_0__"
@ -97,6 +108,8 @@ layers {
padding_y: 0 padding_y: 0
} }
} }
height: 196
width: 196
} }
parameters { parameters {
name: "___conv_0__.w0" name: "___conv_0__.w0"

@ -26,6 +26,8 @@ layers {
filter_size_y: 32 filter_size_y: 32
padding_y: 1 padding_y: 1
stride_y: 1 stride_y: 1
output_y: 227
img_size_y: 256
} }
} }
bias_parameter_name: "___conv_0__.wbias" bias_parameter_name: "___conv_0__.wbias"
@ -43,6 +45,7 @@ layers {
image_conf { image_conf {
channels: 64 channels: 64
img_size: 256 img_size: 256
img_size_y: 256
} }
} }
inputs { inputs {
@ -55,6 +58,8 @@ layers {
} }
bias_parameter_name: "___batch_norm_0__.wbias" bias_parameter_name: "___batch_norm_0__.wbias"
moving_average_fraction: 0.9 moving_average_fraction: 0.9
height: 256
width: 256
} }
layers { layers {
name: "__crmnorm_0__" name: "__crmnorm_0__"
@ -72,8 +77,12 @@ layers {
output_x: 256 output_x: 256
img_size: 256 img_size: 256
blocked: false blocked: false
output_y: 256
img_size_y: 256
} }
} }
height: 256
width: 256
} }
layers { layers {
name: "__pool_0__" name: "__pool_0__"
@ -97,6 +106,8 @@ layers {
padding_y: 0 padding_y: 0
} }
} }
height: 225
width: 225
} }
parameters { parameters {
name: "___conv_0__.w0" name: "___conv_0__.w0"

@ -177,6 +177,8 @@ layers {
filter_size_y: 3 filter_size_y: 3
padding_y: 0 padding_y: 0
stride_y: 1 stride_y: 1
output_y: 30
img_size_y: 32
} }
num_filters: 64 num_filters: 64
} }

@ -26,11 +26,15 @@ layers {
filter_size_y: 3 filter_size_y: 3
padding_y: 1 padding_y: 1
stride_y: 1 stride_y: 1
output_y: 48
img_size_y: 48
} }
} }
bias_parameter_name: "___conv_0__.wbias" bias_parameter_name: "___conv_0__.wbias"
num_filters: 16 num_filters: 16
shared_biases: true shared_biases: true
height: 48
width: 48
} }
layers { layers {
name: "__bilinear_interp_layer_0__" name: "__bilinear_interp_layer_0__"
@ -40,11 +44,17 @@ layers {
inputs { inputs {
input_layer_name: "__conv_0__" input_layer_name: "__conv_0__"
bilinear_interp_conf { bilinear_interp_conf {
image_conf {
channels: 16
img_size: 48
img_size_y: 48
}
out_size_x: 64 out_size_x: 64
out_size_y: 64 out_size_y: 64
num_channels: 16
} }
} }
height: 64
width: 64
} }
layers { layers {
name: "__pool_0__" name: "__pool_0__"
@ -55,19 +65,21 @@ layers {
input_layer_name: "__bilinear_interp_layer_0__" input_layer_name: "__bilinear_interp_layer_0__"
pool_conf { pool_conf {
pool_type: "max-projection" pool_type: "max-projection"
channels: 4 channels: 16
size_x: 2 size_x: 2
stride: 2 stride: 2
output_x: 64 output_x: 32
img_size: 128 img_size: 64
padding: 0 padding: 0
size_y: 2 size_y: 2
stride_y: 2 stride_y: 2
output_y: 64 output_y: 32
img_size_y: 128 img_size_y: 64
padding_y: 0 padding_y: 0
} }
} }
height: 32
width: 32
} }
layers { layers {
name: "__fc_layer_0__" name: "__fc_layer_0__"
@ -78,6 +90,8 @@ layers {
input_layer_name: "__pool_0__" input_layer_name: "__pool_0__"
input_parameter_name: "___fc_layer_0__.w0" input_parameter_name: "___fc_layer_0__.w0"
} }
height: 32
width: 32
} }
parameters { parameters {
name: "___conv_0__.w0" name: "___conv_0__.w0"

@ -4,6 +4,8 @@ layers {
type: "data" type: "data"
size: 2304 size: 2304
active_type: "" active_type: ""
height: 48
width: 48
} }
layers { layers {
name: "__conv_0__" name: "__conv_0__"
@ -26,11 +28,15 @@ layers {
filter_size_y: 3 filter_size_y: 3
padding_y: 1 padding_y: 1
stride_y: 1 stride_y: 1
output_y: 48
img_size_y: 48
} }
} }
bias_parameter_name: "___conv_0__.wbias" bias_parameter_name: "___conv_0__.wbias"
num_filters: 16 num_filters: 16
shared_biases: true shared_biases: true
height: 48
width: 48
} }
layers { layers {
name: "__maxout_layer_0__" name: "__maxout_layer_0__"
@ -40,12 +46,16 @@ layers {
inputs { inputs {
input_layer_name: "__conv_0__" input_layer_name: "__conv_0__"
maxout_conf { maxout_conf {
channels: 16 image_conf {
channels: 16
img_size: 48
img_size_y: 48
}
groups: 2 groups: 2
img_size_x: 0
img_size_y: 0
} }
} }
height: 48
width: 48
} }
layers { layers {
name: "__pool_0__" name: "__pool_0__"
@ -69,48 +79,58 @@ layers {
padding_y: 0 padding_y: 0
} }
} }
height: 24
width: 24
} }
layers { layers {
name: "__conv_1__" name: "__conv_1__"
type: "exconv" type: "exconv"
size: 18432 size: 73728
active_type: "" active_type: ""
inputs { inputs {
input_layer_name: "__pool_0__" input_layer_name: "__pool_0__"
input_parameter_name: "___conv_1__.w0" input_parameter_name: "___conv_1__.w0"
conv_conf { conv_conf {
filter_size: 3 filter_size: 3
channels: 32 channels: 8
stride: 1 stride: 1
padding: 1 padding: 1
groups: 1 groups: 1
filter_channels: 32 filter_channels: 8
output_x: 12 output_x: 24
img_size: 12 img_size: 24
caffe_mode: true caffe_mode: true
filter_size_y: 3 filter_size_y: 3
padding_y: 1 padding_y: 1
stride_y: 1 stride_y: 1
output_y: 24
img_size_y: 24
} }
} }
bias_parameter_name: "___conv_1__.wbias" bias_parameter_name: "___conv_1__.wbias"
num_filters: 128 num_filters: 128
shared_biases: true shared_biases: true
height: 24
width: 24
} }
layers { layers {
name: "__maxout_layer_1__" name: "__maxout_layer_1__"
type: "maxout" type: "maxout"
size: 9216 size: 18432
active_type: "" active_type: ""
inputs { inputs {
input_layer_name: "__conv_0__" input_layer_name: "__conv_1__"
maxout_conf { maxout_conf {
channels: 128 image_conf {
channels: 128
img_size: 24
img_size_y: 24
}
groups: 4 groups: 4
img_size_x: 0
img_size_y: 0
} }
} }
height: 24
width: 24
} }
layers { layers {
name: "__block_expand_layer_0__" name: "__block_expand_layer_0__"
@ -118,7 +138,7 @@ layers {
size: 192 size: 192
active_type: "" active_type: ""
inputs { inputs {
input_layer_name: "__maxout_layer_0__" input_layer_name: "__maxout_layer_1__"
block_expand_conf { block_expand_conf {
channels: 32 channels: 32
stride_x: 1 stride_x: 1
@ -133,6 +153,8 @@ layers {
img_size_y: 0 img_size_y: 0
} }
} }
height: 24
width: 24
} }
layers { layers {
name: "__fc_layer_0__" name: "__fc_layer_0__"
@ -143,6 +165,8 @@ layers {
input_layer_name: "__block_expand_layer_0__" input_layer_name: "__block_expand_layer_0__"
input_parameter_name: "___fc_layer_0__.w0" input_parameter_name: "___fc_layer_0__.w0"
} }
height: 24
width: 24
} }
parameters { parameters {
name: "___conv_0__.w0" name: "___conv_0__.w0"
@ -164,9 +188,9 @@ parameters {
} }
parameters { parameters {
name: "___conv_1__.w0" name: "___conv_1__.w0"
size: 36864 size: 9216
initial_mean: 0.0 initial_mean: 0.0
initial_std: 0.0833333333333 initial_std: 0.166666666667
initial_strategy: 0 initial_strategy: 0
initial_smart: false initial_smart: false
} }

Some files were not shown because too many files have changed in this diff Show More

Loading…
Cancel
Save