You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
graphengine/ge/host_kernels/reshape_kernel.cc

89 lines
3.2 KiB

/**
* Copyright 2020 Huawei Technologies Co., Ltd
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include "host_kernels/reshape_kernel.h"
#include "common/ge_inner_error_codes.h"
#include "common/op/ge_op_utils.h"
#include "common/types.h"
#include "framework/common/debug/ge_log.h"
#include "host_kernels/kernel_utils.h"
#include "inc/kernel_factory.h"
namespace ge {
namespace {
const int kReshapeDataIndex = 0;
const int kOutputDescFirstIndex = 0;
const size_t kReshapeOutputSize = 1;
const size_t kReshapeInputSize = 2;
} // namespace
Status ReshapeKernel::Compute(const NodePtr &node_ptr) {
GELOGI("Reshape dimension kernel in.");
if (node_ptr == nullptr) {
GELOGE(PARAM_INVALID, "Parameter's invalid, Input is nullptr.");
return PARAM_INVALID;
}
Status ret = KernelUtils::CheckDimensionNodeInfo(node_ptr);
if (ret != SUCCESS) {
GELOGW("GetDimensionNodeInfo failed");
return ret;
}
if (!KernelUtils::CheckFormatSupported(node_ptr)) {
GELOGW("CheckFormatSupported failed");
return NOT_CHANGED;
}
GELOGI("Reshape dimension kernel success.");
return SUCCESS;
}
Status ReshapeKernel::Compute(const ge::OpDescPtr op_desc_ptr, const std::vector<ge::ConstGeTensorPtr> &input,
std::vector<ge::GeTensorPtr> &v_output) {
GELOGI("Reshape folding kernel in.");
if (op_desc_ptr == nullptr) {
GELOGE(PARAM_INVALID, "Input opdesc is nullptr.");
return PARAM_INVALID;
}
if ((input.size() != kReshapeInputSize) || (op_desc_ptr->GetOutputsSize() != kReshapeOutputSize)) {
GELOGW("Unexpected Reshape node, node input size: %zu, node output size: %zu, node name: %s", input.size(),
op_desc_ptr->GetOutputsSize(), op_desc_ptr->GetName().c_str());
return NOT_CHANGED;
}
auto output_tensor_desc = op_desc_ptr->GetOutputDesc(kOutputDescFirstIndex);
GeTensorPtr output_ptr = MakeShared<GeTensor>(output_tensor_desc);
if (output_ptr == nullptr) {
GELOGW("Failed to fold node %s, out of memory", op_desc_ptr->GetName().c_str());
return NOT_CHANGED;
}
// print output tensor information, and will be deleted
GELOGI("Reshape op %s output tensor data size is %zu", op_desc_ptr->GetName().c_str(), output_ptr->GetData().size());
size_t data_dim_size = output_ptr->GetTensorDesc().GetShape().GetDims().size();
GELOGI("Reshape op %s output tensor dim size is %zu", op_desc_ptr->GetName().c_str(), data_dim_size);
if (output_ptr->SetData(input.at(kReshapeDataIndex)->GetData()) != GRAPH_SUCCESS) {
GELOGW("Compute: SetData failed");
}
v_output.emplace_back(output_ptr);
GELOGI("Reshape folding kernel success.");
return SUCCESS;
}
REGISTER_KERNEL(RESHAPE, ReshapeKernel);
} // namespace ge