You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
1091 lines
40 KiB
1091 lines
40 KiB
/**
|
|
* Copyright 2020 Huawei Technologies Co., Ltd
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#include "executor/ge_executor.h"
|
|
#include <cce/cce.h>
|
|
#include <ctime>
|
|
#include <iostream>
|
|
#include "common/debug/log.h"
|
|
#include "common/ge/ge_util.h"
|
|
#include "common/helper/model_helper.h"
|
|
#include "common/profiling/profiling_manager.h"
|
|
#include "common/dump/dump_manager.h"
|
|
#include "graph/execute/graph_execute.h"
|
|
#include "graph/load/graph_loader.h"
|
|
#include "graph/load/model_manager/model_manager.h"
|
|
#include "graph/manager/graph_mem_allocator.h"
|
|
#include "single_op/single_op_manager.h"
|
|
#include "graph/load/model_manager/davinci_model.h"
|
|
#include "opskernel_manager/ops_kernel_builder_manager.h"
|
|
#include "graph/opsproto_manager.h"
|
|
#include "ge_local_engine/engine/host_cpu_engine.h"
|
|
|
|
using std::string;
|
|
using std::vector;
|
|
|
|
namespace {
|
|
const size_t kDynamicBatchSizeVecSize = 1;
|
|
const size_t kStaticBatchInfoSize = 1;
|
|
const size_t kDynamicImageSizeVecSize = 2;
|
|
const size_t kDynamicImageSizeInputSize = 2;
|
|
const char *const kBatchLabel = "Batch_";
|
|
|
|
void GetGeTensorDescFromDomiInfo(std::vector<ge::TensorDesc> &ge_descs,
|
|
const std::vector<ge::InputOutputDescInfo> &domi_descs,
|
|
const std::vector<uint32_t> &formats) {
|
|
uint32_t idx = 0;
|
|
for (auto desc_item : domi_descs) {
|
|
ge::TensorDesc ge_desc;
|
|
ge_desc.SetName(desc_item.name.c_str());
|
|
ge_desc.SetDataType(static_cast<ge::DataType>(desc_item.data_type));
|
|
ge_desc.SetFormat(static_cast<ge::Format>(formats[idx]));
|
|
std::vector<int64_t> shape_dims;
|
|
for (auto dim : desc_item.shape_info.dims) {
|
|
shape_dims.push_back(dim);
|
|
}
|
|
ge::Shape ge_shape(shape_dims);
|
|
ge_desc.SetShape(ge_shape);
|
|
ge_desc.SetSize(desc_item.size);
|
|
ge_desc.SetShapeRange(desc_item.shape_info.shape_ranges);
|
|
ge_descs.emplace_back(ge_desc);
|
|
++idx;
|
|
}
|
|
}
|
|
|
|
void GetDomiInputData(const ge::RunModelData &input_data, ge::InputData &inputs) {
|
|
inputs.index = input_data.index;
|
|
inputs.model_id = input_data.modelId;
|
|
inputs.timestamp = input_data.timestamp;
|
|
inputs.timeout = input_data.timeout;
|
|
inputs.request_id = input_data.request_id;
|
|
for (const auto &data_item : input_data.blobs) {
|
|
ge::DataBuffer dataBuf{data_item.data, data_item.length, data_item.isDataSupportMemShare};
|
|
inputs.blobs.emplace_back(dataBuf);
|
|
}
|
|
}
|
|
|
|
void GetDomiOutputData(const ge::RunModelData &output_data, ge::OutputData &outputs) {
|
|
outputs.index = output_data.index;
|
|
outputs.model_id = output_data.modelId;
|
|
for (const auto &data_item : output_data.blobs) {
|
|
ge::DataBuffer dataBuf(data_item.data, data_item.length, data_item.isDataSupportMemShare);
|
|
outputs.blobs.emplace_back(dataBuf);
|
|
}
|
|
}
|
|
|
|
void SetDynamicInputDataFlag(const ge::RunModelData &input_data, const std::vector<std::vector<int64_t>> batch_info,
|
|
ge::InputData &inputs) {
|
|
inputs.is_dynamic_batch = true;
|
|
std::string batch_label;
|
|
size_t match_idx = 0;
|
|
for (size_t i = 0; i < batch_info.size(); ++i) {
|
|
// dynamic_dims
|
|
if (input_data.dynamic_dims.size() != 0) {
|
|
bool is_match = true;
|
|
for (size_t j = 0; j < static_cast<size_t>(input_data.dynamic_dims.size()); ++j) {
|
|
if (static_cast<uint64_t>(batch_info[i][j]) != input_data.dynamic_dims[j]) {
|
|
is_match = false;
|
|
break;
|
|
}
|
|
}
|
|
if (is_match) {
|
|
match_idx = i;
|
|
break;
|
|
}
|
|
// dynamic_batch_size
|
|
} else if (batch_info[i].size() == kDynamicBatchSizeVecSize &&
|
|
batch_info[i][0] == static_cast<int64_t>(input_data.dynamic_batch_size)) {
|
|
match_idx = i;
|
|
break;
|
|
// dynamic_image_size
|
|
} else if (batch_info[i].size() == kDynamicImageSizeVecSize &&
|
|
batch_info[i][0] == static_cast<int64_t>(input_data.dynamic_image_height) &&
|
|
batch_info[i][1] == static_cast<int64_t>(input_data.dynamic_image_width)) {
|
|
match_idx = i;
|
|
break;
|
|
}
|
|
}
|
|
batch_label = kBatchLabel + std::to_string(match_idx);
|
|
inputs.batch_label = batch_label;
|
|
GELOGI("current batch label:%s", batch_label.c_str());
|
|
}
|
|
|
|
bool IsDynamicBatchSizeMatchModel(uint64_t batch_size, const vector<std::vector<int64_t>> &batch_info) {
|
|
if (batch_info.empty()) {
|
|
GELOGE(ge::FAILED, "Dynamic batch info is empty.");
|
|
return false;
|
|
}
|
|
|
|
for (auto batch : batch_info) {
|
|
if (batch.size() != kDynamicBatchSizeVecSize) {
|
|
GELOGE(ge::FAILED, "Dynamic batch param num is %zu, current batch size is %zu.", kDynamicBatchSizeVecSize,
|
|
batch.size());
|
|
return false;
|
|
}
|
|
if (batch[0] == static_cast<int64_t>(batch_size)) {
|
|
return true;
|
|
}
|
|
}
|
|
GELOGE(ge::FAILED, "Dynamic batch %lu can not match the gear of model.", batch_size);
|
|
return false;
|
|
}
|
|
|
|
bool IsDynamicImageSizeMatchModel(uint64_t image_height, uint64_t image_width,
|
|
const vector<std::vector<int64_t>> &batch_info) {
|
|
if (batch_info.empty()) {
|
|
GELOGE(ge::FAILED, "Dynamic batch info is empty.");
|
|
return false;
|
|
}
|
|
|
|
for (auto resolution : batch_info) {
|
|
if (resolution.size() != kDynamicImageSizeVecSize) {
|
|
GELOGE(ge::FAILED, "Dynamic resolution param num is %zu, current resolution size is %zu.",
|
|
kDynamicImageSizeVecSize, resolution.size());
|
|
return false;
|
|
}
|
|
if (resolution[0] == static_cast<int64_t>(image_height) && resolution[1] == static_cast<int64_t>(image_width)) {
|
|
return true;
|
|
}
|
|
}
|
|
|
|
GELOGE(ge::FAILED, "Dynamic resolution (%lu,%lu) can not match the gear of model.", image_height, image_width);
|
|
return false;
|
|
}
|
|
|
|
bool IsDynmaicDimsSizeMatchModel(const vector<uint64_t> cur_dynamic_dims,
|
|
const vector<vector<int64_t>> &batch_info) {
|
|
if (batch_info.empty()) {
|
|
GELOGE(ACL_ERROR_GE_PARAM_INVALID, "Dynamic batch info is empty.");
|
|
return false;
|
|
}
|
|
|
|
bool find_match = false;
|
|
for (auto resolution : batch_info) {
|
|
if (cur_dynamic_dims.size() != resolution.size()) {
|
|
GELOGE(ACL_ERROR_GE_PARAM_INVALID, "Cur dynamic dims param num is %zu, current resolution size is %zu.",
|
|
cur_dynamic_dims.size(), resolution.size());
|
|
return false;
|
|
}
|
|
bool flag = true;
|
|
for (std::size_t i = 0; i < resolution.size(); ++i) {
|
|
if (cur_dynamic_dims[i] != static_cast<uint64_t>(resolution[i])) {
|
|
flag = false;
|
|
break;
|
|
}
|
|
}
|
|
if (flag) {
|
|
find_match = true;
|
|
break;
|
|
}
|
|
}
|
|
if (!find_match) {
|
|
GELOGE(ACL_ERROR_GE_PARAM_INVALID, "choose dynamic dims can not match the gear of model.");
|
|
}
|
|
return find_match;
|
|
}
|
|
} // namespace
|
|
|
|
namespace ge {
|
|
bool GeExecutor::isInit_ = false;
|
|
|
|
static void InitOpsProtoManager() {
|
|
string opsproto_path;
|
|
const char *path_env = std::getenv("ASCEND_OPP_PATH");
|
|
if (path_env != nullptr) {
|
|
string path = path_env;
|
|
string file_path = RealPath(path.c_str());
|
|
if (file_path.empty()) {
|
|
GELOGE(FAILED, "[Check][EnvPath]ASCEND_OPP_PATH path [%s] is invalid.", path.c_str());
|
|
REPORT_INPUT_ERROR("E68016", {"ASCEND_OPP_PATH", path});
|
|
return;
|
|
}
|
|
opsproto_path = (path + "/op_proto/custom/" + ":") + (path + "/op_proto/built-in/");
|
|
GELOGI("Get opsproto so path from env : %s", path.c_str());
|
|
} else {
|
|
string path_base = PluginManager::GetPath();
|
|
GELOGI("path_base is %s", path_base.c_str());
|
|
path_base = path_base.substr(0, path_base.rfind('/'));
|
|
path_base = path_base.substr(0, path_base.rfind('/') + 1);
|
|
opsproto_path = (path_base + "ops/op_proto/custom/" + ":") + (path_base + "ops/op_proto/built-in/");
|
|
}
|
|
GELOGI("Get opsproto path is %s", opsproto_path.c_str());
|
|
OpsProtoManager *manager = OpsProtoManager::Instance();
|
|
map<string, string> option_tmp;
|
|
option_tmp.emplace(std::pair<string, string>(string("ge.opsProtoLibPath"), opsproto_path));
|
|
(void)manager->Initialize(option_tmp);
|
|
}
|
|
|
|
GeExecutor::GeExecutor() {}
|
|
|
|
Status GeExecutor::Initialize() {
|
|
GELOGI("Init GeExecutor begin.");
|
|
if (isInit_) {
|
|
GELOGW("Already initialized, no need to be initialized again.");
|
|
return ge::SUCCESS;
|
|
}
|
|
|
|
OpTilingManager::GetInstance().LoadSo();
|
|
|
|
Status init_hostcpu_engine_status = HostCpuEngine::GetInstance().Initialize();
|
|
if (init_hostcpu_engine_status != SUCCESS) {
|
|
GELOGE(init_hostcpu_engine_status, "Failed to initialize HostCpuEngine");
|
|
return init_hostcpu_engine_status;
|
|
}
|
|
|
|
InitOpsProtoManager();
|
|
|
|
std::vector<rtMemType_t> mem_type(1, RT_MEMORY_HBM);
|
|
mem_type.push_back(RT_MEMORY_P2P_DDR);
|
|
auto ret = MemManager::Instance().Initialize(mem_type);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "Memory Manager init failed.");
|
|
return ret;
|
|
}
|
|
|
|
GE_CHK_STATUS_RET(OpsKernelBuilderManager::Instance().Initialize({}, false),
|
|
"Failed to initialize OpsKernelBuilders.");
|
|
|
|
// Start profiling
|
|
Options profiling_options;
|
|
profiling_options.device_id = 0;
|
|
// job id need to be set, the value is meaningless;
|
|
profiling_options.job_id = "1";
|
|
ProfilingManager::Instance().Init(profiling_options);
|
|
|
|
isInit_ = true;
|
|
GELOGI("Init GeExecutor over.");
|
|
return ge::SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::Finalize() {
|
|
GELOGI("Uninit GeExecutor begin.");
|
|
if (isInit_ == false) {
|
|
GELOGW("GeExecutor has not been initialized.");
|
|
return ge::SUCCESS;
|
|
}
|
|
|
|
(void) OpsKernelBuilderManager::Instance().Finalize();
|
|
|
|
// Stop profiling
|
|
if (ProfilingManager::Instance().ProfilingOn()) {
|
|
ProfilingManager::Instance().StopProfiling();
|
|
ProfilingManager::Instance().PluginUnInit();
|
|
}
|
|
|
|
GELOGI("Uninit GeExecutor over.");
|
|
return ge::SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::SetDynamicBatchSize(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
|
|
uint64_t batch_size) {
|
|
if (dynamic_input_addr == nullptr) {
|
|
GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID, "Dynamic input addr is nullptr!");
|
|
return ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID;
|
|
}
|
|
|
|
uint64_t size = sizeof(uint32_t);
|
|
if (length < size) {
|
|
GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID, "Dynamic input size [%lu] is less than [%lu]!", length, size);
|
|
return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
|
|
}
|
|
if (length >= sizeof(uint64_t)) {
|
|
size = sizeof(uint64_t);
|
|
}
|
|
|
|
// Verify whether the input dynamic batch matches the model gear
|
|
std::vector<std::vector<int64_t>> batch_info;
|
|
std::vector<uint64_t> batch_num{batch_size};
|
|
int32_t dynamic_type = static_cast<int32_t>(FIXED);
|
|
Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "Get dynamic input info failed.");
|
|
return ret;
|
|
}
|
|
|
|
if (!IsDynamicBatchSizeMatchModel(batch_size, batch_info)) {
|
|
GELOGE(ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID, "The current dynamic input does not match the gear of the model.");
|
|
return ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID;
|
|
}
|
|
|
|
ret = GraphExecutor::SetDynamicSize(model_id, batch_num, static_cast<int32_t>(DYNAMIC_BATCH));
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "Set dynamic size failed");
|
|
return ret;
|
|
}
|
|
// memcpy dynamic_batch_size from host to device
|
|
rtError_t rt_ret = rtMemcpy(dynamic_input_addr, length, &batch_size, size, RT_MEMCPY_HOST_TO_DEVICE);
|
|
if (rt_ret != RT_ERROR_NONE) {
|
|
GELOGE(rt_ret, "memcpy dynamic batch input data failed! ret: 0x%X", rt_ret);
|
|
return RT_ERROR_TO_GE_STATUS(rt_ret);
|
|
}
|
|
return SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::SetDynamicImageSize(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
|
|
uint64_t image_height, uint64_t image_width) {
|
|
if (dynamic_input_addr == nullptr) {
|
|
GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID, "Dynamic input addr is nullptr!");
|
|
return ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID;
|
|
}
|
|
|
|
uint64_t dynamic_input_size = kDynamicImageSizeInputSize * sizeof(uint32_t);
|
|
if (length < dynamic_input_size) {
|
|
GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID,
|
|
"Dynamic input size [%lu] is less than [%lu]!", length, dynamic_input_size);
|
|
return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
|
|
}
|
|
uint64_t size = sizeof(uint32_t);
|
|
if (length >= kDynamicImageSizeInputSize * sizeof(uint64_t)) {
|
|
size = sizeof(uint64_t);
|
|
}
|
|
// Verify whether the input dynamic resolution matches the model gear
|
|
std::vector<std::vector<int64_t>> batch_info;
|
|
std::vector<uint64_t> batch_num{image_height, image_width};
|
|
int32_t dynamic_type = static_cast<int32_t>(FIXED);
|
|
Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "Get dynamic input info failed.");
|
|
return ret;
|
|
}
|
|
|
|
if (!IsDynamicImageSizeMatchModel(image_height, image_width, batch_info)) {
|
|
GELOGE(ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID, "The current dynamic input does not match the gear of the model.");
|
|
return ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID;
|
|
}
|
|
|
|
ret = GraphExecutor::SetDynamicSize(model_id, batch_num, static_cast<int32_t>(DYNAMIC_IMAGE));
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "Set dynamic size failed");
|
|
return ret;
|
|
}
|
|
|
|
// Memcpy dynamic resolution height from host to device
|
|
rtError_t rt_ret =
|
|
rtMemcpy(dynamic_input_addr, size, &image_height, size, RT_MEMCPY_HOST_TO_DEVICE);
|
|
if (rt_ret != RT_ERROR_NONE) {
|
|
GELOGE(rt_ret, "memcpy dynamic resolution input data failed! ret: 0x%X", rt_ret);
|
|
return RT_ERROR_TO_GE_STATUS(rt_ret);
|
|
}
|
|
|
|
uint64_t remain_size = length - size;
|
|
// Memcpy dynamic resolution width from host to device
|
|
rt_ret = rtMemcpy(reinterpret_cast<void *>(reinterpret_cast<uint8_t *>(dynamic_input_addr) + size),
|
|
remain_size, &image_width, size, RT_MEMCPY_HOST_TO_DEVICE);
|
|
if (rt_ret != RT_ERROR_NONE) {
|
|
GELOGE(rt_ret, "memcpy dynamic resolution input data failed!");
|
|
return RT_ERROR_TO_GE_STATUS(rt_ret);
|
|
}
|
|
return SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::SetDynamicDims(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
|
|
const vector<uint64_t> &dynamic_dims) {
|
|
if (dynamic_input_addr == nullptr) {
|
|
GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID, "Dynamic input addr is nullptr!");
|
|
return ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID;
|
|
}
|
|
|
|
vector<uint64_t> cur_dynamic_dims;
|
|
Status ret = GetCurDynamicDims(model_id, dynamic_dims, cur_dynamic_dims);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "Set cur gear dynamic dims failed");
|
|
return ret;
|
|
}
|
|
std::vector<std::vector<int64_t>> batch_info;
|
|
int32_t dynamic_type = static_cast<int32_t>(FIXED);
|
|
ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "Get dynamic input info failed.");
|
|
return ret;
|
|
}
|
|
|
|
if (!IsDynmaicDimsSizeMatchModel(cur_dynamic_dims, batch_info)) {
|
|
GELOGE(ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID, "The current dynamic input does not match the gear of the model.");
|
|
return ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID;
|
|
}
|
|
|
|
ret = GraphExecutor::SetDynamicSize(model_id, cur_dynamic_dims, static_cast<int32_t>(DYNAMIC_DIMS));
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "Set dynamic size failed");
|
|
return ret;
|
|
}
|
|
|
|
size_t dynamic_dim_num = cur_dynamic_dims.size();
|
|
uint64_t dynamic_input_size = static_cast<uint64_t>(dynamic_dim_num * sizeof(uint32_t));
|
|
if (length < dynamic_input_size) {
|
|
GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID,
|
|
"Dynamic input size [%lu] is less than [%lu]!", length, dynamic_input_size);
|
|
return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
|
|
}
|
|
uint64_t size = sizeof(uint32_t);
|
|
if (length >= dynamic_dim_num * sizeof(uint64_t)) {
|
|
size = sizeof(uint64_t);
|
|
}
|
|
rtError_t rt_ret;
|
|
for (uint32_t i = 0; i < dynamic_dim_num; ++i) {
|
|
// Memcpy dynamic dim[i] from host to device
|
|
rt_ret = rtMemcpy(reinterpret_cast<void *>(reinterpret_cast<uint8_t *>(dynamic_input_addr) + size * i),
|
|
length - size * i, &cur_dynamic_dims[i], size, RT_MEMCPY_HOST_TO_DEVICE);
|
|
if (rt_ret != RT_ERROR_NONE) {
|
|
GELOGE(rt_ret, "memcpy dynamic resolution input data failed!");
|
|
return RT_ERROR_TO_GE_STATUS(rt_ret);
|
|
}
|
|
}
|
|
return SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::GetCurDynamicDims(uint32_t model_id, const vector<uint64_t> &dynamic_dims,
|
|
vector<uint64_t> &cur_dynamic_dims) {
|
|
cur_dynamic_dims.clear();
|
|
vector<ge::TensorDesc> input_desc;
|
|
vector<ge::TensorDesc> output_desc;
|
|
auto ret = GetModelDescInfo(model_id, input_desc, output_desc);
|
|
if (ret != ge::SUCCESS) {
|
|
GELOGE(ret, "GetModelDescInfo failed.");
|
|
return ret;
|
|
}
|
|
vector<string> user_designate_shape_order;
|
|
vector<int64_t> all_data_dims;
|
|
ret = GetUserDesignateShapeOrder(model_id, user_designate_shape_order);
|
|
if (ret != ge::SUCCESS) {
|
|
GELOGE(ret, "GetUserDesignateShapeOrder failed.");
|
|
return ret;
|
|
}
|
|
for (auto &data_name : user_designate_shape_order) {
|
|
for (auto &desc : input_desc) {
|
|
if (desc.GetName() == data_name) {
|
|
for (auto dim : desc.GetShape().GetDims()) {
|
|
all_data_dims.push_back(dim);
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
if (dynamic_dims.size() != all_data_dims.size()){
|
|
GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID,
|
|
"Dynamic input size [%lu] is not equal with all data dims size [%lu]!",
|
|
dynamic_dims.size(), all_data_dims.size());
|
|
return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
|
|
}
|
|
for (std::size_t i = 0; i < all_data_dims.size(); ++i) {
|
|
if (all_data_dims[i] < 0) {
|
|
cur_dynamic_dims.push_back(dynamic_dims[i]);
|
|
} else if (static_cast<uint64_t>(all_data_dims[i]) != dynamic_dims[i]) {
|
|
GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID,
|
|
"Static dims should be same, index: %zu value: %lu should be %ld",
|
|
i, dynamic_dims[i], all_data_dims[i]);
|
|
return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
|
|
}
|
|
}
|
|
return SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::GetCurShape(const uint32_t model_id, std::vector<int64_t> &batch_info, int32_t &dynamic_type) {
|
|
GELOGI("Begin to get current shape");
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
Status ret = GraphExecutor::GetCurShape(model_id, batch_info, dynamic_type);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "Get current shape failed");
|
|
return ret;
|
|
}
|
|
return SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::SetDynamicAippData(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
|
|
const std::vector<kAippDynamicBatchPara> &aippBatchPara,
|
|
const kAippDynamicPara &aippParms) {
|
|
GELOGI("Enter to SetDynamicAippData.");
|
|
if (dynamic_input_addr == nullptr) {
|
|
GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID, "Dynamic aipp input addr is nullptr!");
|
|
return ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID;
|
|
}
|
|
if (aippBatchPara.empty()) {
|
|
GELOGE(ACL_ERROR_GE_AIPP_BATCH_EMPTY, "aippBatchPara is empty.");
|
|
return ACL_ERROR_GE_AIPP_BATCH_EMPTY;
|
|
}
|
|
uint64_t batch_num = aippBatchPara.size();
|
|
uint64_t real_aippParms_size = sizeof(kAippDynamicPara) - sizeof(kAippDynamicBatchPara);
|
|
uint64_t struct_len = batch_num * sizeof(kAippDynamicBatchPara) + real_aippParms_size;
|
|
GELOGI(
|
|
"Get acl input dynamic aipp data, model_id is %u, length is %lu,"
|
|
"batch num is %lu, struct_len is %lu",
|
|
model_id, length, batch_num, struct_len);
|
|
if (struct_len > length) {
|
|
GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID,
|
|
"input dynamic aipp param len [%lu] is larger than aipp_data size [%lu]", struct_len, length);
|
|
return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
|
|
}
|
|
// Memcpy real kAippDynamicBatchPara from host to device
|
|
rtError_t rt_ret = rtMemcpy(dynamic_input_addr, length, &aippParms, real_aippParms_size, RT_MEMCPY_HOST_TO_DEVICE);
|
|
if (rt_ret != RT_ERROR_NONE) {
|
|
GELOGE(rt_ret, "memcpy real_aippParms_size failed! ret: 0x%X", rt_ret);
|
|
return RT_ERROR_TO_GE_STATUS(rt_ret);
|
|
}
|
|
uint64_t remain_len = length - real_aippParms_size;
|
|
uint8_t *aipp_batch_para_dev = reinterpret_cast<uint8_t *>(dynamic_input_addr) + real_aippParms_size;
|
|
|
|
for (uint64_t i = 0; i < batch_num; ++i) {
|
|
rt_ret = rtMemcpy(reinterpret_cast<void *>(aipp_batch_para_dev + i * sizeof(kAippDynamicBatchPara)),
|
|
(remain_len - i * sizeof(kAippDynamicBatchPara)), &(aippBatchPara[i]),
|
|
sizeof(kAippDynamicBatchPara), RT_MEMCPY_HOST_TO_DEVICE);
|
|
if (rt_ret != RT_ERROR_NONE) {
|
|
GELOGE(rt_ret, "memcpy kAippDynamicBatchPara input data failed! ret: 0x%X", rt_ret);
|
|
return RT_ERROR_TO_GE_STATUS(rt_ret);
|
|
}
|
|
}
|
|
return SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::UnloadModel(uint32_t model_id) {
|
|
GELOGD("unload model %u begin.", model_id);
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
Status ret = GraphLoader::DestroyAicpuSessionForInfer(model_id);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "[GraphLoader] DestroyAicpuSessionForInfer failed. model id: %u", model_id);
|
|
return ret;
|
|
}
|
|
|
|
std::shared_ptr<hybrid::HybridDavinciModel> hybrid_davinci_model =
|
|
ModelManager::GetInstance()->GetHybridModel(model_id);
|
|
if (hybrid_davinci_model != nullptr) {
|
|
uint64_t session_id = hybrid_davinci_model->GetSessionId();
|
|
VarManagerPool::Instance().RemoveVarManager(session_id);
|
|
} else {
|
|
std::shared_ptr<DavinciModel> davinci_model = ModelManager::GetInstance()->GetModel(model_id);
|
|
if (davinci_model != nullptr) {
|
|
uint64_t session_id = davinci_model->GetSessionId();
|
|
VarManagerPool::Instance().RemoveVarManager(session_id);
|
|
}
|
|
}
|
|
ret = GraphLoader::UnloadModel(model_id);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "[GraphLoader] DestroyAicpuSessionForInfer failed. model id: %u", model_id);
|
|
return ret;
|
|
}
|
|
return SUCCESS;
|
|
}
|
|
|
|
// Get input and output descriptor
|
|
Status GeExecutor::GetModelDescInfo(uint32_t model_id, std::vector<ge::TensorDesc> &input_desc,
|
|
std::vector<ge::TensorDesc> &output_desc, bool new_model_desc) {
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
|
|
std::vector<InputOutputDescInfo> input_desc_infos;
|
|
std::vector<InputOutputDescInfo> output_desc_infos;
|
|
std::vector<uint32_t> input_formats;
|
|
std::vector<uint32_t> output_formats;
|
|
|
|
Status ret = GraphExecutor::GetInputOutputDescInfo(model_id, input_desc_infos, output_desc_infos, input_formats,
|
|
output_formats, new_model_desc);
|
|
if (ret != domi::SUCCESS) {
|
|
GELOGE(ret, "GetInputOutputDescInfo failed. ret = %u", ret);
|
|
return ACL_ERROR_GE_GET_TENSOR_INFO;
|
|
}
|
|
|
|
if (input_formats.size() != input_desc_infos.size()) {
|
|
GELOGE(ACL_ERROR_GE_PARAM_INVALID,
|
|
"input_formats size %zu is not equal to input_desc_infos size %zu.",
|
|
input_formats.size(), input_desc_infos.size());
|
|
return ACL_ERROR_GE_PARAM_INVALID;
|
|
}
|
|
|
|
if (output_formats.size() != output_desc_infos.size()) {
|
|
GELOGE(ACL_ERROR_GE_PARAM_INVALID, "output_formats size %zu is not equal to output_desc_infos size %zu.",
|
|
output_formats.size(), output_desc_infos.size());
|
|
return ACL_ERROR_GE_PARAM_INVALID;
|
|
}
|
|
|
|
// Transfer data to TensorDesc
|
|
GetGeTensorDescFromDomiInfo(input_desc, input_desc_infos, input_formats);
|
|
GetGeTensorDescFromDomiInfo(output_desc, output_desc_infos, output_formats);
|
|
|
|
return ge::SUCCESS;
|
|
}
|
|
|
|
///
|
|
/// @ingroup ge
|
|
/// @brief Get dynamic batch_info
|
|
/// @param [in] model_id
|
|
/// @param [out] batch_info
|
|
/// @param [out] dynamic_type
|
|
/// @return execute result
|
|
///
|
|
Status GeExecutor::GetDynamicBatchInfo(uint32_t model_id, std::vector<std::vector<int64_t>> &batch_info,
|
|
int32_t &dynamic_type) {
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
|
|
Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "GetDynamicBatchInfo failed.");
|
|
return ret;
|
|
}
|
|
return SUCCESS;
|
|
}
|
|
|
|
///
|
|
/// @ingroup ge
|
|
/// @brief Get combined dynamic dims info
|
|
/// @param [in] model_id
|
|
/// @param [out] batch_info
|
|
/// @return execute result
|
|
///
|
|
Status GeExecutor::GetCombinedDynamicDims(uint32_t model_id, vector<vector<int64_t>> &batch_info) {
|
|
GELOGI("Begin to get combined dynamic dims info.");
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
|
|
Status ret = GraphExecutor::GetCombinedDynamicDims(model_id, batch_info);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "GetCombinedDynamicDims failed.");
|
|
return ret;
|
|
}
|
|
|
|
GELOGI("Get combined dynamic dims succ.");
|
|
return SUCCESS;
|
|
}
|
|
|
|
///
|
|
/// @ingroup ge
|
|
/// @brief Get user designeate shape order
|
|
/// @param [in] model_id
|
|
/// @param [out] user_designate_shape_order
|
|
/// @return execute result
|
|
///
|
|
Status GeExecutor::GetUserDesignateShapeOrder(uint32_t model_id, vector<string> &user_designate_shape_order) {
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
|
|
Status ret = GraphExecutor::GetUserDesignateShapeOrder(model_id, user_designate_shape_order);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "GetUserDesignateShapeOrder failed.");
|
|
return ret;
|
|
}
|
|
|
|
return SUCCESS;
|
|
}
|
|
|
|
///
|
|
/// @ingroup ge
|
|
/// @brief Get AIPP input format
|
|
/// @param [in] model_id
|
|
/// @param [in] index
|
|
/// @param [out] input_format
|
|
/// @return execute result
|
|
///
|
|
Status GeExecutor::GetAIPPInfo(uint32_t model_id, uint32_t index, AippConfigInfo &aipp_info) {
|
|
GELOGI("Begin to GetAIPPInfo.");
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
Status ret = GraphExecutor::GetAippInfo(model_id, index, aipp_info);
|
|
if (ret != SUCCESS) {
|
|
GELOGW("GetAIPPInfo is not success.");
|
|
return ret;
|
|
}
|
|
GELOGI("GetAIPPInfo succ.");
|
|
return SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::GetAippType(uint32_t model_id, uint32_t index, InputAippType &type, size_t &aipp_index) {
|
|
GELOGI("Begin to get aipp type.");
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
Status ret = GraphExecutor::GetAippType(model_id, index, type, aipp_index);
|
|
if (ret != SUCCESS) {
|
|
GELOGW("Get aipp type is not success.");
|
|
return ret;
|
|
}
|
|
GELOGI("Get aipp type success.");
|
|
return SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::GetModelAttr(uint32_t model_id, std::vector<std::string> &dynamic_output_shape_info) {
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
Status ret = GraphExecutor::GetModelAttr(model_id, dynamic_output_shape_info);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "Get dynamic batch output shape info failed.");
|
|
return ret;
|
|
}
|
|
return SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::CommandHandle(const Command &command) {
|
|
Status ret = GraphLoader::CommandHandle(command);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ACL_ERROR_GE_COMMAND_HANDLE, "CommandHandle: Command Handle failed.");
|
|
return ACL_ERROR_GE_COMMAND_HANDLE;
|
|
}
|
|
return SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::GetMaxUsedMemory(uint32_t model_id, uint32_t &max_size) {
|
|
GELOGI("Get max used memory begin.");
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
|
|
uint64_t max_mem_size = 0;
|
|
Status ret = GraphLoader::GetMaxUsedMemory(model_id, max_mem_size);
|
|
max_size = static_cast<uint32_t>(max_mem_size);
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* @ingroup ge
|
|
* @brief Load data from model file to memory
|
|
* @param [in] const std::string &path: Offline model file path
|
|
* @param [out] domi::ModelData &model_data: Offline model memory data
|
|
* @return SUCCESS handle successfully / others handle failed
|
|
*/
|
|
Status GeExecutor::LoadDataFromFile(const std::string &path, ModelData &model_data) {
|
|
GELOGI("Load data from file begin.");
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
|
|
string filePath = RealPath(path.c_str());
|
|
if (filePath.empty()) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_MODEL_PATH_INVALID,
|
|
"File path is invalid. please check your text file '%s'.", path.c_str());
|
|
return ACL_ERROR_GE_EXEC_MODEL_PATH_INVALID;
|
|
}
|
|
GELOGI("load modelData from file: %s.", path.c_str());
|
|
std::string key_path;
|
|
int32_t priority = 0;
|
|
Status ret = GraphLoader::LoadDataFromFile(path, key_path, priority, model_data);
|
|
if (ret != SUCCESS) {
|
|
if (model_data.model_data != nullptr) {
|
|
delete[] static_cast<char *>(model_data.model_data);
|
|
model_data.model_data = nullptr;
|
|
}
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* @ingroup ge
|
|
* @brief Load model from offline model memory data
|
|
* @param [in] domi::ModelData &model_data: Offline model data
|
|
void *dev_ptr: Input/Output memory start address
|
|
size_t memsize: Input/Output memory length
|
|
void *weight_ptr: Weight memory start address
|
|
size_t weightsize: Weight memory length
|
|
* @param [out] uint32_t &model_id: identification after model loading
|
|
* @return SUCCESS handle successfully / others handle failed
|
|
*/
|
|
Status GeExecutor::LoadModelFromData(uint32_t &model_id, const ModelData &model_data, void *dev_ptr, size_t mem_size,
|
|
void *weight_ptr, size_t weight_size) {
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
|
|
return GraphLoader::LoadModelFromData(model_id, model_data, dev_ptr, mem_size, weight_ptr, weight_size);
|
|
}
|
|
|
|
/**
|
|
* @ingroup ge
|
|
* @brief Load task list from ModelData with queue.
|
|
* @param [out] model_id: model id allocate from manager.
|
|
* @param [in] ge_model_data: Model data load from offline model.
|
|
* @param [in] input_queue_ids: input queue ids create from user.
|
|
* @param [in] output_queue_ids: input queue ids create from user.
|
|
* @return: 0 for success / others for fail
|
|
*/
|
|
Status GeExecutor::LoadModelWithQ(uint32_t &model_id, const ModelData &model_data,
|
|
const std::vector<uint32_t> &input_queue_ids,
|
|
const std::vector<uint32_t> &output_queue_ids) {
|
|
GELOGI("Load model with queue begin.");
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
return GraphLoader::LoadModelWithQ(model_id, model_data, input_queue_ids, output_queue_ids);
|
|
}
|
|
|
|
/**
|
|
* @ingroup ge
|
|
* @brief Synchronous execution of offline model(Do not create thread)
|
|
* @param [in] uint32_t model_id: Model ID to execute
|
|
void* stream: stream to execute
|
|
const domi::InputData *input_data: Model input data
|
|
bool async_mode: is asynchronize mode.
|
|
* @param [out] domi::OutputData *output_data: Model output data
|
|
* @return SUCCESS handle successfully / others handle failed
|
|
*/
|
|
Status GeExecutor::ExecModel(uint32_t model_id, void *stream, const ge::RunModelData &run_input_data,
|
|
ge::RunModelData &run_output_data, bool async_mode) {
|
|
std::vector<GeTensorDesc> input_desc = {};
|
|
std::vector<GeTensorDesc> output_desc = {};
|
|
return ExecModel(model_id, stream, run_input_data, input_desc, run_output_data, output_desc, async_mode);
|
|
}
|
|
|
|
/**
|
|
* @ingroup ge
|
|
* @brief Synchronous execution of offline model(Do not create thread)
|
|
* @param [in] uint32_t model_id: Model ID to execute
|
|
void* stream: stream to execute
|
|
const domi::InputData *input_data: Model input data
|
|
const std::vector<GeTensorDesc> &input_desc: Description of model input data
|
|
bool async_mode: is asynchronize mode
|
|
* @param [out] domi::OutputData *output_data: Model output data
|
|
* @param [out] std::vector<GeTensorDesc> &output_desc: Description of model output data
|
|
* @return SUCCESS handle successfully / others handle failed
|
|
*/
|
|
Status GeExecutor::ExecModel(uint32_t model_id, void *stream, const ge::RunModelData &run_input_data,
|
|
const std::vector<GeTensorDesc> &input_desc, ge::RunModelData &run_output_data,
|
|
std::vector<GeTensorDesc> &output_desc, bool async_mode) {
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
|
|
InputData input_data;
|
|
OutputData output_data;
|
|
GetDomiInputData(run_input_data, input_data);
|
|
GetDomiOutputData(run_output_data, output_data);
|
|
|
|
if ((run_input_data.dynamic_batch_size != 0) || (run_input_data.dynamic_image_width != 0) ||
|
|
(run_input_data.dynamic_image_height != 0) || (run_input_data.dynamic_dims.size() != 0)) {
|
|
std::vector<std::vector<int64_t>> batch_info;
|
|
int32_t dynamic_type = static_cast<int32_t>(FIXED);
|
|
Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "Get dynamic input info failed.");
|
|
return ret;
|
|
}
|
|
if (!batch_info.empty()) {
|
|
SetDynamicInputDataFlag(run_input_data, batch_info, input_data);
|
|
}
|
|
}
|
|
|
|
return GraphLoader::ExecuteModel(model_id, stream, async_mode, input_data, input_desc, output_data, output_desc);
|
|
}
|
|
|
|
/**
|
|
* @ingroup ge
|
|
* @brief Get weight memory size from model file
|
|
* @param [in] const std::string &path: Offline model file path
|
|
* @param [out] size_t &mem_size Execution memory size
|
|
size_t &weight_size Weight memory space size
|
|
* @return SUCCESS handle successfully / others handle failed
|
|
*/
|
|
Status GeExecutor::GetMemAndWeightSize(const std::string &path, size_t &mem_size, size_t &weight_size) {
|
|
GELOGI("Get memory and weight size from file begin.");
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
|
|
ModelData model;
|
|
std::string key;
|
|
Status ret = ge::GraphLoader::LoadDataFromFile(path, key, 0, model);
|
|
if ((ret != SUCCESS) || (model.model_data == nullptr)) {
|
|
GELOGE(ret, "Load data from file failed. ret = %d", ret);
|
|
return ret;
|
|
}
|
|
|
|
ret = ge::ModelManager::GetModelMemAndWeightSize(model, mem_size, weight_size);
|
|
|
|
delete[] static_cast<char *>(model.model_data);
|
|
model.model_data = nullptr;
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* @ingroup ge
|
|
* @brief Get weight memory size from model file
|
|
* @param [in] const void *model_data Offline model buffer
|
|
size_t model_size Offline model buffer length
|
|
* @param [out] size_t &mem_size Execution memory size
|
|
size_t &weight_size Weight memory space size
|
|
* @return SUCCESS handle successfully / others handle failed
|
|
*/
|
|
Status GeExecutor::GetMemAndWeightSize(const void *model_data, size_t model_size, size_t &mem_size,
|
|
size_t &weight_size) {
|
|
GELOGI("Get memory and weight size from data begin.");
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
|
|
if (model_data == nullptr) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_MODEL_ADDR_INVALID, "invalid model data!");
|
|
return ACL_ERROR_GE_EXEC_MODEL_ADDR_INVALID;
|
|
}
|
|
|
|
ModelData model;
|
|
model.model_data = const_cast<void *>(model_data);
|
|
model.model_len = static_cast<uint32_t>(model_size);
|
|
|
|
return ge::ModelManager::GetModelMemAndWeightSize(model, mem_size, weight_size);
|
|
}
|
|
|
|
Status GeExecutor::LoadSingleOp(const std::string &model_name, const ge::ModelData &modelData, void *stream,
|
|
SingleOp **single_op) {
|
|
return LoadSingleOpV2(model_name, modelData, stream, single_op, 0);
|
|
}
|
|
|
|
Status GeExecutor::LoadSingleOpV2(const std::string &model_name, const ge::ModelData &modelData, void *stream,
|
|
SingleOp **single_op, const uint64_t model_id) {
|
|
return SingleOpManager::GetInstance().GetOpFromModel(model_name, modelData, stream, single_op, model_id);
|
|
}
|
|
|
|
Status GeExecutor::LoadDynamicSingleOp(const std::string &model_name, const ge::ModelData &modelData, void *stream,
|
|
DynamicSingleOp **single_op) {
|
|
return LoadDynamicSingleOpV2(model_name, modelData, stream, single_op, 0);
|
|
}
|
|
|
|
Status GeExecutor::LoadDynamicSingleOpV2(const std::string &model_name, const ge::ModelData &modelData, void *stream,
|
|
DynamicSingleOp **single_op, const uint64_t model_id) {
|
|
return SingleOpManager::GetInstance().GetDynamicOpFromModel(model_name, modelData, stream, single_op, model_id);
|
|
}
|
|
|
|
Status GeExecutor::ExecuteAsync(SingleOp *executor, const std::vector<DataBuffer> &inputs,
|
|
std::vector<DataBuffer> &outputs) {
|
|
if (executor == nullptr) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "param is NULL");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
|
|
return executor->ExecuteAsync(inputs, outputs);
|
|
}
|
|
|
|
ge::Status GeExecutor::ExecuteAsync(DynamicSingleOp *executor, const vector<GeTensorDesc> &input_desc,
|
|
const vector<DataBuffer> &inputs, vector<GeTensorDesc> &output_desc,
|
|
vector<DataBuffer> &outputs) {
|
|
GE_CHECK_NOTNULL(executor);
|
|
return executor->ExecuteAsync(input_desc, inputs, output_desc, outputs);
|
|
}
|
|
|
|
Status GeExecutor::ReleaseSingleOpResource(void *stream) {
|
|
ModelManager::GetInstance()->ClearAicpuSo();
|
|
return SingleOpManager::GetInstance().ReleaseResource(stream);
|
|
}
|
|
|
|
Status GeExecutor::GetDeviceIdByModelId(uint32_t model_id, uint32_t &device_id) {
|
|
auto model_manager = ModelManager::GetInstance();
|
|
GE_CHECK_NOTNULL(model_manager);
|
|
auto davinci_model = model_manager->GetModel(model_id);
|
|
if (davinci_model == nullptr) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_MODEL_ID_INVALID, "Model id: %d is invaild or model is not loaded.", model_id);
|
|
return ACL_ERROR_GE_EXEC_MODEL_ID_INVALID;
|
|
}
|
|
|
|
device_id = davinci_model->GetDeviceId();
|
|
return SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::GetBatchInfoSize(uint32_t model_id, size_t &shape_count) {
|
|
std::vector<std::vector<int64_t>> batch_info;
|
|
int32_t dynamic_type = static_cast<int32_t>(FIXED);
|
|
Status ret = GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "Calc batch info size failed. ret = %d", ret);
|
|
return ret;
|
|
}
|
|
if (batch_info.empty()) {
|
|
shape_count = kStaticBatchInfoSize;
|
|
} else {
|
|
shape_count = batch_info.size();
|
|
}
|
|
return SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::GetOrigInputInfo(uint32_t model_id, uint32_t index, OriginInputInfo &orig_input_info) {
|
|
GELOGI("Begin to GetOrigInputInfo.");
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
|
|
Status ret = GraphExecutor::GetOrigInputInfo(model_id, index, orig_input_info);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "GetOrigInputInfo failed.");
|
|
return ret;
|
|
}
|
|
|
|
GELOGI("GetOrigInputInfo succ.");
|
|
return SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::GetAllAippInputOutputDims(uint32_t model_id, uint32_t index,
|
|
std::vector<InputOutputDims> &input_dims,
|
|
std::vector<InputOutputDims> &output_dims) {
|
|
GELOGI("Begin to GetAllAippInputOutputDims.");
|
|
if (!isInit_) {
|
|
GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
|
|
return ACL_ERROR_GE_EXEC_NOT_INIT;
|
|
}
|
|
|
|
Status ret = GraphExecutor::GetAllAippInputOutputDims(model_id, index, input_dims, output_dims);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "GetAllAippInputOutputDims failed.");
|
|
return ret;
|
|
}
|
|
|
|
GELOGI("GetAllAippInputOutputDims succ.");
|
|
return SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::GetOpDescInfo(uint32_t device_id, uint32_t stream_id, uint32_t task_id, OpDescInfo &op_desc_info) {
|
|
GELOGI("Begin to GetOpDescInfo.");
|
|
Status ret = GraphExecutor::GetOpDescInfo(device_id, stream_id, task_id, op_desc_info);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "GetOpDescInfo failed.");
|
|
return ret;
|
|
}
|
|
GELOGI("GetOpDescInfo succ.");
|
|
return SUCCESS;
|
|
}
|
|
|
|
Status GeExecutor::SetDump(const DumpConfig &dump_config) {
|
|
GELOGI("Start to set dump config");
|
|
auto ret = DumpManager::GetInstance().SetDumpConf(dump_config);
|
|
if (ret != SUCCESS) {
|
|
GELOGE(ret, "Set dump conf failed");
|
|
return ret;
|
|
}
|
|
GELOGI("Set dump config successfully");
|
|
return SUCCESS;
|
|
}
|
|
} // namespace ge
|