parent
bed0ecf3d2
commit
622fe6a56b
@ -0,0 +1,24 @@
|
|||||||
|
/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#include "paddle/fluid/operators/load_combine_op.h"
|
||||||
|
|
||||||
|
namespace ops = paddle::operators;
|
||||||
|
|
||||||
|
REGISTER_OP_CUDA_KERNEL(
|
||||||
|
load_combine,
|
||||||
|
ops::LoadCombineOpKernel<paddle::platform::CUDADeviceContext, float>,
|
||||||
|
ops::LoadCombineOpKernel<paddle::platform::CUDADeviceContext, double>,
|
||||||
|
ops::LoadCombineOpKernel<paddle::platform::CUDADeviceContext, int>,
|
||||||
|
ops::LoadCombineOpKernel<paddle::platform::CUDADeviceContext, int64_t>);
|
@ -0,0 +1,102 @@
|
|||||||
|
/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include <fstream>
|
||||||
|
#include <string>
|
||||||
|
#include <vector>
|
||||||
|
|
||||||
|
#include "paddle/fluid/framework/data_type.h"
|
||||||
|
#include "paddle/fluid/framework/data_type_transform.h"
|
||||||
|
#include "paddle/fluid/framework/op_registry.h"
|
||||||
|
#include "paddle/fluid/platform/device_context.h"
|
||||||
|
|
||||||
|
namespace paddle {
|
||||||
|
namespace operators {
|
||||||
|
template <typename DeviceContext, typename T>
|
||||||
|
class LoadCombineOpKernel : public framework::OpKernel<T> {
|
||||||
|
public:
|
||||||
|
void Compute(const framework::ExecutionContext &ctx) const override {
|
||||||
|
auto place = ctx.GetPlace();
|
||||||
|
auto filename = ctx.Attr<std::string>("file_path");
|
||||||
|
auto load_as_fp16 = ctx.Attr<bool>("load_as_fp16");
|
||||||
|
auto model_from_memory = ctx.Attr<bool>("model_from_memory");
|
||||||
|
auto &out_var_names = ctx.Outputs("Out");
|
||||||
|
|
||||||
|
PADDLE_ENFORCE_GT(
|
||||||
|
static_cast<int>(out_var_names.size()), 0,
|
||||||
|
"The number of output variables should be greater than 0.");
|
||||||
|
if (!model_from_memory) {
|
||||||
|
std::ifstream fin(filename, std::ios::binary);
|
||||||
|
PADDLE_ENFORCE(static_cast<bool>(fin),
|
||||||
|
"Cannot open file %s for load_combine op", filename);
|
||||||
|
LoadParamsFromBuffer(ctx, place, &fin, load_as_fp16, out_var_names);
|
||||||
|
} else {
|
||||||
|
PADDLE_ENFORCE(!filename.empty(), "Cannot load file from memory");
|
||||||
|
std::stringstream fin(filename, std::ios::in | std::ios::binary);
|
||||||
|
LoadParamsFromBuffer(ctx, place, &fin, load_as_fp16, out_var_names);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void LoadParamsFromBuffer(
|
||||||
|
const framework::ExecutionContext &context, const platform::Place &place,
|
||||||
|
std::istream *buffer, bool load_as_fp16,
|
||||||
|
const std::vector<std::string> &out_var_names) const {
|
||||||
|
platform::DeviceContextPool &pool = platform::DeviceContextPool::Instance();
|
||||||
|
auto &dev_ctx = *pool.Get(place);
|
||||||
|
auto out_vars = context.MultiOutputVar("Out");
|
||||||
|
|
||||||
|
for (size_t i = 0; i < out_var_names.size(); i++) {
|
||||||
|
PADDLE_ENFORCE(out_vars[i] != nullptr,
|
||||||
|
"Output variable %s cannot be found", out_var_names[i]);
|
||||||
|
|
||||||
|
auto *tensor = out_vars[i]->GetMutable<framework::LoDTensor>();
|
||||||
|
|
||||||
|
// Error checking
|
||||||
|
PADDLE_ENFORCE(static_cast<bool>(*buffer), "Cannot read more");
|
||||||
|
|
||||||
|
// Get data from fin to tensor
|
||||||
|
DeserializeFromStream(*buffer, tensor, dev_ctx);
|
||||||
|
|
||||||
|
auto in_dtype = tensor->type();
|
||||||
|
auto out_dtype =
|
||||||
|
load_as_fp16 ? framework::proto::VarType::FP16 : in_dtype;
|
||||||
|
|
||||||
|
if (in_dtype != out_dtype) {
|
||||||
|
// convert to float16 tensor
|
||||||
|
auto in_kernel_type = framework::OpKernelType(in_dtype, place);
|
||||||
|
auto out_kernel_type = framework::OpKernelType(out_dtype, place);
|
||||||
|
framework::LoDTensor fp16_tensor;
|
||||||
|
// copy LoD info to the new tensor
|
||||||
|
fp16_tensor.set_lod(tensor->lod());
|
||||||
|
framework::TransDataType(in_kernel_type, out_kernel_type, *tensor,
|
||||||
|
&fp16_tensor);
|
||||||
|
|
||||||
|
// reset output tensor
|
||||||
|
out_vars[i]->Clear();
|
||||||
|
tensor = out_vars[i]->GetMutable<framework::LoDTensor>();
|
||||||
|
tensor->set_lod(fp16_tensor.lod());
|
||||||
|
tensor->ShareDataWith(fp16_tensor);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
buffer->peek();
|
||||||
|
PADDLE_ENFORCE(buffer->eof(),
|
||||||
|
"You are not allowed to load partial data via "
|
||||||
|
"load_combine_op, use load_op instead.");
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
} // namespace operators
|
||||||
|
} // namespace paddle
|
@ -0,0 +1,23 @@
|
|||||||
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#include "paddle/fluid/operators/load_op.h"
|
||||||
|
|
||||||
|
namespace ops = paddle::operators;
|
||||||
|
|
||||||
|
REGISTER_OP_CUDA_KERNEL(
|
||||||
|
load, ops::LoadOpKernel<paddle::platform::CUDADeviceContext, float>,
|
||||||
|
ops::LoadOpKernel<paddle::platform::CUDADeviceContext, double>,
|
||||||
|
ops::LoadOpKernel<paddle::platform::CUDADeviceContext, int>,
|
||||||
|
ops::LoadOpKernel<paddle::platform::CUDADeviceContext, int64_t>);
|
@ -0,0 +1,102 @@
|
|||||||
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include <fstream>
|
||||||
|
#include <string>
|
||||||
|
|
||||||
|
#include "paddle/fluid/framework/data_type_transform.h"
|
||||||
|
#include "paddle/fluid/framework/op_registry.h"
|
||||||
|
#include "paddle/fluid/platform/device_context.h"
|
||||||
|
#include "paddle/fluid/platform/profiler.h"
|
||||||
|
|
||||||
|
namespace paddle {
|
||||||
|
namespace operators {
|
||||||
|
template <typename DeviceContext, typename T>
|
||||||
|
class LoadOpKernel : public framework::OpKernel<T> {
|
||||||
|
public:
|
||||||
|
void Compute(const framework::ExecutionContext &ctx) const override {
|
||||||
|
auto place = ctx.GetPlace();
|
||||||
|
// FIXME(yuyang18): We save variable to local file now, but we should change
|
||||||
|
// it to save an output stream.
|
||||||
|
auto filename = ctx.Attr<std::string>("file_path");
|
||||||
|
std::ifstream fin(filename, std::ios::binary);
|
||||||
|
PADDLE_ENFORCE(static_cast<bool>(fin), "Cannot open file %s for load op",
|
||||||
|
filename);
|
||||||
|
|
||||||
|
auto out_var_name = ctx.Outputs("Out").data();
|
||||||
|
auto *out_var = ctx.OutputVar("Out");
|
||||||
|
|
||||||
|
PADDLE_ENFORCE(out_var != nullptr, "Output variable %s cannot be found ",
|
||||||
|
out_var_name);
|
||||||
|
|
||||||
|
PADDLE_ENFORCE(out_var != nullptr, "Output variable cannot be found ");
|
||||||
|
|
||||||
|
if (out_var->IsType<framework::LoDTensor>()) {
|
||||||
|
LoadLodTensor(fin, place, out_var, ctx);
|
||||||
|
} else if (out_var->IsType<framework::SelectedRows>()) {
|
||||||
|
LoadSelectedRows(fin, place, out_var);
|
||||||
|
} else {
|
||||||
|
PADDLE_ENFORCE(
|
||||||
|
false,
|
||||||
|
"Load only support LoDTensor and SelectedRows, %s has wrong type",
|
||||||
|
out_var_name);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void LoadLodTensor(std::istream &fin, const platform::Place &place,
|
||||||
|
framework::Variable *var,
|
||||||
|
const framework::ExecutionContext &ctx) const {
|
||||||
|
// get device context from pool
|
||||||
|
platform::DeviceContextPool &pool = platform::DeviceContextPool::Instance();
|
||||||
|
auto &dev_ctx = *pool.Get(place);
|
||||||
|
auto *tensor = var->GetMutable<framework::LoDTensor>();
|
||||||
|
DeserializeFromStream(fin, tensor, dev_ctx);
|
||||||
|
|
||||||
|
auto load_as_fp16 = ctx.Attr<bool>("load_as_fp16");
|
||||||
|
auto in_dtype = tensor->type();
|
||||||
|
auto out_dtype = load_as_fp16 ? framework::proto::VarType::FP16 : in_dtype;
|
||||||
|
|
||||||
|
if (in_dtype != out_dtype) {
|
||||||
|
// convert to float16 tensor
|
||||||
|
auto in_kernel_type = framework::OpKernelType(in_dtype, place);
|
||||||
|
auto out_kernel_type = framework::OpKernelType(out_dtype, place);
|
||||||
|
framework::LoDTensor fp16_tensor;
|
||||||
|
// copy LoD info to the new tensor
|
||||||
|
fp16_tensor.set_lod(tensor->lod());
|
||||||
|
framework::TransDataType(in_kernel_type, out_kernel_type, *tensor,
|
||||||
|
&fp16_tensor);
|
||||||
|
|
||||||
|
// reset output tensor
|
||||||
|
var->Clear();
|
||||||
|
tensor = var->GetMutable<framework::LoDTensor>();
|
||||||
|
tensor->set_lod(fp16_tensor.lod());
|
||||||
|
tensor->ShareDataWith(fp16_tensor);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void LoadSelectedRows(std::istream &fin, const platform::Place &place,
|
||||||
|
framework::Variable *var) const {
|
||||||
|
auto *selectedRows = var->GetMutable<framework::SelectedRows>();
|
||||||
|
// get device context from pool
|
||||||
|
platform::DeviceContextPool &pool = platform::DeviceContextPool::Instance();
|
||||||
|
auto &dev_ctx = *pool.Get(place);
|
||||||
|
framework::DeserializeFromStream(fin, selectedRows, dev_ctx);
|
||||||
|
selectedRows->SyncIndex();
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
} // namespace operators
|
||||||
|
} // namespace paddle
|
@ -0,0 +1,24 @@
|
|||||||
|
/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#include "paddle/fluid/operators/save_combine_op.h"
|
||||||
|
|
||||||
|
namespace ops = paddle::operators;
|
||||||
|
|
||||||
|
REGISTER_OP_CUDA_KERNEL(
|
||||||
|
save_combine,
|
||||||
|
ops::SaveCombineOpKernel<paddle::platform::CUDADeviceContext, float>,
|
||||||
|
ops::SaveCombineOpKernel<paddle::platform::CUDADeviceContext, double>,
|
||||||
|
ops::SaveCombineOpKernel<paddle::platform::CUDADeviceContext, int>,
|
||||||
|
ops::SaveCombineOpKernel<paddle::platform::CUDADeviceContext, int64_t>);
|
@ -0,0 +1,95 @@
|
|||||||
|
/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include <stdint.h>
|
||||||
|
#include <fstream>
|
||||||
|
#include <numeric>
|
||||||
|
#include <sstream>
|
||||||
|
#include <string>
|
||||||
|
|
||||||
|
#include "paddle/fluid/framework/data_type.h"
|
||||||
|
#include "paddle/fluid/framework/data_type_transform.h"
|
||||||
|
#include "paddle/fluid/framework/framework.pb.h"
|
||||||
|
#include "paddle/fluid/framework/lod_tensor.h"
|
||||||
|
#include "paddle/fluid/framework/op_registry.h"
|
||||||
|
#include "paddle/fluid/platform/device_context.h"
|
||||||
|
#include "paddle/fluid/platform/port.h"
|
||||||
|
|
||||||
|
namespace paddle {
|
||||||
|
namespace operators {
|
||||||
|
template <typename DeviceContext, typename T>
|
||||||
|
class SaveCombineOpKernel : public framework::OpKernel<T> {
|
||||||
|
public:
|
||||||
|
void Compute(const framework::ExecutionContext &ctx) const override {
|
||||||
|
auto place = ctx.GetPlace();
|
||||||
|
auto filename = ctx.Attr<std::string>("file_path");
|
||||||
|
auto overwrite = ctx.Attr<bool>("overwrite");
|
||||||
|
auto save_as_fp16 = ctx.Attr<bool>("save_as_fp16");
|
||||||
|
|
||||||
|
bool is_present = FileExists(filename);
|
||||||
|
if (is_present && !overwrite) {
|
||||||
|
PADDLE_THROW("%s exists!, cannot save_combine to it when overwrite=false",
|
||||||
|
filename, overwrite);
|
||||||
|
}
|
||||||
|
|
||||||
|
MkDirRecursively(DirName(filename).c_str());
|
||||||
|
std::ofstream fout(filename, std::ios::binary);
|
||||||
|
PADDLE_ENFORCE(static_cast<bool>(fout), "Cannot open %s to write",
|
||||||
|
filename);
|
||||||
|
|
||||||
|
auto &inp_var_names = ctx.Inputs("X");
|
||||||
|
auto &inp_vars = ctx.MultiInputVar("X");
|
||||||
|
PADDLE_ENFORCE_GT(static_cast<int>(inp_var_names.size()), 0,
|
||||||
|
"The number of input variables should be greater than 0");
|
||||||
|
|
||||||
|
// get device context from pool
|
||||||
|
platform::DeviceContextPool &pool = platform::DeviceContextPool::Instance();
|
||||||
|
auto &dev_ctx = *pool.Get(place);
|
||||||
|
|
||||||
|
for (size_t i = 0; i < inp_var_names.size(); i++) {
|
||||||
|
PADDLE_ENFORCE(inp_vars[i] != nullptr,
|
||||||
|
"Cannot find variable %s for save_combine_op",
|
||||||
|
inp_var_names[i]);
|
||||||
|
PADDLE_ENFORCE(inp_vars[i]->IsType<framework::LoDTensor>(),
|
||||||
|
"SaveCombineOp only supports LoDTensor, %s has wrong type",
|
||||||
|
inp_var_names[i]);
|
||||||
|
|
||||||
|
auto &tensor = inp_vars[i]->Get<framework::LoDTensor>();
|
||||||
|
// Serialize tensors one by one
|
||||||
|
|
||||||
|
// Check types to see if a fp16 transformation is required
|
||||||
|
auto in_dtype = tensor.type();
|
||||||
|
auto out_dtype =
|
||||||
|
save_as_fp16 ? framework::proto::VarType::FP16 : in_dtype;
|
||||||
|
|
||||||
|
if (in_dtype != out_dtype) {
|
||||||
|
auto in_kernel_type = framework::OpKernelType(in_dtype, place);
|
||||||
|
auto out_kernel_type = framework::OpKernelType(out_dtype, place);
|
||||||
|
framework::LoDTensor out;
|
||||||
|
// copy LoD info to the new tensor
|
||||||
|
out.set_lod(tensor.lod());
|
||||||
|
framework::TransDataType(in_kernel_type, out_kernel_type, tensor, &out);
|
||||||
|
framework::SerializeToStream(fout, out, dev_ctx);
|
||||||
|
} else {
|
||||||
|
framework::SerializeToStream(fout, tensor, dev_ctx);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
fout.close();
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
} // namespace operators
|
||||||
|
} // namespace paddle
|
@ -0,0 +1,26 @@
|
|||||||
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#include "paddle/fluid/operators/save_op.h"
|
||||||
|
#include "paddle/fluid/platform/float16.h"
|
||||||
|
|
||||||
|
namespace ops = paddle::operators;
|
||||||
|
|
||||||
|
REGISTER_OP_CUDA_KERNEL(
|
||||||
|
save, ops::SaveOpKernel<paddle::platform::CUDADeviceContext, float>,
|
||||||
|
ops::SaveOpKernel<paddle::platform::CUDADeviceContext, double>,
|
||||||
|
ops::SaveOpKernel<paddle::platform::CUDADeviceContext, int>,
|
||||||
|
ops::SaveOpKernel<paddle::platform::CUDADeviceContext, int64_t>,
|
||||||
|
ops::SaveOpKernel<paddle::platform::CUDADeviceContext,
|
||||||
|
paddle::platform::float16>);
|
@ -0,0 +1,133 @@
|
|||||||
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include <stdint.h>
|
||||||
|
#include <fstream>
|
||||||
|
#include <numeric>
|
||||||
|
#include <string>
|
||||||
|
#include <vector>
|
||||||
|
|
||||||
|
#include "paddle/fluid/framework/data_type.h"
|
||||||
|
#include "paddle/fluid/framework/data_type_transform.h"
|
||||||
|
#include "paddle/fluid/framework/framework.pb.h"
|
||||||
|
#include "paddle/fluid/framework/lod_tensor.h"
|
||||||
|
#include "paddle/fluid/framework/op_registry.h"
|
||||||
|
#include "paddle/fluid/framework/selected_rows.h"
|
||||||
|
#include "paddle/fluid/framework/variable.h"
|
||||||
|
|
||||||
|
namespace paddle {
|
||||||
|
namespace operators {
|
||||||
|
// define LOOKUP_TABLE_PATH for checkpoint notify to save lookup table variables
|
||||||
|
// to directory specified.
|
||||||
|
constexpr char LOOKUP_TABLE_PATH[] = "kLookupTablePath";
|
||||||
|
template <typename DeviceContext, typename T>
|
||||||
|
class SaveOpKernel : public framework::OpKernel<T> {
|
||||||
|
public:
|
||||||
|
void Compute(const framework::ExecutionContext &ctx) const override {
|
||||||
|
auto place = ctx.GetPlace();
|
||||||
|
|
||||||
|
auto *input_var = ctx.InputVar("X");
|
||||||
|
auto iname = ctx.Inputs("X").data();
|
||||||
|
PADDLE_ENFORCE(input_var != nullptr, "Cannot find variable %s for save_op",
|
||||||
|
iname);
|
||||||
|
|
||||||
|
if (input_var->IsType<framework::LoDTensor>()) {
|
||||||
|
SaveLodTensor(ctx, place, input_var);
|
||||||
|
} else if (input_var->IsType<framework::SelectedRows>()) {
|
||||||
|
SaveSelectedRows(ctx, place, input_var);
|
||||||
|
} else {
|
||||||
|
PADDLE_ENFORCE(
|
||||||
|
false,
|
||||||
|
"SaveOp only support LoDTensor and SelectedRows, %s has wrong type",
|
||||||
|
iname);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void SaveLodTensor(const framework::ExecutionContext &ctx,
|
||||||
|
const platform::Place &place,
|
||||||
|
const framework::Variable *var) const {
|
||||||
|
auto filename = ctx.Attr<std::string>("file_path");
|
||||||
|
auto overwrite = ctx.Attr<bool>("overwrite");
|
||||||
|
|
||||||
|
if (FileExists(filename) && !overwrite) {
|
||||||
|
PADDLE_THROW("%s is existed, cannot save to it when overwrite=false",
|
||||||
|
filename, overwrite);
|
||||||
|
}
|
||||||
|
|
||||||
|
MkDirRecursively(DirName(filename).c_str());
|
||||||
|
|
||||||
|
auto &tensor = var->Get<framework::LoDTensor>();
|
||||||
|
|
||||||
|
// get device context from pool
|
||||||
|
platform::DeviceContextPool &pool = platform::DeviceContextPool::Instance();
|
||||||
|
auto &dev_ctx = *pool.Get(place);
|
||||||
|
|
||||||
|
// FIXME(yuyang18): We save variable to local file now, but we should change
|
||||||
|
// it to save an output stream.
|
||||||
|
std::ofstream fout(filename, std::ios::binary);
|
||||||
|
PADDLE_ENFORCE(static_cast<bool>(fout), "Cannot open %s to write",
|
||||||
|
filename);
|
||||||
|
|
||||||
|
auto save_as_fp16 = ctx.Attr<bool>("save_as_fp16");
|
||||||
|
auto in_dtype = tensor.type();
|
||||||
|
auto out_dtype = save_as_fp16 ? framework::proto::VarType::FP16 : in_dtype;
|
||||||
|
|
||||||
|
if (in_dtype != out_dtype) {
|
||||||
|
auto in_kernel_type = framework::OpKernelType(in_dtype, place);
|
||||||
|
auto out_kernel_type = framework::OpKernelType(out_dtype, place);
|
||||||
|
framework::LoDTensor out;
|
||||||
|
framework::TransDataType(in_kernel_type, out_kernel_type, tensor, &out);
|
||||||
|
// copy LoD info to the new tensor
|
||||||
|
out.set_lod(tensor.lod());
|
||||||
|
framework::SerializeToStream(fout, out, dev_ctx);
|
||||||
|
} else {
|
||||||
|
framework::SerializeToStream(fout, tensor, dev_ctx);
|
||||||
|
}
|
||||||
|
fout.close();
|
||||||
|
}
|
||||||
|
|
||||||
|
void SaveSelectedRows(const framework::ExecutionContext &ctx,
|
||||||
|
const platform::Place &place,
|
||||||
|
const framework::Variable *var) const {
|
||||||
|
framework::Variable *out_put_var = ctx.OutputVar(LOOKUP_TABLE_PATH);
|
||||||
|
PADDLE_ENFORCE(
|
||||||
|
out_put_var != nullptr,
|
||||||
|
"Can not find variable kLookupTablePath for SaveSelectedRows");
|
||||||
|
auto *lt_var = out_put_var->GetMutable<std::string>();
|
||||||
|
|
||||||
|
std::string filename = lt_var->data();
|
||||||
|
VLOG(4) << "SaveSelectedRows get File name: " << filename;
|
||||||
|
|
||||||
|
MkDirRecursively(DirName(filename).c_str());
|
||||||
|
|
||||||
|
auto &selectedRows = var->Get<framework::SelectedRows>();
|
||||||
|
|
||||||
|
// get device context from pool
|
||||||
|
platform::DeviceContextPool &pool = platform::DeviceContextPool::Instance();
|
||||||
|
auto &dev_ctx = *pool.Get(place);
|
||||||
|
|
||||||
|
// FIXME(yuyang18): We save variable to local file now, but we should change
|
||||||
|
// it to save an output stream.
|
||||||
|
std::ofstream fout(filename, std::ios::binary);
|
||||||
|
PADDLE_ENFORCE(static_cast<bool>(fout), "Cannot open %s to write",
|
||||||
|
filename);
|
||||||
|
framework::SerializeToStream(fout, selectedRows, dev_ctx);
|
||||||
|
fout.close();
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
} // namespace operators
|
||||||
|
} // namespace paddle
|
Loading…
Reference in new issue