|
|
|
@ -40,21 +40,21 @@ class Tensor {
|
|
|
|
|
|
|
|
|
|
template <typename T>
|
|
|
|
|
const T* data() const {
|
|
|
|
|
CheckDims<T>();
|
|
|
|
|
EnforceSufficientMemory<T>();
|
|
|
|
|
return reinterpret_cast<const T*>(
|
|
|
|
|
reinterpret_cast<uintptr_t>(holder_->ptr()) + offset_);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
template <typename T>
|
|
|
|
|
T* raw_data() const {
|
|
|
|
|
CheckDims<T>();
|
|
|
|
|
EnforceSufficientMemory<T>();
|
|
|
|
|
return reinterpret_cast<T*>(reinterpret_cast<uintptr_t>(holder_->ptr()) +
|
|
|
|
|
offset_);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
template <typename T>
|
|
|
|
|
T* mutable_data(DDim dims, platform::Place place) {
|
|
|
|
|
set_dims(dims);
|
|
|
|
|
Resize(dims);
|
|
|
|
|
return mutable_data<T>(place);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
@ -147,11 +147,9 @@ class Tensor {
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
template <typename T>
|
|
|
|
|
void ShareDataFrom(const Tensor& src) {
|
|
|
|
|
src.CheckDims<T>();
|
|
|
|
|
holder_ = src.holder_;
|
|
|
|
|
set_dims(src.dims());
|
|
|
|
|
offset_ = src.offset_;
|
|
|
|
|
void ShareDataWith(const Tensor& src) {
|
|
|
|
|
src.EnforceSufficientMemory<T>();
|
|
|
|
|
*this = src;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
template <typename T>
|
|
|
|
@ -159,9 +157,9 @@ class Tensor {
|
|
|
|
|
PADDLE_ENFORCE(platform::is_cpu_place(src.holder_->place()) &&
|
|
|
|
|
platform::is_cpu_place(dst_place),
|
|
|
|
|
"Tensor::CopyFrom only support CPU now.");
|
|
|
|
|
src.CheckDims<T>();
|
|
|
|
|
src.EnforceSufficientMemory<T>();
|
|
|
|
|
size_t size = product(src.dims_) * sizeof(T);
|
|
|
|
|
set_dims(src.dims());
|
|
|
|
|
Resize(src.dims());
|
|
|
|
|
const void* src_ptr = static_cast<const void*>(src.data<T>());
|
|
|
|
|
void* dst_ptr = static_cast<void*>(mutable_data<T>(dst_place));
|
|
|
|
|
memcpy(dst_ptr, src_ptr, size);
|
|
|
|
@ -169,34 +167,25 @@ class Tensor {
|
|
|
|
|
|
|
|
|
|
template <typename T>
|
|
|
|
|
Tensor Slice(const int& begin_idx, const int& end_idx) const {
|
|
|
|
|
CheckDims<T>();
|
|
|
|
|
PADDLE_ENFORCE(begin_idx >= 0 && end_idx <= dims_[0],
|
|
|
|
|
"Slice index is less than zero or out of bound.");
|
|
|
|
|
EnforceSufficientMemory<T>();
|
|
|
|
|
PADDLE_ENFORCE(begin_idx >= 0, "Slice begin index is less than zero.");
|
|
|
|
|
PADDLE_ENFORCE(end_idx <= dims_[0], "Slice end index is out of bound.");
|
|
|
|
|
PADDLE_ENFORCE(begin_idx < end_idx,
|
|
|
|
|
"Begin index must be less than end index.");
|
|
|
|
|
PADDLE_ENFORCE(dims_[0] != 1, "Can not slice a tensor with dims_[0] = 1.");
|
|
|
|
|
std::vector<int> d = vectorize(dims_);
|
|
|
|
|
int base = 1;
|
|
|
|
|
for (size_t i = 1; i < d.size(); ++i) {
|
|
|
|
|
base *= d[i];
|
|
|
|
|
}
|
|
|
|
|
int base = product(dims_) / dims_[0];
|
|
|
|
|
Tensor dst;
|
|
|
|
|
dst.holder_ = holder_;
|
|
|
|
|
DDim dst_dims = dims_;
|
|
|
|
|
dst_dims[0] = end_idx - begin_idx;
|
|
|
|
|
dst.set_dims(dst_dims);
|
|
|
|
|
dst.Resize(dst_dims);
|
|
|
|
|
dst.offset_ = offset_ + begin_idx * base * sizeof(T);
|
|
|
|
|
return dst;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void set_dims(const DDim& dims) {
|
|
|
|
|
if (dims == dims_) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
dims_ = dims;
|
|
|
|
|
}
|
|
|
|
|
void Resize(const DDim& dims) { dims_ = dims; }
|
|
|
|
|
|
|
|
|
|
DDim dims() const { return dims_; }
|
|
|
|
|
const DDim& dims() const { return dims_; }
|
|
|
|
|
|
|
|
|
|
private:
|
|
|
|
|
// Placeholder hides type T, so it doesn't appear as a template
|
|
|
|
@ -211,21 +200,9 @@ class Tensor {
|
|
|
|
|
|
|
|
|
|
template <typename T, typename PlaceType>
|
|
|
|
|
struct PlaceholderImpl : public Placeholder {
|
|
|
|
|
private:
|
|
|
|
|
template <typename PType>
|
|
|
|
|
class Deleter {
|
|
|
|
|
public:
|
|
|
|
|
Deleter(PType place) : place_(place) {}
|
|
|
|
|
void operator()(T* ptr) { memory::Free(place_, static_cast<void*>(ptr)); }
|
|
|
|
|
|
|
|
|
|
private:
|
|
|
|
|
PType place_;
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
public:
|
|
|
|
|
PlaceholderImpl(PlaceType place, size_t size)
|
|
|
|
|
: ptr_(static_cast<T*>(memory::Alloc(place, size)),
|
|
|
|
|
Deleter<PlaceType>(place)),
|
|
|
|
|
memory::PodDeleter<T, PlaceType>(place)),
|
|
|
|
|
place_(place),
|
|
|
|
|
size_(size) {}
|
|
|
|
|
|
|
|
|
@ -234,13 +211,13 @@ class Tensor {
|
|
|
|
|
virtual paddle::platform::Place place() const { return place_; }
|
|
|
|
|
virtual std::type_index type() const { return std::type_index(typeid(T)); }
|
|
|
|
|
|
|
|
|
|
std::unique_ptr<T, Deleter<PlaceType>> ptr_;
|
|
|
|
|
std::unique_ptr<T, memory::PodDeleter<T, PlaceType>> ptr_;
|
|
|
|
|
platform::Place place_; // record the place of ptr_.
|
|
|
|
|
size_t size_; // size of the memory block.
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
template <typename T>
|
|
|
|
|
inline void CheckDims() const {
|
|
|
|
|
inline void EnforceSufficientMemory() const {
|
|
|
|
|
PADDLE_ENFORCE(holder_ != nullptr,
|
|
|
|
|
"Tenosr holds no memory. Call Tensor::mutable_data first.");
|
|
|
|
|
PADDLE_ENFORCE(holder_->size() >= product(dims_) * sizeof(T) + offset_,
|
|
|
|
@ -250,7 +227,11 @@ class Tensor {
|
|
|
|
|
|
|
|
|
|
std::shared_ptr<Placeholder> holder_; // holds the memory block if allocated.
|
|
|
|
|
DDim dims_;
|
|
|
|
|
size_t offset_; // marks the begin of tensor data area.
|
|
|
|
|
// A PlaceHolder may be shared by more than one tensor. Some of them may be
|
|
|
|
|
// slices of the others. So the offset_ is introduced here to indicate the
|
|
|
|
|
// byte offset between PlaceHolder::ptr_ and where tensor's data really
|
|
|
|
|
// begins.
|
|
|
|
|
size_t offset_;
|
|
|
|
|
template <bool less, size_t i, typename... args>
|
|
|
|
|
friend struct paddle::pybind::details::CastToPyBufferImpl;
|
|
|
|
|
};
|
|
|
|
|