commit
ff000ae76a
@ -0,0 +1,181 @@
|
|||||||
|
package pserver
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"errors"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/PaddlePaddle/Paddle/go/utils/networkhelper"
|
||||||
|
"github.com/coreos/etcd/clientv3"
|
||||||
|
"github.com/coreos/etcd/clientv3/concurrency"
|
||||||
|
log "github.com/sirupsen/logrus"
|
||||||
|
)
|
||||||
|
|
||||||
|
// EtcdClient is the etcd client that the pserver uses for fault
|
||||||
|
// tolerance, service registry and coordination.
|
||||||
|
type EtcdClient struct {
|
||||||
|
numPservers int
|
||||||
|
etcdEndpoints string
|
||||||
|
etcdClient *clientv3.Client
|
||||||
|
// etcdTimeout is also used as retry intervals.
|
||||||
|
etcdTimeout time.Duration
|
||||||
|
// FIXME: ensure GetExternalIP gets the correct ip for trainers to connect.
|
||||||
|
externalIP string
|
||||||
|
// desired number of pservers in the job.
|
||||||
|
// assume desired will not change during one training job.
|
||||||
|
desired int
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewEtcdClient creates an EtcdClient
|
||||||
|
func NewEtcdClient(endpoints string, numPservers int, timeout time.Duration) *EtcdClient {
|
||||||
|
return &EtcdClient{
|
||||||
|
etcdTimeout: timeout,
|
||||||
|
numPservers: numPservers,
|
||||||
|
etcdEndpoints: endpoints,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Register registers the pserver on etcd
|
||||||
|
//
|
||||||
|
// Register returns the index of the current pserver.
|
||||||
|
func (e *EtcdClient) Register() (int, error) {
|
||||||
|
|
||||||
|
var err error
|
||||||
|
e.externalIP, err = networkhelper.GetExternalIP()
|
||||||
|
if err != nil {
|
||||||
|
return 0, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// initialize connection to etcd.
|
||||||
|
ep := strings.Split(e.etcdEndpoints, ",")
|
||||||
|
for {
|
||||||
|
cli, err := clientv3.New(clientv3.Config{
|
||||||
|
Endpoints: ep,
|
||||||
|
DialTimeout: e.etcdTimeout,
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("connect to etcd error: %v", err)
|
||||||
|
time.Sleep(e.etcdTimeout)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
e.etcdClient = cli
|
||||||
|
log.Debugf("inited client to %s", e.etcdEndpoints)
|
||||||
|
break
|
||||||
|
}
|
||||||
|
// init /ps_desired using transaction, for multiple pservers may want to write
|
||||||
|
// it at the same time.
|
||||||
|
for {
|
||||||
|
ctx, cancel := context.WithTimeout(context.Background(), time.Second)
|
||||||
|
_, err := e.initDesiredPsercers(ctx, e.numPservers)
|
||||||
|
cancel()
|
||||||
|
if err != nil {
|
||||||
|
log.Warn(err)
|
||||||
|
time.Sleep(e.etcdTimeout)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
break
|
||||||
|
}
|
||||||
|
// TODO: when implementing extending or reducing pservers, /ps_desired is
|
||||||
|
// changed, then we need to watch /ps_desired node for events. For now, just
|
||||||
|
// write once when init and read from it.
|
||||||
|
// wait and set s.desired init value
|
||||||
|
for {
|
||||||
|
ctx, cancel := context.WithTimeout(context.Background(), time.Second)
|
||||||
|
resp, err := e.etcdClient.Get(ctx, PsDesired)
|
||||||
|
cancel()
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("getting %s error: %v", PsDesired, err)
|
||||||
|
time.Sleep(e.etcdTimeout)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if len(resp.Kvs) != 0 {
|
||||||
|
e.desired, err = strconv.Atoi(string(resp.Kvs[0].Value))
|
||||||
|
if err != nil {
|
||||||
|
log.Errorf("value of %s invalid %v\n", PsDesired, err)
|
||||||
|
time.Sleep(e.etcdTimeout)
|
||||||
|
// NOTE: wait util ps_desired value change
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
var pserverIdx int
|
||||||
|
// try register pserver node on etcd
|
||||||
|
for {
|
||||||
|
ctx, cancel := context.WithTimeout(context.Background(), time.Second)
|
||||||
|
var err error
|
||||||
|
pserverIdx, err = e.registerPserverEtcd(ctx)
|
||||||
|
cancel()
|
||||||
|
if err != nil {
|
||||||
|
log.Warn(err)
|
||||||
|
time.Sleep(e.etcdTimeout)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
return pserverIdx, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (e *EtcdClient) initDesiredPsercers(ctx context.Context, numPservers int) (*clientv3.TxnResponse, error) {
|
||||||
|
return concurrency.NewSTM(e.etcdClient, func(c concurrency.STM) error {
|
||||||
|
dsStr := c.Get(PsDesired)
|
||||||
|
if dsStr == "" {
|
||||||
|
c.Put(PsDesired, strconv.Itoa(numPservers))
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}, concurrency.WithAbortContext(ctx), concurrency.WithIsolation(concurrency.RepeatableReads))
|
||||||
|
}
|
||||||
|
|
||||||
|
// registerPserverEtcd registers pserver node on etcd using transaction.
|
||||||
|
func (e *EtcdClient) registerPserverEtcd(ctx context.Context) (int, error) {
|
||||||
|
var idx int
|
||||||
|
_, err := concurrency.NewSTM(e.etcdClient, func(c concurrency.STM) error {
|
||||||
|
registered := false
|
||||||
|
for i := 0; i < e.desired; i++ {
|
||||||
|
psKey := "/ps/" + strconv.Itoa(i)
|
||||||
|
log.Debugf("checking %s", psKey)
|
||||||
|
ps := c.Get(psKey)
|
||||||
|
log.Debugf("got value (%s) for key: %s", ps, psKey)
|
||||||
|
|
||||||
|
if ps == "" {
|
||||||
|
resp, err := e.etcdClient.Grant(context.TODO(), 5)
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
// find the first id and write info
|
||||||
|
c.Put(psKey, e.externalIP, clientv3.WithLease(resp.ID))
|
||||||
|
log.Debugf("set pserver node %s with value %s", psKey, e.externalIP)
|
||||||
|
ch, kaerr := e.etcdClient.KeepAlive(context.TODO(), resp.ID)
|
||||||
|
if kaerr != nil {
|
||||||
|
log.Errorf("keepalive etcd node error: %v", kaerr)
|
||||||
|
return kaerr
|
||||||
|
}
|
||||||
|
|
||||||
|
// Eat the keep alive message so etcd
|
||||||
|
// will not expire the lease.
|
||||||
|
go func(ch <-chan *clientv3.LeaseKeepAliveResponse) {
|
||||||
|
ka := <-ch
|
||||||
|
log.Debugf("keepalive: %d\n", ka.TTL)
|
||||||
|
}(ch)
|
||||||
|
log.Debug("register finished")
|
||||||
|
idx = i
|
||||||
|
registered = true
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if registered == true {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return errors.New("not registerd, may due to already have enough pservers")
|
||||||
|
}, concurrency.WithAbortContext(ctx), concurrency.WithIsolation(concurrency.RepeatableReads))
|
||||||
|
|
||||||
|
if err != nil {
|
||||||
|
return 0, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return idx, nil
|
||||||
|
}
|
@ -0,0 +1,5 @@
|
|||||||
|
---
|
||||||
|
Language: Cpp
|
||||||
|
BasedOnStyle: Google
|
||||||
|
Standard: Cpp11
|
||||||
|
...
|
@ -0,0 +1 @@
|
|||||||
|
add_subdirectory(detail)
|
@ -0,0 +1,7 @@
|
|||||||
|
if(${WITH_GPU})
|
||||||
|
nv_library(system_allocator SRCS system_allocator.cc DEPS gflags)
|
||||||
|
nv_test(system_allocator_test SRCS system_allocator_test.cc DEPS system_allocator gflags)
|
||||||
|
else(${WITH_GPU})
|
||||||
|
cc_library(system_allocator SRCS system_allocator.cc DEPS gflags)
|
||||||
|
cc_test(system_allocator_test SRCS system_allocator_test.cc DEPS system_allocator gflags)
|
||||||
|
endif(${WITH_GPU})
|
@ -0,0 +1,35 @@
|
|||||||
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include "paddle/memory/detail/buddy_allocator.h"
|
||||||
|
|
||||||
|
namespace paddle {
|
||||||
|
namespace memory {
|
||||||
|
namespace detail {
|
||||||
|
|
||||||
|
BuddyAllocator::BuddyAllocator(size_t pool_size, size_t max_pools,
|
||||||
|
SystemAllocator* system_allocator)
|
||||||
|
: pool_size_(pool_size),
|
||||||
|
max_pools_(max_pools),
|
||||||
|
system_allocator_(system_allocator) {
|
||||||
|
PADDLE_ASSERT(pool_size > 0);
|
||||||
|
PADDLE_ASSERT(max_pools > 0);
|
||||||
|
PADDLE_ASSERT(system_allocator != nullptr);
|
||||||
|
}
|
||||||
|
|
||||||
|
} // namespace detail
|
||||||
|
} // namespace memory
|
||||||
|
} // namespace paddle
|
@ -0,0 +1,86 @@
|
|||||||
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include "paddle/memory/detail/system_allocator.h"
|
||||||
|
|
||||||
|
#include <mutex>
|
||||||
|
#include <vector>
|
||||||
|
|
||||||
|
namespace paddle {
|
||||||
|
namespace memory {
|
||||||
|
namespace detail {
|
||||||
|
|
||||||
|
class BuddyAllocator {
|
||||||
|
public:
|
||||||
|
BuddyAllocator(size_t pool_size, size_t max_pools,
|
||||||
|
SystemAllocator* system_allocator);
|
||||||
|
~BuddyAllocator();
|
||||||
|
|
||||||
|
void* Alloc(size_t size);
|
||||||
|
void Free(void*);
|
||||||
|
size_t Used();
|
||||||
|
|
||||||
|
private:
|
||||||
|
struct Block {
|
||||||
|
size_t size_;
|
||||||
|
Block* left_; // left buddy
|
||||||
|
Block* right_; // right buddy
|
||||||
|
};
|
||||||
|
|
||||||
|
// Initially, there is only one pool. If a Alloc founds not enough
|
||||||
|
// memory from that pool, and there has not been max_num_pools_,
|
||||||
|
// create a new pool by calling system_allocator_.Alloc(pool_size_).
|
||||||
|
std::vector<void*> pools_;
|
||||||
|
|
||||||
|
size_t pool_size_; // the size of each pool;
|
||||||
|
size_t max_num_pools_; // the size of all pools;
|
||||||
|
|
||||||
|
SystemAllocator* system_allocator_;
|
||||||
|
|
||||||
|
std::mutex mutex_;
|
||||||
|
|
||||||
|
// Disable copy and assignment.
|
||||||
|
BuddyAllocator(const BuddyAllocator&) = delete;
|
||||||
|
BuddyAllocator& operator=(const BuddyAllocator&) = delete;
|
||||||
|
};
|
||||||
|
|
||||||
|
BuddyAllocator<CPUAllocator>* GetCPUBuddyAllocator() {
|
||||||
|
static BuddyAllocator<CPUAllocator>* a = nullptr;
|
||||||
|
if (a == nullptr) {
|
||||||
|
a = new BuddyAllocator<CPUAllocator>();
|
||||||
|
}
|
||||||
|
return a;
|
||||||
|
}
|
||||||
|
|
||||||
|
#ifndef PADDLE_ONLY_CPU // The following code are for CUDA.
|
||||||
|
|
||||||
|
BuddyAllocator<GPUAllocator>* GetGPUBuddyAllocator(int gpu_id) {
|
||||||
|
static BuddyAllocator<GPUAllocator>** as = NULL;
|
||||||
|
if (as == NULL) {
|
||||||
|
int gpu_num = platform::GetDeviceCount();
|
||||||
|
as = new BuddyAllocator<GPUAllocator>*[gpu_num];
|
||||||
|
for (int gpu = 0; gpu < gpu_num; gpu++) {
|
||||||
|
as[gpu] = new BuddyAllocator<GPUAllocator>();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return as[gpu_id];
|
||||||
|
}
|
||||||
|
|
||||||
|
#endif // PADDLE_ONLY_CPU
|
||||||
|
|
||||||
|
} // namespace detail
|
||||||
|
} // namespace memory
|
||||||
|
} // namespace paddle
|
@ -0,0 +1,90 @@
|
|||||||
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#include "paddle/memory/detail/system_allocator.h"
|
||||||
|
|
||||||
|
#include <stdlib.h> // for malloc and free
|
||||||
|
#include <sys/mman.h> // for mlock and munlock
|
||||||
|
|
||||||
|
#include "gflags/gflags.h"
|
||||||
|
#include "paddle/platform/assert.h"
|
||||||
|
#include "paddle/platform/cuda.h"
|
||||||
|
|
||||||
|
// If use_pinned_memory is true, CPUAllocator calls mlock, which
|
||||||
|
// returns pinned and locked memory as staging areas for data exchange
|
||||||
|
// between host and device. Allocates too much would reduce the amount
|
||||||
|
// of memory available to the system for paging. So, by default, we
|
||||||
|
// should set false to use_pinned_memory.
|
||||||
|
DEFINE_bool(use_pinned_memory, false,
|
||||||
|
"If set, allocate cpu/gpu pinned memory.");
|
||||||
|
|
||||||
|
namespace paddle {
|
||||||
|
namespace memory {
|
||||||
|
namespace detail {
|
||||||
|
|
||||||
|
void* CPUAllocator::Alloc(size_t size) {
|
||||||
|
// According to http://www.cplusplus.com/reference/cstdlib/malloc/,
|
||||||
|
// malloc might not return nullptr if size is zero, but the returned
|
||||||
|
// pointer shall not be dereferenced -- so we make it nullptr.
|
||||||
|
if (size <= 0) return nullptr;
|
||||||
|
|
||||||
|
void* p = malloc(size);
|
||||||
|
if (p != nullptr && FLAGS_use_pinned_memory) {
|
||||||
|
mlock(p, size);
|
||||||
|
}
|
||||||
|
return p;
|
||||||
|
}
|
||||||
|
|
||||||
|
void CPUAllocator::Free(void* p, size_t size) {
|
||||||
|
if (p != nullptr && FLAGS_use_pinned_memory) {
|
||||||
|
munlock(p, size);
|
||||||
|
}
|
||||||
|
free(p);
|
||||||
|
}
|
||||||
|
|
||||||
|
#ifndef PADDLE_ONLY_CPU
|
||||||
|
|
||||||
|
void* GPUAllocator::Alloc(size_t size) {
|
||||||
|
// CUDA documentation doesn't explain if cudaMalloc returns nullptr
|
||||||
|
// if size is 0. We just make sure it does.
|
||||||
|
if (size <= 0) {
|
||||||
|
return nullptr;
|
||||||
|
}
|
||||||
|
|
||||||
|
void* p = 0;
|
||||||
|
cudaError_t result =
|
||||||
|
FLAGS_use_pinned_memory ? cudaMallocHost(&p, size) : cudaMalloc(&p, size);
|
||||||
|
if (result != cudaSuccess) {
|
||||||
|
cudaGetLastError(); // clear error if there is any.
|
||||||
|
}
|
||||||
|
return result == cudaSuccess ? p : nullptr;
|
||||||
|
}
|
||||||
|
|
||||||
|
void GPUAllocator::Free(void* p, size_t size) {
|
||||||
|
// Purposefully allow cudaErrorCudartUnloading, because
|
||||||
|
// that is returned if you ever call cudaFree after the
|
||||||
|
// driver has already shutdown. This happens only if the
|
||||||
|
// process is terminating, in which case we don't care if
|
||||||
|
// cudaFree succeeds.
|
||||||
|
cudaError_t err = FLAGS_use_pinned_memory ? cudaFreeHost(p) : cudaFree(p);
|
||||||
|
if (err != cudaErrorCudartUnloading) {
|
||||||
|
platform::throw_on_error(err, "cudaFree{Host} failed");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#endif // PADDLE_ONLY_CPU
|
||||||
|
|
||||||
|
} // namespace detail
|
||||||
|
} // namespace memory
|
||||||
|
} // namespace paddle
|
@ -0,0 +1,53 @@
|
|||||||
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include <stddef.h> // for size_t
|
||||||
|
|
||||||
|
namespace paddle {
|
||||||
|
namespace memory {
|
||||||
|
namespace detail {
|
||||||
|
|
||||||
|
// SystemAllocator is the parent class of CPUAllocator and
|
||||||
|
// GPUAllocator. A BuddyAllocator object uses a SystemAllocator*
|
||||||
|
// pointing to the underlying system allocator. An alternative to
|
||||||
|
// this class hierarchy is to pass a system allocator class to
|
||||||
|
// BuddyAllocator as a template parameter. This approach makes
|
||||||
|
// BuddyAllocator a class template, and it's very complicated
|
||||||
|
// algorithm would make the buddy_allocator.h messy.
|
||||||
|
class SystemAllocator {
|
||||||
|
public:
|
||||||
|
virtual ~SystemAllocator() {}
|
||||||
|
virtual void* Alloc(size_t size) = 0;
|
||||||
|
virtual void Free(void* p, size_t size) = 0;
|
||||||
|
};
|
||||||
|
|
||||||
|
class CPUAllocator : public SystemAllocator {
|
||||||
|
public:
|
||||||
|
virtual void* Alloc(size_t size);
|
||||||
|
virtual void Free(void* p, size_t size);
|
||||||
|
};
|
||||||
|
|
||||||
|
#ifndef PADDLE_ONLY_CPU
|
||||||
|
class GPUAllocator : public SystemAllocator {
|
||||||
|
public:
|
||||||
|
virtual void* Alloc(size_t size);
|
||||||
|
virtual void Free(void* p, size_t size);
|
||||||
|
};
|
||||||
|
#endif // PADDLE_ONLY_CPU
|
||||||
|
|
||||||
|
} // namespace detail
|
||||||
|
} // namespace memory
|
||||||
|
} // namespace paddle
|
@ -0,0 +1,71 @@
|
|||||||
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#include "paddle/memory/detail/system_allocator.h"
|
||||||
|
|
||||||
|
#include <memory>
|
||||||
|
#include <vector>
|
||||||
|
|
||||||
|
#include "gflags/gflags.h"
|
||||||
|
#include "gtest/gtest.h"
|
||||||
|
|
||||||
|
DECLARE_bool(use_pinned_memory);
|
||||||
|
|
||||||
|
void TestAllocator(paddle::memory::detail::SystemAllocator& a, size_t size) {
|
||||||
|
bool freed = false;
|
||||||
|
{
|
||||||
|
void* p = a.Alloc(size);
|
||||||
|
if (size > 0) {
|
||||||
|
EXPECT_NE(p, nullptr);
|
||||||
|
} else {
|
||||||
|
EXPECT_EQ(p, nullptr);
|
||||||
|
}
|
||||||
|
|
||||||
|
int* i = static_cast<int*>(p);
|
||||||
|
std::shared_ptr<int> ptr(i, [&](void* p) {
|
||||||
|
freed = true;
|
||||||
|
a.Free(p, size);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
EXPECT_TRUE(freed);
|
||||||
|
}
|
||||||
|
|
||||||
|
TEST(CPUAllocator, NoLockMem) {
|
||||||
|
FLAGS_use_pinned_memory = false;
|
||||||
|
paddle::memory::detail::CPUAllocator a;
|
||||||
|
TestAllocator(a, 2048);
|
||||||
|
TestAllocator(a, 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
TEST(CPUAllocator, LockMem) {
|
||||||
|
FLAGS_use_pinned_memory = true;
|
||||||
|
paddle::memory::detail::CPUAllocator a;
|
||||||
|
TestAllocator(a, 2048);
|
||||||
|
TestAllocator(a, 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
#ifndef PADDLE_ONLY_CPU
|
||||||
|
TEST(GPUAllocator, NoStaging) {
|
||||||
|
FLAGS_use_pinned_memory = false;
|
||||||
|
paddle::memory::detail::GPUAllocator a;
|
||||||
|
TestAllocator(a, 2048);
|
||||||
|
TestAllocator(a, 0);
|
||||||
|
}
|
||||||
|
TEST(GPUAllocator, Staging) {
|
||||||
|
FLAGS_use_pinned_memory = true;
|
||||||
|
paddle::memory::detail::GPUAllocator a;
|
||||||
|
TestAllocator(a, 2048);
|
||||||
|
TestAllocator(a, 0);
|
||||||
|
}
|
||||||
|
#endif // PADDLE_ONLY_CPU
|
@ -0,0 +1,59 @@
|
|||||||
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#include "paddle/memory/memory.h"
|
||||||
|
#include "paddle/memory/detail/buddy_allocator.h"
|
||||||
|
#include "paddle/memory/detail/system_allocator.h"
|
||||||
|
#include "paddle/platform/assert.h"
|
||||||
|
|
||||||
|
#include <boost/variant.hpp>
|
||||||
|
|
||||||
|
namespace paddle {
|
||||||
|
namespace memory {
|
||||||
|
|
||||||
|
void* Alloc(platform::Place pl, size_t size) {
|
||||||
|
#ifndef PADDLE_ONLY_CPU
|
||||||
|
if (paddle::platform::is_gpu_place(pl)) {
|
||||||
|
size_t gpu_id = boost::get<platform::GPUPlace>(pl).device;
|
||||||
|
return detail::GetGPUBuddyAllocator(gpu_id)->Alloc(size);
|
||||||
|
}
|
||||||
|
#endif // PADDLE_ONLY_CPU
|
||||||
|
PADDLE_ASSERT(paddle::platform::is_cpu_place(pl));
|
||||||
|
return detail::GetCPUBuddyAllocator()->Alloc(size);
|
||||||
|
}
|
||||||
|
|
||||||
|
void Free(paddle::platform::Place pl, void* p) {
|
||||||
|
#ifndef PADDLE_ONLY_CPU
|
||||||
|
if (paddle::platform::is_gpu_place(pl)) {
|
||||||
|
size_t gpu_id = boost::get<platform::GPUPlace>(pl).device;
|
||||||
|
detail::GetGPUBuddyAllocator(gpu_id)->Free(p);
|
||||||
|
}
|
||||||
|
#endif // PADDLE_ONLY_CPU
|
||||||
|
PADDLE_ASSERT(paddle::platform::is_cpu_place(pl));
|
||||||
|
detail::GetCPUBuddyAllocator()->Free(p);
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t Used(paddle::platform::Place pl) {
|
||||||
|
#ifndef PADDLE_ONLY_CPU
|
||||||
|
if (paddle::platform::is_gpu_place(pl)) {
|
||||||
|
size_t gpu_id = boost::get<platform::GPUPlace>(pl).device;
|
||||||
|
return detail::GetGPUBuddyAllocator(gpu_id)->Used();
|
||||||
|
}
|
||||||
|
#endif // PADDLE_ONLY_CPU
|
||||||
|
PADDLE_ASSERT(paddle::platform::is_cpu_place(pl));
|
||||||
|
return detail::GetCPUBuddyAllocator()->Used();
|
||||||
|
}
|
||||||
|
|
||||||
|
} // namespace memory
|
||||||
|
} // namespace paddle
|
@ -0,0 +1,40 @@
|
|||||||
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#ifndef PADDLE_ONLY_CPU
|
||||||
|
|
||||||
|
#include <thrust/system/cuda/error.h>
|
||||||
|
#include <thrust/system_error.h>
|
||||||
|
|
||||||
|
namespace paddle {
|
||||||
|
namespace platform {
|
||||||
|
|
||||||
|
inline void throw_on_error(cudaError_t e, const char* message) {
|
||||||
|
if (e) {
|
||||||
|
throw thrust::system_error(e, thrust::cuda_category(), message);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
int GetDeviceCount(void) {
|
||||||
|
int count;
|
||||||
|
throw_on_error(cudaGetDeviceCount(&count), "cudaGetDeviceCount failed");
|
||||||
|
return count;
|
||||||
|
}
|
||||||
|
|
||||||
|
} // namespace platform
|
||||||
|
} // namespace paddle
|
||||||
|
|
||||||
|
#endif // PADDLE_ONLY_CPU
|
@ -1,10 +0,0 @@
|
|||||||
#include <gtest/gtest.h>
|
|
||||||
#include <paddle/platform/must_check.h>
|
|
||||||
|
|
||||||
int __must_check SomeFunctionMustCheck() { return 0; }
|
|
||||||
|
|
||||||
TEST(MustCheck, all) {
|
|
||||||
// This line should not be compiled, because the
|
|
||||||
// return value of SomeFunctionMustCheck marked as __must_check
|
|
||||||
// SomeFunctionMustCheck();
|
|
||||||
}
|
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in new issue