Merge branch 'develop' of https://github.com/PaddlePaddle/Paddle into cross_channel_norm
commit
77b0bf41be
@ -0,0 +1,81 @@
|
|||||||
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include "paddle/framework/data_layout.h"
|
||||||
|
#include "paddle/framework/data_type.h"
|
||||||
|
#include "paddle/framework/library_type.h"
|
||||||
|
#include "paddle/platform/device_context.h"
|
||||||
|
#include "paddle/platform/place.h"
|
||||||
|
|
||||||
|
namespace paddle {
|
||||||
|
namespace framework {
|
||||||
|
|
||||||
|
struct OpKernelType {
|
||||||
|
struct Hash {
|
||||||
|
size_t operator()(const OpKernelType& key) const {
|
||||||
|
int place = key.place_.which() + (1 << LEFT_SHIFT);
|
||||||
|
int data_type =
|
||||||
|
static_cast<int>(key.data_type_) + (1 << (LEFT_SHIFT + 1));
|
||||||
|
int data_layout =
|
||||||
|
static_cast<int>(key.data_layout_) + (1 << (LEFT_SHIFT + 2));
|
||||||
|
int library_type =
|
||||||
|
static_cast<int>(key.library_type_) + (1 << (LEFT_SHIFT + 3));
|
||||||
|
std::hash<int> hasher;
|
||||||
|
return hasher(place + data_type + data_layout + library_type);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
// place, data_type, library_type kinds less than 2^8
|
||||||
|
constexpr static int LEFT_SHIFT = 8;
|
||||||
|
proto::DataType data_type_;
|
||||||
|
DataLayout data_layout_;
|
||||||
|
platform::Place place_;
|
||||||
|
LibraryType library_type_;
|
||||||
|
|
||||||
|
OpKernelType(proto::DataType data_type, platform::Place place,
|
||||||
|
DataLayout data_layout = DataLayout::kAnyLayout,
|
||||||
|
LibraryType library_type = LibraryType::kPlain)
|
||||||
|
: data_type_(data_type),
|
||||||
|
data_layout_(data_layout),
|
||||||
|
place_(place),
|
||||||
|
library_type_(library_type) {}
|
||||||
|
|
||||||
|
OpKernelType(proto::DataType data_type,
|
||||||
|
const platform::DeviceContext& dev_ctx,
|
||||||
|
DataLayout data_layout = DataLayout::kAnyLayout,
|
||||||
|
LibraryType library_type = LibraryType::kPlain)
|
||||||
|
: data_type_(data_type),
|
||||||
|
data_layout_(data_layout),
|
||||||
|
place_(dev_ctx.GetPlace()),
|
||||||
|
library_type_(library_type) {}
|
||||||
|
|
||||||
|
bool operator==(const OpKernelType& o) const {
|
||||||
|
return platform::places_are_same_class(place_, o.place_) &&
|
||||||
|
data_type_ == o.data_type_ && data_layout_ == o.data_layout_ &&
|
||||||
|
library_type_ == o.library_type_;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
inline std::ostream& operator<<(std::ostream& os,
|
||||||
|
const OpKernelType& kernel_key) {
|
||||||
|
os << "data_type[" << kernel_key.data_type_ << "]:data_layout["
|
||||||
|
<< kernel_key.data_layout_ << "]:place[" << kernel_key.place_
|
||||||
|
<< "]:library_type[" << kernel_key.library_type_ << "]";
|
||||||
|
return os;
|
||||||
|
}
|
||||||
|
|
||||||
|
} // namespace framework
|
||||||
|
} // namespace paddle
|
@ -0,0 +1,51 @@
|
|||||||
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#include "paddle/framework/op_kernel_type.h"
|
||||||
|
#include <gtest/gtest.h>
|
||||||
|
#include <iostream>
|
||||||
|
|
||||||
|
TEST(OpKernelType, ToString) {
|
||||||
|
using OpKernelType = paddle::framework::OpKernelType;
|
||||||
|
using DataType = paddle::framework::proto::DataType;
|
||||||
|
using CPUPlace = paddle::platform::CPUPlace;
|
||||||
|
using DataLayout = paddle::framework::DataLayout;
|
||||||
|
using LibraryType = paddle::framework::LibraryType;
|
||||||
|
|
||||||
|
OpKernelType op_kernel_type(DataType::FP32, CPUPlace(), DataLayout::kNCHW,
|
||||||
|
LibraryType::kCUDNN);
|
||||||
|
|
||||||
|
std::ostringstream stream;
|
||||||
|
stream << op_kernel_type;
|
||||||
|
ASSERT_EQ(
|
||||||
|
stream.str(),
|
||||||
|
"data_type[5]:data_layout[NCHW]:place[CPUPlace]:library_type[CUDNN]");
|
||||||
|
}
|
||||||
|
|
||||||
|
TEST(OpKernelType, Hash) {
|
||||||
|
using OpKernelType = paddle::framework::OpKernelType;
|
||||||
|
using DataType = paddle::framework::proto::DataType;
|
||||||
|
using CPUPlace = paddle::platform::CPUPlace;
|
||||||
|
using CUDAPlace = paddle::platform::CUDAPlace;
|
||||||
|
using DataLayout = paddle::framework::DataLayout;
|
||||||
|
using LibraryType = paddle::framework::LibraryType;
|
||||||
|
|
||||||
|
OpKernelType op_kernel_type_1(DataType::FP32, CPUPlace(), DataLayout::kNCHW,
|
||||||
|
LibraryType::kCUDNN);
|
||||||
|
OpKernelType op_kernel_type_2(DataType::FP32, CUDAPlace(0), DataLayout::kNCHW,
|
||||||
|
LibraryType::kCUDNN);
|
||||||
|
|
||||||
|
OpKernelType::Hash hasher;
|
||||||
|
ASSERT_NE(hasher(op_kernel_type_1), hasher(op_kernel_type_2));
|
||||||
|
}
|
@ -0,0 +1,161 @@
|
|||||||
|
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License. */
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
#include <condition_variable>
|
||||||
|
#include <cstdio>
|
||||||
|
#include <functional>
|
||||||
|
#include <iostream>
|
||||||
|
#include <mutex>
|
||||||
|
#include <queue>
|
||||||
|
#include <thread>
|
||||||
|
|
||||||
|
#include "paddle/platform/call_once.h"
|
||||||
|
#include "paddle/platform/enforce.h"
|
||||||
|
|
||||||
|
namespace paddle {
|
||||||
|
namespace framework {
|
||||||
|
|
||||||
|
typedef std::function<void()> Task;
|
||||||
|
|
||||||
|
class ThreadPool {
|
||||||
|
public:
|
||||||
|
/**
|
||||||
|
* @brief Get a instance of threadpool, the thread number will
|
||||||
|
* be specified as the number of hardware thread contexts
|
||||||
|
*/
|
||||||
|
static ThreadPool* GetInstance() {
|
||||||
|
std::call_once(init_flag, &ThreadPool::Init);
|
||||||
|
return threadpool.get();
|
||||||
|
}
|
||||||
|
|
||||||
|
~ThreadPool() {
|
||||||
|
{
|
||||||
|
// notify all threads to stop running
|
||||||
|
running_ = false;
|
||||||
|
scheduled_.notify_all();
|
||||||
|
}
|
||||||
|
|
||||||
|
for (auto& t : threads_) {
|
||||||
|
t->join();
|
||||||
|
t.reset(nullptr);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
int GetNumThreads() const { return num_threads_; }
|
||||||
|
|
||||||
|
int GetAvailable() {
|
||||||
|
std::unique_lock<std::mutex> lock(mutex_);
|
||||||
|
return available_;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Push a function to the queue, and will be scheduled and
|
||||||
|
* executed if a thread is available.
|
||||||
|
* @param[in] Task will be pushed to the task queue.
|
||||||
|
*/
|
||||||
|
void Run(const Task& fn) {
|
||||||
|
std::unique_lock<std::mutex> lock(mutex_);
|
||||||
|
tasks_.push(fn);
|
||||||
|
lock.unlock();
|
||||||
|
scheduled_.notify_one();
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Wait until all the tasks are completed.
|
||||||
|
*/
|
||||||
|
void Wait() {
|
||||||
|
std::unique_lock<std::mutex> lock(mutex_);
|
||||||
|
completed_.wait(lock, [=] { return Done() == true; });
|
||||||
|
}
|
||||||
|
|
||||||
|
private:
|
||||||
|
ThreadPool& operator=(const ThreadPool&) = delete;
|
||||||
|
ThreadPool(const ThreadPool&) = delete;
|
||||||
|
|
||||||
|
ThreadPool(int num_threads)
|
||||||
|
: num_threads_(num_threads), available_(num_threads), running_(true) {
|
||||||
|
threads_.resize(num_threads);
|
||||||
|
for (auto& thread : threads_) {
|
||||||
|
// TODO(Yancey1989): binding the thread on the specify CPU number
|
||||||
|
thread.reset(new std::thread(std::bind(&ThreadPool::TaskLoop, this)));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief If the task queue is empty and avaialbe
|
||||||
|
* is equal to the number of threads, means that
|
||||||
|
* all tasks are completed.
|
||||||
|
*
|
||||||
|
* Note: this function is not thread-safe.
|
||||||
|
*
|
||||||
|
* @return true if all tasks are completed.
|
||||||
|
*/
|
||||||
|
bool Done() { return tasks_.empty() && available_ == num_threads_; }
|
||||||
|
|
||||||
|
void TaskLoop() {
|
||||||
|
while (running_) {
|
||||||
|
std::unique_lock<std::mutex> lock(mutex_);
|
||||||
|
scheduled_.wait(lock, [=] { return !tasks_.empty() || !running_; });
|
||||||
|
|
||||||
|
if (!running_) {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
// pop a task from the task queue
|
||||||
|
auto task = tasks_.front();
|
||||||
|
tasks_.pop();
|
||||||
|
|
||||||
|
--available_;
|
||||||
|
lock.unlock();
|
||||||
|
|
||||||
|
// run the task
|
||||||
|
task();
|
||||||
|
|
||||||
|
{
|
||||||
|
std::unique_lock<std::mutex> lock(mutex_);
|
||||||
|
++available_;
|
||||||
|
if (Done()) {
|
||||||
|
completed_.notify_all();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static void Init() {
|
||||||
|
if (threadpool.get() == nullptr) {
|
||||||
|
// TODO(Yancey1989): specify the max threads number
|
||||||
|
int num_threads = std::thread::hardware_concurrency();
|
||||||
|
PADDLE_ENFORCE_GT(num_threads, 0);
|
||||||
|
threadpool.reset(new ThreadPool(num_threads));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private:
|
||||||
|
static std::unique_ptr<ThreadPool> threadpool;
|
||||||
|
static std::once_flag init_flag;
|
||||||
|
|
||||||
|
int num_threads_;
|
||||||
|
int available_;
|
||||||
|
bool running_;
|
||||||
|
std::queue<Task> tasks_;
|
||||||
|
std::vector<std::unique_ptr<std::thread>> threads_;
|
||||||
|
std::mutex mutex_;
|
||||||
|
std::condition_variable scheduled_;
|
||||||
|
std::condition_variable completed_;
|
||||||
|
};
|
||||||
|
|
||||||
|
std::unique_ptr<ThreadPool> ThreadPool::threadpool(nullptr);
|
||||||
|
std::once_flag ThreadPool::init_flag;
|
||||||
|
} // namespace framework
|
||||||
|
} // namespace paddle
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in new issue