!7483 Add GPU-UniformSampler and nn.UniformCandidateSampler
Merge pull request !7483 from 34bunny/GPU-ucspull/7483/MERGE
commit
0d5e119fa4
@ -0,0 +1,36 @@
|
||||
/**
|
||||
* Copyright 2020 Huawei Technologies Co., Ltd
|
||||
*
|
||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||
* you may not use this file except in compliance with the License.
|
||||
* You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
#include "backend/kernel_compiler/gpu/cuda_impl/uniform_sampler_impl.cuh"
|
||||
|
||||
template <typename S>
|
||||
__global__ void AssignToOutput(const int size, const S prob_val, S *output_array) {
|
||||
for (size_t pos = blockIdx.x * blockDim.x + threadIdx.x; pos < size; pos += blockDim.x * gridDim.x) {
|
||||
output_array[pos] = prob_val;
|
||||
}
|
||||
}
|
||||
|
||||
template <typename S>
|
||||
void CalUniformSampler(const int true_size, const int num_sampled, const S prob_val, S *true_expected_count,
|
||||
S *sampled_expected_count, cudaStream_t cuda_stream) {
|
||||
AssignToOutput<<<GET_BLOCKS(true_size), GET_THREADS, 0, cuda_stream>>>(true_size, prob_val, true_expected_count);
|
||||
AssignToOutput<<<GET_BLOCKS(num_sampled), GET_THREADS, 0, cuda_stream>>>(num_sampled, prob_val,
|
||||
sampled_expected_count);
|
||||
}
|
||||
|
||||
template void CalUniformSampler<float>(const int true_size, const int num_sampled, const float prob_val,
|
||||
float *true_expected_count, float *sampled_expected_count,
|
||||
cudaStream_t cuda_stream);
|
@ -0,0 +1,26 @@
|
||||
/**
|
||||
* Copyright 2020 Huawei Technologies Co., Ltd
|
||||
*
|
||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||
* you may not use this file except in compliance with the License.
|
||||
* You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
#ifndef MINDSPORE_CCSRC_BACKEND_KERNEL_COMPILER_GPU_CUDA_IMPL_UNIFORM_SAMPLER_IMPL_CUH_
|
||||
#define MINDSPORE_CCSRC_BACKEND_KERNEL_COMPILER_GPU_CUDA_IMPL_UNIFORM_SAMPLER_IMPL_CUH_
|
||||
#include <cuda_runtime.h>
|
||||
#include "runtime/device/gpu/cuda_common.h"
|
||||
|
||||
template <typename S>
|
||||
void CalUniformSampler(const int true_size, const int num_sampled, const S prob_val, S *true_expected_count,
|
||||
S *sampled_expected_count, cudaStream_t cuda_stream);
|
||||
|
||||
#endif // MINDSPORE_CCSRC_BACKEND_KERNEL_COMPILER_GPU_CUDA_IMPL_UNIFORM_SAMPLER_IMPL_CUH_
|
@ -0,0 +1,29 @@
|
||||
/**
|
||||
* Copyright 2020 Huawei Technologies Co., Ltd
|
||||
*
|
||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||
* you may not use this file except in compliance with the License.
|
||||
* You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
#include "backend/kernel_compiler/gpu/nn/uniform_sampler_gpu_kernel.h"
|
||||
|
||||
namespace mindspore {
|
||||
namespace kernel {
|
||||
MS_REG_GPU_KERNEL_TWO(UniformSampler,
|
||||
KernelAttr()
|
||||
.AddInputAttr(kNumberTypeInt32)
|
||||
.AddOutputAttr(kNumberTypeInt32)
|
||||
.AddOutputAttr(kNumberTypeFloat32)
|
||||
.AddOutputAttr(kNumberTypeFloat32),
|
||||
UniformSamplerGpuKernel, int, float)
|
||||
} // namespace kernel
|
||||
} // namespace mindspore
|
@ -0,0 +1,144 @@
|
||||
/**
|
||||
* Copyright 2020 Huawei Technologies Co., Ltd
|
||||
*
|
||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||
* you may not use this file except in compliance with the License.
|
||||
* You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
#ifndef MINDSPORE_CCSRC_BACKEND_KERNEL_COMPILER_GPU_NN_UNIFORM_SAMPLER_GPU_KERNEL_H_
|
||||
#define MINDSPORE_CCSRC_BACKEND_KERNEL_COMPILER_GPU_NN_UNIFORM_SAMPLER_GPU_KERNEL_H_
|
||||
|
||||
#include <cmath>
|
||||
#include <set>
|
||||
#include <vector>
|
||||
#include <random>
|
||||
#include "backend/kernel_compiler/gpu/gpu_kernel.h"
|
||||
#include "backend/kernel_compiler/gpu/gpu_kernel_factory.h"
|
||||
#include "backend/kernel_compiler/gpu/cuda_impl/uniform_sampler_impl.cuh"
|
||||
|
||||
namespace mindspore {
|
||||
namespace kernel {
|
||||
template <typename T, typename S>
|
||||
class UniformSamplerGpuKernel : public GpuKernel {
|
||||
public:
|
||||
UniformSamplerGpuKernel() : num_true_(0), num_sampled_(0), unique_(false), range_max_(0), input_size_(0) {}
|
||||
~UniformSamplerGpuKernel() override = default;
|
||||
|
||||
const std::vector<size_t> &GetInputSizeList() const override { return input_size_list_; }
|
||||
const std::vector<size_t> &GetOutputSizeList() const override { return output_size_list_; }
|
||||
const std::vector<size_t> &GetWorkspaceSizeList() const override { return workspace_size_list_; }
|
||||
|
||||
bool Launch(const std::vector<AddressPtr> &inputs, const std::vector<AddressPtr> &workspaces,
|
||||
const std::vector<AddressPtr> &outputs, void *stream_ptr) override {
|
||||
VARIABLE_NOT_USED(workspaces);
|
||||
T *sampled_candidates = GetDeviceAddress<T>(outputs, 0);
|
||||
S *true_expected_count = GetDeviceAddress<S>(outputs, 1);
|
||||
S *sampled_expected_count = GetDeviceAddress<S>(outputs, 2);
|
||||
int counter = Sampling();
|
||||
float prob = Probability();
|
||||
size_t sampled_candidates_size = num_sampled_ * sizeof(T);
|
||||
S value = ApproximateExpectedCount(prob, num_sampled_, counter);
|
||||
CHECK_CUDA_RET_WITH_EXCEPT(cudaMemcpyAsync(sampled_candidates, &sampled_candidates_[0], sampled_candidates_size,
|
||||
cudaMemcpyHostToDevice, reinterpret_cast<cudaStream_t>(stream_ptr)),
|
||||
"cudaMemcpyAsync sampled_candidates failed");
|
||||
CalUniformSampler(static_cast<int>(input_size_), num_sampled_, value, true_expected_count, sampled_expected_count,
|
||||
reinterpret_cast<cudaStream_t>(stream_ptr));
|
||||
return true;
|
||||
}
|
||||
|
||||
bool Init(const CNodePtr &kernel_node) override {
|
||||
size_t input_num = AnfAlgo::GetInputTensorNum(kernel_node);
|
||||
if (input_num != 1) {
|
||||
MS_LOG(ERROR) << "Input number is " << input_num << ", but UniformSampler needs 1 input.";
|
||||
return false;
|
||||
}
|
||||
size_t output_num = AnfAlgo::GetOutputTensorNum(kernel_node);
|
||||
if (output_num != 3) {
|
||||
MS_LOG(ERROR) << "Output number is " << output_num << ", but UniformSampler has 3 outputs.";
|
||||
return false;
|
||||
}
|
||||
// getting attrs
|
||||
num_true_ = GetAttr<int>(kernel_node, "num_true");
|
||||
num_sampled_ = GetAttr<int>(kernel_node, "num_sampled");
|
||||
unique_ = GetAttr<bool>(kernel_node, "unique");
|
||||
range_max_ = GetAttr<int>(kernel_node, "range_max");
|
||||
int seed = GetAttr<int>(kernel_node, "seed");
|
||||
if (seed == 0) seed = time(NULL);
|
||||
generator_.seed(seed);
|
||||
auto input_shape = AnfAlgo::GetInputDeviceShape(kernel_node, 0);
|
||||
if (input_shape.size() != 2) {
|
||||
MS_LOG(ERROR) << "Input is " << input_shape.size() << "-D, but UniformSampler supports only 2-D inputs.";
|
||||
return false;
|
||||
}
|
||||
input_size_ = input_shape[0] * input_shape[1];
|
||||
InitSizeLists();
|
||||
return true;
|
||||
}
|
||||
|
||||
protected:
|
||||
void InitSizeLists() override {
|
||||
input_size_list_.push_back(input_size_ * sizeof(T));
|
||||
output_size_list_.push_back(num_sampled_ * sizeof(T));
|
||||
output_size_list_.push_back(input_size_ * sizeof(S));
|
||||
output_size_list_.push_back(num_sampled_ * sizeof(S));
|
||||
}
|
||||
|
||||
int Sampling() {
|
||||
int counter = 0;
|
||||
int tmp;
|
||||
int picked;
|
||||
std::set<int> set_container;
|
||||
// pick between [0, range_max_-1]
|
||||
std::uniform_int_distribution<int> distribution(0, range_max_ - 1);
|
||||
sampled_candidates_.clear();
|
||||
if (unique_) {
|
||||
picked = 0;
|
||||
while (picked < num_sampled_) {
|
||||
tmp = distribution(generator_);
|
||||
counter++;
|
||||
if (set_container.find(tmp) == set_container.end()) {
|
||||
set_container.insert(tmp);
|
||||
sampled_candidates_.push_back(tmp);
|
||||
picked++;
|
||||
}
|
||||
}
|
||||
} else {
|
||||
for (int i = 0; i < num_sampled_; i++) {
|
||||
sampled_candidates_.push_back(distribution(generator_));
|
||||
}
|
||||
counter = num_sampled_;
|
||||
}
|
||||
return counter;
|
||||
}
|
||||
|
||||
S Probability() { return static_cast<S>(1.0f / range_max_); }
|
||||
|
||||
S ApproximateExpectedCount(S p, int sampled_size, int counter) {
|
||||
if (sampled_size == counter) return p * sampled_size;
|
||||
return -std::expm1(counter * std::log1p(-p));
|
||||
}
|
||||
|
||||
private:
|
||||
int num_true_;
|
||||
int num_sampled_;
|
||||
bool unique_;
|
||||
int range_max_;
|
||||
size_t input_size_;
|
||||
std::default_random_engine generator_;
|
||||
std::vector<int> sampled_candidates_;
|
||||
std::vector<size_t> input_size_list_;
|
||||
std::vector<size_t> output_size_list_;
|
||||
std::vector<size_t> workspace_size_list_;
|
||||
};
|
||||
} // namespace kernel
|
||||
} // namespace mindspore
|
||||
#endif // MINDSPORE_CCSRC_BACKEND_KERNEL_COMPILER_GPU_NN_UNIFORM_SAMPLER_GPU_KERNEL_H_
|
@ -0,0 +1,116 @@
|
||||
# Copyright 2020 Huawei Technologies Co., Ltd
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
# ============================================================================
|
||||
|
||||
import numpy as np
|
||||
import pytest
|
||||
|
||||
from mindspore import Tensor
|
||||
from mindspore.ops import operations as P
|
||||
import mindspore.nn as nn
|
||||
import mindspore.context as context
|
||||
|
||||
class UniformSamplerNet(nn.Cell):
|
||||
def __init__(self, num_true, num_sampled, unique, range_max):
|
||||
super(UniformSamplerNet, self).__init__()
|
||||
self.sampler = P.UniformSampler(num_true, num_sampled, unique, range_max)
|
||||
|
||||
def construct(self, x):
|
||||
return self.sampler(x)
|
||||
|
||||
|
||||
def uniform_sampler(x, num_true, num_sampled, unique, range_max):
|
||||
uniform_sampler_net = UniformSamplerNet(num_true, num_sampled, unique, range_max)
|
||||
out1, out2, out3 = uniform_sampler_net(Tensor(x.astype(np.int32)))
|
||||
return out1.shape, out2.shape, out3.shape
|
||||
|
||||
@pytest.mark.level0
|
||||
@pytest.mark.platform_x86_gpu_training
|
||||
@pytest.mark.env_onecard
|
||||
def test_uniform_sampler_unique_1_true():
|
||||
context.set_context(mode=context.GRAPH_MODE, device_target="GPU")
|
||||
ms1, ms2, ms3 = uniform_sampler(np.array([[1], [3], [4], [6], [3]]), 1, 3, True, 4)
|
||||
expected_1 = (3,)
|
||||
expected_2 = (5, 1)
|
||||
expected_3 = (3,)
|
||||
np.testing.assert_array_equal(ms1, expected_1)
|
||||
np.testing.assert_array_equal(ms2, expected_2)
|
||||
np.testing.assert_array_equal(ms3, expected_3)
|
||||
|
||||
@pytest.mark.level0
|
||||
@pytest.mark.platform_x86_gpu_training
|
||||
@pytest.mark.env_onecard
|
||||
def test_uniform_sampler_not_unique_1_true():
|
||||
context.set_context(mode=context.GRAPH_MODE, device_target="GPU")
|
||||
ms1, ms2, ms3 = uniform_sampler(np.array([[1], [3], [4], [6], [3]]), 1, 3, False, 4)
|
||||
expected_1 = (3,)
|
||||
expected_2 = (5, 1)
|
||||
expected_3 = (3,)
|
||||
np.testing.assert_array_equal(ms1, expected_1)
|
||||
np.testing.assert_array_equal(ms2, expected_2)
|
||||
np.testing.assert_array_equal(ms3, expected_3)
|
||||
|
||||
@pytest.mark.level0
|
||||
@pytest.mark.platform_x86_gpu_training
|
||||
@pytest.mark.env_onecard
|
||||
def test_uniform_sampler_unique_2_true():
|
||||
context.set_context(mode=context.GRAPH_MODE, device_target="GPU")
|
||||
ms1, ms2, ms3 = uniform_sampler(np.array([[1, 2], [3, 2], [4, 2], [6, 2], [3, 2]]), 2, 3, True, 4)
|
||||
expected_1 = (3,)
|
||||
expected_2 = (5, 2)
|
||||
expected_3 = (3,)
|
||||
np.testing.assert_array_equal(ms1, expected_1)
|
||||
np.testing.assert_array_equal(ms2, expected_2)
|
||||
np.testing.assert_array_equal(ms3, expected_3)
|
||||
|
||||
@pytest.mark.level0
|
||||
@pytest.mark.platform_x86_gpu_training
|
||||
@pytest.mark.env_onecard
|
||||
def test_uniform_sampler_not_unique_2_true():
|
||||
context.set_context(mode=context.GRAPH_MODE, device_target="GPU")
|
||||
ms1, ms2, ms3 = uniform_sampler(np.array([[1, 2], [3, 2], [4, 2], [6, 2], [3, 2]]), 2, 3, False, 4)
|
||||
expected_1 = (3,)
|
||||
expected_2 = (5, 2)
|
||||
expected_3 = (3,)
|
||||
np.testing.assert_array_equal(ms1, expected_1)
|
||||
np.testing.assert_array_equal(ms2, expected_2)
|
||||
np.testing.assert_array_equal(ms3, expected_3)
|
||||
|
||||
@pytest.mark.level0
|
||||
@pytest.mark.platform_x86_gpu_training
|
||||
@pytest.mark.env_onecard
|
||||
def test_uniform_sampler_large():
|
||||
context.set_context(mode=context.GRAPH_MODE, device_target="GPU")
|
||||
ms1, ms2, ms3 = uniform_sampler(np.array([[12221, 41414], [3312, 5125152], [3312454, 51252],
|
||||
[65125, 225125], [35125, 5125122]]), 2, 5, False, 100)
|
||||
expected_1 = (5,)
|
||||
expected_2 = (5, 2)
|
||||
expected_3 = (5,)
|
||||
np.testing.assert_array_equal(ms1, expected_1)
|
||||
np.testing.assert_array_equal(ms2, expected_2)
|
||||
np.testing.assert_array_equal(ms3, expected_3)
|
||||
|
||||
|
||||
@pytest.mark.level0
|
||||
@pytest.mark.platform_x86_gpu_training
|
||||
@pytest.mark.env_onecard
|
||||
def test_uniform_sampler_large_random():
|
||||
context.set_context(mode=context.GRAPH_MODE, device_target="GPU")
|
||||
ms1, ms2, ms3 = uniform_sampler(np.arange(2142).reshape(34, 63), 63, 10, False, 12)
|
||||
expected_1 = (10,)
|
||||
expected_2 = (34, 63)
|
||||
expected_3 = (10,)
|
||||
np.testing.assert_array_equal(ms1, expected_1)
|
||||
np.testing.assert_array_equal(ms2, expected_2)
|
||||
np.testing.assert_array_equal(ms3, expected_3)
|
Loading…
Reference in new issue