You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
154 lines
4.8 KiB
154 lines
4.8 KiB
#include "paddle/operators/math/math_function.h"
|
|
#include "gtest/gtest.h"
|
|
|
|
TEST(math_function, gemm_notrans_cblas) {
|
|
paddle::framework::Tensor input1;
|
|
paddle::framework::Tensor input2;
|
|
paddle::framework::Tensor input3;
|
|
|
|
int m = 2;
|
|
int n = 3;
|
|
int k = 3;
|
|
auto* cpu_place = new paddle::platform::CPUPlace();
|
|
float* input1_ptr = input1.mutable_data<float>({2, 3}, *cpu_place);
|
|
float arr1[6] = {0, 1, 2, 3, 4, 5};
|
|
memcpy(input1_ptr, arr1, 6 * sizeof(float));
|
|
float* input2_ptr = input2.mutable_data<float>({3, 4}, *cpu_place);
|
|
float arr2[12] = {0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11};
|
|
memcpy(input2_ptr, arr2, 12 * sizeof(float));
|
|
float* input3_ptr = input3.mutable_data<float>({2, 4}, *cpu_place);
|
|
float arr3[8] = {0, 1, 2, 3, 4, 5, 6, 7};
|
|
memcpy(input3_ptr, arr3, 8 * sizeof(float));
|
|
|
|
paddle::platform::CPUDeviceContext context(*cpu_place);
|
|
paddle::operators::math::gemm<paddle::platform::CPUPlace, float>(
|
|
context, false, false, m, n, k, 1, input1_ptr, 3, input2_ptr + 1, 4, 1,
|
|
input3_ptr + 1, 4);
|
|
|
|
EXPECT_EQ(input3_ptr[0], 0);
|
|
EXPECT_EQ(input3_ptr[1], 24);
|
|
EXPECT_EQ(input3_ptr[2], 28);
|
|
EXPECT_EQ(input3_ptr[3], 32);
|
|
EXPECT_EQ(input3_ptr[4], 4);
|
|
EXPECT_EQ(input3_ptr[5], 73);
|
|
EXPECT_EQ(input3_ptr[6], 86);
|
|
EXPECT_EQ(input3_ptr[7], 99);
|
|
}
|
|
|
|
TEST(math_function, gemm_trans_clbas) {
|
|
paddle::framework::Tensor input1;
|
|
paddle::framework::Tensor input2;
|
|
paddle::framework::Tensor input3;
|
|
|
|
int m = 2;
|
|
int n = 3;
|
|
int k = 3;
|
|
auto* cpu_place = new paddle::platform::CPUPlace();
|
|
float* input1_ptr = input1.mutable_data<float>({2, 3}, *cpu_place);
|
|
float arr1[6] = {0, 1, 2, 3, 4, 5};
|
|
memcpy(input1_ptr, arr1, 6 * sizeof(float));
|
|
float* input2_ptr = input2.mutable_data<float>({4, 3}, *cpu_place);
|
|
float arr2[12] = {0, 4, 8, 1, 5, 9, 2, 6, 10, 3, 7, 11};
|
|
memcpy(input2_ptr, arr2, 12 * sizeof(float));
|
|
float* input3_ptr = input3.mutable_data<float>({2, 4}, *cpu_place);
|
|
float arr3[8] = {0, 1, 2, 3, 4, 5, 6, 7};
|
|
memcpy(input3_ptr, arr3, 8 * sizeof(float));
|
|
|
|
paddle::platform::CPUDeviceContext context(*cpu_place);
|
|
paddle::operators::math::gemm<paddle::platform::CPUPlace, float>(
|
|
context, false, true, m, n, k, 1, input1_ptr, 3, input2_ptr + 3, 3, 1,
|
|
input3_ptr + 1, 4);
|
|
|
|
EXPECT_EQ(input3_ptr[0], 0);
|
|
EXPECT_EQ(input3_ptr[1], 24);
|
|
EXPECT_EQ(input3_ptr[2], 28);
|
|
EXPECT_EQ(input3_ptr[3], 32);
|
|
EXPECT_EQ(input3_ptr[4], 4);
|
|
EXPECT_EQ(input3_ptr[5], 73);
|
|
EXPECT_EQ(input3_ptr[6], 86);
|
|
EXPECT_EQ(input3_ptr[7], 99);
|
|
}
|
|
|
|
TEST(math_function, zero) {
|
|
paddle::framework::Tensor tensor;
|
|
auto* cpu_place = new paddle::platform::CPUPlace();
|
|
float* t = tensor.mutable_data<float>({2, 2}, *cpu_place);
|
|
paddle::platform::CPUDeviceContext context(*cpu_place);
|
|
paddle::operators::math::SetConstant<paddle::platform::CPUPlace, float>
|
|
functor;
|
|
functor(context, &tensor, 0);
|
|
EXPECT_EQ(t[0], 0);
|
|
EXPECT_EQ(t[1], 0);
|
|
EXPECT_EQ(t[2], 0);
|
|
EXPECT_EQ(t[3], 0);
|
|
|
|
functor(context, &tensor, 1);
|
|
|
|
EXPECT_EQ(t[0], 1);
|
|
EXPECT_EQ(t[1], 1);
|
|
EXPECT_EQ(t[2], 1);
|
|
EXPECT_EQ(t[3], 1);
|
|
}
|
|
|
|
template <typename T>
|
|
void GemvTest(int m, int n, bool trans) {
|
|
paddle::framework::Tensor mat_a;
|
|
paddle::framework::Tensor vec_b;
|
|
paddle::framework::Tensor vec_c;
|
|
auto* cpu_place = new paddle::platform::CPUPlace();
|
|
int b_num = trans ? m : n;
|
|
int c_num = trans ? n : m;
|
|
|
|
T* data_a = mat_a.mutable_data<T>({m, n}, *cpu_place);
|
|
T* data_b = vec_b.mutable_data<T>({b_num}, *cpu_place);
|
|
T* data_c = vec_c.mutable_data<T>({c_num}, *cpu_place);
|
|
for (int i = 0; i < mat_a.numel(); ++i) {
|
|
data_a[i] = static_cast<T>(i);
|
|
}
|
|
for (int i = 0; i < vec_b.numel(); ++i) {
|
|
data_b[i] = static_cast<T>(i);
|
|
}
|
|
|
|
paddle::platform::CPUDeviceContext context(*cpu_place);
|
|
paddle::operators::math::gemv<paddle::platform::CPUPlace, T>(
|
|
context, trans, static_cast<int>(m), static_cast<int>(n), 1., data_a,
|
|
data_b, 0., data_c);
|
|
|
|
if (!trans) {
|
|
for (int i = 0; i < m; ++i) {
|
|
T sum = 0.0;
|
|
for (int j = 0; j < n; ++j) {
|
|
sum += data_a[i * n + j] * data_b[j];
|
|
}
|
|
ASSERT_FLOAT_EQ(data_c[i], sum);
|
|
}
|
|
} else {
|
|
for (int i = 0; i < n; ++i) {
|
|
T sum = 0.0;
|
|
for (int j = 0; j < m; ++j) {
|
|
sum += data_a[j * n + i] * data_b[j];
|
|
}
|
|
ASSERT_FLOAT_EQ(data_c[i], sum);
|
|
}
|
|
}
|
|
}
|
|
|
|
TEST(math_function, gemv) {
|
|
GemvTest<float>(3, 13, false);
|
|
GemvTest<double>(4, 5, false);
|
|
GemvTest<float>(12, 7, true);
|
|
GemvTest<double>(7, 9, true);
|
|
}
|
|
|
|
TEST(math_funciton, set_constant) {
|
|
paddle::framework::Tensor t;
|
|
t.Resize({10, 10});
|
|
t.mutable_data<int>(paddle::platform::CPUPlace());
|
|
auto* ctx = new paddle::platform::CPUDeviceContext();
|
|
paddle::operators::math::set_constant(*ctx, &t, 10);
|
|
for (int64_t i = 0; i < t.numel(); ++i) {
|
|
PADDLE_ENFORCE_EQ(10, t.data<int>()[i]);
|
|
}
|
|
delete ctx;
|
|
}
|