|
|
|
@ -1,4 +1,4 @@
|
|
|
|
|
# Copyright 2019 Huawei Technologies Co., Ltd
|
|
|
|
|
# Copyright 2019-2021 Huawei Technologies Co., Ltd
|
|
|
|
|
#
|
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
@ -31,20 +31,17 @@ class NetMul(nn.Cell):
|
|
|
|
|
return self.mul(x, y)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.level0
|
|
|
|
|
@pytest.mark.platform_x86_gpu_training
|
|
|
|
|
@pytest.mark.env_onecard
|
|
|
|
|
def test_mul():
|
|
|
|
|
x0_np = np.random.uniform(-2, 2, (2, 3, 4, 4)).astype(np.float32)
|
|
|
|
|
y0_np = np.random.uniform(-2, 2, (2, 3, 4, 4)).astype(np.float32)
|
|
|
|
|
x1_np = np.random.uniform(-2, 2, (2, 3, 4, 4)).astype(np.float32)
|
|
|
|
|
y1_np = np.random.uniform(-2, 2, (2, 1, 4, 4)).astype(np.float32)
|
|
|
|
|
x2_np = np.random.uniform(-2, 2, (2, 1, 1, 4)).astype(np.float32)
|
|
|
|
|
y2_np = np.random.uniform(-2, 2, (2, 3, 4, 4)).astype(np.float32)
|
|
|
|
|
x3_np = np.random.uniform(-2, 2, 1).astype(np.float32)
|
|
|
|
|
y3_np = np.random.uniform(-2, 2, 1).astype(np.float32)
|
|
|
|
|
x4_np = np.array(768).astype(np.float32)
|
|
|
|
|
y4_np = np.array(3072.5).astype(np.float32)
|
|
|
|
|
def mul(nptype):
|
|
|
|
|
x0_np = np.random.uniform(-2, 2, (2, 3, 4, 4)).astype(nptype)
|
|
|
|
|
y0_np = np.random.uniform(-2, 2, (2, 3, 4, 4)).astype(nptype)
|
|
|
|
|
x1_np = np.random.uniform(-2, 2, (2, 3, 4, 4)).astype(nptype)
|
|
|
|
|
y1_np = np.random.uniform(-2, 2, (2, 1, 4, 4)).astype(nptype)
|
|
|
|
|
x2_np = np.random.uniform(-2, 2, (2, 1, 1, 4)).astype(nptype)
|
|
|
|
|
y2_np = np.random.uniform(-2, 2, (2, 3, 4, 4)).astype(nptype)
|
|
|
|
|
x3_np = np.random.uniform(-2, 2, 1).astype(nptype)
|
|
|
|
|
y3_np = np.random.uniform(-2, 2, 1).astype(nptype)
|
|
|
|
|
x4_np = np.array(78).astype(nptype)
|
|
|
|
|
y4_np = np.array(37.5).astype(nptype)
|
|
|
|
|
|
|
|
|
|
x0 = Tensor(x0_np)
|
|
|
|
|
y0 = Tensor(y0_np)
|
|
|
|
@ -58,36 +55,36 @@ def test_mul():
|
|
|
|
|
y4 = Tensor(y4_np)
|
|
|
|
|
|
|
|
|
|
context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU")
|
|
|
|
|
mul = NetMul()
|
|
|
|
|
output0 = mul(x0, y0)
|
|
|
|
|
mul_net = NetMul()
|
|
|
|
|
output0 = mul_net(x0, y0)
|
|
|
|
|
expect0 = np.multiply(x0_np, y0_np)
|
|
|
|
|
diff0 = output0.asnumpy() - expect0
|
|
|
|
|
error0 = np.ones(shape=expect0.shape) * 1.0e-5
|
|
|
|
|
assert np.all(diff0 < error0)
|
|
|
|
|
assert output0.shape == expect0.shape
|
|
|
|
|
|
|
|
|
|
output1 = mul(x1, y1)
|
|
|
|
|
output1 = mul_net(x1, y1)
|
|
|
|
|
expect1 = np.multiply(x1_np, y1_np)
|
|
|
|
|
diff1 = output1.asnumpy() - expect1
|
|
|
|
|
error1 = np.ones(shape=expect1.shape) * 1.0e-5
|
|
|
|
|
assert np.all(diff1 < error1)
|
|
|
|
|
assert output1.shape == expect1.shape
|
|
|
|
|
|
|
|
|
|
output2 = mul(x2, y2)
|
|
|
|
|
output2 = mul_net(x2, y2)
|
|
|
|
|
expect2 = np.multiply(x2_np, y2_np)
|
|
|
|
|
diff2 = output2.asnumpy() - expect2
|
|
|
|
|
error2 = np.ones(shape=expect2.shape) * 1.0e-5
|
|
|
|
|
assert np.all(diff2 < error2)
|
|
|
|
|
assert output2.shape == expect2.shape
|
|
|
|
|
|
|
|
|
|
output3 = mul(x3, y3)
|
|
|
|
|
output3 = mul_net(x3, y3)
|
|
|
|
|
expect3 = np.multiply(x3_np, y3_np)
|
|
|
|
|
diff3 = output3.asnumpy() - expect3
|
|
|
|
|
error3 = np.ones(shape=expect3.shape) * 1.0e-5
|
|
|
|
|
assert np.all(diff3 < error3)
|
|
|
|
|
assert output3.shape == expect3.shape
|
|
|
|
|
|
|
|
|
|
output4 = mul(x4, y4)
|
|
|
|
|
output4 = mul_net(x4, y4)
|
|
|
|
|
expect4 = np.multiply(x4_np, y4_np)
|
|
|
|
|
diff4 = output4.asnumpy() - expect4
|
|
|
|
|
error4 = np.ones(shape=expect4.shape) * 1.0e-5
|
|
|
|
@ -95,42 +92,72 @@ def test_mul():
|
|
|
|
|
assert output4.shape == expect4.shape
|
|
|
|
|
|
|
|
|
|
context.set_context(mode=context.GRAPH_MODE, device_target="GPU")
|
|
|
|
|
mul = NetMul()
|
|
|
|
|
output0 = mul(x0, y0)
|
|
|
|
|
mul_net = NetMul()
|
|
|
|
|
output0 = mul_net(x0, y0)
|
|
|
|
|
expect0 = np.multiply(x0_np, y0_np)
|
|
|
|
|
diff0 = output0.asnumpy() - expect0
|
|
|
|
|
error0 = np.ones(shape=expect0.shape) * 1.0e-5
|
|
|
|
|
assert np.all(diff0 < error0)
|
|
|
|
|
assert output0.shape == expect0.shape
|
|
|
|
|
|
|
|
|
|
output1 = mul(x1, y1)
|
|
|
|
|
output1 = mul_net(x1, y1)
|
|
|
|
|
expect1 = np.multiply(x1_np, y1_np)
|
|
|
|
|
diff1 = output1.asnumpy() - expect1
|
|
|
|
|
error1 = np.ones(shape=expect1.shape) * 1.0e-5
|
|
|
|
|
assert np.all(diff1 < error1)
|
|
|
|
|
assert output1.shape == expect1.shape
|
|
|
|
|
|
|
|
|
|
output2 = mul(x2, y2)
|
|
|
|
|
output2 = mul_net(x2, y2)
|
|
|
|
|
expect2 = np.multiply(x2_np, y2_np)
|
|
|
|
|
diff2 = output2.asnumpy() - expect2
|
|
|
|
|
error2 = np.ones(shape=expect2.shape) * 1.0e-5
|
|
|
|
|
assert np.all(diff2 < error2)
|
|
|
|
|
assert output2.shape == expect2.shape
|
|
|
|
|
|
|
|
|
|
output3 = mul(x3, y3)
|
|
|
|
|
output3 = mul_net(x3, y3)
|
|
|
|
|
expect3 = np.multiply(x3_np, y3_np)
|
|
|
|
|
diff3 = output3.asnumpy() - expect3
|
|
|
|
|
error3 = np.ones(shape=expect3.shape) * 1.0e-5
|
|
|
|
|
assert np.all(diff3 < error3)
|
|
|
|
|
assert output3.shape == expect3.shape
|
|
|
|
|
|
|
|
|
|
output4 = mul(x4, y4)
|
|
|
|
|
output4 = mul_net(x4, y4)
|
|
|
|
|
expect4 = np.multiply(x4_np, y4_np)
|
|
|
|
|
diff4 = output4.asnumpy() - expect4
|
|
|
|
|
error4 = np.ones(shape=expect4.shape) * 1.0e-5
|
|
|
|
|
assert np.all(diff4 < error4)
|
|
|
|
|
assert output4.shape == expect4.shape
|
|
|
|
|
|
|
|
|
|
@pytest.mark.level0
|
|
|
|
|
@pytest.mark.platform_x86_gpu_training
|
|
|
|
|
@pytest.mark.env_onecard
|
|
|
|
|
def test_mul_float64():
|
|
|
|
|
mul(np.float64)
|
|
|
|
|
|
|
|
|
|
@pytest.mark.level0
|
|
|
|
|
@pytest.mark.platform_x86_gpu_training
|
|
|
|
|
@pytest.mark.env_onecard
|
|
|
|
|
def test_mul_float32():
|
|
|
|
|
mul(np.float32)
|
|
|
|
|
|
|
|
|
|
@pytest.mark.level0
|
|
|
|
|
@pytest.mark.platform_x86_gpu_training
|
|
|
|
|
@pytest.mark.env_onecard
|
|
|
|
|
def test_mul_float16():
|
|
|
|
|
mul(np.float16)
|
|
|
|
|
|
|
|
|
|
@pytest.mark.level0
|
|
|
|
|
@pytest.mark.platform_x86_gpu_training
|
|
|
|
|
@pytest.mark.env_onecard
|
|
|
|
|
def test_mul_int64():
|
|
|
|
|
mul(np.int64)
|
|
|
|
|
|
|
|
|
|
@pytest.mark.level0
|
|
|
|
|
@pytest.mark.platform_x86_gpu_training
|
|
|
|
|
@pytest.mark.env_onecard
|
|
|
|
|
def test_mul_int32():
|
|
|
|
|
mul(np.int32)
|
|
|
|
|
|
|
|
|
|
class NetMul_dynamic(nn.Cell):
|
|
|
|
|
def __init__(self):
|
|
|
|
|
super(NetMul_dynamic, self).__init__()
|
|
|
|
@ -143,14 +170,12 @@ class NetMul_dynamic(nn.Cell):
|
|
|
|
|
out = self.mul(x, y)
|
|
|
|
|
return out
|
|
|
|
|
|
|
|
|
|
@pytest.mark.level0
|
|
|
|
|
@pytest.mark.platform_x86_gpu_training
|
|
|
|
|
@pytest.mark.env_onecard
|
|
|
|
|
def test_mul_dynamic():
|
|
|
|
|
x1_np = np.array([768]).astype(np.float32)
|
|
|
|
|
y1_np = np.array([3072.5]).astype(np.float32)
|
|
|
|
|
x2_np = np.random.uniform(-2, 2, (2, 1, 1, 4)).astype(np.float32)
|
|
|
|
|
y2_np = np.random.uniform(-2, 2, (2, 3, 4, 4)).astype(np.float32)
|
|
|
|
|
|
|
|
|
|
def mul_dynamic(nptype):
|
|
|
|
|
x1_np = np.array([78]).astype(nptype)
|
|
|
|
|
y1_np = np.array([37.5]).astype(nptype)
|
|
|
|
|
x2_np = np.random.uniform(-2, 2, (2, 1, 1, 4)).astype(nptype)
|
|
|
|
|
y2_np = np.random.uniform(-2, 2, (2, 3, 4, 4)).astype(nptype)
|
|
|
|
|
|
|
|
|
|
x1 = Tensor(x1_np)
|
|
|
|
|
y1 = Tensor(y1_np)
|
|
|
|
@ -159,10 +184,10 @@ def test_mul_dynamic():
|
|
|
|
|
|
|
|
|
|
context.set_context(mode=context.GRAPH_MODE, device_target="GPU")
|
|
|
|
|
|
|
|
|
|
mul = NetMul_dynamic()
|
|
|
|
|
mul_net = NetMul_dynamic()
|
|
|
|
|
|
|
|
|
|
output1 = mul(x1, y1)
|
|
|
|
|
output2 = mul(x2, y2)
|
|
|
|
|
output1 = mul_net(x1, y1)
|
|
|
|
|
output2 = mul_net(x2, y2)
|
|
|
|
|
expect1 = np.multiply(x1_np, y1_np)
|
|
|
|
|
expect2 = np.multiply(x2_np, y2_np)
|
|
|
|
|
diff1 = output1.asnumpy() - expect1
|
|
|
|
@ -173,3 +198,33 @@ def test_mul_dynamic():
|
|
|
|
|
error2 = np.ones(shape=expect2.shape) * 1.0e-5
|
|
|
|
|
assert np.all(diff2 < error2)
|
|
|
|
|
assert output2.shape == expect2.shape
|
|
|
|
|
|
|
|
|
|
@pytest.mark.level0
|
|
|
|
|
@pytest.mark.platform_x86_gpu_training
|
|
|
|
|
@pytest.mark.env_onecard
|
|
|
|
|
def test_mul_dynamic_float64():
|
|
|
|
|
mul_dynamic(np.float64)
|
|
|
|
|
|
|
|
|
|
@pytest.mark.level0
|
|
|
|
|
@pytest.mark.platform_x86_gpu_training
|
|
|
|
|
@pytest.mark.env_onecard
|
|
|
|
|
def test_mul_dynamic_float32():
|
|
|
|
|
mul_dynamic(np.float32)
|
|
|
|
|
|
|
|
|
|
@pytest.mark.level0
|
|
|
|
|
@pytest.mark.platform_x86_gpu_training
|
|
|
|
|
@pytest.mark.env_onecard
|
|
|
|
|
def test_mul_dynamic_float16():
|
|
|
|
|
mul_dynamic(np.float16)
|
|
|
|
|
|
|
|
|
|
@pytest.mark.level0
|
|
|
|
|
@pytest.mark.platform_x86_gpu_training
|
|
|
|
|
@pytest.mark.env_onecard
|
|
|
|
|
def test_mul_dynamic_int64():
|
|
|
|
|
mul_dynamic(np.int64)
|
|
|
|
|
|
|
|
|
|
@pytest.mark.level0
|
|
|
|
|
@pytest.mark.platform_x86_gpu_training
|
|
|
|
|
@pytest.mark.env_onecard
|
|
|
|
|
def test_mul_dynamic_int32():
|
|
|
|
|
mul_dynamic(np.int32)
|
|
|
|
|