|
|
|
@ -17,6 +17,9 @@ from __future__ import print_function
|
|
|
|
|
import unittest
|
|
|
|
|
import numpy as np
|
|
|
|
|
import math
|
|
|
|
|
# import paddle.fluid as fluid
|
|
|
|
|
# import paddle.fluid.core as core
|
|
|
|
|
# from op_builder import OpBuilder
|
|
|
|
|
from op_test import OpTest
|
|
|
|
|
|
|
|
|
|
np.random.seed(100)
|
|
|
|
@ -51,7 +54,7 @@ class CodeTableWithCustomTree(object):
|
|
|
|
|
|
|
|
|
|
def get_length(self):
|
|
|
|
|
length = 0
|
|
|
|
|
for ele in self.ptable_[self.index_]:
|
|
|
|
|
for ele in self.ptable_[self.index_]: # find the first -1 to stop trace
|
|
|
|
|
|
|
|
|
|
if ele >= 0:
|
|
|
|
|
length = length + 1
|
|
|
|
@ -71,12 +74,10 @@ def hsigmoid(x, w, label, bias, num_classes):
|
|
|
|
|
pre_sum = np.zeros((batch_size, 1))
|
|
|
|
|
out = np.zeros((batch_size, 1)).astype("float32")
|
|
|
|
|
for i in range(batch_size):
|
|
|
|
|
#print("\n leaf {leaf}: \n".format(leaf = label[i]))
|
|
|
|
|
code_table = CodeTable(num_classes, label[i])
|
|
|
|
|
length = code_table.get_length()
|
|
|
|
|
for j in range(length):
|
|
|
|
|
idx = code_table.cal_index(j)
|
|
|
|
|
#print("index {index} ".format(index = j))
|
|
|
|
|
pre_output[i][j] += bias[0][idx]
|
|
|
|
|
for i in range(batch_size):
|
|
|
|
|
code_table = CodeTable(num_classes, label[i])
|
|
|
|
@ -87,13 +88,12 @@ def hsigmoid(x, w, label, bias, num_classes):
|
|
|
|
|
# clip[-40.0, 40.0]
|
|
|
|
|
pre_output = np.clip(pre_output, -40.0, 40.0)
|
|
|
|
|
# out(i, 0) = \sum_j bit(i, j) * preout(i, j)
|
|
|
|
|
pre_output = -1 * pre_output
|
|
|
|
|
for i in range(batch_size):
|
|
|
|
|
#print("\n leaf {leaf}: \n".format(leaf = label[i]))
|
|
|
|
|
code_table = CodeTable(num_classes, label[i])
|
|
|
|
|
length = code_table.get_length()
|
|
|
|
|
sum = 0.0
|
|
|
|
|
for j in range(length):
|
|
|
|
|
#print("bit {bit} ".format(bit = code_table.cal_bit(j)))
|
|
|
|
|
if code_table.cal_bit(j):
|
|
|
|
|
sum += pre_output[i][j]
|
|
|
|
|
out[i] = -1.0 * sum
|
|
|
|
@ -108,6 +108,7 @@ def hsigmoidWithCustomTree(x, w, ptable, pcode, label, bias, num_classes):
|
|
|
|
|
batch_size = x.shape[0]
|
|
|
|
|
code_length = len(ptable[0])
|
|
|
|
|
code_table = [0 for _ in range(code_length)]
|
|
|
|
|
# init pre_out with shape [N, code_length]
|
|
|
|
|
pre_output = np.zeros((batch_size, code_length))
|
|
|
|
|
pre_sum = np.zeros((batch_size, 1))
|
|
|
|
|
out = np.zeros((batch_size, 1)).astype("float32")
|
|
|
|
@ -125,6 +126,7 @@ def hsigmoidWithCustomTree(x, w, ptable, pcode, label, bias, num_classes):
|
|
|
|
|
pre_output[i][j] += np.dot(w[idx], x[i])
|
|
|
|
|
# clip[-40.0, 40.0]
|
|
|
|
|
pre_output = np.clip(pre_output, -40.0, 40.0)
|
|
|
|
|
pre_output = -1 * pre_output
|
|
|
|
|
# out(i, 0) = \sum_j bit(i, j) * preout(i, j)
|
|
|
|
|
for i in range(batch_size):
|
|
|
|
|
code_table = CodeTableWithCustomTree(ptable, pcode, i)
|
|
|
|
@ -141,26 +143,27 @@ def hsigmoidWithCustomTree(x, w, ptable, pcode, label, bias, num_classes):
|
|
|
|
|
return pre_output, out
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
# class TestHSigmoidOp(OpTest):
|
|
|
|
|
# def setUp(self):
|
|
|
|
|
# self.op_type = "hierarchical_sigmoid"
|
|
|
|
|
# num_classes = 6
|
|
|
|
|
# feature_size = 8
|
|
|
|
|
# batch_size = 7
|
|
|
|
|
# x = np.random.random((batch_size, feature_size)).astype("float32")
|
|
|
|
|
# w = np.random.random((num_classes - 1, feature_size)).astype("float32")
|
|
|
|
|
# label = np.random.randint(0, num_classes, (batch_size, 1))
|
|
|
|
|
# bias = np.random.random((1, num_classes - 1)).astype("float32")
|
|
|
|
|
# self.attrs = {'num_classes': num_classes}
|
|
|
|
|
# self.inputs = {'X': x, 'W': w, 'Label': label, 'Bias': bias}
|
|
|
|
|
# pre_output, out = hsigmoid(x, w, label, bias, num_classes)
|
|
|
|
|
# self.outputs = {'PreOut': pre_output, 'Out': out}
|
|
|
|
|
class TestHSigmoidOp(OpTest):
|
|
|
|
|
def setUp(self):
|
|
|
|
|
self.op_type = "hierarchical_sigmoid"
|
|
|
|
|
num_classes = 6
|
|
|
|
|
feature_size = 8
|
|
|
|
|
batch_size = 4
|
|
|
|
|
x = np.random.random((batch_size, feature_size)).astype("float32") * 2
|
|
|
|
|
w = np.random.random(
|
|
|
|
|
(num_classes - 1, feature_size)).astype("float32") * 2
|
|
|
|
|
label = np.random.randint(0, num_classes, (batch_size, 1))
|
|
|
|
|
bias = np.random.random((1, num_classes - 1)).astype("float32")
|
|
|
|
|
self.attrs = {'num_classes': num_classes}
|
|
|
|
|
self.inputs = {'X': x, 'W': w, 'Label': label, 'Bias': bias}
|
|
|
|
|
pre_output, out = hsigmoid(x, w, label, bias, num_classes)
|
|
|
|
|
self.outputs = {'PreOut': pre_output, 'Out': out}
|
|
|
|
|
|
|
|
|
|
# def test_check_output(self):
|
|
|
|
|
# self.check_output()
|
|
|
|
|
def test_check_output(self):
|
|
|
|
|
self.check_output()
|
|
|
|
|
|
|
|
|
|
# def test_check_grad(self):
|
|
|
|
|
# self.check_grad(['Bias', 'X', 'W'], ['Out'], no_grad_set=set('Label'))
|
|
|
|
|
def test_check_grad(self):
|
|
|
|
|
self.check_grad(['Bias', 'X', 'W'], ['Out'], no_grad_set=set('Label'))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
class TestHSigmoidOpWithCostumTree(OpTest):
|
|
|
|
@ -169,9 +172,9 @@ class TestHSigmoidOpWithCostumTree(OpTest):
|
|
|
|
|
num_classes = 6 #using 1,2,3,4,5,6 to build a huffman tree and select 1,2,5,6 as sample
|
|
|
|
|
feature_size = 8
|
|
|
|
|
batch_size = 4
|
|
|
|
|
x = np.random.random((batch_size, feature_size)).astype("float32") * 10
|
|
|
|
|
x = np.random.random((batch_size, feature_size)).astype("float32") * 2
|
|
|
|
|
w = np.random.random(
|
|
|
|
|
(num_classes - 1, feature_size)).astype("float32") * 10
|
|
|
|
|
(num_classes - 1, feature_size)).astype("float32") * 2
|
|
|
|
|
label = np.array([0, 1, 4, 5])
|
|
|
|
|
ptable = np.array(
|
|
|
|
|
[(0, 2, -1, -1, -1), (0, 1, 3, -1, -1), (0, 1, 4, -1, -1),
|
|
|
|
|