|
|
@ -57,6 +57,9 @@ class AttentionHead(nn.Layer):
|
|
|
|
else:
|
|
|
|
else:
|
|
|
|
targets = paddle.zeros(shape=[batch_size], dtype="int32")
|
|
|
|
targets = paddle.zeros(shape=[batch_size], dtype="int32")
|
|
|
|
probs = None
|
|
|
|
probs = None
|
|
|
|
|
|
|
|
char_onehots = None
|
|
|
|
|
|
|
|
outputs = None
|
|
|
|
|
|
|
|
alpha = None
|
|
|
|
|
|
|
|
|
|
|
|
for i in range(num_steps):
|
|
|
|
for i in range(num_steps):
|
|
|
|
char_onehots = self._char_to_onehot(
|
|
|
|
char_onehots = self._char_to_onehot(
|
|
|
|