diff --git a/python/paddle/nn/functional/activation.py b/python/paddle/nn/functional/activation.py index bd975106bd..6acb806403 100644 --- a/python/paddle/nn/functional/activation.py +++ b/python/paddle/nn/functional/activation.py @@ -665,7 +665,7 @@ def selu(x, paddle.disable_static() - x = paddle.to_tensor(np.array([[0, 1],[2, 3]])) + x = paddle.to_tensor(np.array([[0.0, 1.0],[2.0, 3.0]])) out = F.selu(x) # [[0, 1.050701],[2.101402, 3.152103]] """ if in_dygraph_mode(): diff --git a/python/paddle/nn/layer/activation.py b/python/paddle/nn/layer/activation.py index 6373b05883..bb0bd5f70f 100644 --- a/python/paddle/nn/layer/activation.py +++ b/python/paddle/nn/layer/activation.py @@ -570,7 +570,7 @@ class SELU(layers.Layer): paddle.disable_static() - x = paddle.to_tensor(np.array([[0, 1],[2, 3]])) + x = paddle.to_tensor(np.array([[0.0, 1.0],[2.0, 3.0]])) m = paddle.nn.SELU() out = m(x) # [[0, 1.050701],[2.101402, 3.152103]] """