|
|
@ -1310,9 +1310,10 @@ class BatchNorm(layers.Layer):
|
|
|
|
self._fuse_with_relu, "use_global_stats",
|
|
|
|
self._fuse_with_relu, "use_global_stats",
|
|
|
|
self._use_global_stats, 'trainable_statistics',
|
|
|
|
self._use_global_stats, 'trainable_statistics',
|
|
|
|
self._trainable_statistics)
|
|
|
|
self._trainable_statistics)
|
|
|
|
batch_norm_out, _, _, _, _ = core.ops.batch_norm(
|
|
|
|
batch_norm_out, _, _, _, _, _ = core.ops.batch_norm(
|
|
|
|
input, self.weight, self.bias, self._mean, self._variance,
|
|
|
|
input, self.weight, self.bias, self._mean, self._variance,
|
|
|
|
mean_out, variance_out, *attrs)
|
|
|
|
mean_out, variance_out, *attrs)
|
|
|
|
|
|
|
|
|
|
|
|
return dygraph_utils._append_activation_in_dygraph(
|
|
|
|
return dygraph_utils._append_activation_in_dygraph(
|
|
|
|
batch_norm_out, act=self._act)
|
|
|
|
batch_norm_out, act=self._act)
|
|
|
|
|
|
|
|
|
|
|
|