|
|
|
@ -72,12 +72,12 @@ class ParallelExecutor(unittest.TestCase):
|
|
|
|
|
first_loss = numpy.array(fluid.global_scope().find_var('fetched_var')
|
|
|
|
|
.get_lod_tensor_array()[0])
|
|
|
|
|
print first_loss
|
|
|
|
|
#
|
|
|
|
|
# for i in xrange(10):
|
|
|
|
|
# exe.run([], 'fetched_var')
|
|
|
|
|
# exe.run([loss.name], 'fetched_var')
|
|
|
|
|
# last_loss = numpy.array(fluid.global_scope().find_var('fetched_var')
|
|
|
|
|
# .get_lod_tensor_array()[0])
|
|
|
|
|
#
|
|
|
|
|
# print first_loss, last_loss
|
|
|
|
|
# self.assertGreater(first_loss[0], last_loss[0])
|
|
|
|
|
|
|
|
|
|
for i in xrange(10):
|
|
|
|
|
exe.run([], 'fetched_var')
|
|
|
|
|
exe.run([loss.name], 'fetched_var')
|
|
|
|
|
last_loss = numpy.array(fluid.global_scope().find_var('fetched_var')
|
|
|
|
|
.get_lod_tensor_array()[0])
|
|
|
|
|
|
|
|
|
|
print first_loss, last_loss
|
|
|
|
|
self.assertGreater(first_loss[0], last_loss[0])
|
|
|
|
|