change API name

test=develop
move-code
sneaxiy 7 years ago
parent 161b8ddcaa
commit 3a09693f5c

@ -61,10 +61,10 @@ paddle.fluid.io.load_params (ArgSpec(args=['executor', 'dirname', 'main_program'
paddle.fluid.io.load_persistables (ArgSpec(args=['executor', 'dirname', 'main_program', 'filename'], varargs=None, keywords=None, defaults=(None, None)), ('document', '28df5bfe26ca7a077f91156abb0fe6d2')) paddle.fluid.io.load_persistables (ArgSpec(args=['executor', 'dirname', 'main_program', 'filename'], varargs=None, keywords=None, defaults=(None, None)), ('document', '28df5bfe26ca7a077f91156abb0fe6d2'))
paddle.fluid.io.save_inference_model (ArgSpec(args=['dirname', 'feeded_var_names', 'target_vars', 'executor', 'main_program', 'model_filename', 'params_filename', 'export_for_deployment'], varargs=None, keywords=None, defaults=(None, None, None, True)), ('document', '582d87b8df75a5a639a107db8ff86f9c')) paddle.fluid.io.save_inference_model (ArgSpec(args=['dirname', 'feeded_var_names', 'target_vars', 'executor', 'main_program', 'model_filename', 'params_filename', 'export_for_deployment'], varargs=None, keywords=None, defaults=(None, None, None, True)), ('document', '582d87b8df75a5a639a107db8ff86f9c'))
paddle.fluid.io.load_inference_model (ArgSpec(args=['dirname', 'executor', 'model_filename', 'params_filename', 'pserver_endpoints'], varargs=None, keywords=None, defaults=(None, None, None)), ('document', '7a5255386075dac3c75b7058254fcdcb')) paddle.fluid.io.load_inference_model (ArgSpec(args=['dirname', 'executor', 'model_filename', 'params_filename', 'pserver_endpoints'], varargs=None, keywords=None, defaults=(None, None, None)), ('document', '7a5255386075dac3c75b7058254fcdcb'))
paddle.fluid.io.PyReader.__init__ (ArgSpec(args=['self', 'feed_list', 'capacity', 'use_double_buffer', 'iterable'], varargs=None, keywords=None, defaults=(True, False)), ('document', '18211b287474b401bc460d3f73dbc1c7')) paddle.fluid.io.PyReader.__init__ (ArgSpec(args=['self', 'feed_list', 'capacity', 'use_double_buffer', 'iterable'], varargs=None, keywords=None, defaults=(True, False)), ('document', 'b3d72958b2568aae3f90f72abdcb7d1a'))
paddle.fluid.io.PyReader.decorate_paddle_reader (ArgSpec(args=['self', 'reader', 'places'], varargs=None, keywords=None, defaults=(None,)), ('document', 'faef298f73e91aedcfaf5d184f3109b7')) paddle.fluid.io.PyReader.decorate_batch_generator (ArgSpec(args=['self', 'reader', 'places'], varargs=None, keywords=None, defaults=(None,)), ('document', 'd10224fef1095247063b6976da793021'))
paddle.fluid.io.PyReader.decorate_sample_generator (ArgSpec(args=['self', 'sample_generator', 'batch_size', 'drop_last', 'places'], varargs=None, keywords=None, defaults=(True, None)), ('document', 'd3fe49fc342e7778ed086e965f41bf12')) paddle.fluid.io.PyReader.decorate_sample_generator (ArgSpec(args=['self', 'sample_generator', 'batch_size', 'drop_last', 'places'], varargs=None, keywords=None, defaults=(True, None)), ('document', '7abd9cf7d695bab5bb6cf7ded5903cb2'))
paddle.fluid.io.PyReader.decorate_tensor_provider (ArgSpec(args=['self', 'reader', 'places'], varargs=None, keywords=None, defaults=(None,)), ('document', 'd10224fef1095247063b6976da793021')) paddle.fluid.io.PyReader.decorate_sample_list_generator (ArgSpec(args=['self', 'reader', 'places'], varargs=None, keywords=None, defaults=(None,)), ('document', 'faef298f73e91aedcfaf5d184f3109b7'))
paddle.fluid.io.PyReader.reset (ArgSpec(args=['self'], varargs=None, keywords=None, defaults=None), ('document', 'ff1cc1e2beb8824d453656c72c28ddfb')) paddle.fluid.io.PyReader.reset (ArgSpec(args=['self'], varargs=None, keywords=None, defaults=None), ('document', 'ff1cc1e2beb8824d453656c72c28ddfb'))
paddle.fluid.io.PyReader.start (ArgSpec(args=['self'], varargs=None, keywords=None, defaults=None), ('document', 'b7ea0a548991924e4cfe61a577b8e56d')) paddle.fluid.io.PyReader.start (ArgSpec(args=['self'], varargs=None, keywords=None, defaults=None), ('document', 'b7ea0a548991924e4cfe61a577b8e56d'))
paddle.fluid.initializer.ConstantInitializer.__init__ (ArgSpec(args=['self', 'value', 'force_cpu'], varargs=None, keywords=None, defaults=(0.0, False)), ('document', '6adf97f83acf6453d4a6a4b1070f3754')) paddle.fluid.initializer.ConstantInitializer.__init__ (ArgSpec(args=['self', 'value', 'force_cpu'], varargs=None, keywords=None, defaults=(0.0, False)), ('document', '6adf97f83acf6453d4a6a4b1070f3754'))

@ -127,7 +127,9 @@ class CompiledProgram(object):
program in the given places. Otherwise, the places used when compiled program in the given places. Otherwise, the places used when compiled
is determined by the Executor, and the places used are controlled is determined by the Executor, and the places used are controlled
by environment variables: FLAGS_selected_gpus or CUDA_VISIBLE_DEVICES by environment variables: FLAGS_selected_gpus or CUDA_VISIBLE_DEVICES
if using GPU; or CPU_NUM if using CPU. if using GPU; or CPU_NUM if using CPU. For example, if you want to
run on GPU 0 and 1, set places=[fluid.CUDAPlace(0), fluid.CUDAPlace(1)].
If you want to run on 2 CPU cores, set places=[fluid.CPUPlace()]*2.
Returns: Returns:
self self

@ -632,6 +632,9 @@ def _py_reader(capacity,
reader.reset = __reset__ reader.reset = __reset__
reader.decorate_tensor_provider = __set_tensor_provider__ reader.decorate_tensor_provider = __set_tensor_provider__
reader.decorate_paddle_reader = __set_paddle_reader__ reader.decorate_paddle_reader = __set_paddle_reader__
reader.decorate_batch_generator = __set_tensor_provider__
reader.decorate_sample_list_generator = __set_paddle_reader__
reader.start = __start__ reader.start = __start__
return reader return reader

@ -82,7 +82,7 @@ class PyReader(object):
reader = fluid.io.PyReader(feed_list=[image, label], reader = fluid.io.PyReader(feed_list=[image, label],
capacity=4, iterable=False) capacity=4, iterable=False)
reader.decorate_paddle_reader(user_defined_reader) reader.decorate_sample_list_generator(user_defined_reader)
... # definition of network is omitted ... # definition of network is omitted
executor.run(fluid.default_main_program()) executor.run(fluid.default_main_program())
for _ in range(EPOCH_NUM): for _ in range(EPOCH_NUM):
@ -109,7 +109,7 @@ class PyReader(object):
reader = fluid.io.PyReader(feed_list=[image, label], reader = fluid.io.PyReader(feed_list=[image, label],
capacity=4, iterable=True) capacity=4, iterable=True)
reader.decorate_paddle_reader(user_defined_reader, reader.decorate_sample_list_generator(user_defined_reader,
places=fluid.cuda_places()) places=fluid.cuda_places())
... # definition of network is omitted ... # definition of network is omitted
executor.run(fluid.default_main_program()) executor.run(fluid.default_main_program())
@ -287,7 +287,7 @@ class PyReader(object):
:code:`places` must be set when the PyReader object is iterable. :code:`places` must be set when the PyReader object is iterable.
If all inputs have no lods, this method is faster than If all inputs have no lods, this method is faster than
:code:`decorate_paddle_reader(paddle.batch(sample_generator, ...))` . :code:`decorate_sample_list_generator(paddle.batch(sample_generator, ...))` .
Args: Args:
sample_generator (generator): Python generator that yields sample_generator (generator): Python generator that yields
@ -306,7 +306,7 @@ class PyReader(object):
break break
if has_lod: if has_lod:
self.decorate_paddle_reader( self.decorate_sample_list_generator(
paddle.batch( paddle.batch(
sample_generator, sample_generator,
batch_size=batch_size, batch_size=batch_size,
@ -319,9 +319,9 @@ class PyReader(object):
batch_size=batch_size, batch_size=batch_size,
generator=sample_generator, generator=sample_generator,
drop_last=drop_last) drop_last=drop_last)
self.decorate_tensor_provider(reader, places=places) self.decorate_batch_generator(reader, places=places)
def decorate_paddle_reader(self, reader, places=None): def decorate_sample_list_generator(self, reader, places=None):
''' '''
Set the data source of the PyReader object. Set the data source of the PyReader object.
@ -347,9 +347,9 @@ class PyReader(object):
for slots in paddle_reader(): for slots in paddle_reader():
yield [slots[var.name] for var in self._feed_list] yield [slots[var.name] for var in self._feed_list]
self.decorate_tensor_provider(__tensor_reader_impl__, places) self.decorate_batch_generator(__tensor_reader_impl__, places)
def decorate_tensor_provider(self, reader, places=None): def decorate_batch_generator(self, reader, places=None):
''' '''
Set the data source of the PyReader object. Set the data source of the PyReader object.

@ -25,6 +25,7 @@ CLASS_NUM = 10
def random_reader(): def random_reader():
np.random.seed(1)
for i in range(BATCH_SIZE * 40): for i in range(BATCH_SIZE * 40):
image = np.random.random([784]) image = np.random.random([784])
label = np.random.random_integers(low=0, high=CLASS_NUM - 1) label = np.random.random_integers(low=0, high=CLASS_NUM - 1)
@ -79,7 +80,8 @@ class TestBase(unittest.TestCase):
reader = paddle.batch(random_reader, batch_size=BATCH_SIZE) reader = paddle.batch(random_reader, batch_size=BATCH_SIZE)
ps = places if use_double_buffer else fluid.cpu_places(len(places)) ps = places if use_double_buffer else fluid.cpu_places(len(places))
py_reader.decorate_paddle_reader(
py_reader.decorate_sample_list_generator(
reader, places=ps if py_reader.iterable else None) reader, places=ps if py_reader.iterable else None)
exe = fluid.Executor(place=places[0]) exe = fluid.Executor(place=places[0])
@ -92,6 +94,7 @@ class TestBase(unittest.TestCase):
step = 0 step = 0
step_list = [] step_list = []
loss_list = []
start_t = time.time() start_t = time.time()
if not py_reader.iterable: if not py_reader.iterable:
for _ in six.moves.range(EPOCH_NUM): for _ in six.moves.range(EPOCH_NUM):
@ -102,6 +105,7 @@ class TestBase(unittest.TestCase):
L, = exe.run(program=prog, L, = exe.run(program=prog,
fetch_list=[loss], fetch_list=[loss],
use_program_cache=True) use_program_cache=True)
loss_list.append(np.mean(L))
step += 1 step += 1
except fluid.core.EOFException: except fluid.core.EOFException:
py_reader.reset() py_reader.reset()
@ -123,10 +127,15 @@ class TestBase(unittest.TestCase):
feed=d, feed=d,
fetch_list=[loss], fetch_list=[loss],
use_program_cache=True) use_program_cache=True)
loss_list.append(np.mean(L))
step += 1 step += 1
step_list.append(step) step_list.append(step)
end_t = time.time() end_t = time.time()
ret = {"time": end_t - start_t, "step": step_list} ret = {
"time": end_t - start_t,
"step": step_list,
"loss": np.array(loss_list)
}
return ret return ret
def prepare_places(self, with_data_parallel, with_cpu=True, with_gpu=True): def prepare_places(self, with_data_parallel, with_cpu=True, with_gpu=True):
@ -148,12 +157,18 @@ class TestBase(unittest.TestCase):
for with_data_parallel in [True, False]: for with_data_parallel in [True, False]:
for p in self.prepare_places(with_data_parallel): for p in self.prepare_places(with_data_parallel):
for use_double_buffer in [False, True]: for use_double_buffer in [False, True]:
results = []
for use_legacy_py_reader in [False, True]: for use_legacy_py_reader in [False, True]:
ret = self.run_main( ret = self.run_main(
use_legacy_py_reader=use_legacy_py_reader, use_legacy_py_reader=use_legacy_py_reader,
with_data_parallel=with_data_parallel, with_data_parallel=with_data_parallel,
places=p, places=p,
use_double_buffer=use_double_buffer) use_double_buffer=use_double_buffer)
results.append(ret)
if not use_double_buffer:
diff = np.max(
np.abs(results[0]['loss'] - results[1]['loss']))
self.assertLess(diff, 1e-3)
if __name__ == '__main__': if __name__ == '__main__':

@ -62,7 +62,7 @@ class TestCaseBase(unittest.TestCase):
all_datas = self.generate_all_data(batch_reader) all_datas = self.generate_all_data(batch_reader)
if not use_sample_generator: if not use_sample_generator:
py_reader.decorate_paddle_reader( py_reader.decorate_sample_list_generator(
batch_reader, places=fluid.cpu_places()) batch_reader, places=fluid.cpu_places())
else: else:
py_reader.decorate_sample_generator( py_reader.decorate_sample_generator(

Loading…
Cancel
Save