Commit Graph

26 Commits (824a79d383531e804e7274ef2141c30c7532e2c2)

Author SHA1 Message Date
tangwei12 25f80fd304
Fix/distributed proto (#29981)
4 years ago
tangwei12 032414ca2a
[Feature] one ps (3/4) (#29604)
4 years ago
zhang wenhui 5a83496c8d
Multi task (#26002)
4 years ago
Chen Weihang c34c80d302
Polish framework error message part3 (#25701)
5 years ago
123malin 9d2bd0ac38
downpour_worker增加try_catch机制,打印program所有参数 (#24700)
5 years ago
hutuxian 0ec3a42e97
Random Dump (#24477)
5 years ago
xujiaqi01 93ea9dd27a
fix stat var in hogwild worker (#23367)
5 years ago
123malin 00594c1c88
support dumping params/grads in transpiler mode (#22490)
5 years ago
tangwei12 82bc814a57
integrated HALF_ASYNC to communicator (#21869)
5 years ago
yaoxuefeng 10ca3f9609
add thread scope stat accurate metrics test=develop (#19480)
6 years ago
jiaqi d18aabb472
support patch data, add load_one_table, fix bug (#18509)
6 years ago
guru4elephant 7f8bc49d00
polish_executor_and_add_ctx_cache (#17536)
6 years ago
jiaqi 66d51206b1
add save/load model, shrink table, cvm, config file & fix pull dense bug (#17118)
6 years ago
dongdaxiang d739bab844 fix async_executor problem and remove some unnecessary testcase, fix trainer_desc import problem
6 years ago
dongdaxiang 60b7bf6fa6 add infer_from_dataset for inference
6 years ago
dongdaxiang 6bf796df14 refine print fetch list
6 years ago
dongdaxiang 68d7bf3de5 add fetch var function
6 years ago
dongdaxiang 73b1f396d7 add data_generator into paddle.fluid.incubate.data_generator, add op run log in hogwild_device_worker and downpour_device_worker
6 years ago
dongdaxiang 73544e8b8d add training speed log
6 years ago
dongdaxiang 9419de521f add IO percent for multi_trainer
6 years ago
dongdaxiang 6af697adb0 add trainfileswithprofiler for downpour worker
6 years ago
dongdaxiang b66f0074b6 fix data reading bugs in api, add VLOG(3) log for setup
6 years ago
dongdaxiang cf1360643f add printer for fetch variable
6 years ago
dongdaxiang 39014b9f9f fix class register problem
6 years ago
dongdaxiang c165012031 refine device_worker and trainer code
6 years ago
dongdaxiang 855bf579d2 add dist_multi_trainer for distributed training, add trainer_factory and device_worker_factory so that we can easily extend new training mode, add pull dense worker which is a singleton for parameter fetching
6 years ago