Commit Graph

26 Commits (1ab4101d6cc11e27c5ba35f2d5a3f50983435e91)

Author SHA1 Message Date
hutuxian 0ec3a42e97
Random Dump (#24477)
5 years ago
xujiaqi01 3a45767d49
add fleet pslib pull and push sparse op and push dense op (#23139)
5 years ago
xujiaqi01 371f377bea
add GeneralRoleMaker (#22295)
5 years ago
xujiaqi01 0eb4d990c4
fix compiled error when with_pslib=on (#21769)
5 years ago
Tao Luo c0656dcb1a
remove -Wno-error=sign-compare, make warning as error (#21358)
5 years ago
Chengmo 940c6ff1c8
Fix communicator slow bug & fix communicator stop bug (#20366)
5 years ago
Thunderbrook f76a32df4a
dump fix dov vec file num (#20539)
5 years ago
xujiaqi01 cedc04775c
support change shuffle and train thread num (#19841)
5 years ago
yaoxuefeng 10ca3f9609
add thread scope stat accurate metrics test=develop (#19480)
6 years ago
Thunderbrook 1fe468d319
support debug each output of each ins (#19004)
6 years ago
jiaqi 3f8031e256
dataset (#17973)
6 years ago
dongdaxiang ba15d6b164 move root_scope->DropKids() into Finalize() so that we do not have to drop all the kids
6 years ago
xujiaqi01 a5b1a0e12b support multi dataset && add init model && fix bug
6 years ago
dongdaxiang 6af697adb0 add trainfileswithprofiler for downpour worker
6 years ago
dongdaxiang 2644b88685 add comment for MPI Symetric role maker
6 years ago
dongdaxiang cf45c54340 add distributed optimizer factory
6 years ago
xujiaqi01 39449ba0b9 fix bug && add DestroyReaders in trainer
6 years ago
dongdaxiang 328f11b8b6 refactor downpour optimization
6 years ago
dongdaxiang b66f0074b6 fix data reading bugs in api, add VLOG(3) log for setup
6 years ago
dongdaxiang b415ec27e8 make Dataset* as an argument
6 years ago
xjqbest dd67ad08a2 modify c++ and python dataset related code & fix bug
6 years ago
dongdaxiang 2486389793 add RunFromDataset in executor
6 years ago
xjqbest 824b84d185 add DataSet and InMemoryDataFeed, support load data into memory and shuffle data
6 years ago
dongdaxiang 39014b9f9f fix class register problem
6 years ago
dongdaxiang c165012031 refine device_worker and trainer code
6 years ago
dongdaxiang 855bf579d2 add dist_multi_trainer for distributed training, add trainer_factory and device_worker_factory so that we can easily extend new training mode, add pull dense worker which is a singleton for parameter fetching
6 years ago