Commit Graph

43 Commits (bd8dfe38caef4d4769fe2a57263fbe90a6891592)

Author SHA1 Message Date
Thunderbrook 0cb60c700d
add heter ps mode (#25682)
5 years ago
xujiaqi01 3448455b71
fix dataset py3 (#25012)
5 years ago
hutuxian 0ec3a42e97
Random Dump (#24477)
5 years ago
hutuxian 65f495c9fa
remove print statement to make code clean(#23907)
5 years ago
hutuxian 6f7077db25
Add logs to debug random fail (#23731)
5 years ago
xujiaqi01 3a45767d49
add fleet pslib pull and push sparse op and push dense op (#23139)
5 years ago
yaoxuefeng 2235ee1a5e
multi-loss optimization by adding a DownpourOpt worker (#22025)
5 years ago
tangwei12 82bc814a57
integrated HALF_ASYNC to communicator (#21869)
5 years ago
Thunderbrook 349e82d669
support general embedding params (#21217)
5 years ago
xujiaqi01 9e045170c0
add copy table (#21086)
5 years ago
Thunderbrook 59bcdc8a19
support dump param of model into afs (#20302)
5 years ago
xujiaqi01 5223b0dd9d
add check nan / inf in downpour worker (#20694)
5 years ago
Thunderbrook f76a32df4a
dump fix dov vec file num (#20539)
5 years ago
Thunderbrook 1fe468d319
support debug each output of each ins (#19004)
6 years ago
jiaqi 768059b3a0
adjust ins weight according to nid slot (#18784)
6 years ago
fuyinno4 c167a4b4dd
Fix shrink-dense and add scale-datanorm (#18746)
6 years ago
Thunderbrook d8396281ef
add slot to sparse table (#18686)
6 years ago
jiaqi d18aabb472
support patch data, add load_one_table, fix bug (#18509)
6 years ago
hutuxian cf15c3ff1e
fix errors in python3 (#18239)
6 years ago
hutuxian 969e6378b9
Pipeline Concurrency (#17402)
6 years ago
jiaqi 66d51206b1
add save/load model, shrink table, cvm, config file & fix pull dense bug (#17118)
6 years ago
dongdaxiang 8257136012 fix bug
6 years ago
dongdaxiang ceac9df87a fix code style for incubator
6 years ago
dongdaxiang 93c3c7f9b3 fix dataset testcase problem
6 years ago
dongdaxiang d739bab844 fix async_executor problem and remove some unnecessary testcase, fix trainer_desc import problem
6 years ago
dongdaxiang 60b7bf6fa6 add infer_from_dataset for inference
6 years ago
dongdaxiang 88880d9b69 fix import trainer_desc_pb2 error
6 years ago
dongdaxiang 0030eb2a61 fix distributed building
6 years ago
dongdaxiang 8e14d8f900 add data_generator package into setup.py
6 years ago
dongdaxiang d87ba58c14 refine document of python API, make device_worker and trainer's API private
6 years ago
dongdaxiang b95b80bc76 add doc string for executor and update API.spec
6 years ago
dongdaxiang 6bf796df14 refine print fetch list
6 years ago
dongdaxiang 68d7bf3de5 add fetch var function
6 years ago
dongdaxiang 6af697adb0 add trainfileswithprofiler for downpour worker
6 years ago
dongdaxiang b7a202aa38 add distributed optimizer factory
6 years ago
dongdaxiang e657c127a8 hide opt_info in distirbuted optimizer
6 years ago
xujiaqi01 ecfc7df913 add dataset factory && fix style
6 years ago
dongdaxiang ff87698a44 refactor downpour optimization
6 years ago
heqiaozhi 9bca1926c1 refactor & fix bug
6 years ago
dongdaxiang 6de9ebc65c refine VLOG in fleet_wrapper.h
6 years ago
dongdaxiang c165012031 refine device_worker and trainer code
6 years ago
dongdaxiang 8a335b50be add downpour device_worker pb configuration
6 years ago
dongdaxiang caf0c10e71 add dist_multi_trainer for distributed training, add trainer_factory and device_worker_factory so that we can easily extend new training mode, add pull dense worker which is a singleton for parameter fetching
6 years ago