Commit Graph

259 Commits (16ba0abc79c6d6f5c642d99638ee0f77d8a45a1e)

Author SHA1 Message Date
JZ-LIANG 16ba0abc79
Recompute Offload: fixed bug in memcpy (#30484)
5 years ago
guofei 11e78ebaa3
Modify the calculation logic of LambOptimizer (#29313)
5 years ago
JZ-LIANG 75936d838f
Recompute Offload (#30233)
5 years ago
WangXi ee16006b5d
Optimization grad merge performance (#29784)
5 years ago
Leo Chen a4b9daf97c
fix optimizer dtype (#29917)
5 years ago
lilong12 01950ceb42
fix the bug in pipeline data parallelism (#29731)
5 years ago
Aurelius84 5d530c9319
fix amp support fleet (#29491)
5 years ago
Zhou Wei c0a991c874
accumulate gradient for leaf tensor with previous graph and expose leaf tensor concept (#28429)
5 years ago
ShenLiang e2d01eb650
Support dynamic graph distributed (#28997)
5 years ago
Leo Chen 3815d7aa40
Upgrade string literals to raw string (#28989)
5 years ago
lilong12 f77a78cdee
enable pipeline to run with Executor.run() (#28373)
5 years ago
Zhou Wei e122e16456
fix english doc, unittest, and remove useless alias of 2.0 lr_scheduler (#27686)
5 years ago
WangXi 0a1862d1d2
fleet combine amp dgc recompute meta optimizer (#27643)
5 years ago
123malin 54c368db1e
[API 2.0: doc] fix doc of nonzero (#27685)
5 years ago
Chen Weihang dec53a9c79
Remove DataParallel.scale_loss & apply_collective_grads (#27603)
5 years ago
Chen Weihang bb84f0e646
Add new paddle.save/load APIs (#27331)
5 years ago
Yibing Liu 34091533c9
Fix bug in continuous apply, test=develop (#27337)
5 years ago
JZ-LIANG 5d039f4086
modified the implement of Lars optimizer (#26733)
5 years ago
Chen Weihang 9cb57f94c6
Update set_dict method name & add aliases (#26700)
5 years ago
Yang Zhang 6129b0e246
Revert `no_grad` changes and add new implementation (#26826)
5 years ago
Zhou Wei 407de03905
[2.0API] Reconstruct all API related to LR Scheduler, unify dygraph and static (#26550)
5 years ago
MRXLT eeda90d674
[WIP] update optimizer for 2.0 (#26288)
5 years ago
mapingshuo 7ae10900fa
fix slow var initialize, test=develop (#26516)
5 years ago
Dong Daxiang cbf8ba1591
add check approval (#26284)
5 years ago
Yang Zhang 617eb67f29
Upgrade `no_grad` decorator (#25472)
5 years ago
WangXi 2c9d0f3cb9
【paddle.fleet】Add dgc to fleet meta optimizer (#25738)
5 years ago
lilong12 8a68d2c213
Revert "add device attr for regularizer, test=develop (#24981)" (#25375)
5 years ago
mapingshuo c70f592002
add gradient Merge optimizer to meta (#25763)
5 years ago
tangwei12 caa90a6510
Integrated Trainer of Parameter Server (API add `fluid.contrib.layers.sparse_embedding` only) (#22957)
5 years ago
mapingshuo 3e2a348886
add string variable support for RecomputeOptimizer (#25728)
5 years ago
mapingshuo ed72406558
add gradient Merge Optimizer (#25625)
5 years ago
mapingshuo ea60e64470
correct the LookaheadOptimizer programDesc, test=develop (#25688)
5 years ago
leesusu 856e6d3348
Correct parameter l2 passed to ftrl op (#25223)
5 years ago
Zhou Wei 914ff10a8f
fix state dict to save/load learning rate scheduler (#25403)
5 years ago
hong fed0588571
Fix parameter list iterator bug (#25089)
5 years ago
lilong12 e39aa70ec7
add the support for pipeline (#24560)
5 years ago
lilong12 3d96601b82
modify pipeline optimizer to only support the mode of sync pipeline training (#25065)
5 years ago
Zhou Wei c505c4dbea
add new API: optimizer.set_lr (#24455)
5 years ago
lilong12 ab5a1fb853
add device attr for regularizer, test=develop (#24981)
5 years ago
Zhou Wei 98da8a295d
add new learing rate strategy to reduce lr when loss reach on plateau (#24322)
5 years ago
swtkiwi f5c6dd6def
test=develop (#24522)
6 years ago
hong 04e9d721a2
unitize name in optimizer; test=develop (#24008)
6 years ago
Zhou Wei 8002b2beb4
Avoid logging.info be printed many times in dygraph_mode,test=develop (#23932)
6 years ago
Zhou Wei 66dc8e30f0
move the initialize position of grad_clip to optimizer(__init__),and speed up clip (#23782)
6 years ago
mapingshuo f0e743f136
fix AMP and recompute (#23551)
6 years ago
Zhou Wei 629b6c7896
add the prompt message of repeated settings of regularization,test=develop (#23355)
6 years ago
qingqing01 6162cf2f2e
Make optimizer consistent in dygraph and static-graph and remove some LOG-INFO. (#23426)
6 years ago
Zhou Wei e8efaee92d
update gradient clip english doc for new gradient clipping strategy
6 years ago
Leo Chen a62599a888
[feature] prune program by feed and fetch_list automatically (#22474)
6 years ago
Zhou Wei 7fda333ac1
add new method of gradient_clip, better to use,test=develop (#23224)
6 years ago