Commit Graph

32 Commits (17030ff28b9a54bb57779e9b8448a6d222110ec5)

Author SHA1 Message Date
lilong12 c3974d0e2a
[3D-parallel] Reformat pipeline parallel (#31786)
4 years ago
arlesniak 7ccf6b6030
[oneDNN] Initial bf16 amp integration (#31093)
4 years ago
huangxu96 138620084c
Add fleet amp_init() (#30572)
4 years ago
huangxu96 342d62de60
add amp example document (#30314)
4 years ago
Zhen Wang 7f7dfccf20
Support pure fp16 training for AMP API. (#29544)
4 years ago
WangXi ab04997846
[fleet] combine amp and gradient merge, test=develop (#30086)
4 years ago
huangxu96 c05170d3d8
add alias for fluid.contrib.mixed_precision (#29562)
4 years ago
Aurelius84 5d530c9319
fix amp support fleet (#29491)
4 years ago
Zhen Wang be3777a50a
Add pure fp16 training with master weights. (#27712)
4 years ago
furnace 7584bb5096
Layer norm fp16 (#29169)
4 years ago
WangXi 0c2a51d240
optimizer amp, all use fp16 communication, overlap last comm and compute (#28957)
4 years ago
Leo Chen 11e32baf1e
Add matmtl_v2 to amp list (#28693)
4 years ago
Leo Chen 71d6220772
Skip reader op in mixed_precision decorator (#28353)
4 years ago
WangXi 0a1862d1d2
fleet combine amp dgc recompute meta optimizer (#27643)
4 years ago
Zhang Ting 906e7f921e
add fuse_bn_act op (#27230)
4 years ago
Zhen Wang d708b21074
Update amp_check_finite_and_scale_op and add an updating_loss_scaling op for static graph amp training. (#26240)
4 years ago
Zhen Wang bcdbac1753
fix some cast error. (#26884)
5 years ago
mapingshuo f0e743f136
fix AMP and recompute (#23551)
5 years ago
gongweibao 5e07db15e6
fix init scaling value test=develop (#22145)
5 years ago
Zhen Wang be2e3e67d9
Fix some typos in AMP. (#21354)
5 years ago
gongweibao 3255fe69bb Add custom black variable name set in amp interface. (#20875)
5 years ago
gongweibao 1d82025e89
Add interface so user can get scaled loss when they use customized loss. (#20571)
5 years ago
gongweibao 7b9e33972a
delete backward return list test=develop (#20294)
5 years ago
Jie Fang d9db94d752 Optimize amp for multi-gpu to enable FP16 gradients transfer across gpus. (#19714)
5 years ago
gongweibao 6c2bc29cc0
Fix float16 optimizer. (#19682)
6 years ago
Jie Fang c6a598a276 init new amp, optimize inserting cast op for batchnorm (#18596)
6 years ago
gongweibao abaf87be2b
Change backward_guard to optimize_guard to maximize the allreduce overlap. (#19506)
6 years ago
Zeng Jinle 5dce1da680 remove reset recordio usage (#19519)
6 years ago
Jie Fang 2b4ef509ea init custom black white list (#18377)
6 years ago
Jie Fang 172c2facef init black/white lists (#17847)
6 years ago
Jie Fang 30e178fa2c init auto loss scaling (#17194)
6 years ago
Yibing Liu beda78258f
Init mixed precision training interface (#16856)
6 years ago