Commit Graph

379 Commits (adaec0073d02c0ea55bcabc4671ebfc8dbd3182c)

Author SHA1 Message Date
cc 88e9d74a75
Collecting concat output threshold, test=develop (#24742)
6 years ago
ShenLiang 950892044f
fix conflict, test=develop (#24238)
6 years ago
cc 6c89ca2157
Add output threshold for ops that have several output activations, test=develop (#24726)
6 years ago
lidanqing 8ef3c02e90
Update DNNL QAT document 2.0-alpha (#24494)
6 years ago
cc 4d35112255
[Fix bug] Init scale node in OutScaleForTrainingPass and enable test_quantization_scale_pass UT (#24393)
6 years ago
joanna.wozna.intel 53125c2f6f
Model converter to dot file (#23169)
6 years ago
Wojciech Uss db052009c7
Enabled quantize all and skip missing in QAT (#24281)
6 years ago
Leo Chen 381492fca3
add try finally, test=develop (#24243)
6 years ago
lidanqing 61ec30f030
Update QAT INT8 2.0 doc (#24127)
6 years ago
Sylwester Fraczek e1a7a88057
added reshape transpose matmul fuse pass (#23754)
6 years ago
ShenLiang 0fb9b208ab
Add batch_fc op in contrib (#24017)
6 years ago
arlesniak d31a174f51
added fusing matmul-transpose-reshape pass (#23866)
6 years ago
Wojciech Uss 3d744162dd
QAT: support for new models (#23928)
6 years ago
zhangchunle 6bd200db66
remove high level api (#23854)
6 years ago
ShenLiang 30bd7e1c83
Add rank_attention_op attributes for GPU memory in contrib (#23915)
6 years ago
cc 40aa14ec77
Weight quantization support channel_wise_abs_max method to achieve higher accuracy (#23629)
6 years ago
mapingshuo f0e743f136
fix AMP and recompute (#23551)
6 years ago
joanna.wozna.intel 12ba05ce0c
Add scale-matmul fuse pass (#23734)
6 years ago
Wojciech Uss 2383a9f7ee
[Doc update] Update for QAT INT8 MKL-DNN document (#23361)
6 years ago
Chengmo 8c0bdde934
Add Tdm sampler op in Contrib (#23290)
6 years ago
Wojciech Uss 1753860dd0
Enable matmul and cleanup in QAT2 (#23657)
6 years ago
silingtong123 cec234b1aa
test=develop, error message of tree_conv OP enhancement (#23574)
6 years ago
cc 25628587f1
Collect output scale for quantized op and fused op (#23369)
6 years ago
Bai Yifan 9bc223c8a2
fix test_graph_wrapper failure on cudnnv7, test=develop (#23451)
6 years ago
ShenLiang c706ff20a3
fix conflict, test=develop (#23298)
6 years ago
Chengmo a2e9af5663
Add Tdm child OP in contrib (#23241)
6 years ago
cc 3ea7c59f76
Set fuse_all_reduce_ops=false for quantization test, test=develop (#23413)
6 years ago
cc 7c55a94de5
Disable test_quantization_scale_pass unittest for random error, test=develop (#23441)
6 years ago
Yiqun Liu bc2981e998
Disable test_code_generator and test_post_training_quantization_mobilenetv1 (#23440)
6 years ago
Wojciech Uss 9fd9067455
handle conv2d activations in older QAT models (#23202)
6 years ago
Wojciech Uss be2ac9cc3a
separated QAT1 and QAT2 (#23284)
6 years ago
lidanqing c524b930e7
Update QAT INT8 related code (#23104)
6 years ago
Wojciech Uss f836c8aa8f
add check for scales and a message (#23119)
6 years ago
cc bd80903333
Add activation_type in AddQuantDequantPass to be compatible with paddleslim, test=develop (#23221)
6 years ago
cc 589cd8782f
Post_training_quantizaion supports min_max methon (#23078)
6 years ago
lidanqing 432a4b2789
Changes QAT MKL-DNN documents (#22840)
6 years ago
cc b6717faf80
Added an option to use external FP32 model in QAT comparison test (#22858)
6 years ago
Sylwester Fraczek 5ff2439f51
fixed save qat2 model resnet50 and ernie (#22822)
6 years ago
hong f05c213f98
fix basic gru lstm parameter attr bug; test=develop (#22508)
6 years ago
tianshuo78520a 433cef03e5
fix typo word (#22784)
6 years ago
ShenLiang 3132681e8a
add partial_sum op in contrib (#22292)
6 years ago
ShenLiang e136661304
add partial_concat op in contrib (#22528)
6 years ago
tianshuo78520a d2ba91aad1
fix typo words (#22653)
6 years ago
Wojciech Uss 405bb94bd1
Fix batch_num default value (#22682)
6 years ago
bingyanghuang 6abeb5c54d
fix ernie accuracy doc (#22671)
6 years ago
Wojciech Uss fce37bc593
Doc update with Ernie QAT INT8 benchmarking (#22519)
6 years ago
Wojciech Uss 4cddb43c5c
Add support for Ernie NLP model to the Slim QAT (#22506)
6 years ago
cc d143f70a09 Post_training_quantization support set quant 8/16 bits (#22492)
6 years ago
cc 197913ebe1
Add weight quantization in post_training_quanzitaion (#22445)
6 years ago
joanna.wozna.intel 3099d9d47c Restore requantize squash (#22399)
6 years ago
juncaipeng b339dff231
fix use cache file, test=develop (#22240)
6 years ago
juncaipeng 8f7372ca81
add mul and matmul quantization, test=develop (#22054)
6 years ago
Bai Yifan a46bb2e6ab
Fix slim log format, test=develop (#22281)
6 years ago
songyouwei 83037e55ca named_sublayers and named_parameters (#21868)
6 years ago
gongweibao 5e07db15e6
fix init scaling value test=develop (#22145)
6 years ago
Chengmo 418abc92f4
Update pyramid related OP (#21372)
6 years ago
zhoushiyu cee2ccb078
add shuffle batch op (#21674)
6 years ago
lidanqing 9dff56e8e2 change qat_performance with mobilenet, change batch_size of qat2_resnet50 (#21895)
6 years ago
juncaipeng 8b74fc4fa7
Fix post training quantization (#21745)
6 years ago
Youwei Song f6144d8463 remove build_once & name_scope (#21131)
6 years ago
juncaipeng 5c4106023c
disable qat int8 test for now, test=develop (#21696)
6 years ago
juncaipeng 52f38877e4 Fix ci bug for deleting data files when other test is running (#21661)
6 years ago
lidanqing fbf9eca0d3 QAT Int8 document (#21360)
6 years ago
juncaipeng 1f57ac1241
delete concat in AddQuantDequantPass, test=develop (#21454)
7 years ago
lidanqing c0aa13672e Fp32 vs int8 qat C++ performance (#21244)
7 years ago
Liufang Sang 1840c1652c add config file to avoid load checkpoint test=develop (#21373)
7 years ago
itminner 07e6a94268 paddleslim quantization skip pattern support list of string (#21141)
7 years ago
Zhen Wang be2e3e67d9
Fix some typos in AMP. (#21354)
7 years ago
juncaipeng 84865b806b add resnet50 test for post trainint quantization, test=develop (#21272)
7 years ago
Zeng Jinle b9f8ae8494
Add global value getter setter (#21285)
7 years ago
juncaipeng 29b63f0aa1
support set model_filename and params_filename in post_training_quantization, test=develop (#21213)
7 years ago
Liufang Sang c91cb6c550 fix load checkpoint error in test_reader (#20924)
7 years ago
juncaipeng 00b11a4a1e
Support more ops in post training quantization, test=develop (#21073)
7 years ago
joanna.wozna.intel 37e0e7a96b QAT int8 accuracy little improvement (#21074)
7 years ago
Bai Yifan bd8b0ebaba fix distiller typo, test=develop (#21070)
7 years ago
juncaipeng 2c07727fb0
delete test resnet50 in post train quantization to avoid timeout error, test=develop (#21081)
7 years ago
juncaipeng fa522dffa0
Fix bug in add_quant_dequant_pass, test=develop (#21018)
7 years ago
juncaipeng 175ba39c03
Add post_training_quantization (#20800)
7 years ago
gongweibao 3255fe69bb Add custom black variable name set in amp interface. (#20875)
7 years ago
Aurelius84 aacd16dbb4 add pyramid_hash_op (#20698)
7 years ago
Bai Yifan fd5321b3f3 modify slim print precision to round(,6), test=develop (#20833)
7 years ago
whs fa67e6e83e
Fix unitest of pruning in python3 env. (#20825)
7 years ago
bingyanghuang fd49ebcbd8 update int8 benchmark with 6271 data, test=develop test=document_fix (#20736)
7 years ago
wangguanzhong 9a3e22aad4
move nms2 to contrib, test=develop (#20709)
7 years ago
juncaipeng f201b465ec
Move pool2d to add_quant_dequant_pass, test=develop (#20586)
7 years ago
Bai Yifan ffec9195e4
soft_label_distiller fix, test=develop (#20645)
7 years ago
Liufang Sang bc4af386f4 [PaddleSlim] refine slim reader to support dataloader (#20604)
7 years ago
gongweibao 1d82025e89
Add interface so user can get scaled loss when they use customized loss. (#20571)
7 years ago
翟飞跃 36acfaeeda Add fused_embedding_seq layer into fluid.contrib (#19771)
7 years ago
Guo Sheng bd99df715a
Fix basic_gru and docs of gru_unit and dynamic_gru (#19393)
7 years ago
bingyanghuang a5f530a8f9 Update QAT MKL-DNN readme doc (#20029)
7 years ago
Liufang Sang 86c2c362ae fix fuse_reduce_op quantization bug (#20306)
7 years ago
Michał Gallus 540935a825 [Bug-fix][1.6] Improve QAT accuracy (#20174)
7 years ago
gongweibao 7b9e33972a
delete backward return list test=develop (#20294)
7 years ago
xiteng1988 3785f08f68 fix next_tokens of controller (#20060)
7 years ago
bingyanghuang 9de6772510 Follow comment of Merged QAT PR 18970 (#19979)
7 years ago
Bai Yifan 28a559de12
fix PyReader._iterable usage, test=develop (#20045)
7 years ago
whs 3ea2b661c0
Make PaddleSlim support PyReader (#19995)
7 years ago
Wojciech Uss 4286a6270d Add support for new QAT models (#18970)
7 years ago
jhjiangcs 766bd529d1 add optimizer:dpsgd,test=develop (#19915)
7 years ago
whs bdb3e376d0
[PaddleSlim] Enhence compressor api in PaddleSlim (#19894)
7 years ago
juncaipeng b0ceed6fb4
add fake_quant_dequant_op for average pool2d, test=develop (#19880)
7 years ago
Tao Luo a4919d3688
move tree_conv to fluid.contrib.layers (#19918)
7 years ago
Aurelius84 418a0967f3
move match_matrix var_conv2d et.al api into fluid.contrib test=develop (#19859)
7 years ago
Jie Fang d9db94d752 Optimize amp for multi-gpu to enable FP16 gradients transfer across gpus. (#19714)
7 years ago
Bai Yifan e2c6bada36 Support dispensable student_loss in PaddleSlim distillation (#19824)
7 years ago
Bai Yifan aa63d5ac6d Make the distillation process not save teacher variables in PaddleSlim (#19633)
7 years ago
gongweibao 6c2bc29cc0
Fix float16 optimizer. (#19682)
7 years ago
Chen Weihang 73daa3d6c0
Code Cleanup: delete three useless raw variables in Conv2D (#19644)
7 years ago
Jie Fang c6a598a276 init new amp, optimize inserting cast op for batchnorm (#18596)
7 years ago
123malin 2f037c3189
fix the diff between async mode and async_half mode (#19535)
7 years ago
whs 1c2aae567a
Skip start epoch and end epoch when dumping strategy in PaddleSlim (#19580)
7 years ago
gongweibao abaf87be2b
Change backward_guard to optimize_guard to maximize the allreduce overlap. (#19506)
7 years ago
Zeng Jinle 5dce1da680 remove reset recordio usage (#19519)
7 years ago
lidanqing ba368bf696 clean up intel labeled TODOs (#19476)
7 years ago
Leo Chen 6fb310ae29 Fix bug of getting bool Flags from os.environ (#19349)
7 years ago
liu zhengxi 32598ffd8f
Python infer api update and add unit test (#19353)
7 years ago
Jiancheng Li 1799c257ad Update Light-NAS to support latency-aware search (#19050)
7 years ago
Zhen Wang 0fe72469ea
Add the max-pool2d quantization support and the partial quantization support. (#19310)
7 years ago
Tao Luo 2f8c7e021f
remove unused inference_transpiler unit-tests (#19130)
7 years ago
yaoxuefeng 9150cf50fc
add save cache model api in fleet& add slots shuffle in dataset module & add metric op to calculate ctr related metrics (#18871)
7 years ago
chengduo 01c7daade7
Add checking for the fetch_list of Executor.run (#18957)
7 years ago
whs c92b78b060
Fix unitest of light nas. (#18931)
7 years ago
tangwei12 d845848341
do some odd jobs (#18641)
7 years ago
guru4elephant 5ed713d519
remove ctr reader, all functions are satisfied in dataset (#18672)
7 years ago
bingyanghuang a25be53cb5 QAT int8 MKL-DNN transformation pass with MUL (#18322)
7 years ago
Jie Fang 2b4ef509ea init custom black white list (#18377)
7 years ago
Wojciech Uss 8ed819d8e4 Call the test_slim_int8_* tests through absolute path (#18386)
7 years ago
翟飞跃 19da59ed3f Remove all the code, API and doc of MKL-DNN INT8v1 (#18347)
7 years ago
whs 1bdfd2eb85
Fix checkpoint of Light-NAS (#18330)
7 years ago
Hongyu Liu df2eee71d8
Sequence mask support tensor (#18249)
7 years ago
Jie Fang 172c2facef init black/white lists (#17847)
7 years ago
songhao 6b3d96254d fix some bug when merge sparse embedding parameters, test=develop (#18223)
7 years ago
Wojciech Uss 36b60e240a Enable MKL-DNN for slim FP32 vs. INT8 tests (#18214)
7 years ago
bingyanghuang 976cf46010 slim QAT MKL-DNN readme (#18187)
7 years ago
chengduo 5588b923f3
Add multi process reader (#18115)
7 years ago
翟飞跃 802ea50956 fix spelling errors (#17941)
7 years ago
翟飞跃 78441c5449 add mkldnn Int8v2 slim doc (#17909)
7 years ago
翟飞跃 c2fb9b906a change dirname mobilenet to mobilenetv1 (#18160)
7 years ago
bingyanghuang 5cf7e7415d update mkldnn int8v2 doc (#18177)
7 years ago
Wojciech Uss ca5642c850 unify FP32 vs. INT8 comparison tests output (#18111)
7 years ago
Sylwester Fraczek accb132f0f fix slim int8 mkldnn multithreading issue (#18009)
7 years ago
chengduo b5a1c1463d
Update CPU_NUM config (#18059)
7 years ago
Wojciech Uss 78e932862c Added unit test for QAT FP32 & INT8 comparison (#17814)
7 years ago
Huihuang Zheng 0bf2535158
Cherry-pick: fix random CI failure. (#18011)
7 years ago
Kaipeng Deng 96ee528e3e
fix logging basicConfig cannot be setting after import paddle (#17786)
7 years ago
whs 82bfa0a5ba
Enhence auto pruning and fix save/load graph when using py_reader
7 years ago
bingyanghuang 90ebce9ead QAT int8 MKL-DNN transformation pass (#17819)
7 years ago
翟飞跃 209a3f4e09 add log into slim mkldnn int8 test (#17926)
7 years ago
whs 5df65e506d
Add Ligth-NAS for PaddleSlim (#17679)
7 years ago