Commit Graph

77 Commits (a6beb96dd0235c236336f2db31df875b33db6635)

Author SHA1 Message Date
cc 75eec3d1f6
Post training quantization supports optimize model by fusing (#24822)
6 years ago
Wojciech Uss 78d4f0cc91
add option to exclude ops by id from quantization (#24689)
6 years ago
cc dbcd7c69e9
Update sigmoid output from Y to out, test=develop (#24765)
6 years ago
cc 88e9d74a75
Collecting concat output threshold, test=develop (#24742)
6 years ago
cc 6c89ca2157
Add output threshold for ops that have several output activations, test=develop (#24726)
6 years ago
cc 4d35112255
[Fix bug] Init scale node in OutScaleForTrainingPass and enable test_quantization_scale_pass UT (#24393)
6 years ago
Wojciech Uss db052009c7
Enabled quantize all and skip missing in QAT (#24281)
6 years ago
Sylwester Fraczek e1a7a88057
added reshape transpose matmul fuse pass (#23754)
6 years ago
arlesniak d31a174f51
added fusing matmul-transpose-reshape pass (#23866)
6 years ago
Wojciech Uss 3d744162dd
QAT: support for new models (#23928)
6 years ago
cc 40aa14ec77
Weight quantization support channel_wise_abs_max method to achieve higher accuracy (#23629)
6 years ago
joanna.wozna.intel 12ba05ce0c
Add scale-matmul fuse pass (#23734)
6 years ago
Wojciech Uss 1753860dd0
Enable matmul and cleanup in QAT2 (#23657)
6 years ago
cc 25628587f1
Collect output scale for quantized op and fused op (#23369)
6 years ago
Wojciech Uss 9fd9067455
handle conv2d activations in older QAT models (#23202)
6 years ago
Wojciech Uss be2ac9cc3a
separated QAT1 and QAT2 (#23284)
6 years ago
Wojciech Uss f836c8aa8f
add check for scales and a message (#23119)
6 years ago
cc bd80903333
Add activation_type in AddQuantDequantPass to be compatible with paddleslim, test=develop (#23221)
6 years ago
cc 589cd8782f
Post_training_quantizaion supports min_max methon (#23078)
6 years ago
tianshuo78520a d2ba91aad1
fix typo words (#22653)
6 years ago
Wojciech Uss 4cddb43c5c
Add support for Ernie NLP model to the Slim QAT (#22506)
6 years ago
cc d143f70a09 Post_training_quantization support set quant 8/16 bits (#22492)
6 years ago
cc 197913ebe1
Add weight quantization in post_training_quanzitaion (#22445)
6 years ago
juncaipeng 8f7372ca81
add mul and matmul quantization, test=develop (#22054)
6 years ago
juncaipeng 8b74fc4fa7
Fix post training quantization (#21745)
6 years ago
juncaipeng 1f57ac1241
delete concat in AddQuantDequantPass, test=develop (#21454)
6 years ago
lidanqing c0aa13672e Fp32 vs int8 qat C++ performance (#21244)
6 years ago
itminner 07e6a94268 paddleslim quantization skip pattern support list of string (#21141)
6 years ago
juncaipeng 84865b806b add resnet50 test for post trainint quantization, test=develop (#21272)
6 years ago
juncaipeng 29b63f0aa1
support set model_filename and params_filename in post_training_quantization, test=develop (#21213)
6 years ago
juncaipeng 00b11a4a1e
Support more ops in post training quantization, test=develop (#21073)
6 years ago
joanna.wozna.intel 37e0e7a96b QAT int8 accuracy little improvement (#21074)
6 years ago
juncaipeng fa522dffa0
Fix bug in add_quant_dequant_pass, test=develop (#21018)
6 years ago
juncaipeng 175ba39c03
Add post_training_quantization (#20800)
7 years ago
juncaipeng f201b465ec
Move pool2d to add_quant_dequant_pass, test=develop (#20586)
7 years ago
Liufang Sang 86c2c362ae fix fuse_reduce_op quantization bug (#20306)
7 years ago
Michał Gallus 540935a825 [Bug-fix][1.6] Improve QAT accuracy (#20174)
7 years ago
bingyanghuang 9de6772510 Follow comment of Merged QAT PR 18970 (#19979)
7 years ago
Wojciech Uss 4286a6270d Add support for new QAT models (#18970)
7 years ago
whs bdb3e376d0
[PaddleSlim] Enhence compressor api in PaddleSlim (#19894)
7 years ago
juncaipeng b0ceed6fb4
add fake_quant_dequant_op for average pool2d, test=develop (#19880)
7 years ago
lidanqing ba368bf696 clean up intel labeled TODOs (#19476)
7 years ago
liu zhengxi 32598ffd8f
Python infer api update and add unit test (#19353)
7 years ago
Zhen Wang 0fe72469ea
Add the max-pool2d quantization support and the partial quantization support. (#19310)
7 years ago
bingyanghuang a25be53cb5 QAT int8 MKL-DNN transformation pass with MUL (#18322)
7 years ago
翟飞跃 802ea50956 fix spelling errors (#17941)
7 years ago
Kaipeng Deng 96ee528e3e
fix logging basicConfig cannot be setting after import paddle (#17786)
7 years ago
bingyanghuang 90ebce9ead QAT int8 MKL-DNN transformation pass (#17819)
7 years ago
翟飞跃 993c703bcc INT8 MKL-DNN v2 integrate to slim (#17634)
7 years ago
Zhen Wang 3398f99608
Adding AddQuantDequantPass for TensorRT int8 (#17529)
7 years ago