Commit Graph

39 Commits (b6d5c74f698f0c9fdcbe7d910c77387edd8bb2bc)

Author SHA1 Message Date
Jiabin Yang 667f88f9a6
Fix/gcc 4.8 ubt link error (#18558)
6 years ago
Physher 0caa08ea40 Add mkldnn int8 mul-op kernel (#17834)
6 years ago
Leo Zhao 8f5fffca0a rename mkldnn set/get_cur_thread_id() to set/get_cur_mkldnn_session_id() (#18453)
6 years ago
Brian Liu 4bc2987d2f Fix bug in quantize kernel which cause crash in vgg16/19 model (#17964)
6 years ago
Leo Zhao 681d3553f1 Fix potential mkldnn concat/pool/conv kernel issues (#18393)
6 years ago
lidanqing f8ecc3de89 refactor the function ConvFwdPrimitiveDesc (#17897)
6 years ago
Wojciech Uss 78e932862c Added unit test for QAT FP32 & INT8 comparison (#17814)
6 years ago
Jacek Czaja 84bb45c054 [MKL-DNN] Thread-Safety for MKL-DNN reusing Part 1 (#17965)
6 years ago
Zeng Jinle 3ece61f71e
Remove attribute in Allocator::Allocate (#17878)
6 years ago
Yihua Xu 14a32bf0c4 Fix the accuracy issue while using float precision to get the scale. (#17884)
6 years ago
lidanqing d7c5c2bd64 Add input format in Transpose GetHash (#17737)
6 years ago
lidanqing 04b6c29ee0 Improve mobilenetv2 INT8 performance by using INT8 relu as post-op (#17570)
6 years ago
Michał Gallus 0c39b97b4e [MKL-DNN] Add Fully Connected Op for inference only(#15226)
6 years ago
Sylwester Fraczek bccb0ba49a fix quantize_squash_pass segfault when no tensor linked to Bias (#17292)
6 years ago
guomingz 2281ebf0f3 Enable the convolution/relu6(bounded_relu) fusion for FP32 on Intel platform. (#17130)
6 years ago
Jacek Czaja 87a44b1149 [MKL-DNN] Added reusing of primitive descriptors (fp32) (#16667)
6 years ago
Jacek Czaja 2632327429 [MKL-DNN] Tensor modifications revert (#16462)
6 years ago
xiaolil1 e235882c18 Enable MKL-DNN INT8 Concat Kernel. (#16156)
6 years ago
xiaolil1 e818fa1004 Enable INT8 transpose kernel for MobileNet-SSD improvement. (#16159)
6 years ago
luotao1 d9f0e7252a refine with comments
6 years ago
luotao1 6fa52f83ba Merge branch 'develop' into fc_infershape
6 years ago
Wojciech Uss 2579ade45f Add cpu_quantize_pass for C-API quantization (#16127)
6 years ago
luotao1 721c2c00ef refine fc_infershape
6 years ago
xiaolil1 a177d48217 Add Requantize OP (#15318)
6 years ago
xiaolil1 1abddd8d97 Optimize Quantize Op with primitive reuse. (#15929)
6 years ago
xiaoli.liu@intel.com c4187dbd7c Merge branch 'develop' of https://github.com/PaddlePaddle/Paddle into dequantize-reuse
6 years ago
xiaolil1 6724be2b0d INT8 Pool kernel Key Creation Optimization. (#15883)
6 years ago
Tao Luo ba90e05281
Merge pull request #15917 from jczaja/prv-tensor-mkldnn-ops
6 years ago
Tao Luo effec86600
Merge pull request #15913 from liangan1/func_coverage
6 years ago
Jacek Czaja c63f6b2039 - MKL-DNN pooling updated to set_prim_desc
6 years ago
xiaoli.liu@intel.com 70759d181b Optimize INT8 DeQuantize Op with primitive reuse.
6 years ago
Michal Gallus 6ebe9877bb Improve code reuse at MKL-DNN sum
6 years ago
liangan1 4acc522087 Enable function coverage for U8/S8 ConvMKLDNNOpKernel
6 years ago
Jacek Czaja dec9cf53c8 [MKL-DNN] MKL-DNN specific Tensor modification (#15429)
6 years ago
Sylwester Fraczek 74672d1aff Change *(smart_ptr.get()) -> *smart_ptr
6 years ago
Krzysztof Binias 1578c60bdd Add new ut and remove unnecessary code
6 years ago
Michal Gallus 7a8eff36a6 Fix old FC backward weights descriptor creation
6 years ago
Krzysztof Binias 69b7c595d6 Small fix
6 years ago
Krzysztof Binias b1bdcd4de8 Make separate folders for mkldnn codes
6 years ago