Commit Graph

673 Commits (3d015f1cf529915ab52cb8aef7c475f67fb128b5)

Author SHA1 Message Date
Pei Yang 8a4f85feb9
Add unittests and OP version registry for quant_conv2d_dequant_fuse_pass (#27689)
4 years ago
joanna.wozna.intel b0ee1405f7
Add conv2d bfloat16 support (#27325)
4 years ago
cc c5c13473c6
Add compatibility check for four mkldnn pass (#27364)
4 years ago
Wilber 3d5522146e
register seq_concat_fc_fuse pass. (#27479)
4 years ago
wanghuancoder df43905f12
use iwyu clean include (#27267)
4 years ago
Pei Yang 8182337096
clear pass logs (#27434)
4 years ago
Shang Zhizhou d93661942e
fix bug sequececonv_eltadd_relu_fuse_pass (#27404)
4 years ago
Leo Chen aba759ba16
[Feature] Enhance inplace addto strategy for gradient accumulation in static graph (#27112)
4 years ago
Wilber 39546aa2f3
Add pass compatible and unit test. (#27377)
4 years ago
Pei Yang fd7ab4e63c
register pass compatibility (#27357)
4 years ago
haozech 7e6dfcf9b2
Add 3 pass version check (#27283)
4 years ago
Shang Zhizhou 3c11717988
add op version checker to ir passes (#27329)
4 years ago
Wilber f827665ae6
[Pass Compatible] Bind python compatible. (#27262)
4 years ago
joanna.wozna.intel 1483ea2304
Add bfloat16 passes (#26999)
4 years ago
Pei Yang 5fb8c92054
fix multihead matmul shared params (#27121)
4 years ago
Adam 8bcb1f29d9
Add conv+affine_channel fuse pass to MKLDNN pass strategy and fix it (#26779)
5 years ago
Pei Yang e3f8e5cf5c
trt int8 support conv2d_transpose (#26636)
5 years ago
joanna.wozna.intel eb097d64f6
Fix int8 performace drop cpu_quantize_placement_pass (#26715)
5 years ago
joanna.wozna.intel 559e43eee4
Small change in conv2d and quantize pass (#26671)
5 years ago
Pei Yang b757466b0d
fix trt dynamic ernie serialization unit test (#26228)
5 years ago
cc 3f816bc8b4
[Quantization] Conv2d_transpose and mul support channnelwise quantization (#25639)
5 years ago
Yiqun Liu 1be6bf45ae
Add assign to fusion_group and enhance inplace execution in fusion_group. (#26121)
5 years ago
joanna.wozna.intel 734cf1c3e9
Change use_quantizer attribute name and data type (#25838)
5 years ago
Zhou Wei e0a9115e28
fix random compile failure due to missing file (#25661)
5 years ago
arlesniak e52df3b125
Added DNNL cache management for DyGraph (#25624)
5 years ago
Adam 98899b73d2
Fix FC + GRU fuse pass (#25687)
5 years ago
wanghuancoder 1917b38099
fix some errmsg report,in framework/ir/, about 21 files (#25525)
5 years ago
wanghuancoder 9b46fe0440
fix some errmsg report,in framework/ir/, about 5 files (#25539)
5 years ago
wanghuancoder e65c5b8e83
fix some errmsg report, in framework/ir/ (#25471)
5 years ago
wanghuancoder 6c0982b942
fix some errmsg report, in framework/ir/mkldnn (#25467)
5 years ago
wanghuancoder fce6466217
fix some errmsg report, in framework/ir/ subdir(memory,optimizer,multi_device) (#25460)
5 years ago
Zhaolong Xing 7b7e605189
[Fix BUGs]: fix multhead matmul pass's instable bug (#25123)
5 years ago
Wojciech Uss d0a921ba98
Quant2 updates and fixes (#25313)
5 years ago
Jacek Czaja 17c751bec6
[oneDNN] Fix to #25078 (#25256)
5 years ago
Sylwester Fraczek 36abeff44f
adding elementwiseadd quantization (#25178)
5 years ago
Pei Yang b2f5a149e7
[Paddle-TRT] Better Paddle-TensorRT support for PaddleSlim quant models (#25097)
5 years ago
Shibo Tao 19c4db1b56
don't re-generate header file if content doesn't change (#25130)
5 years ago
Jacek Czaja a7944904d3
[oneDNN]elementwise_add and elementwise_mul int8 support (#24984)
5 years ago
Sylwester Fraczek 53d563a0fe
Reshape transpose matmul coverage (#24970)
5 years ago
Sylwester Fraczek a7ee634b45
fix WARNING: ThreadSanitizer: heap-use-after-free (#24929)
5 years ago
Jacek Czaja 40a5f3fd86
[oneDNN] Clearing mkldnn cache in naiveexecutor destructor (#24756)
5 years ago
Chen Weihang d1062d5278
Replace all errors thrown by LOG(FATAL) with PADDLE_THROW (#24759)
5 years ago
Wojciech Uss 78d4f0cc91
add option to exclude ops by id from quantization (#24689)
5 years ago
Wilber ba2f8f0ce4
fix embedding_eltwise_layernorm_fuse_pass. test=develop (#24592)
5 years ago
Yiqun Liu 6b464f969a
Add an operator node in unittest to make the fusing result unique. (#24617)
5 years ago
Yiqun Liu 560c815390
Add some check for CUDA Driver API and NVRTC (#22719)
5 years ago
Jacek Czaja 8b88cd5167
[oneDNN] Fix to inplace pass (#24442)
5 years ago
Chen Weihang aa0f254fbe
Add macro BOOST_GET to enrich the error information of boost :: get (#24175)
5 years ago
Wojciech Uss db052009c7
Enabled quantize all and skip missing in QAT (#24281)
5 years ago
Huihuang Zheng 8a1a2af82e
Add Assert Op (#24280)
5 years ago
joanna.wozna.intel 356f5ee220
[Refactoring] Unify op-dequant squashes (#24277)
5 years ago
joanna.wozna.intel b43b46e619
[INT8] Add requant-op squash (#24143)
5 years ago
Sylwester Fraczek e1a7a88057
added reshape transpose matmul fuse pass (#23754)
5 years ago
wangchaochaohu fa43d74a3a
fix the intermediate node of graph for fusion group test=develop (#24184)
5 years ago
liuwei1031 9a93f6aae0
improve efficiency of runtime InferVarType (#22778)
5 years ago
wangchaochaohu 2270864019
Fusion group optimize for cuda codegen(#23940)
5 years ago
Jacek Czaja eb411613e9
[DNNL] refine activations Inplace support (#24145)
5 years ago
Jacek Czaja 461e6a01ec
[DNNL] activations Inplace support (#24123)
5 years ago
arlesniak d31a174f51
added fusing matmul-transpose-reshape pass (#23866)
5 years ago
Zeng Jinle acef55df04
fix isolated var fetch bug, test=develop (#24070)
5 years ago
Jacek Czaja c6c65c65c7
[DNNL] Added elementwise_add mkl-dnn inplace (#23477)
5 years ago
Yiqun Liu 071a702060
Fix the error misjudgment when there are control nodes in graph. (#23943)
5 years ago
Zeng Jinle c49791362f
Correct reader device index (#23802)
5 years ago
joanna.wozna.intel 12ba05ce0c
Add scale-matmul fuse pass (#23734)
5 years ago
Chen Weihang 532079a222
API (CompiledProgram) error message enhancement (#23559)
5 years ago
chenhaoze 9b06dd8628
Add three passes and api reference of paddle_pass_builder. test=develop (#23741)
5 years ago
joanna.wozna.intel 5ee099ca57
Op-requant squash (#23665)
5 years ago
mozga-intel 3baaee9aab
Remove: NGraph engine from PDPD repository (#23545)
5 years ago
joanna.wozna.intel 3cb5623dad
Add matmul dequant squash (#23505)
5 years ago
wangchaochaohu c1187cd6f4
Fp16 refine for fusion group (#23472)
5 years ago
joanna.wozna.intel ce08fdcf2b
Add support for INT8 matmul in C-API quantization (#23463)
5 years ago
wangchaochaohu d085f79228
fix untime fail for output var stop_gradient=True for fusion group (#23317)
5 years ago
Kaipeng Deng d223a24904
Fix inplace_abn compile error on Windows (#23464)
5 years ago
wangchaochaohu 5c60778731
polish the code of fusion group test=develop (#23370)
5 years ago
Yiqun Liu bc2981e998
Disable test_code_generator and test_post_training_quantization_mobilenetv1 (#23440)
5 years ago
joanna.wozna.intel 8c463700e1
Add default pass attributes (#23042)
5 years ago
Kaipeng Deng 21d95be0db
Add inplace abn op (#22806)
5 years ago
Zeng Jinle 3a21980b78
add reader dependency pass, test=develop (#23301)
5 years ago
wangchaochaohu d280106007
Add support for attr type Op and add fill_constant Op and scale Op (#23163)
5 years ago
Jacek Czaja 2bb1b0e89e
[DNNL] Added MKL-DNN inplace pass for C-API inference (#23315)
5 years ago
Wojciech Uss f836c8aa8f
add check for scales and a message (#23119)
5 years ago
Tao Luo c00d427d52
simplify the cmake log of ir/CMakeLists.txt (#23262)
5 years ago
Zeng Jinle bae5930ba1
fix graph attr copy issues, test=develop (#23191)
5 years ago
Zeng Jinle acfc9b8a70
Reader sequential and inference partial feed (#22699)
5 years ago
Wilber 95b356a069
update embedding_eltwise_layernorm fuse and kernel. test=develop (#23114)
5 years ago
Yiqun Liu 3af4771122
Add the detection and code-generation of sqrt and square in fusion_group (#23095)
5 years ago
Sylwester Fraczek abee05a8c8
added mkldnn swish activation (#23041)
5 years ago
wangchaochaohu 3757e0687c
Add Unittest for backward of fusion group (#22932)
5 years ago
wangchaochaohu f0d193a23c
Cast fusion for fusion group (#22876)
5 years ago
Wilber ff3ddbb502
add skip_layernorm pass. test=develop (#22895)
5 years ago
Zhaolong Xing 8d6dc102fe
[Ernie GPU Optimize]: Embedding_eltwise_layernorm Fuse (#22494)
5 years ago
liu zhengxi 61fef9754b
Fix fc padding bug during inference fusion (#22860)
5 years ago
wangchaochaohu ca9e77a8d4
add sum op support for fusion group (#22771)
5 years ago
tianshuo78520a 433cef03e5
fix typo word (#22784)
5 years ago
GaoWei8 cdf5f6fb8c
Add an inference interface to disable FC padding (#22097)
5 years ago
tianshuo78520a d2ba91aad1
fix typo words (#22653)
5 years ago
Yiqun Liu 22bbd54719
Add the support of fp16 in fusion_group (#22239)
5 years ago
Wilber 9a8203aa25
fix fc_lstm_fuse when multi sub-graph use same fc_bias. test=develop (#22551)
5 years ago
Zhaolong Xing 8acd745c25
[Ernie GPU Optim]: Fuse three fc to multihtead matmul (#22486)
5 years ago
Yiqun Liu dcfb603897
Enable the detection of subgraph composed of grad ops (#21223)
5 years ago