Commit Graph

23241 Commits (a446d26e8ad805fd6c37a7f2a44b01fe28ffbd9e)
 

Author SHA1 Message Date
phlrain 0e40298949 fix matmul shape check; test=develop
6 years ago
phlrain 56c2d384c7 add elementwise floordiv, mod; test=develop
6 years ago
Wu Yi b7baeed7bb fix win gpu build test=develop (#16334)
6 years ago
liuwei1031 df5d19aa9d
temoprarily disable the code of use kCUDNN, test=develop (#16205)
6 years ago
ruri 09e05a110b
Merge pull request #16217 from ceci3/doc
6 years ago
zhhsplendid 124f1df481 Add flags for init and re-alloc gpu
6 years ago
Zhen Wang 8965819fbb rewrite the cuda kernels of channel_wise_quant_op and channe_wise_dequant_op. test=develop
6 years ago
Wu Yi 8bebfe5640
add resnet nccl2 dist training, mp training unit test (#16167)
6 years ago
Wojciech Uss cbe2dbf0db Add enabling quantization (#16326)
6 years ago
lujun 09442fb27e checkpoint pr be moved here, test=develop
6 years ago
Tao Luo 9a05859179
Merge pull request #16322 from wojtuss/wojtuss/fix_cpu_quantize_pass
6 years ago
qingqing01 8caa785e83
Enhance affine_channel_op infer-shape check (#16317)
6 years ago
flame 08838f3909
Fix save inference model bug (#16242)
6 years ago
Kaipeng Deng 957ea995fc
Merge pull request #16243 from heavengate/batch_norm_not_persistent
6 years ago
nhzlx 4f4daa4b66 cherry-pick from feature/anakin-engine: add data type for zero copy #16313
6 years ago
nhzlx 07dcf2856c git cherry-pick from feature/anakin-engine: update anakin subgraph #16278
6 years ago
nhzlx c407dfa3cb cherry-pick from feature/anakin-engine: refine paddle-anakin to new interface. #16276
6 years ago
nhzlx a25331bc26 cherry-pick from feature/anakin-engine: deal the changing shape when using anakin #16189
6 years ago
nhzlx c79f06d3d8 cherry-pick from feature/anakin-engine: add batch interface for pd-anakin #16178
6 years ago
nhzlx 69d37f81d7 cherry-pick from feature/anakin-engine: refine anakin subgraph. #16157
6 years ago
nhzlx a1d200a5de cherry-pick from feature/anakin-engine: Anakin support facebox #16111
6 years ago
flame a32d420043 cherry-pick from feature/anakin-engine: batch norm (#16110)
6 years ago
flame 0945b97f07 cherry-pick feature/anakin-engine: add anakin softmax/transpose/batch_norm/flatten/reshape op (#16020)
6 years ago
nhzlx b21770a2aa cherry-pick from feature/anakin-engine: Add subgraph fuse support and anakin engine #16018
6 years ago
nhzlx 084310f536 paddle-anakin: concat, split, pool2d converter#16003
6 years ago
flame be523baad2 Add anakin conv2d/relu/sigmoid/tanh converter (#15997)
6 years ago
Yan Chunwei d0ce6a9044 fix anakin converter registry (#15993)
6 years ago
Tao Luo a5124ee0bb
Merge pull request #16301 from luotao1/runtime_context_pass
6 years ago
lujun 622fe6a56b checkpoint pr be moved here, test=develop
6 years ago
baojun 2de263a5d9 Add softmax_with_cross_entropy_op to ngraph engine (#16304)
6 years ago
sneaxiy bb166a1e10 fix API.spec
6 years ago
ruri a3b8028d46
Merge pull request #16202 from shippingwang/add_sqrt_doc
6 years ago
phlrain dd080b17c3 Merge branch 'develop' of https://github.com/PaddlePaddle/Paddle into fix_sequence_pad_2
6 years ago
phlrain 1580be5d6c fix sequence pad; test=develop
6 years ago
dengkaipeng aba2713ffc fix comment. test=develop
6 years ago
chengduo f26ba5bddd
Fuse AllReduce (#15921)
6 years ago
Zeng Jinle d0ef682552
Merge pull request #16274 from sneaxiy/fix_grad_maker
6 years ago
baojun 804afc51db Minor ngraph fix (#16270)
6 years ago
Tao Luo 9195c3bb03
Merge pull request #16280 from luotao1/cos_sim_infershape
6 years ago
Wojciech Uss 104a9f1e27 fix pattern maching conv2d with(out) ResidualData
6 years ago
Wu Yi 6382b62f6b
Collective ops (#15572)
6 years ago
lujun bed0ecf3d2 checkpoint pr be moved here, test=develop
6 years ago
lujun 5bb04ea47d
Merge pull request #12 from PaddlePaddle/develop
6 years ago
sneaxiy 023a3a3d62 fix op grad maker
6 years ago
luotao1 82af8031d9 add runtime_context_cache_pass
6 years ago
Zhen Wang ec88b6cc5a add channel wise quantization in ir pass.
6 years ago
Tao Luo b9fc80a133
Merge pull request #16287 from PaddlePaddle/revert-16002-runtime_context
6 years ago
whs 18911b6eea
[enhence] Make step_input of dynamic_rnn support custom lod level. (#15972)
6 years ago
zhhsplendid 22715487dc add allocator flags
6 years ago
luotao1 c05af910bc refine cos_sim infershape
6 years ago