Commit Graph

10986 Commits (a373aa76451b83a9eeb7617f92bdc9c1ea6e9ff6)

Author SHA1 Message Date
jakpiase d834f4e6e8
Added vanilla LSTM and LSTM with peepholes oneDNN fp32 kernel (#30661)
4 years ago
arlesniak 5bf25d1e8b
More precise mkldnn kernel rules in GetExpectedKernelType (#29840)
4 years ago
Jacek Czaja 173660be7b
[oneDNN] Cache oneDNN stream not to recreate in each oneDNN op (#30358)
4 years ago
Shang Zhizhou ae0f88a988
add DLA support:C++&&Python api (#30165)
4 years ago
chentianyu03 fb7fbc7a5d
fix abs bug and add abs test case (#30637)
4 years ago
ShenLiang 9514b4aa5f
Fix scatter grad bug (#30604)
4 years ago
Pei Yang cf9bdb9404
extend trt ut timeout threshold (#30537)
4 years ago
Thunderbrook 1bebc09253
solve build gpu task core (#30626)
4 years ago
石晓伟 33bf6eb753
revert external gflags, test=develop (#30623)
4 years ago
Jacek Czaja dfdb0359ea
- Disabling oneDNN inplace pass (#30588)
4 years ago
TTerror 10271ddfc4
support reduce_max op on kunlun (#30581)
4 years ago
QingshuChen 5013c67644
fix softmax bug for multi_card in kunlun (#30600)
4 years ago
wuhuanzhou 7e671c07b6
optimize unity build (#30195)
4 years ago
liuyuhui e5b0d9e1fc
[Kunlun] Add condition_variable and notify() in BindThreadedSSAGraphExecutor (#30586)
4 years ago
wanghuancoder 90773473a0
use nvtx push pop in timeline (#30567)
4 years ago
chentianyu03 358106fcb0
make abs op support complex types (#30375)
4 years ago
Wilber 2d5758c456
update. (#30585)
4 years ago
Tao Luo 9dd71c74df
disable test_analyzer_detect (#30541)
4 years ago
tangwei12 c9e78a22c5
add trainers for pserver (#30523)
4 years ago
wanghuancoder d1b25ed9d7
add some RecordEvent, for dygraph timeline (#30299)
4 years ago
liym27 ff25c5b36f
Fix bug: GetAttrValue should deal with attr with attrType vector<double> (#30536)
4 years ago
WangXi 572c466d19
[Prepare for MultiProcess xpu] unified gen nccl id, refine imperative reducer (#30455)
4 years ago
ykkk2333 549855ac20
add rmsprop_op_xpu test=kunlun (#30493)
4 years ago
Zhou Wei fb20ec9a4e
fix bug of multicard grad ncclAllReduce (#30553)
4 years ago
Zhen Wang f30d00553a
Fix the compiling error of update_loss_scaling when using cuda9. (#30538)
4 years ago
Leo Chen 81217a94d8
unify calling cudaSetDevice (#30470)
4 years ago
pangyoki 00554b3f6b
fix error message of Inplace strategy (#30520)
4 years ago
Leo Chen 7043b8cfc6
support layer_norm fp16 in dygraph amp (#30430)
4 years ago
wanghuancoder 59ad6ff3e3
delete empty line of pybing.cc, test=develop (#30529)
4 years ago
hutuxian e207fe6385
Ascend Framework Part2: pybind files (#30410)
4 years ago
hutuxian 40ede12631
Ascend Framework Part1: OP & Wrapper (#30281)
4 years ago
liuyuhui 843dc3cdbd
[Kunlun]PR3: add xpu executor, multi xpu card train function optimization (#30317)
4 years ago
QingshuChen 8489d4f76f
optimize batch_norm & pool op for kunlun (#30490)
4 years ago
wanghuancoder bd97192274
if pybind.cc changed, generate total report, test=develop (#30514)
4 years ago
taixiurong 5e5c2827a3
fix range op crash in dygraph xpu place (#30469)
4 years ago
JZ-LIANG 16ba0abc79
Recompute Offload: fixed bug in memcpy (#30484)
4 years ago
guofei 11e78ebaa3
Modify the calculation logic of LambOptimizer (#29313)
4 years ago
Adam Osewski c5ffad126c
[oneDNN] Refactor fuse pass helper functions to one place. (#30460)
4 years ago
Zhang Ting c9a334e1b3
add VecCastCUDAKernel (#30296)
4 years ago
pangyoki 13d757362c
Add Inplace strategy (Output reuse Input Varbase) in dygraph (#30103)
4 years ago
Yang Zhang 008b0a8b56
Fix float64 bug in layer norm (#30452)
4 years ago
石晓伟 715d862868
export global google flags to users, test=develop (#30448)
4 years ago
Wojciech Uss 88fc7a7d68
fix cache key for inplaced elementwise ops (#30404)
4 years ago
wawltor 3d49882e2c
fix the rnn mask memory bug for out of read (#30459)
4 years ago
taixiurong 6a3c8725b0
support transformer v2.0 (#30381)
4 years ago
ShenLiang e85be1b1b2
fix flatten api grad (#30426)
4 years ago
yaoxuefeng 6e0da01c61
Heter ps new (#30198)
4 years ago
123malin 2a98e9323a
test=develop, add distributed_infer (#30300)
4 years ago
QingshuChen cf786d22ec
fix bug that cann't find mkldnn(kunlun) (#30394)
4 years ago
cc 8e3a294045
skip quantizing ops in cpu inference (#30342)
4 years ago
alncat 7bbf3ac5ab
Added support for inference using quantization aware trained dygraph (#30288)
4 years ago
GaoWei8 180877e988
Softmax backward optimize (#30249)
4 years ago
Zhang Jun 10a8f3e5c3
fix bug on compiling inference shared lib with crypto;test=develop (#30269)
4 years ago
Huihuang Zheng 28e156c27f
Fix Sleep Error in enforce.h (#30335)
4 years ago
Leo Chen 3d015f1cf5
Set expected place in child thread for dataloader to avoid costing cuda memory on other card (#30338)
4 years ago
QingshuChen 2c1bba02e4
optimize memcpy perf for kunlun (#30291)
4 years ago
ShenLiang a60f17b89d
Support unused parameters in dynamic graph distributed (#30224)
4 years ago
JZ-LIANG 75936d838f
Recompute Offload (#30233)
4 years ago
lidanqing a60893f6b5
correct the allowed dimension size (#30326)
4 years ago
Chen Weihang c8c8f205ba
remove c++ stacktrace hint (#30325)
4 years ago
tangwei12 5e839e4da5
add sparse embedding & load vars for 2.0 & gloo bug fix (#30306)
4 years ago
tangwei12 25f80fd304
Fix/distributed proto (#29981)
4 years ago
Chengmo d479ae1725
【Paddle.Fleet】Support local save sparse param (#30175)
4 years ago
Double_V 231501fefc
fix elugradgrad test fail & error message opt (#30171)
4 years ago
Zhen Wang fb49ea388e
Fix the accuracy problem of allclose op when using float64 data type in static mode. (#29890)
4 years ago
yaoxuefeng 4656525e24
fix datanorm error msg (#30294)
4 years ago
furnace 77051cc9f0
add fp16 support for tril_triu op (#30186)
4 years ago
石晓伟 efa54629fb
fix header file paths of gflags, commit 3, test=develop (#30273)
4 years ago
Chengmo 5b2c15afcd
Fix server.h include device_context (#30243)
4 years ago
石晓伟 a0ee09148e
enhance error msgs of fusion_seqpool_cvm_concat_op.cc, test=develop (#30240)
4 years ago
石晓伟 a66eebab5c
fix header file paths of gflags, commit 4, test=develop (#30274)
4 years ago
石晓伟 8c4500ff6d
fix header file paths of gflags, commit 2, test=develop (#30272)
4 years ago
liym27 b4989fb744
Support vector<double> as type of op attribute and op set_value suppport vector<double> as value (#30126)
4 years ago
wangchaochaohu 8dcae0c55d
register OPMaker and Infer Shape Check for fused_elementwise_add (#30259)
4 years ago
AshburnLee 924aac2216
Add tf32 switch for cuDNN (#29192)
4 years ago
石晓伟 8ce2482b80
fix header file paths of gflags, commit 1, test=develop (#30271)
4 years ago
chentianyu03 c7371b7b20
type promotion for grad (#30177)
4 years ago
liym27 3ce878f309
Check the rank of input in kernel of set_value op (#30147)
4 years ago
WeiXin 66dc4ac77b
modify error message based on comments (#30189)
4 years ago
wawltor fee424411a
just add the op error message for the matmul xpu (#30246)
4 years ago
GaoWei8 0a21924a8d
optimize softmax forward (#30217)
4 years ago
wangchaochaohu af80859dd6
reduce the occupied size of memory for the fused pattern of elementwise_add Op and activation Op(relu Op for example) (#29885)
4 years ago
zhang wenhui 5932fee60a
enhance error message, test=develop (#30220)
4 years ago
pangyoki da16b33f2e
add View(reuse allocation) strategy on squeeze, unsqueeze, reshape, flatten op (#29913)
4 years ago
Jacek Czaja 4aba17b5db
[oneDNN] Added UT for testing elementwise_mul caching (#30203)
4 years ago
Zhen Wang 7f7dfccf20
Support pure fp16 training for AMP API. (#29544)
4 years ago
Leo Chen 789743e190
use cuda generator in bernoulli cuda kernel (#30199)
4 years ago
Leo Chen 8696335f86
Fix dtype of ungenerated grad var (#28511)
4 years ago
Wilber 609c022222
shape op support int8 and uint8 tensor (#30201)
4 years ago
Wilber 01a287bf0a
fix windows compile when WITH_PYTHON=ON and WITH_TENSORRT=ON (#30194)
4 years ago
ruri e42e1e80dc
Add version checking, test=op_version (#30129)
4 years ago
Leo Chen 1f97d61c68
Add callback after TensorCopy (#30123)
4 years ago
Chengmo 528e03fc08
【Paddle.Fleet】Fix tensor table (#30075)
4 years ago
Wilber ade244948c
disable mkldnn inplace pass on windows (#30164)
4 years ago
joanna.wozna.intel 907262ee15
Fix analysis predictor test (#30191)
4 years ago
lijianshe02 2dc7ee276b
enhance error message of nll_loss op test=develop (#30125)
4 years ago
Huihuang Zheng 54bf3f5a56
Refine PADDLE_ENFORCE Error Messages. test=develop (#30149)
4 years ago
Chen Weihang d0fb06b27f
[Complex] Simplify prepared op impl to improve performance (#30153)
4 years ago
123malin c5b415bfd9
Improve Index select cuda kernel (#30139)
4 years ago
wangchaochaohu 7dd551e08b
refine the paddle place support using str (#28769)
4 years ago
WeiXin 404c16763a
Add detailed error message for curandStatus_t, cublasStatus_t, cusolverStatus_t (#30161)
4 years ago
Wilber 91a8a25721
enhance error info for py_func (#30138)
4 years ago
weihaoji b8207af6bc
[XPU] Remove lite_xpu ut lite_resnet50_test since fusion pass changes introduced precision diff. test=develop (#30122)
4 years ago
liuyuhui 15fac5e7fa
fix assign_op_xpu concat_op_xpu warining (#30120)
4 years ago
Jack Zhou f5428eca4f
fix enforce msg of sum xpu op (#30113)
4 years ago
123malin 198fbdfb60
Add Lookahead and ModelAverage Optimizer (#30004)
4 years ago
Leo Chen adac38c506
add dispenable input for core.ops.reshape2/expand/slice (#30072)
4 years ago
ShenLiang becf99d2e8
fix error message (#30135)
4 years ago
Zhou Wei 30888ca343
Polish and Optimize the print/repr information of Layer (#29998)
4 years ago
wangguanzhong 69839f8a9a
fix error message for distribute_fpn_proposals_op (#30116)
4 years ago
QingshuChen 8e1c3ddf15
add aarch64 and sunway kunlun lib (#30027)
4 years ago
Shang Zhizhou 05b27695f1
add inference api: DisableTensorRtOps (#30109)
4 years ago
石晓伟 53bb126510
fix a bug in op_version_registry, test=develop, test=op_version (#29994)
4 years ago
xiemoyuan 3e0c492910
Optimize the error message of framework. (#30134)
4 years ago
liym27 9922bd4125
Fix bug: In dynamic mode, if start or end is negetive, __getitem__ return wrong result(#30003)
4 years ago
chentianyu03 666e665132
change the kron gradient when complex types (#29995)
4 years ago
chentianyu03 a5e422c85d
add trace op_register_version and fix version bug; test=op_version (#30000)
4 years ago
cc 9f34374b48
Fix the formate of raising error in randperm op (#30108)
4 years ago
liuyuhui 254ad61959
fix xpu pe sync, test=notest (#30095)
4 years ago
Thunderbrook 0b8e1fadc5
add topo-aware in heter-ps (#30087)
4 years ago
hong 297fff1a79
support dygraph in xpu place (#30051)
4 years ago
wangchaochaohu d0a5620575
fix the compiler error when gcc4 cuda9.0 (#29997)
4 years ago
WangXi ee16006b5d
Optimization grad merge performance (#29784)
4 years ago
yongqiangma e891f4da1b
Add p_norm op version info (#30042)
4 years ago
tangwei12 7d1c149e09
for inference checkpoint (#30081)
4 years ago
tangwei12 7d4bdff07d
fix large scale memory (#30035)
4 years ago
Shang Zhizhou 08dc5bc27e
fix op version checker of pass bug (#30028)
4 years ago
cc 68398abce9
[Inference] zero_copy_tensor supports int8_t (#30053)
4 years ago
whs 1b999d2b5d
Add version checking (#30040)
4 years ago
ceci3 85b2f05ab0
register ModifyAttr for instance_norm, test=op_version (#30065)
4 years ago
channings ddcff254db
fix op_register_version for compare ops, test=op_version (#30007)
4 years ago
Wilber 66e16b7e99
update lite subgraph. (#30056)
4 years ago
GaoWei8 a64822589f
add REGISTER_OP_VERSION for LSTM (#30038)
4 years ago
yinhaofeng 6e93fb92f9
Register op version for linspace,test=op_version (#30025)
4 years ago
123malin d0056c324d
test=develop, add op_register_version for roll_op (#30023)
4 years ago
chentianyu03 e012930aa3
complex gradient matmul (#29966)
4 years ago
ShenLiang 893d37e5c6
Fix rank_attention op_version, test=op_version (#30006)
4 years ago
Adam Osewski 13aef97043
operator checkpoints for new attributes. (#29832)
4 years ago
wangguanzhong 844d8e0c2c
add REGISTER_OP_VERSION for generate_proposals, roi_align, roi_pool test=op_version (#30034)
4 years ago
cc c3c064a8fc
Add mkldnn nearest_interp and bilinear_interp op (#30016)
4 years ago
chalsliu c053bf2a57
Revert "register ModifyAttr for instance_norm, test=op_version (#29938)"
4 years ago
wawltor cc2f94620c
add the support the op version check for matmul, test=op_version (#30011)
4 years ago
wawltor b33aaea86c
add the op version check for the elementwise ops, test=op_version (#30010)
4 years ago
Chengmo 4cbcc9b6da
fix momentum op register (#29941)
4 years ago
hutuxian 7c1f69bdf0
add op_version for flip op [test=op_version] (#30019)
4 years ago
ceci3 77c1684397
register ModifyAttr for instance_norm, test=op_version (#29938)
4 years ago
Leo Chen 47d10c55d5
Enhance debugging (#30001)
4 years ago
FlyingQianMM d42f93e504
add op_register_version for allclose op; test=op_version (#29968)
4 years ago
wawltor 8f49f9d5c9
change the elementwise ops version check, test=op_version
4 years ago
guofei b23faf37be
Add moving_average_abs_max_scale op_register_version test=develop (#29957)
4 years ago