Commit Graph

1905 Commits (d0f719f78730fb77af3c54b32de54520be290e54)

Author SHA1 Message Date
qiaolongfei ffe1b69229 Merge branch 'develop' of https://github.com/PaddlePaddle/Paddle into add_compile_time_infershape
7 years ago
kexinzhao 087addaa76 Merge pull request #4558 from kexinzhao/adagrad_op
7 years ago
Kexin Zhao 78f4c803f3 change learning rate and fix format
7 years ago
Kavya Srinet 0336304176 Merge branch 'develop' of https://github.com/PaddlePaddle/Paddle into rmsprop
7 years ago
qiaolongfei 628715d602 clean code
7 years ago
qiaolongfei 352af966d7 add python unit test
7 years ago
kavyasrinet 3e2be065b9 Merge pull request #4604 from kavyasrinet/activations
7 years ago
Abhinav Arora 828c5b3e1d Adding Adadelta optimization operator (#4576)
7 years ago
Kavya Srinet 11070e5f36 Updated the reltive error
7 years ago
Kavya Srinet 60af56c1b8 Added Leaky Relu activation
7 years ago
qiaolongfei 5917e09cde tmp work
7 years ago
Kavya Srinet fa12e51675 Adding the default attribute test case
7 years ago
Kavya Srinet 94855f4af0 Fixed changes proposed in the review
7 years ago
Abhinav Arora eed2c1e1d6 Changing SGD inputs and outputs to conform to Operator naming convention (#4586)
7 years ago
Abhinav Arora 324876bbbf Changing learning rate from type Input(float) to Input(tensor) (#4578)
7 years ago
zchen0211 94b94e5b68 Merge branch 'develop' of https://github.com/PaddlePaddle/Paddle into develop
7 years ago
Abhinav Arora 42e7fe05a2 Changing learning rate from attribute to input(float) (#4568)
7 years ago
Kavya Srinet 163d287143 Made learning rate the input
7 years ago
Kexin Zhao d1de7ec630 Change learning rate from attribute to input tensor
7 years ago
Kavya Srinet 61c03f9d59 Adding the implementation for rmsprop operator
7 years ago
zchen0211 58174b12f7 Merge branch 'develop' of https://github.com/PaddlePaddle/Paddle into develop
7 years ago
Kexin Zhao 1ac654a69f Implementing the Adagrad optimizer step operator
7 years ago
qiaolongfei 32f5c9dd93 recurrent_op pass the unit test
7 years ago
zchen0211 15941dbd8c solve conflict for cond_op and scatter
7 years ago
qiaolongfei 7163dd0413 revert code
7 years ago
caoying03 be8bef9bdd Merge branch 'develop' into add_config_helper_for_resize_layer
7 years ago
qiaolongfei af6f3c0423 use float32 in cond_op
7 years ago
Yu Yang 0900aedfa0 Merge pull request #4514 from reyoung/feature/remove_add_op
7 years ago
caoying03 480154896c add configuration helper for resize layer.
7 years ago
chengduo 4f5491b2b4 Merge pull request #4146 from chengduoZH/Add_pool_op
7 years ago
chengduoZH 6abcb74c8f fix unit test class name
7 years ago
Yu Yang aa52fa1c64 Merge pull request #4491 from reyoung/feature/stable_lstm
7 years ago
chengduoZH 2d8a5b97cc fix unit test
7 years ago
Qiao Longfei 7fe0297e64 remove Runtime InferShape for cond op (#4518)
7 years ago
Yu Yang 6164b8986e Fix CI
7 years ago
Yu Yang 762a99cc06 Remove add_op since it can be replaced by sum_op
7 years ago
guosheng 46641c6361 Fix infer when input is empty in v2/inference.py
7 years ago
chengduoZH df59889984 remove conflict
7 years ago
Luo Tao bb7f555803 remove rowwise_add_op
7 years ago
Cao Ying 99130c6e94 Merge pull request #4498 from pengli09/fix-random-seed
7 years ago
Peng Li 4dfc10ccf7 a patch for fixing random seeds in gradient checkers
7 years ago
Liu Yiqun 8bafdda0ad Merge branch 'develop' into core_add_sequence_softmax_op
7 years ago
Yu Yang f60f0eae11 Using double precision to stablize lstm gradient check
7 years ago
Yu Yang 9fbf94b61a Merge pull request #4487 from abhinavarora/softsign_activation
7 years ago
Abhinav Arora 0c3eee09ff Implementing the SoftSign activation operator
7 years ago
Yu Yang 279178e457 Fix bug in test_prelu and test_xe
7 years ago
Yu Yang 54892c0797 Simplify op_test
7 years ago
Yu Yang 61cc3ae4d1 Stablize elementwise_mul by using double precision
7 years ago
Yu Yang 6efcbc4fcb Fix bug in test_prelu and test_xe
7 years ago
zchen0211 88a8eedda1 scatter gather gpu
7 years ago