Commit Graph

948 Commits (ad477b9183fa3c9164efa5cc51bb3ba551d30658)

Author SHA1 Message Date
Yancey1989 ad477b9183 update
7 years ago
Yancey1989 e880a356fe update
7 years ago
Yancey1989 462579c416 update
7 years ago
Yancey1989 a4d410aec8 Merge branch 'develop' of github.com:PaddlePaddle/Paddle into seqconcat_op
7 years ago
Yancey1989 d211b51bd4 update comment
7 years ago
QI JUN 9efd5422f9 Merge pull request #4655 from abhinavarora/fill_constant_op
7 years ago
Tao Luo ceefb555f7 Merge pull request #4500 from luotao1/interp
7 years ago
Abhinav Arora 6efacc14d8 Implementing the fill constant op for the executor
7 years ago
Yu Yang dcb09e932d Use PROTO_LITE when refactoring Paddle
7 years ago
Yu Yang 92add2a29b Fix compile error in develop branch
7 years ago
Qiao Longfei e12ec95ac1 Merge pull request #4630 from jacquesqiao/merge-infershapecontext
7 years ago
Abhinav Arora 4cb5bd9021 Implementing the Adamax optimizer operator (#4538)
7 years ago
kavyasrinet f30a1f42f0 Adding relu6 activation function (#4607)
7 years ago
Luo Tao 597299074e fix bug in REGISTER_OP(reduce_min)
7 years ago
Luo Tao a06f099d9f refine comment of interp_op
7 years ago
Luo Tao 707d144c93 Unify Reduce functions and simplify register code
7 years ago
Luo Tao 5b862fedf1 remove debug log in interp_op.cc
7 years ago
Luo Tao 4724bdbe68 Merge branch 'develop' into interp
7 years ago
qiaolongfei c0a34e1c64 rename InferShapeContextBase to InferShapeContext
7 years ago
Yi Wang 99895730f7 Merge pull request #4609 from kavyasrinet/tanhshrink
7 years ago
Yi Wang 097f533bca Resolve conflict
7 years ago
kexinzhao 087addaa76 Merge pull request #4558 from kexinzhao/adagrad_op
7 years ago
Kexin Zhao 78f4c803f3 change learning rate and fix format
7 years ago
Kavya Srinet f52cdaa0ce Updated RMSProp to have learning rate as an input and work with GPU
7 years ago
Kavya Srinet 0336304176 Merge branch 'develop' of https://github.com/PaddlePaddle/Paddle into rmsprop
7 years ago
Kavya Srinet 154a6ed29c Implementing tanhshrink operator
7 years ago
kavyasrinet 3e2be065b9 Merge pull request #4604 from kavyasrinet/activations
7 years ago
Abhinav Arora 828c5b3e1d Adding Adadelta optimization operator (#4576)
7 years ago
Kavya Srinet 60af56c1b8 Added Leaky Relu activation
7 years ago
Yi Wang 1172f24929 Merge pull request #4590 from wangkuiyi/paddle_only_cpu
7 years ago
qiaolongfei 8ebc31d935 optimize the dsize
7 years ago
qiaolongfei 775c60246b remove using in sgd header file
7 years ago
Yu Yang 2594a50245 Polish code
7 years ago
Yu Yang c4effc7d2d Fix CI Test
7 years ago
qiaolongfei ee7b3ed09e use EigenScalar to get learning_rate from GPU device
7 years ago
Yi Wang 4558807c48 Use PADDLE_WITH_CUDA instead of PADDLE_WITH_GPU
7 years ago
Yi Wang e79d2f1b65 Merge pull request #4584 from reyoung/feature/change_macro_paddle_no_gpu
7 years ago
Kavya Srinet fa12e51675 Adding the default attribute test case
7 years ago
Kavya Srinet 94855f4af0 Fixed changes proposed in the review
7 years ago
Yu Yang e119177a8c Use unique_ptr
7 years ago
Yu Yang 84500f9487 Change `PADDLE_ONLY_CPU` to `PADDLE_WITH_GPU`
7 years ago
Abhinav Arora eed2c1e1d6 Changing SGD inputs and outputs to conform to Operator naming convention (#4586)
7 years ago
Abhinav Arora 324876bbbf Changing learning rate from type Input(float) to Input(tensor) (#4578)
7 years ago
Yu Yang 14a59d2e6b Merge branch 'develop' of github.com:baidu/Paddle into feature/grad_reg_mechanism_cont2
7 years ago
zchen0211 94b94e5b68 Merge branch 'develop' of https://github.com/PaddlePaddle/Paddle into develop
7 years ago
zchen0211 2d876b8643 gather scatter fix according to google style
7 years ago
Abhinav Arora 42e7fe05a2 Changing learning rate from attribute to input(float) (#4568)
7 years ago
Kavya Srinet 163d287143 Made learning rate the input
7 years ago
Kexin Zhao d1de7ec630 Change learning rate from attribute to input tensor
7 years ago
zchen0211 2ccaec4f57 gather scatter cond
7 years ago