split different comm method for mnist distributed training (#18715)
	
		
	
				
					
				
			* split different comm method for mnist distributed trainingDDDivano-patch-1
							parent
							
								
									ccf06a48b0
								
							
						
					
					
						commit
						ebf9797ec3
					
				@ -0,0 +1,35 @@
 | 
				
			||||
# Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
 | 
				
			||||
#
 | 
				
			||||
# Licensed under the Apache License, Version 2.0 (the "License");
 | 
				
			||||
# you may not use this file except in compliance with the License.
 | 
				
			||||
# You may obtain a copy of the License at
 | 
				
			||||
#
 | 
				
			||||
#     http://www.apache.org/licenses/LICENSE-2.0
 | 
				
			||||
#
 | 
				
			||||
# Unless required by applicable law or agreed to in writing, software
 | 
				
			||||
# distributed under the License is distributed on an "AS IS" BASIS,
 | 
				
			||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
				
			||||
# See the License for the specific language governing permissions and
 | 
				
			||||
# limitations under the License.
 | 
				
			||||
 | 
				
			||||
from __future__ import print_function
 | 
				
			||||
import unittest
 | 
				
			||||
from test_dist_base import TestDistBase
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
class TestDistMnistNCCL2BackWardDeps(TestDistBase):
 | 
				
			||||
    def _setup_config(self):
 | 
				
			||||
        self._sync_mode = True
 | 
				
			||||
        self._use_reduce = False
 | 
				
			||||
        self._use_reader_alloc = False
 | 
				
			||||
        self._nccl2_mode = True
 | 
				
			||||
        self._enable_backward_deps = True
 | 
				
			||||
 | 
				
			||||
    def test_dist_train(self):
 | 
				
			||||
        import paddle.fluid as fluid
 | 
				
			||||
        if fluid.core.is_compiled_with_cuda():
 | 
				
			||||
            self.check_with_place("dist_mnist.py", delta=1e-5)
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
if __name__ == "__main__":
 | 
				
			||||
    unittest.main()
 | 
				
			||||
@ -0,0 +1,35 @@
 | 
				
			||||
# Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
 | 
				
			||||
#
 | 
				
			||||
# Licensed under the Apache License, Version 2.0 (the "License");
 | 
				
			||||
# you may not use this file except in compliance with the License.
 | 
				
			||||
# You may obtain a copy of the License at
 | 
				
			||||
#
 | 
				
			||||
#     http://www.apache.org/licenses/LICENSE-2.0
 | 
				
			||||
#
 | 
				
			||||
# Unless required by applicable law or agreed to in writing, software
 | 
				
			||||
# distributed under the License is distributed on an "AS IS" BASIS,
 | 
				
			||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
				
			||||
# See the License for the specific language governing permissions and
 | 
				
			||||
# limitations under the License.
 | 
				
			||||
 | 
				
			||||
from __future__ import print_function
 | 
				
			||||
import unittest
 | 
				
			||||
from test_dist_base import TestDistBase
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
class TestDistMnistNCCL2DGC(TestDistBase):
 | 
				
			||||
    def _setup_config(self):
 | 
				
			||||
        self._sync_mode = True
 | 
				
			||||
        self._use_reduce = False
 | 
				
			||||
        self._use_reader_alloc = False
 | 
				
			||||
        self._nccl2_mode = True
 | 
				
			||||
        self._use_dgc = True
 | 
				
			||||
 | 
				
			||||
    def test_dist_train(self):
 | 
				
			||||
        import paddle.fluid as fluid
 | 
				
			||||
        if fluid.core.is_compiled_with_cuda():
 | 
				
			||||
            self.check_with_place("dist_mnist.py", delta=1e-5)
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
if __name__ == "__main__":
 | 
				
			||||
    unittest.main()
 | 
				
			||||
@ -0,0 +1,35 @@
 | 
				
			||||
# Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
 | 
				
			||||
#
 | 
				
			||||
# Licensed under the Apache License, Version 2.0 (the "License");
 | 
				
			||||
# you may not use this file except in compliance with the License.
 | 
				
			||||
# You may obtain a copy of the License at
 | 
				
			||||
#
 | 
				
			||||
#     http://www.apache.org/licenses/LICENSE-2.0
 | 
				
			||||
#
 | 
				
			||||
# Unless required by applicable law or agreed to in writing, software
 | 
				
			||||
# distributed under the License is distributed on an "AS IS" BASIS,
 | 
				
			||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
				
			||||
# See the License for the specific language governing permissions and
 | 
				
			||||
# limitations under the License.
 | 
				
			||||
 | 
				
			||||
from __future__ import print_function
 | 
				
			||||
import unittest
 | 
				
			||||
from test_dist_base import TestDistBase
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
class TestDistMnistNCCL2HAllreduce(TestDistBase):
 | 
				
			||||
    def _setup_config(self):
 | 
				
			||||
        self._sync_mode = True
 | 
				
			||||
        self._use_reduce = False
 | 
				
			||||
        self._use_reader_alloc = False
 | 
				
			||||
        self._nccl2_mode = True
 | 
				
			||||
        self._use_hallreduce = True
 | 
				
			||||
 | 
				
			||||
    def test_dist_train(self):
 | 
				
			||||
        import paddle.fluid as fluid
 | 
				
			||||
        if fluid.core.is_compiled_with_cuda():
 | 
				
			||||
            self.check_with_place("dist_mnist.py", delta=1e-5)
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
if __name__ == "__main__":
 | 
				
			||||
    unittest.main()
 | 
				
			||||
@ -0,0 +1,35 @@
 | 
				
			||||
# Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
 | 
				
			||||
#
 | 
				
			||||
# Licensed under the Apache License, Version 2.0 (the "License");
 | 
				
			||||
# you may not use this file except in compliance with the License.
 | 
				
			||||
# You may obtain a copy of the License at
 | 
				
			||||
#
 | 
				
			||||
#     http://www.apache.org/licenses/LICENSE-2.0
 | 
				
			||||
#
 | 
				
			||||
# Unless required by applicable law or agreed to in writing, software
 | 
				
			||||
# distributed under the License is distributed on an "AS IS" BASIS,
 | 
				
			||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
				
			||||
# See the License for the specific language governing permissions and
 | 
				
			||||
# limitations under the License.
 | 
				
			||||
 | 
				
			||||
from __future__ import print_function
 | 
				
			||||
import unittest
 | 
				
			||||
from test_dist_base import TestDistBase
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
class TestDistMnistNCCL2MultiNCCLComm(TestDistBase):
 | 
				
			||||
    def _setup_config(self):
 | 
				
			||||
        self._sync_mode = True
 | 
				
			||||
        self._use_reduce = False
 | 
				
			||||
        self._use_reader_alloc = False
 | 
				
			||||
        self._nccl2_mode = True
 | 
				
			||||
        self._nccl_comm_num = 3
 | 
				
			||||
 | 
				
			||||
    def test_dist_train(self):
 | 
				
			||||
        import paddle.fluid as fluid
 | 
				
			||||
        if fluid.core.is_compiled_with_cuda():
 | 
				
			||||
            self.check_with_place("dist_mnist.py", delta=1e-5)
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
if __name__ == "__main__":
 | 
				
			||||
    unittest.main()
 | 
				
			||||
@ -1,90 +0,0 @@
 | 
				
			||||
# Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
 | 
				
			||||
#
 | 
				
			||||
# Licensed under the Apache License, Version 2.0 (the "License");
 | 
				
			||||
# you may not use this file except in compliance with the License.
 | 
				
			||||
# You may obtain a copy of the License at
 | 
				
			||||
#
 | 
				
			||||
#     http://www.apache.org/licenses/LICENSE-2.0
 | 
				
			||||
#
 | 
				
			||||
# Unless required by applicable law or agreed to in writing, software
 | 
				
			||||
# distributed under the License is distributed on an "AS IS" BASIS,
 | 
				
			||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
				
			||||
# See the License for the specific language governing permissions and
 | 
				
			||||
# limitations under the License.
 | 
				
			||||
 | 
				
			||||
from __future__ import print_function
 | 
				
			||||
import unittest
 | 
				
			||||
from test_dist_base import TestDistBase
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
class TestDistMnistNCCL2(TestDistBase):
 | 
				
			||||
    def _setup_config(self):
 | 
				
			||||
        self._sync_mode = True
 | 
				
			||||
        self._use_reduce = False
 | 
				
			||||
        self._use_reader_alloc = False
 | 
				
			||||
        self._nccl2_mode = True
 | 
				
			||||
 | 
				
			||||
    def test_dist_train(self):
 | 
				
			||||
        import paddle.fluid as fluid
 | 
				
			||||
        if fluid.core.is_compiled_with_cuda():
 | 
				
			||||
            self.check_with_place("dist_mnist.py", delta=1e-5)
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
class TestDistMnistNCCL2MultiNCCLComm(TestDistBase):
 | 
				
			||||
    def _setup_config(self):
 | 
				
			||||
        self._sync_mode = True
 | 
				
			||||
        self._use_reduce = False
 | 
				
			||||
        self._use_reader_alloc = False
 | 
				
			||||
        self._nccl2_mode = True
 | 
				
			||||
        self._nccl_comm_num = 3
 | 
				
			||||
 | 
				
			||||
    def test_dist_train(self):
 | 
				
			||||
        import paddle.fluid as fluid
 | 
				
			||||
        if fluid.core.is_compiled_with_cuda():
 | 
				
			||||
            self.check_with_place("dist_mnist.py", delta=1e-5)
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
class TestDistMnistNCCL2DGC(TestDistBase):
 | 
				
			||||
    def _setup_config(self):
 | 
				
			||||
        self._sync_mode = True
 | 
				
			||||
        self._use_reduce = False
 | 
				
			||||
        self._use_reader_alloc = False
 | 
				
			||||
        self._nccl2_mode = True
 | 
				
			||||
        self._use_dgc = True
 | 
				
			||||
 | 
				
			||||
    def test_dist_train(self):
 | 
				
			||||
        import paddle.fluid as fluid
 | 
				
			||||
        if fluid.core.is_compiled_with_cuda():
 | 
				
			||||
            self.check_with_place("dist_mnist.py", delta=1e-5)
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
class TestDistMnistNCCL2BackWardDeps(TestDistBase):
 | 
				
			||||
    def _setup_config(self):
 | 
				
			||||
        self._sync_mode = True
 | 
				
			||||
        self._use_reduce = False
 | 
				
			||||
        self._use_reader_alloc = False
 | 
				
			||||
        self._nccl2_mode = True
 | 
				
			||||
        self._enable_backward_deps = True
 | 
				
			||||
 | 
				
			||||
    def test_dist_train(self):
 | 
				
			||||
        import paddle.fluid as fluid
 | 
				
			||||
        if fluid.core.is_compiled_with_cuda():
 | 
				
			||||
            self.check_with_place("dist_mnist.py", delta=1e-5)
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
class TestDistMnistNCCL2HAllreduce(TestDistBase):
 | 
				
			||||
    def _setup_config(self):
 | 
				
			||||
        self._sync_mode = True
 | 
				
			||||
        self._use_reduce = False
 | 
				
			||||
        self._use_reader_alloc = False
 | 
				
			||||
        self._nccl2_mode = True
 | 
				
			||||
        self._use_hallreduce = True
 | 
				
			||||
 | 
				
			||||
    def test_dist_train(self):
 | 
				
			||||
        import paddle.fluid as fluid
 | 
				
			||||
        if fluid.core.is_compiled_with_cuda():
 | 
				
			||||
            self.check_with_place("dist_mnist.py", delta=1e-5)
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
if __name__ == "__main__":
 | 
				
			||||
    unittest.main()
 | 
				
			||||
@ -0,0 +1,34 @@
 | 
				
			||||
# Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
 | 
				
			||||
#
 | 
				
			||||
# Licensed under the Apache License, Version 2.0 (the "License");
 | 
				
			||||
# you may not use this file except in compliance with the License.
 | 
				
			||||
# You may obtain a copy of the License at
 | 
				
			||||
#
 | 
				
			||||
#     http://www.apache.org/licenses/LICENSE-2.0
 | 
				
			||||
#
 | 
				
			||||
# Unless required by applicable law or agreed to in writing, software
 | 
				
			||||
# distributed under the License is distributed on an "AS IS" BASIS,
 | 
				
			||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
				
			||||
# See the License for the specific language governing permissions and
 | 
				
			||||
# limitations under the License.
 | 
				
			||||
 | 
				
			||||
from __future__ import print_function
 | 
				
			||||
import unittest
 | 
				
			||||
from test_dist_base import TestDistBase
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
class TestDistMnistNCCL2(TestDistBase):
 | 
				
			||||
    def _setup_config(self):
 | 
				
			||||
        self._sync_mode = True
 | 
				
			||||
        self._use_reduce = False
 | 
				
			||||
        self._use_reader_alloc = False
 | 
				
			||||
        self._nccl2_mode = True
 | 
				
			||||
 | 
				
			||||
    def test_dist_train(self):
 | 
				
			||||
        import paddle.fluid as fluid
 | 
				
			||||
        if fluid.core.is_compiled_with_cuda():
 | 
				
			||||
            self.check_with_place("dist_mnist.py", delta=1e-5)
 | 
				
			||||
 | 
				
			||||
 | 
				
			||||
if __name__ == "__main__":
 | 
				
			||||
    unittest.main()
 | 
				
			||||
					Loading…
					
					
				
		Reference in new issue