|
|
|
@ -11,7 +11,8 @@ export POD_IP=127.0.0.1
|
|
|
|
|
export PADDLE_TRAINERS=127.0.0.1,127.0.0.2
|
|
|
|
|
export PADDLE_TRAINER_ID=0
|
|
|
|
|
|
|
|
|
|
distributed_args="--use_paddlecloud True --cluster_node_ips ${cluster_node_ips} --node_ip ${node_ip} --selected_gpus=0,1 --log_dir testlog"
|
|
|
|
|
distributed_args="--use_paddlecloud --cluster_node_ips=${cluster_node_ips} --node_ip=${node_ip}
|
|
|
|
|
--selected_gpus=0,1 --log_dir=testlog"
|
|
|
|
|
python -m paddle.distributed.launch ${distributed_args} multi_process.py
|
|
|
|
|
|
|
|
|
|
str1="selected_gpus:0 worker_endpoints:127.0.0.1:6170,127.0.0.1:6171,127.0.0.2:6170,127.0.0.2:6171 trainers_num:4 current_endpoint:127.0.0.1:6170 trainer_id:0"
|
|
|
|
|