|
|
@ -141,9 +141,9 @@ message DistributedStrategy {
|
|
|
|
optional bool fuse_all_reduce_ops = 18 [ default = true ];
|
|
|
|
optional bool fuse_all_reduce_ops = 18 [ default = true ];
|
|
|
|
optional int32 fuse_grad_size_in_MB = 19 [ default = 32 ];
|
|
|
|
optional int32 fuse_grad_size_in_MB = 19 [ default = 32 ];
|
|
|
|
optional float fuse_grad_size_in_TFLOPS = 20 [ default = 50 ];
|
|
|
|
optional float fuse_grad_size_in_TFLOPS = 20 [ default = 50 ];
|
|
|
|
optional bool cudnn_exhaustive_search = 21 [ default = true ];
|
|
|
|
optional bool cudnn_exhaustive_search = 21 [ default = false ];
|
|
|
|
optional int32 conv_workspace_size_limit = 22 [ default = 512 ];
|
|
|
|
optional int32 conv_workspace_size_limit = 22 [ default = 512 ];
|
|
|
|
optional bool cudnn_batchnorm_spatial_persistent = 23 [ default = true ];
|
|
|
|
optional bool cudnn_batchnorm_spatial_persistent = 23 [ default = false ];
|
|
|
|
optional bool adaptive_localsgd = 24 [ default = false ];
|
|
|
|
optional bool adaptive_localsgd = 24 [ default = false ];
|
|
|
|
optional bool fp16_allreduce = 25 [ default = false ];
|
|
|
|
optional bool fp16_allreduce = 25 [ default = false ];
|
|
|
|
optional bool sharding = 26 [ default = false ];
|
|
|
|
optional bool sharding = 26 [ default = false ];
|
|
|
|