DistributeTranspilerConfig¶
- class paddle.fluid.transpiler.distribute_transpiler. DistributeTranspilerConfig [source]
- 
         - api_attr
- 
             Static Graph 
 A configuration class that provide support for transpiler distributed jobs. Some important parameters are explained as follows: - slice_var_up ( bool ) ¶
- 
           Whether to do Tensor slice for parameter servers, default is True. 
 - split_method ( PSDispatcher ) ¶
- 
           Methods of dispatching parameters for server, api_fluid_transpiler_RoundRobin or api_fluid_transpiler_HashName can be used and default is RoundRobin. Try to choose the best method to balance loads for parameter servers. 
 - min_block_size ( int ) ¶
- 
           Minimum number of split elements in block, default is 8192. According to : https://github.com/PaddlePaddle/Paddle/issues/8638#issuecomment-369912156 We can use bandwidth efficiently when data size is larger than 2MB.If you want to change it, please be sure you have read the slice_variable function. You can find the definition of slice_variable in https://github.com/PaddlePaddle/Paddle/blob/develop/python/paddle/fluid/transpiler/distribute_transpiler.py . 
 Examples from paddle.fluid.transpiler.ps_dispatcher import RoundRobin import paddle.fluid as fluid config = fluid.DistributeTranspilerConfig() config.slice_var_up = True config.split_method = RoundRobin config.min_block_size = 81920 
