ParallelMode¶
- class paddle.distributed. ParallelMode [source]
- 
         There are all the parallel modes currently supported: - DATA_PARALLEL: Distribute input data to different devices. 
- TENSOR_PARALLEL: Shards tensors in the network to different devices. 
- PIPELINE_PARALLEL: Place different layers of the network on different devices. 
- SHARDING_PARALLEL: Segment the model parameters, parameter gradients and optimizer states corresponding to the parameters to each device. 
 Examples import paddle parallel_mode = paddle.distributed.ParallelMode print(parallel_mode.DATA_PARALLEL) # 0 
