Usage Constraints During Operator Parallel
API name |
constraints |
Config layout constraints |
---|---|---|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Support config layout. Tensor Layout of the inputs should be the same or broadcastable. Output Layout is not open for configuration. |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
When the input_x is splited on the axis dimension, the distributed result may be inconsistent with that on the single machine. |
Not support config layout |
|
When the input_x is splited on the axis dimension, the distributed result may be inconsistent with that on the single machine. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
1. The data format only supports ‘NCHW’; |
Not support config layout |
|
1. The data format only supports ‘NCDHW’; |
Not support config layout |
|
Not support config layout |
||
None |
Not support config layout |
|
It does not support GPU. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Support config layout. The second input, bias, should have ths same tensor layout as the last dimension of input_x. Output Layout is not open for configuration. |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
1. The first dimension of input (anchor_box) and input (groundtruth_box) can be split; |
Not support config layout |
|
None |
Not support config layout |
|
The shard strategy is ignored in the Auto Parallel and Semi Auto Parallel mode. |
Not support config layout |
|
1. The strategy for ‘B’ dimension must be the same; |
Not support config layout |
|
None |
Not support config layout |
|
The input_x can’t be split into the dimension of axis, otherwise it’s inconsistent with the single machine in the mathematical logic. |
Not support config layout |
|
1. The data format only supports ‘NCHW’; |
Not support config layout |
|
1. The data format only supports ‘NCDHW’; |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
1. Sharding of the H/W dimension of input (x) and the second dimension of input (boxes) is not supported. |
Not support config layout |
|
The |
Not support config layout |
|
The same as CumProd. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
The same as Gather. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
1. Set the strategy for |
Not support config layout |
|
1. When batch_dims > 0: |
Not support config layout |
|
The dimension corresponding to dim cannot be segmented; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The first input can’t be split, and the last dimension of the second input can’t be split; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
None |
Support config input layout. Output Layout is not open for configuration. |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
The first dimension of |
Not support config layout |
|
The same as InplaceAdd. |
Not support config layout |
|
The same as InplaceAdd. |
Not support config layout |
|
None |
Not support config layout |
|
The first dimension of the |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
The strategy for the second input gamma and the third input beta needs to be equal to the input_x_strategy[begin_params.axis:], input_x_strategy is the strategy for the first input. |
Support config layout. The layout configuration for the second input gamma and the third input beta needs to be equal to the input_x_layout_tuple[begin_params.axis:], input_x_layout_tuple is the layout configuration for the first input. |
|
None |
Not support config layout |
|
The input_x can’t be split into the dimension of axis, otherwise it’s inconsistent with the single machine in the mathematical logic. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
You don’t need to configure strategy for |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
The logits can’t be split into the dimension of axis, otherwise it’s inconsistent with the single machine in the mathematical logic. |
Not support config layout |
|
None |
Not support config layout |
|
1. When |
Support config layout. |
|
None |
Not support config layout |
|
1. The data format only supports ‘NCHW’; |
Not support config layout |
|
The same as AvgPool3D. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
Only support 1-dim indices. Must configure strategy for the output and the first and second inputs. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
When the shape of weight is not [1], the shard strategy in channel dimension of input_x should be consistent with weight. |
Not support config layout |
|
Only the all-1 strategy is supported. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
When the input_x is splited on the axis dimension, the distributed result may be inconsistent with that on the single machine. |
Not support config layout |
|
When the input_x is splited on the axis dimension, the distributed result may be inconsistent with that on the single machine. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
Configuring sharding strategy is not supported. In auto parallel mode, if multiple operators are followed by the reshape operator, different shard strategys are not allowed to be configured for these operators. |
Not support config layout |
|
None |
Not support config layout |
|
When |
Not support config layout |
|
Sharding the H/W dimension of the input(features) and the second dimension of input(rois) is not supported. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
The second input cannot be split, and the top n dimensions of the third input (n is the dimension of the second input) cannot be split; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The second input cannot be split, and the top n dimensions of the third input (n is the dimension of the second input) cannot be split; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The first dimension of the first input cannot be split, the second input cannot be split, and the top n dimensions of the third input (n is the dimension of the second input) cannot be split; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The first dimension of the first input cannot be split, the second input cannot be split, and the top n dimensions of the third input (n is the dimension of the second input) cannot be split; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The second input cannot be split, and the top n dimensions of the third input (n is the dimension of the second input) cannot be split; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The second input cannot be split, the top n-1 dimension of the third input (n is the dimension of the second input) cannot be split, and the remaining k dimensions (excluding the top n-1 dimension) of the third input are consistent with the last k partitions of the first input; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The second input cannot be split, the top n-1 dimension of the third input (n is the dimension of the second input) cannot be split, and the remaining k dimensions (excluding the top n-1 dimension) of the third input are consistent with the last k partitions of the first input; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The top m dimension of the first input cannot be cut (m is the value of the last dimension of the second input indexes [- 1]). The second input cannot be split. The top n-1 dimension of the third input (n is the dimension of the second input) cannot be split. The partitions of the remaining k dimensions (excluding the top n-1 dimension) of the third input is consistent with the last k partitions of the first input; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The second input cannot be split, and the top n dimensions of the third input (n is the dimension of the second input) cannot be split; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The first dimension of first input can not be split, the second input can not be split, and the first n dimensions (n is the dimension size of the second input) of the third input can not be split; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The second input cannot be split, the top n-1 dimension of the third input (n is the dimension of the second input) cannot be split, and the remaining k dimensions (excluding the top n-1 dimension) of the third input are consistent with the last k partitions of the first input; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The second input cannot be split, the top n-1 dimension of the third input (n is the dimension of the second input) cannot be split, and the remaining k dimensions (excluding the top n-1 dimension) of the third input are consistent with the last k partitions of the first input; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The top m dimension of the first input cannot be cut (m is the value of the last dimension of the second input indexes [- 1]). The second input cannot be split. The top n-1 dimension of the third input (n is the dimension of the second input) cannot be split. The partitions of the remaining k dimensions (excluding the top n-1 dimension) of the third input is consistent with the last k partitions of the first input; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The top m dimension of the first input cannot be cut (m is the value of the last dimension of the second input indexes [- 1]). The second input cannot be split. The top n-1 dimension of the third input (n is the dimension of the second input) cannot be split. The partitions of the remaining k dimensions (excluding the top n-1 dimension) of the third input is consistent with the last k partitions of the first input; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The second input cannot be split, the top n-1 dimension of the third input (n is the dimension of the second input) cannot be split, and the remaining k dimensions (excluding the top n-1 dimension) of the third input are consistent with the last k partitions of the first input; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The second input cannot be split, the top n-1 dimension of the third input (n is the dimension of the second input) cannot be split, and the remaining k dimensions (excluding the top n-1 dimension) of the third input are consistent with the last k partitions of the first input; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
The top m dimension of the first input cannot be cut (m is the value of the last dimension of the second input indexes [- 1]). The second input cannot be split. The top n-1 dimension of the third input (n is the dimension of the second input) cannot be split. The partitions of the remaining k dimensions (excluding the top n-1 dimension) of the third input is consistent with the last k partitions of the first input; In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
In auto_parallel mode, the dual recursive algorithm is not supported. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
The logits can’t be split into the dimension of axis, otherwise it’s inconsistent with the single machine in the mathematical logic. |
Not support config layout |
|
The last dimension of logits and labels can’t be splited; Only supports using output[0]. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
The same as Gather. |
Not support config layout |
|
The input_x can’t be split into the dimension of axis, otherwise it’s inconsistent with the single machine in the mathematical logic. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
Only support mask with all 0 values; The dimension needs to be split should be all extracted; Split is supported when the strides of dimension is 1. |
Not support config layout |
|
The dimension needs to be split should be all extracted. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
Only support configuring shard strategy for multiples. |
Not support config layout |
|
The input_x can’t be split into the last dimension, otherwise it’s inconsistent with the single machine in the mathematical logic. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
Only support the repeat calculate shard strategy (1,). |
Not support config layout |
|
The shard of input_x and segment_ids must be the same as the dimension of segment_ids. |
Not support config layout |
|
The shard of input_x and segment_ids must be the same as the dimension of segment_ids. Note that if the segment id i is missing, then the output[i] will be filled with the maximum of the input type. The user needs to mask the maximum value to avoid value overflow. The communication operation such as AllReudce will raise an Run Task Error due to overflow. |
Not support config layout |
|
The shard of input_x and segment_ids must be the same as the dimension of segment_ids. Note that if the segment id i is missing, then the output[i] will be filled with the minimum of the input type. The user needs to mask the minimum value to avoid value overflow. The communication operation such as AllReudce will raise an Run Task Error due to overflow. |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
|
None |
Not support config layout |
Repeated calculation means that the device is not fully used. For example, the cluster has 8 devices to run distributed training, the splitting strategy only cuts the input into 4 copies. In this case, double counting will occur.