# Copyright 2024 Huawei Technologies Co., Ltd
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# ============================================================================
"""
Defines communication operators with functional form.
"""
from mindspore.communication import GlobalComm, get_group_rank_from_world_rank, get_group_size
from mindspore.communication.management import _get_group
from mindspore.communication._comm_helper import _get_group_rank_from_world_rank_from_cache_helper
from mindspore.common.tensor import Tensor
from mindspore._c_expression import Tensor as Tensor_
from mindspore.ops import ReduceOp, cat
from mindspore.ops._primitive_cache import _get_cache_prim
from mindspore.ops.primitive import _primexpr
from mindspore.ops.auto_generate.gen_ops_prim import (inner_comm_all_reduce_op, inner_comm_all_gather_op,
inner_comm_all_to_all_v_op, inner_comm_irecv_op,
inner_comm_isend_op, inner_comm_reduce_scatter_op)
from mindspore._c_expression import CommHandle as CommHandle_
from mindspore._c_expression.typing import Type
from mindspore import jit_class
__all__ = [
'all_reduce',
'all_gather_into_tensor',
'all_to_all_with_output_shape',
'all_to_all_single_with_output_shape',
'barrier',
'broadcast',
'gather_into_tensor',
'isend',
'irecv',
'reduce_scatter_tensor',
'reduce',
'scatter_tensor',
'send',
'recv',
'P2POp',
'batch_isend_irecv',
]
import mindspore.ops.operations as P
_GROPU_SIZE_CACHE = {}
@jit_class
class CommHandle(CommHandle_):
r"""
Usually, handles are created in C++during the execution of communication operators and returned to the Python
layer. It will not be created directly in Python. Only in scenarios where graph patterns are compatible,
handles will be created using Python.
"""
def wait(self):
r"""
The wait for asynchronous handles will not take effect for handles created on the Python side.
>>> import numpy as np
>>> from mindspore.communication import init
>>> from mindspore.communication.comm_func import all_reduce
>>> from mindspore import Tensor
>>>
>>> init()
>>> input_tensor = Tensor(np.ones([2, 8]).astype(np.float32))
>>> output, handle = all_reduce(input_tensor, async_op=True)
>>> handle.wait()
>>> print(output)
[[2. 2. 2. 2. 2. 2. 2. 2.]
[2. 2. 2. 2. 2. 2. 2. 2.]]
"""
default_handle = CommHandle()
def _check_split_sizes_sequence(tensor, sequence):
if not sequence:
raise TypeError(f"sequence can not be empty list.")
element0 = sequence[0]
for idx in range(1, len(sequence)):
if sequence[idx] != element0:
raise TypeError(f"sequence containing different elements is not supported yet. "
f"Elements must be the same.")
if sum(sequence) != tensor.shape[0]:
raise TypeError(f" The sum of sequence should equal to tensor.shape[0].")
def _check_compute_split_count(tensor, output_split_sizes, input_split_sizes, group):
"""
Check the output_split_sizes and input_split_sizes by the rules in _check_split_sizes_sequence,
compute the split count and return it.
"""
group_size = get_group_size(group)
if output_split_sizes:
_check_split_sizes_sequence(tensor, output_split_sizes)
output_split_value = output_split_sizes[0]
else:
output_split_value = None
if input_split_sizes:
_check_split_sizes_sequence(tensor, input_split_sizes)
input_split_value = input_split_sizes[0]
else:
input_split_value = None
split_count = 0
if input_split_value and output_split_value is None:
split_count = tensor.shape[0] // input_split_value
elif input_split_value is None and output_split_value:
split_count = tensor.shape[0] // output_split_value
elif input_split_value and output_split_value:
if input_split_value != output_split_value:
raise TypeError(f"input_split_value should equal to output_split_value.")
split_count = tensor.shape[0] // input_split_value
else:
split_count = group_size
return split_count
@_primexpr
def _check_all_tensors(tensor_list):
"""check all elements in tensor_list are type of Tensor"""
if not isinstance(tensor_list, (list, tuple)):
raise TypeError(f"Expected list or tuple, but got {type(tensor_list)}.")
for t in tensor_list:
if not isinstance(t, Tensor):
raise TypeError(f"Expected tensor, but got {type(t)}")
@_primexpr
def _check_all_tensors_or_tuple(tensor_list):
"""check all elements in tensor_list are type of Tensor or tuple or list"""
if not isinstance(tensor_list, (list, tuple)):
raise TypeError(f"Expected list or tuple, but got {type(tensor_list)}.")
for t in tensor_list:
if not isinstance(t, (Tensor, tuple, list)):
raise TypeError(f"Expected tensor or tuple, but got {type(t)}")
@_primexpr
def _check_all_tensor_same_dtype(*tensor_lists):
"""check all the input tensor has same dtype"""
consistent_dtype = None
for list_ in tensor_lists:
if not isinstance(list_, (list, tuple)):
list_ = [list_]
for tensor_ in list_:
if not isinstance(tensor_, Tensor):
continue
dtype = tensor_.dtype
if consistent_dtype is None:
consistent_dtype = dtype
else:
if dtype != consistent_dtype:
raise TypeError("all_to_all input dtype must be the same, "
f"but got {consistent_dtype} and {dtype}.")
def _get_size(shape):
numel = 1
for s in shape:
numel *= s
return numel
def _is_split_sizes_empty(split_sizes):
return split_sizes is None or not split_sizes
def _contiguous(tensor):
if not tensor.is_contiguous() or tensor.storage_offset() != 0:
tensor = tensor.contiguous()
return tensor
[文档]def all_reduce(tensor, op=ReduceOp.SUM, group=GlobalComm.WORLD_COMM_GROUP, async_op=False):
"""
Reduce tensors across all devices in such a way that all deviceswill get the same final result,
returns the tensor which is all reduced.
Note:
The tensors must have the same shape and format in all processes of the collection.
Args:
tensor (Tensor): The input tensor to be all reduced. The shape of tensor is :math:`(x_1, x_2, ..., x_R)`.
op (str, optional): Specifies an operation used for element-wise reductions, like sum, prod, max, and min.
On the CPU, only 'sum' is supported. Default: ``ReduceOp.SUM`` .
group (str, optional): The communication group to work on. Default: ``GlobalComm.WORLD_COMM_GROUP`` , which
means ``"hccl_world_group"`` in Ascend, and ``"nccl_world_group"`` in GPU.
async_op (bool, optional): Whether this operator should be an async operator. Default: ``False`` .
Returns:
Tuple(Tensor, CommHandle), the output tensor has the same shape of the input,
i.e., :math:`(x_1, x_2, ..., x_R)`. The contents depend on the specified operation.
CommHandle is an async work handle, if `async_op` is set to True. CommHandle will be None,
when `async_op` is False.
Raises:
TypeError: If the type of the first input parameter is not Tensor, or any of `op` and `group` is not a str.
RuntimeError: If device target is invalid, or backend is invalid, or distributed initialization fails.
Supported Platforms:
``Ascend``
Examples:
.. note::
Before running the following examples, you need to configure the communication environment variables.
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
without any third-party or configuration file dependencies.
Please see the `msrun start up
<https://www.mindspore.cn/docs/zh-CN/master/model_train/parallel/msrun_launcher.html>`_
for more details.
This example should be run with 2 devices.
>>> import numpy as np
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
>>> comm.init()
>>> input_tensor = ms.Tensor(np.ones([2, 8]).astype(np.float32))
>>> output = comm.comm_func.all_reduce(input_tensor)
>>> print(output)
[[2. 2. 2. 2. 2. 2. 2. 2.]
[2. 2. 2. 2. 2. 2. 2. 2.]]
"""
if not isinstance(tensor, (Tensor, Tensor_)):
raise TypeError("For all_reduce, the input tensor must be tensor")
if not isinstance(op, str):
raise TypeError("For all_reduce, the input op type must be str")
if op not in ('sum', 'prod', 'min', 'max'):
raise TypeError("For all_reduce, the input op value must be one of sum, prod, min, max")
group = _get_group(group)
tensor = _contiguous(tensor)
output = inner_comm_all_reduce_op(tensor, op, group)
return _deal_comm_outputs(output, async_op)
[文档]def all_gather_into_tensor(tensor, group=GlobalComm.WORLD_COMM_GROUP, async_op=False):
"""
Gathers tensors from the specified communication group and returns the tensor which is all gathered.
Note:
- The tensors must have the same shape and format in all processes of the collection.
Args:
tensor (Tensor): The input tensor to be all gathered into tensor.
The shape of tensor is :math:`(x_1, x_2, ..., x_R)`.
group (str, optional): The communication group to work on. Default: ``GlobalComm.WORLD_COMM_GROUP`` , which
means ``"hccl_world_group"`` in Ascend, and ``"nccl_world_group"`` in GPU.
async_op (bool, optional): Whether this operator should be an async operator. Default: ``False`` .
Returns:
Tuple(Tensor, CommHandle), if the number of devices in the group is N,
then the shape of output tensor is :math:`(N, x_1, x_2, ..., x_R)`.
CommHandle is an async work handle, if `async_op` is set to True.
CommHandle will be None, when `async_op` is False.
Raises:
TypeError: If the type of the first input parameter is not Tensor, or `group` is not a str.
ValueError: If the local rank id of the calling process in the group
is larger than the group's rank size.
RuntimeError: If device target is invalid, or backend is invalid, or distributed initialization fails.
Supported Platforms:
``Ascend``
Examples:
.. note::
Before running the following examples, you need to configure the communication environment variables.
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
without any third-party or configuration file dependencies.
Please see the `msrun start up
<https://www.mindspore.cn/docs/zh-CN/master/model_train/parallel/msrun_launcher.html>`_
for more details.
This example should be run with 2 devices.
>>> import numpy as np
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
>>> ms.set_context(mode=ms.GRAPH_MODE)
>>> comm.init()
>>> input_tensor = ms.Tensor(np.ones([2, 8]).astype(np.float32))
>>> output = comm.comm_func.all_gather_into_tensor(input_tensor)
>>> print(output)
[[1. 1. 1. 1. 1. 1. 1. 1.]
[1. 1. 1. 1. 1. 1. 1. 1.]
[1. 1. 1. 1. 1. 1. 1. 1.]
[1. 1. 1. 1. 1. 1. 1. 1.]]
"""
if not isinstance(tensor, (Tensor, Tensor_)):
raise TypeError("For all_gather_into_tensor, the input tensor must be tensor")
group = _get_group(group)
global _GROPU_SIZE_CACHE
if group not in _GROPU_SIZE_CACHE:
_GROPU_SIZE_CACHE[group] = get_group_size(group)
group_size = _GROPU_SIZE_CACHE[group]
tensor = _contiguous(tensor)
output = inner_comm_all_gather_op(tensor, group_size, group)
return _deal_comm_outputs(output, async_op)
[文档]def reduce_scatter_tensor(tensor, op=ReduceOp.SUM, group=GlobalComm.WORLD_COMM_GROUP, async_op=False):
r"""
Reduces and scatters tensors from the specified communication group and
returns the tensor which is reduced and scattered.
Note:
The tensors must have the same shape and format in all processes of the collection.
Args:
tensor(Tensor): The input tensor to be reduced and scattered, suppose it has a shape :math:`(N, *)`, where `*`
means any number of additional dimensions. N must be divisible by rank_size.
rank_size refers to the number of cards in the communication group.
op (str, optional): Specifies an operation used for element-wise reductions,
like SUM and MAX. Default: ``ReduceOp.SUM`` .
group (str, optional): The communication group to work on. Default: ``GlobalComm.WORLD_COMM_GROUP`` , which
means ``"hccl_world_group"`` in Ascend, and ``"nccl_world_group"`` in GPU.
async_op (bool, optional): Whether this operator should be an async operator. Default: ``False`` .
Returns:
Tuple(Tensor, CommHandle), the output tensor has the same dtype as `input_x` with a shape of
:math:`(N/rank\_size, *)`. CommHandle is an async work handle, if `async_op` is set to True.
CommHandle will be None, when `async_op` is False.
Raises:
TypeError: If the type of the first input parameter is not Tensor, or any of `op` and `group` is not a str.
ValueError: If the first dimension of the input cannot be divided by the rank_size.
RuntimeError: If device target is invalid, or backend is invalid, or distributed initialization fails.
Supported Platforms:
``Ascend``
Examples:
.. note::
Before running the following examples, you need to configure the communication environment variables.
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
without any third-party or configuration file dependencies.
Please see the `msrun start up
<https://www.mindspore.cn/docs/zh-CN/master/model_train/parallel/msrun_launcher.html>`_
for more details.
This example should be run with 2 devices.
>>> import numpy as np
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
>>> ms.set_context(mode=ms.GRAPH_MODE)
>>> comm.init()
>>> input_tensor = ms.Tensor(np.ones([8, 8]).astype(np.float32))
>>> output = comm.comm_func.reduce_scatter_tensor(input_tensor)
>>> print(output)
[[2. 2. 2. 2. 2. 2. 2. 2.]
[2. 2. 2. 2. 2. 2. 2. 2.]
[2. 2. 2. 2. 2. 2. 2. 2.]
[2. 2. 2. 2. 2. 2. 2. 2.]]
"""
if not isinstance(tensor, (Tensor, Tensor_)):
raise TypeError("For reduce_scatter_tensor, the input tensor must be tensor")
group = _get_group(group)
global _GROPU_SIZE_CACHE
if group not in _GROPU_SIZE_CACHE:
_GROPU_SIZE_CACHE[group] = get_group_size(group)
rank_size = _GROPU_SIZE_CACHE[group]
tensor = _contiguous(tensor)
output = inner_comm_reduce_scatter_op(tensor, rank_size, op, group)
return _deal_comm_outputs(output, async_op)
[文档]def reduce(tensor, dst, op=ReduceOp.SUM, group=GlobalComm.WORLD_COMM_GROUP):
"""
Reduces tensors across the processes in the specified communication group, sends the result
to the target dst(global rank), and returns the tensor which is sent to the target process.
Note:
Only process with destination rank receives the reduced output.
Only support PyNative mode, Graph mode is not currently supported.
Other processes only get a tensor with shape [1], which has no mathematical meaning.
Args:
tensor (Tensor): The input tensor to be reduced. The shape of tensor is :math:`(x_1, x_2, ..., x_R)`.
dst (int): The target rank of the process(global rank) that receives the reduced output.
op (str, optional): Specifies an operation used for element-wise reductions, like sum, prod, max, and min.
On the CPU, only 'sum' is supported. Default: ``ReduceOp.SUM`` .
group (str, optional): The communication group to work on. Default: ``GlobalComm.WORLD_COMM_GROUP`` , which
means ``"hccl_world_group"`` in Ascend, and ``"nccl_world_group"`` in GPU.
Returns:
Tensor. Return the tensor in the specific rank of the process after reduction.
The shape of tensor is :math:`(x_1, x_2, ..., x_R)`.
Raises:
TypeError: If the type of the first input parameter is not Tensor, or any of `op` and `group` is not a str.
RuntimeError: If device target is invalid, or backend is invalid, or distributed initialization fails.
Supported Platforms:
``Ascend``
Examples:
.. note::
Before running the following examples, you need to configure the communication environment variables.
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
without any third-party or configuration file dependencies.
Please see the `msrun start up
<https://www.mindspore.cn/docs/zh-CN/master/model_train/parallel/msrun_launcher.html>`_
for more details.
This example should be run with 4 devices.
>>> import numpy as np
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
>>> # Launch 4 processes.
>>> comm.init()
>>> dest_rank=1
>>> input_tensor = ms.Tensor(np.ones([2, 8]).astype(np.float32))
>>> output = comm.comm_func.reduce(input_tensor)
>>> print(output)
Process with rank 1: [[4. 4. 4. 4. 4. 4. 4. 4.]
[4. 4. 4. 4. 4. 4. 4. 4.]],
Other proesses: [0.].
"""
if not isinstance(tensor, (Tensor, Tensor_)):
raise TypeError("For reduce, the input tensor must be tensor")
group_rank = get_group_rank_from_world_rank(dst, group)
reduce_op = _get_cache_prim(P.Reduce)(dest_rank=group_rank, op=op, group=group)
return reduce_op(tensor)
[文档]class P2POp:
"""
Object for `batch_isend_irecv` input, to store information of ``"isend"`` and ``"irecv"``.
Note:
- Allow pass-in recv shape rather than tensor when `op` is 'irecv'.
- `tensor` will not be modified in-place by final result.
Args:
op(Union[str, function]): Only string of ``"isend"`` and ``"irecv"`` are allow.
Or function of ``comm_func.isend`` and ``comm_func.irecv`` are allow.
tensor(Union[Tensor, Tuple(int)]): tensor for sending/receiving or receive tensor shape
when `op` is ``"irecv"``.
peer(int): remote global rank for send/receive.
group (str, optional): The communication group to work on. Default: ``GlobalComm.WORLD_COMM_GROUP`` , which
means ``"hccl_world_group"`` in Ascend, and ``"nccl_world_group"`` in GPU.
tag(int, optional): currently not supported yet. default: ``0``.
Keyword Args:
recv_dtype (mindspore.dtype, optional): when `tensor` is a tuple shape, this arg will be used and has
to be configured. default: ``None``
Returns:
P2POp Object.
Raises:
ValueError: when `op` is not string or function of 'isend' and 'irecv'.
TypeError: when `tensor` is not type of Tensor or Tuple.
NotImplementedError: when `tag` is not 0.
Supported Platforms:
``Ascend``
Examples:
>>> import numpy as np
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
>>> send_tensor = ms.Tensor(1.)
>>> send_op = comm.comm_func.P2POp('isend', send_tensor, 1)
>>> send_op = comm.comm_func.P2POp(isend, send_tensor, 1)
>>> recv_tensor = Tensor(0.)
>>> recv_op = comm.comm_func.P2POp('irecv', recv_tensor, 0)
>>> recv_op = comm.comm_func.P2POp(irecv, recv_tensor, 0)
>>> recv_op = comm.comm_func.P2POp('irecv', (), 0, recv_dtype=mindspore.float32)
"""
def __init__(self, op, tensor, peer, group=None, tag=0, *, recv_dtype=None):
self.op = op
self.tensor = tensor
if not isinstance(peer, int):
raise TypeError(f"peer must be type of int, but got type of {type(peer)}")
if recv_dtype and not isinstance(recv_dtype, Type):
raise TypeError(f"recv_dtype must be type of mindspore dtype, but got type of {type(recv_dtype)}")
self.peer = peer
self.group = group
self.tag = tag
self.recv_dtype = recv_dtype
def __new__(cls, op, tensor, peer, group=None, tag=0, recv_dtype=None):
if not (isinstance(op, str) or callable(op)):
raise TypeError(f"op must be type of string or function, but got type of {type(op)}")
if isinstance(op, str):
op_name = op
else:
op_name = op.__name__
if op_name not in ['isend', 'irecv']:
raise ValueError(f"Expected ``op`` to be of type ``isend`` or `irecv``, but got {op_name}")
if not isinstance(tensor, (Tensor, tuple)):
raise TypeError(f"Expected ``tensor`` to be type of tuple or Tensor, but got {type(tensor)}.")
if tag != 0:
raise NotImplementedError("``tag`` not support yet.")
return object.__new__(cls)
[文档]def batch_isend_irecv(p2p_op_list):
"""
Batch send and recv tensors asynchronously.
Note:
- The 'isend' and 'irecv' of `P2POp` in `p2p_op_list` between ranks need to match each other.
- `P2POp` in `p2p_op_list` can only use the same communication group.
- `tag` of `P2POp` in `p2p_op_list` is not support yet.
- `tensor` of `P2POp` in `p2p_op_list` will not be modified by result inplace.
- Only support PyNative mode, Graph mode is not currently supported.
Args:
p2p_op_list(P2POp): list contains `P2POp`. `P2POp` is type of :class:`mindspore.communication.comm_func.P2POp`
Returns:
tuple(Tensor). Output tensors is corresponding to `p2p_op_list`.
At `P2POp` with 'isend' position, output tensor is a fake tensor with scalar, which has no meaning.
At `P2POp` with 'irecv' position, output tensor is a tensor received from remote device.
Raises:
TypeError: If `p2p_op_list` are not all type of `P2POp`.
Supported Platforms:
``Ascend``
Examples:
.. note::
Before running the following examples, you need to configure the communication environment variables.
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
without any third-party or configuration file dependencies.
Please see the `msrun start up
<https://www.mindspore.cn/docs/zh-CN/master/model_train/parallel/msrun_launcher.html>`_
for more details.
This example should be run with 2 devices.
>>> import numpy as np
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
>>> comm.init()
>>> this_rank = comm.get_rank()
>>> world_size = comm.get_group_size()
>>> next_rank = (this_rank + 1) % world_size
>>> prev_rank = (this_rank + world_size - 1) % world_size
>>>
>>> send_tensor = ms.Tensor(this_rank + 1, dtype=mindspore.float32)
>>> recv_tensor = ms.Tensor(0., dtype=mindspore.float32)
>>>
>>> send_op = comm.comm_func.P2POp('isend', send_tensor, next_rank)
>>> recv_op = comm.comm_func.P2POp('irecv', recv_tensor, prev_rank)
>>>
>>> p2p_op_list = [send_op, recv_op]
>>> output = comm.comm_func.batch_isend_irecv(p2p_op_list)
>>> print(output)
rank 0:
(Tensor(shape=[], dtype=Float32, value= 0), Tensor(shape=[], dtype=Float32, value= 2))
rank 1:
(Tensor(shape=[], dtype=Float32, value= 0), Tensor(shape=[], dtype=Float32, value= 1))
"""
send_tensors = []
op_types = []
remotes_ranks = []
receive_shapes = []
receive_dtypes = []
tags = []
if not isinstance(p2p_op_list, list):
raise TypeError(f"p2p_op_list must be type of list, but got type of {p2p_op_list}.")
if not p2p_op_list:
raise TypeError(f"p2p_op_list can not be empty list.")
group = p2p_op_list[0].group
if group is None:
group = GlobalComm.WORLD_COMM_GROUP
type_ = None
for i, p2p_op in enumerate(p2p_op_list):
if not isinstance(p2p_op, P2POp):
raise TypeError("must be type of P2POp")
if isinstance(p2p_op.op, str):
type_ = p2p_op.op
else:
type_ = p2p_op.op.__name__
rank_ = p2p_op.peer if p2p_op.group is None else \
get_group_rank_from_world_rank(p2p_op.peer, p2p_op.group)
remotes_ranks.append(rank_)
tags.append(p2p_op.tag)
if type_ == "isend":
send_tensors.append(p2p_op.tensor)
elif type_ == "irecv":
if isinstance(p2p_op.tensor, Tensor):
receive_shapes.append(p2p_op.tensor.shape)
receive_dtypes.append(p2p_op.tensor.dtype)
elif isinstance(p2p_op.tensor, tuple):
receive_shapes.append(p2p_op.tensor)
if p2p_op.recv_dtype is None:
raise ValueError(f"'recv_dtype' of {i}th P2POp in p2p_op_list is None but op_types is"
"'irecv' and P2POp.tensor is a tuple type.")
receive_dtypes.append(p2p_op.recv_dtype)
else:
raise TypeError("p2p_op.tensor must be tensor or shape")
else:
raise TypeError("p2p_op.op must be isend or irecv")
op_types.append(type_)
_op = _get_cache_prim(P.BatchISendIRecv)(op_types,
remotes_ranks,
receive_shapes,
receive_dtypes,
group)
output = _op(send_tensors)
return output
[文档]def scatter_tensor(tensor, src=0, group=GlobalComm.WORLD_COMM_GROUP):
r"""
Scatter tensor evently across the processes in the specified communication group.
Note:
The interface behavior only support Tensor input and scatter evenly, which
is different from that of `pytoch.distributed.scatter`.
Only the tensor in process `src` (global rank) will do scatter.
Only support PyNative mode, Graph mode is not currently supported.
Args:
tensor (Tensor): The input tensor to be scattered. The shape of tensor is :math:`(x_1, x_2, ..., x_R)`.
src (int, optional): Specifies the rank(global rank) of the process that send the tensor.
And only process `src` will send the tensor.
group (str, optional): The communication group to work on.
Default: "GlobalComm.WORLD_COMM_GROUP".
Returns:
Tensor, the shape of output is :math:`(x_1/src\_rank, x_2, ..., x_R)`. The dimension 0 of data is equal to
the dimension of input tensor divided by `src`, and the other dimension keep the same.
Raise:
TypeError: If the type of the first input parameter is not Tensor, or any of `op` and `group` is not a str.
RuntimeError: If device target is invalid, or backend is invalid, or distributed initialization fails.
Supported Platforms:
``Ascend`` ``GPU``
Examples:
.. note::
Before running the following examples, you need to configure the communication environment variables.
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
without any third-party or configuration file dependencies.
Please see the `msrun start up
<https://www.mindspore.cn/docs/zh-CN/master/model_train/parallel/msrun_launcher.html>`_
for more details.
This example should be run with 2 devices.
>>> import numpy as np
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
>>> # Launch 2 processes.
>>>
>>> comm.init()
>>> input = ms.Tensor(np.arange(8).reshape([4, 2]).astype(np.float32))
>>> out = comm.comm_func.scatter_tensor(tensor=data, src=0)
>>> print(out)
# rank_0
[[0. 1.]
[2. 3.]]
# rank_1
[[4. 5.]
[6. 7.]]
"""
if not isinstance(tensor, (Tensor, Tensor_)):
raise TypeError("For scatter_tensor, the input tensor must be tensor")
if not isinstance(src, int):
raise TypeError("For scatter_tensor, the src must be int")
_src = get_group_rank_from_world_rank(src, group)
_op = _get_cache_prim(P.CollectiveScatter)(_src, group)
return _op(tensor)
[文档]def gather_into_tensor(tensor, dst=0, group=GlobalComm.WORLD_COMM_GROUP):
r"""
Gathers tensors from the specified communication group. The operation will gather the tensor
from processes according to dimension 0.
Note:
Only the tensor in process `dst` (global rank) will keep the gathered tensor. The other process
will keep a tensor with shape [1], which has no mathematical meaning.
Only support PyNative mode, Graph mode is not currently supported.
Args:
tensor (Tensor): The tensor to be gathered. The shape of tensor is :math:`(x_1, x_2, ..., x_R)`.
dst(int, optional): Specifies the rank(global rank) of the process that receive the tensor.
And only process `dst` will receive the gathered tensor. Default: 0.
group (str, optional): The communication group to work on. Default: ``GlobalComm.WORLD_COMM_GROUP``.
Returns:
Tensor, the shape of output is :math:`(\sum x_1, x_2, ..., x_R)`. The dimension 0 of data is equal to
sum of the dimension of input tensor, and the other dimension keep the same.
Raise:
TypeError: If the type of the first input parameter is not Tensor, or any of `op` and `group` is not a str.
RuntimeError: If device target is invalid, or backend is invalid, or distributed initialization fails.
Supported Platforms:
``Ascend``
Examples:
.. note::
Before running the following examples, you need to configure the communication environment variables.
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
without any third-party or configuration file dependencies.
Please see the `msrun start up
<https://www.mindspore.cn/docs/zh-CN/master/model_train/parallel/msrun_launcher.html>`_
for more details.
This example should be run with 2 devices.
>>> import numpy as np
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
>>> # Launch 2 processes.
>>>
>>> comm.init()
>>> input = ms.Tensor(np.arange(4).reshape([2, 2]).astype(np.float32))
>>> output = comm.comm_func.gather_into_tensor(tensor=data, dst=0)
>>> print(output)
Process with rank 0: [[0. 1.],
[2. 3.],
[0. 1.],
[2. 3.]]
Process with rank 1: [0]
"""
if not isinstance(tensor, (Tensor, Tensor_)):
raise TypeError("For gather_into_tensor, the input tensor must be tensor")
if not isinstance(dst, int):
raise TypeError("For gather_into_tensor, the dst must be int")
_dst = get_group_rank_from_world_rank(dst, group)
_op = _get_cache_prim(P.CollectiveGather)(_dst, group)
return _op(tensor)
[文档]def broadcast(tensor, src=0, group=GlobalComm.WORLD_COMM_GROUP):
"""
Broadcasts the tensor to the whole group.
Note:
The tensors must have the same shape and format in all processes of the collection.
Only support PyNative mode, Graph mode is not currently supported.
Args:
tensor (Tensor): The tensor to be broadcasted. The shape of tensor is :math:`(x_1, x_2, ..., x_R)`.
src (int, optional): Specifies the rank(global rank) of the process that broadcast the tensor.
And only process `src` will broadcast the tensor.
group (str, optional): The communication group to work on. Default: ``GlobalComm.WORLD_COMM_GROUP``.
Returns:
Tensor, tensor has the same shape as input tensor :math:`(x_1, x_2, ..., x_R)`.
Raises:
TypeError: If src is not an integer or group is not a string.
RuntimeError: If device target is invalid, or backend is invalid, or distributed initialization fails.
Supported Platforms:
``Ascend`` ``GPU``
Examples:
.. note::
Before running the following examples, you need to configure the communication environment variables.
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
without any third-party or configuration file dependencies.
Please see the `msrun start up
<https://www.mindspore.cn/docs/zh-CN/master/model_train/parallel/msrun_launcher.html>`_
for more details.
This example should be run with 2 devices.
>>> import numpy as np
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
>>> # Launch 2 processes.
>>>
>>> comm.init()
>>> data = ms.Tensor(np.arange(8).reshape([2, 4]).astype(np.float32))
>>> out = comm.comm_func.broadcast(tensor=data, src=0)
[[0. 1. 2. 3.]
[4. 5. 6. 7.]]
Tutorial Examples:
- `Distributed Set Communication Primitives - Broadcast
<https://www.mindspore.cn/docs/en/master/api_python/samples/ops/communicate_ops.html#broadcast>`_
"""
if not isinstance(tensor, (Tensor, Tensor_)):
raise TypeError("For broadcast, the input tensor must be tensor")
if not isinstance(src, int):
raise TypeError("For broadcast, the src must be int")
_src = get_group_rank_from_world_rank(src, group)
_op = _get_cache_prim(P.Broadcast)(_src, group)
return _op((tensor,))[0]
[文档]def barrier(group=GlobalComm.WORLD_COMM_GROUP):
"""
Synchronizes all processes in the specified group. Once the process call this operation, it will be blocked until
all processes call this operation. After all processes finish calling the operations, the blocked processes
will be woken and continue their task.
Args:
group (str, optional): The communication group to work on. Default: ``GlobalComm.WORLD_COMM_GROUP``.
Raises:
RuntimeError: If backend is invalid, or distributed initialization fails.
Supported Platforms:
``Ascend``
Examples:
.. note::
Before running the following examples, you need to configure the communication environment variables.
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
without any third-party or configuration file dependencies.
Please see the `msrun start up
<https://www.mindspore.cn/docs/zh-CN/master/model_train/parallel/msrun_launcher.html>`_
for more details.
This example should be run with 2 devices.
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
>>> # Launch 2 processes.
>>> comm.init()
>>> comm.comm_func.barrier()
Tutorial Examples:
- `Distributed Set Communication Primitives - Barrier
<https://www.mindspore.cn/docs/en/master/api_python/samples/ops/communicate_ops.html#barrier>`_
"""
if not isinstance(group, str):
raise TypeError(f"group must be type of string, but got {type(group)}")
_op = _get_cache_prim(P.Barrier)(group)
return _op()
def _deal_comm_outputs(output, async_op):
if isinstance(output, tuple):
if not async_op:
output[1].wait()
return (output[0], None)
return output
if not async_op:
return (output, None)
return (output, default_handle)
[文档]def send(tensor, dst=0, group=GlobalComm.WORLD_COMM_GROUP, tag=0):
"""
Send tensors to the specified dest_rank.
Note:
Send and Receive must be used in combination and have same tag.
Args:
tensor (Tensor): The shape of tensor is :math:`(x_1, x_2, ..., x_R)`.
dst (int, optional): A required integer identifying the destination rank(global rank). Default: 0.
group (str, optional): The communication group to work on.
Default: "hccl_world_group" on Ascend, "nccl_world_group" on GPU.
tag (int, optional): A required integer identifying the send/recv message tag. The message will
be received by the Receive op with the same "tag". Default: 0.
Raises:
TypeError: `dst` is not an int or `group` is not a str。
ValueError: If the rank ID of the process is greater than the rank size of the communication group.
Supported Platforms:
``Ascend``
Examples:
.. note::
Before running the following examples, you need to configure the communication environment variables.
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
without any third-party or configuration file dependencies.
Please see the `msrun start up
<https://www.mindspore.cn/docs/zh-CN/master/model_train/parallel/msrun_launcher.html>`_
for more details.
This example should be run with 2 devices.
>>> import numpy as np
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
>>> comm.init()
>>> input_ = ms.Tensor(np.ones([2, 8]).astype(np.float32))
>>> comm.comm_func.send(input_, 0)
"""
if not isinstance(tensor, (Tensor, Tensor_)):
raise TypeError("For send, the input tensor must be tensor")
group = _get_group(group)
_dst = _get_group_rank_from_world_rank_from_cache_helper(dst, group)
tensor = _contiguous(tensor)
output = inner_comm_isend_op(tensor, _dst, group, tag)
_deal_comm_outputs(output, False)
[文档]def recv(tensor, src=0, group=GlobalComm.WORLD_COMM_GROUP, tag=0):
"""
Receive tensors from src.
Note:
Send and Receive must be used in combination and have same tag.
The shape and dtype of input `tensor` is used to receive tensor, but the value
of input `tensor` would not take effect.
Only support PyNative mode, Graph mode is not currently supported.
Args:
tensor (Tensor): The shape of tensor is :math:`(x_1, x_2, ..., x_R)`. The shape and dtype of this
tensor is used to receive tensor, but the value of input `tensor` would not take effect.
src (int, optional): A required integer identifying the source rank(global rank). Default: 0.
group (str, optional): The communication group to work on.
Default: "hccl_world_group" on Ascend, "nccl_world_group" on GPU.
tag (int, optional): A required integer identifying the send/recv message tag. The message will
be received by the Send op with the same "tag". Default: 0.
Returns:
Tensor, the shape of output is :math:`(x_1, x_2, ..., x_R)`.
Raises:
TypeError: If `src` is not an int or `group` is not a str.
ValueError: If the rank ID of the process is greater than the rank size of the communication group.
Supported Platforms:
``Ascend``
Examples:
.. note::
Before running the following examples, you need to configure the communication environment variables.
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
without any third-party or configuration file dependencies.
Please see the `msrun start up
<https://www.mindspore.cn/docs/zh-CN/master/model_train/parallel/msrun_launcher.html>`_
for more details.
This example should be run with 2 devices.
>>> import numpy as np
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
# Launch 2 processes.
Process 0 send the following array to Process 1
[[ 0. 1.]
[ 2. 3.]]
>>> comm.init()
>>> x = ms.Tensor(np.zeros([2, 2]))
# Process 1 receive tensor from Process 0.
>>> out = comm.comm_func.recv(x, src=0)
>>> print(out)
[[ 0. 1.]
[ 2. 3.]]
"""
if not isinstance(tensor, (Tensor, Tensor_)):
raise TypeError("For recv, the input tensor must be tensor")
if not isinstance(src, int):
raise TypeError("For recv, the src must be int")
group = _get_group(group)
_src = _get_group_rank_from_world_rank_from_cache_helper(src, group)
tensor = _contiguous(tensor)
shape = tensor.shape
dtype = tensor.dtype
output, _ = _deal_comm_outputs(inner_comm_irecv_op(tag, _src, shape, group, dtype), False)
return output
[文档]def isend(tensor, dst=0, group=GlobalComm.WORLD_COMM_GROUP, tag=0):
"""
Send tensors to the specified dest_rank asynchronously.
Note:
Send and Receive must be used in combination and have same tag.
Only support PyNative mode, Graph mode is not currently supported.
Args:
tensor (Tensor): The shape of tensor is :math:`(x_1, x_2, ..., x_R)`.
dst (int, optional): A required integer identifying the destination rank(global rank). Default: 0.
group (str, optional): The communication group to work on.
Default: "hccl_world_group" on Ascend, "nccl_world_group" on GPU.
tag (int, optional): A required integer identifying the send/recv message tag. The message will
be received by the Receive op with the same "tag". Default: 0.
Returns:
CommHandle, it is an async work handle.
Raises:
TypeError: `dst` is not an int or `group` is not a str。
ValueError: If the rank ID of the process is greater than the rank size of the communication group.
Supported Platforms:
``Ascend``
Examples:
.. note::
Before running the following examples, you need to configure the communication environment variables.
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
without any third-party or configuration file dependencies.
Please see the `msrun start up
<https://www.mindspore.cn/docs/zh-CN/master/model_train/parallel/msrun_launcher.html>`_
for more details.
This example should be run with 2 devices.
>>> import numpy as np
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
>>> comm.init()
>>> input_ = ms.Tensor(np.ones([2, 8]).astype(np.float32))
>>> handle = comm.comm_func.isend(input_, 0)
>>> handle.wait()
"""
if not isinstance(tensor, (Tensor, Tensor_)):
raise TypeError("For isend, the input tensor must be tensor")
group = _get_group(group)
_dst = _get_group_rank_from_world_rank_from_cache_helper(dst, group)
tensor = _contiguous(tensor)
output = inner_comm_isend_op(tensor, _dst, group, tag)
_, handle = _deal_comm_outputs(output, True)
return handle
[文档]def irecv(tensor, src=0, group=GlobalComm.WORLD_COMM_GROUP, tag=0):
"""
Receive tensors from src asynchronously.
Note:
Send and Receive must be used in combination and have same tag.
The shape and dtype of input `tensor` is used to receive tensor, but the value
of input `tensor` would not take effect.
Only support PyNative mode, Graph mode is not currently supported.
Args:
tensor (Tensor): The shape of tensor is :math:`(x_1, x_2, ..., x_R)`. The shape and dtype of this
tensor is used to receive tensor, but the value of input `tensor` would not take effect.
src (int, optional): A required integer identifying the source rank(global rank). Default: 0.
group (str, optional): The communication group to work on.
Default: "hccl_world_group" on Ascend, "nccl_world_group" on GPU.
tag (int, optional): A required integer identifying the send/recv message tag. The message will
be received by the Send op with the same "tag". Default: 0.
Returns:
Tuple(Tensor, CommHandle), the shape of output is :math:`(x_1, x_2, ..., x_R)`.
CommHandle is an async work handle, if `async_op` is set to True.
CommHandle will be None, when `async_op` is False.
Raises:
TypeError: If `src` is not an int or `group` is not a str.
ValueError: If the rank ID of the process is greater than the rank size of the communication group.
Supported Platforms:
``Ascend``
Examples:
.. note::
Before running the following examples, you need to configure the communication environment variables.
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
without any third-party or configuration file dependencies.
Please see the `msrun start up
<https://www.mindspore.cn/docs/zh-CN/master/model_train/parallel/msrun_launcher.html>`_
for more details.
This example should be run with 2 devices.
>>> import numpy as np
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
# Launch 2 processes.
# Process 0 send the following array to Process 1
# [[ 0. 1.]
# [ 2. 3.]]
>>> comm.init()
>>> x = ms.Tensor(np.zeros([2, 2]))
# Process 1 receive tensor from Process 0.
>>> out, handle = comm.comm_func.irecv(x, src=0)
>>> handle.wait()
>>> print(out)
[[ 0. 1.]
[ 2. 3.]]
"""
group = _get_group(group)
_src = _get_group_rank_from_world_rank_from_cache_helper(src, group)
tensor = _contiguous(tensor)
shape = tensor.shape
dtype = tensor.dtype
output = inner_comm_irecv_op(tag, _src, shape, group, dtype)
return _deal_comm_outputs(output, True)
[文档]def all_to_all_with_output_shape(output_shape_list, input_tensor_list, group=None, async_op=False):
"""
scatter and gather list of tensor to/from all rank according to input/output tensor list.
Note:
tensor shape in `output_shape_list` and `input_tensor_list` should be match across ranks.
Only support PyNative mode, Graph mode is not currently supported.
Args:
output_shape_list (Union[Tuple(Tensor), List(Tensor), Tuple(Tuple(int))]): List of shape
that indicate the gathered tensors shape from remote ranks.
input_tensor_list (Union[Tuple(Tensor), List(Tensor)]):
List of tensors to scatter to the remote rank.
group (str, optional): The communication group to work on.
Default: None, which means "hccl_world_group" on Ascend, "nccl_world_group" on GPU.
async_op (bool, optional): Whether this operator should be an async operator. Default: ``False`` .
Returns:
Tuple(Tuple(Tensor), CommHandle), the tensors is gathered from remote ranks.
CommHandle is an async work handle, if `async_op` is set to True.
CommHandle will be None, when `async_op` is False.
Raises:
TypeError: If `input_tensor_list` is not list of tensors.
TypeError: If `output_shape_list` is not list of tuple or tensors.
TypeError: If tensors in `input_tensor_list` are not the same type.
Supported Platforms:
``Ascend``
Examples:
.. note::
Before running the following examples, you need to configure the communication environment variables.
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
without any third-party or configuration file dependencies.
Please see the `msrun start up
<https://www.mindspore.cn/docs/zh-CN/master/model_train/parallel/msrun_launcher.html>`_
for more details.
This example should be run with 2 devices.
>>> import numpy as np
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
>>> comm.init()
>>> this_rank = comm.get_rank()
>>> if this_rank == 0:
>>> send_tensor_list = [ms.Tensor(1.), ms.Tensor([[2, 3], [4, 5.]])]
>>> recv_tensor_list = [(), (2,)]
>>> if this_rank == 1:
>>> send_tensor_list = [ms.Tensor([2, 2.]), ms.Tensor([4, 5, 6, 7.])]
>>> recv_tensor_list = [(2, 2), (4,)]
>>> output = comm.comm_func.all_to_all_with_output_shape(recv_tensor_list, send_tensor_list)
>>> print(output)
rank 0:
(Tensor(shape=[], dtype=Float32, value= 1),
Tensor(shape=[2], dtype=Float32, value= [2.00000000e+00, 2.00000000e+00]))
rank 1:
(Tensor(shape=[2, 2], dtype=Float32, value=
[[2.00000000e+00, 3.00000000e+00],
[4.00000000e+00, 5.00000000e+00]]),
Tensor(shape=[4], dtype=Float32, value=[4.00000000e+00, 5.00000000e+00, 6.00000000e+00, 7.00000000e+00]))
"""
_check_all_tensors(input_tensor_list)
_check_all_tensors_or_tuple(output_shape_list)
_check_all_tensor_same_dtype(input_tensor_list)
send_numel_list = []
send_flatten_tensor = []
recv_numel_list = []
recv_shape_list = []
for tensor in input_tensor_list:
send_numel_list.append(tensor.size)
send_flatten_tensor.append(tensor.reshape(-1))
for tensor in output_shape_list:
if isinstance(tensor, Tensor):
recv_numel_list.append(tensor.size)
recv_shape_list.append(tensor.shape)
else:
_shape = tensor
recv_numel_list.append(_get_size(_shape))
recv_shape_list.append(_shape)
send_flatten_tensor = cat(send_flatten_tensor)
send_flatten_tensor = _contiguous(send_flatten_tensor)
group = GlobalComm.WORLD_COMM_GROUP if group is None else _get_group(group)
global _GROPU_SIZE_CACHE
if group not in _GROPU_SIZE_CACHE:
_GROPU_SIZE_CACHE[group] = get_group_size(group)
rank_size = _GROPU_SIZE_CACHE[group]
output = inner_comm_all_to_all_v_op(send_flatten_tensor, group, send_numel_list, recv_numel_list,
rank_size, False)
output, handle = _deal_comm_outputs(output, async_op)
result = []
offset = 0
for numel, shape in zip(recv_numel_list, recv_shape_list):
result.append(output[offset:offset + numel].reshape(shape))
offset = offset + numel
return (tuple(result), handle)
def _get_all_to_all_single_numel_list(tensor, output_shape, output_split_sizes, input_split_sizes, group):
"""get numel list for all_to_all_single."""
global _GROPU_SIZE_CACHE
if _is_split_sizes_empty(input_split_sizes):
if group not in _GROPU_SIZE_CACHE:
_GROPU_SIZE_CACHE[group] = get_group_size(group)
_world_size = _GROPU_SIZE_CACHE[group]
if tensor.shape[0] % _world_size != 0:
raise ValueError("input shape at dim 0 must be divided by world_size, "
f"but got {tensor.shape[0]} and {_world_size}.")
_split_size = tensor.shape[0] // _world_size
input_split_sizes = (_split_size,) * _world_size
if _is_split_sizes_empty(output_split_sizes):
if group not in _GROPU_SIZE_CACHE:
_GROPU_SIZE_CACHE[group] = get_group_size(group)
_world_size = _GROPU_SIZE_CACHE[group]
shape_dim_0 = None
if isinstance(output_shape, Tensor):
shape_dim_0 = output_shape.shape[0]
else:
shape_dim_0 = output_shape[0]
if shape_dim_0 % _world_size != 0:
raise ValueError("output shape at dim 0 must be divided by world_size, "
f"but got {shape_dim_0} and {_world_size}.")
_split_size = shape_dim_0 // _world_size
output_split_sizes = (_split_size,) * _world_size
send_size_without_first_dim = _get_size(tensor.shape[1:])
send_numel_list = [size * send_size_without_first_dim for size in input_split_sizes]
recv_size_without_first_dim = None
recv_shape_without_first_dim = None
if isinstance(output_shape, Tensor):
recv_shape_without_first_dim = output_shape.shape[1:]
recv_size_without_first_dim = _get_size(recv_shape_without_first_dim)
else:
recv_shape_without_first_dim = output_shape[1:]
recv_size_without_first_dim = _get_size(recv_shape_without_first_dim)
recv_numel_list = [size * recv_size_without_first_dim for size in output_split_sizes]
return send_numel_list, recv_numel_list, recv_shape_without_first_dim
[文档]def all_to_all_single_with_output_shape(output_shape, tensor, output_split_sizes=None,
input_split_sizes=None, group=None, async_op=False):
"""
scatter and gather input with split size to/from all rank, and return result in a single tensor.
Note:
'output_shape' and 'tensor' shape should be match across ranks.
Only support PyNative mode, Graph mode is not currently supported.
Args:
output_shape (Union(Tensor, Tuple(int))): shape to indicate the shape
of tensor gathered concatenated from remote rank.
tensor (Tensor): tensor to be scattered to remote rank.
output_split_sizes (Union(Tuple(int), List(int))): output split size at dim 0. If set to None,
it means equally split by ``world_size``. Default: None.
input_split_sizes (Union(Tuple(int), List(int))): input split size at dim 0. If set to None,
it means equally split by ``world_size``. Default: None.
group (str, optional): The communication group to work on.
Default: None, which means "hccl_world_group" on Ascend, "nccl_world_group" on GPU.
async_op (bool, optional): Whether this operator should be an async operator. Default: ``False`` .
Returns:
Tuple(Tensor, CommHandle), the output tensor is gathered concatenated from remote ranks.
If the numel of tensor gathered from remote is zero, it will return a Tensor will value 0,
which has no actual meanning. CommHandle is an async work handle, if `async_op` is set to True.
CommHandle will be None, when `async_op` is False.
Raises:
TypeError: If `tensor` is not tensor.
TypeError: If `output_shape` is not tuple or tensors.
Supported Platforms:
``Ascend``
Examples:
.. note::
Before running the following examples, you need to configure the communication environment variables.
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
without any third-party or configuration file dependencies.
Please see the `msrun start up
<https://www.mindspore.cn/docs/zh-CN/master/model_train/parallel/msrun_launcher.html>`_
for more details.
This example should be run with 2 devices.
>>> import numpy as np
>>> import mindspore as ms
>>> import mindspore.communication as comm
>>>
>>> comm.init()
>>> this_rank = comm.get_rank()
>>> if this_rank == 0:
>>> output_shape = (3, 3)
>>> tensor = ms.Tensor([[0, 1, 2.], [3, 4, 5], [6, 7, 8]])
>>> result = comm.comm_func.all_to_all_single_with_output_shape(output_shape, tensor, [2, 1], [2, 1])
>>> if this_rank == 1:
>>> output_shape = (2, 3)
>>> tensor = ms.Tensor([[9, 10., 11], [12, 13, 14]])
>>> result = comm.comm_func.all_to_all_single_with_output_shape(output_shape, tensor)
>>> print(result)
rank 0:
[[ 0. 1. 2.]
[ 3. 4. 5.]
[ 9. 10. 11.]]
rank 1:
[[ 6. 7. 8.]
[12. 13. 14.]]
"""
_check_all_tensors([tensor])
_check_all_tensors_or_tuple([output_shape])
if group is None:
group = GlobalComm.WORLD_COMM_GROUP
split_sizes_empty = _is_split_sizes_empty(output_split_sizes) and _is_split_sizes_empty(input_split_sizes)
send_numel_list, recv_numel_list, recv_shape_without_first_dim = \
_get_all_to_all_single_numel_list(tensor, output_shape, output_split_sizes, input_split_sizes, group)
tensor = _contiguous(tensor)
_input = tensor.reshape(-1)
group = GlobalComm.WORLD_COMM_GROUP if group is None else _get_group(group)
global _GROPU_SIZE_CACHE
if group not in _GROPU_SIZE_CACHE:
_GROPU_SIZE_CACHE[group] = get_group_size(group)
rank_size = _GROPU_SIZE_CACHE[group]
result = inner_comm_all_to_all_v_op(_input, group, send_numel_list, recv_numel_list, rank_size, split_sizes_empty)
result, handle = _deal_comm_outputs(result, async_op)
if any(recv_numel_list):
result = result.reshape((-1,) + recv_shape_without_first_dim)
return result, handle