mindspore.communication.comm_func.isend

View Source On Gitee
mindspore.communication.comm_func.isend(tensor, dst=0, group=GlobalComm.WORLD_COMM_GROUP, tag=0)[source]

Send tensors to the specified dest_rank asynchronously.

Note

Send and Receive must be used in combination and have same tag. Only support PyNative mode, Graph mode is not currently supported.

Parameters
  • tensor (Tensor) – The shape of tensor is \((x_1, x_2, ..., x_R)\).

  • dst (int, optional) – A required integer identifying the destination rank(global rank). Default: 0.

  • group (str, optional) – The communication group to work on. Default: "hccl_world_group" on Ascend, "nccl_world_group" on GPU.

  • tag (int, optional) – A required integer identifying the send/recv message tag. The message will be received by the Receive op with the same "tag". Default: 0.

Returns

CommHandle, it is an async work handle.

Raises
  • TypeErrordst is not an int or group is not a str。

  • ValueError – If the rank ID of the process is greater than the rank size of the communication group.

Supported Platforms:

Ascend

Examples

Note

Before running the following examples, you need to configure the communication environment variables.

For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method without any third-party or configuration file dependencies. Please see the msrun start up for more details.

This example should be run with 2 devices.

>>> import numpy as np
>>> import mindspore as ms
>>> from mindspore.communication import init
>>> from mindspore.communication.comm_func import isend, irecv
>>> from mindspore.communication import get_rank, get_group_size
>>>
>>> np.random.seed(1)
>>> init()
>>> rank = get_rank()
>>> size = get_group_size()
>>> x = np.ones([2, 2]).astype(np.float32) * 0.01 * (rank + 1)
>>> x2 = np.ones([2, 2]).astype(np.float32)
>>>
>>>
>>> if rank < size / 2:
>>>     _x = ms.Tensor(x)
>>>     isend(_x, rank + size // 2)
>>> else:
>>>     _x2 = ms.Tensor(x2)
>>>     output, handle = irecv(_x2, rank - size // 2)
>>>     handle.wait()
>>>     print(output)
[[0.01  0.01]
 [0.01  0.01]]