# Dataset Slicing [](https://gitee.com/mindspore/docs/blob/master/tutorials/source_en/parallel/dataset_slice.md) ## Overview When performing distributed training, taking image data as an example, when the size of a single image is too large, such as large-format images of remote sensing satellites, when an image is too large, it is necessary to slice the image and read a portion of each card to perform distributed training. Scenarios that deal with dataset slicing need to be combined with model parallelism to achieve the desired effect of reducing video memory, so this feature is provided based on automatic parallelism. The sample used in this tutorial is not a large-format network, and is intended as an example only. Real-life applications to large-format networks often require detailed design of parallel strategies. > Dataset sharding is only supported in fully-automatic mode and semi-automatic mode, and is not involved in data parallel mode. ### Related Interfaces 1. `mindspore.dataset.vision.SlicePatches(num_height=1, num_width=1)`: Slices the Tensor into multiple blocks horizontally and vertically. Suitable for scenarios where the Tensor has a large height and width. `num_height` is the number of slices in vertical direction and `num_width` is the number of slices in horizontal direction. More parameters can be found in [SlicePatches](https://www.mindspore.cn/docs/en/master/api_python/dataset_vision/mindspore.dataset.vision.SlicePatches.html). 2. `dataset_strategy(config=((1, 1, 1, 8), (8,)))`: indicates dataset slicing strategy. For more details, [AutoParallel Parallel Configuration](https://www.mindspore.cn/docs/en/master/api_python/parallel/mindspore.parallel.auto_parallel.AutoParallel.html). The `dataset_strategy` interface has the following limitations: - Each input is allowed to be sliced in at most one dimension. If `dataset_strategy(config=((1, 1, 1, 8), (8,)))` or `config=((1, 1, 1, 8), (1,))` is supported, each input is sliced in just one dimension at most, but not `config=((1, 1, 4, 2), (1,))`, whose first input is sliced into two dimensions. - The input with the highest dimension, the number of slices, must be more than the other dimensions. If `config=((1, 1, 1, 8), (8,))` or `config=((1, 1, 1, 1), (1,))` is supported, the input with the most dimensions is the first one, with the number of slices of 8, and the rest of the inputs have no more than 8 slices, but not `config=((1, 1, 1, 1), (8,))`, whose input with the most dimensions is the first input with a slice of 1, but the second input has a cut of 8, which exceeds the slices of the first input. ## Operation Practices ### Sample Code Description > Download the full sample code here: [dataset_slice](https://gitee.com/mindspore/docs/tree/master/docs/sample_code/dataset_slice). The directory structure is as follows: ```text └─ sample_code ├─ dataset_slice ├── train.py └── run.sh ... ``` `train.py` is the script that defines the network structure and the training process. `run.sh` is the execution script. ### Configuring a Distributed Environment Initialize the communication with init. ```python import mindspore as ms from mindspore.communication import init ms.set_context(mode=ms.GRAPH_MODE) init() ``` ### Loading the Dataset When using dataset slicing, you need to call the `SlicePatches` interface to construct the dataset at the same time. To ensure that the read-in data is consistent across cards, the dataset needs to be fixed with a random number seed. ```python import os import mindspore.dataset as ds from mindspore import nn slice_h_num = 1 slice_w_num = 4 ds.config.set_seed(1000) # set dataset seed to make sure that all cards read the same data def create_dataset(batch_size): dataset_path = os.getenv("DATA_PATH") dataset = ds.MnistDataset(dataset_path) image_transforms = [ ds.vision.Rescale(1.0 / 255.0, 0), ds.vision.Normalize(mean=(0.1307,), std=(0.3081,)), ds.vision.HWC2CHW() ] label_transform = ds.transforms.TypeCast(ms.int32) dataset = dataset.map(image_transforms, 'image') dataset = dataset.map(label_transform, 'label') # slice image slice_patchs_img_op = ds.vision.SlicePatches(slice_h_num, slice_w_num) img_cols = ['img' + str(x) for x in range(slice_h_num * slice_w_num)] dataset = dataset.map(operations=slice_patchs_img_op, input_columns="image", output_columns=img_cols) dataset = dataset.project([img_cols[get_rank() % (slice_h_num * slice_w_num)], "label"]) dataset = dataset.batch(batch_size) return dataset data_set = create_dataset(32) ``` ### Network Definition The network definition here is consistent with the single-card model and the initialization of network parameters and optimizer parameters is deferred through the `no_init_parameters` interface: ```python from mindspore import nn from mindspore.nn.utils import no_init_parameters class Network(nn.Cell): def __init__(self): super().__init__() self.flatten = nn.Flatten() self.layer1 = nn.Dense(28*28, 512) self.relu1 = nn.ReLU() self.layer2 = nn.Dense(512, 512) self.relu2 = nn.ReLU() self.layer3 = nn.Dense(512, 10) def construct(self, x): x = self.flatten(x) x = self.layer1(x) x = self.relu1(x) x = self.layer2(x) x = self.relu2(x) logits = self.layer3(x) return logits with no_init_parameters(): net = Network() optimizer = nn.SGD(net.trainable_params(), 1e-2) ``` ### Training the Network In this step, the loss function as well as the training process needs to be defined, and the parallel settings are implemented by wrapping grad_fn with the top-level `AutoParallel` class, setting the parallel mode to semi-automatic parallel mode `semi-auto`. In addition, configure the dataset `dataset_strategy` slicing strategy `config` to ((1, 1, 1, 4), (1,)), which represents no slice vertically, and 4 chunks horizontally. This example is written in a functional manner, which is partially consistent with the single-card model: ```python from mindspore import nn import mindspore as ms from mindspore.parallel.auto_parallel import AutoParallel loss_fn = nn.CrossEntropyLoss() def forward_fn(data, target): logits = net(data) loss = loss_fn(logits, target) return loss, logits grad_fn = ms.value_and_grad(forward_fn, None, net.trainable_params(), has_aux=True) # Set parallelism, data slicing strategies grad_fn = AutoParallel(grad_fn, parallel_mode="semi_auto") grad_fn.dataset_strategy(config=((1, 1, slice_h_num, slice_w_num), (1,))) for epoch in range(1): i = 0 for image, label in data_set: (loss_value, _), grads = grad_fn(image, label) optimizer(grads) if i % 10 == 0: print("epoch: %s, step: %s, loss is %s" % (epoch, i, loss_value)) i += 1 ``` ### Running Stand-alone 8-card Script Next, the corresponding script is called by the command. Take the `mpirun` startup method, the 8-card distributed training script as an example, and perform the distributed training: ```bash bash run.sh ``` After training, the log file is saved to the `log_output` directory, and the part of results about the Loss are saved in `log_output/worker_*.log`. The example is as follows: ```text epoch: 0, step: 0, loss is 2.281521 epoch: 0, step: 10, loss is 2.185312 epoch: 0, step: 20, loss is 1.9531741 epoch: 0, step: 30, loss is 1.6952474 epoch: 0, step: 40, loss is 1.2967496 ... ```