Source code for mindspore.dataset.transforms.transforms

# Copyright 2019-2022 Huawei Technologies Co., Ltd
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# ==============================================================================
"""
The module transforms provides common operations, including Compose, OneHot and TypeCast.
"""
import json
from abc import ABC

import sys
from enum import IntEnum
import numpy as np

import mindspore._c_dataengine as cde
from mindspore._c_expression import typing
from mindspore.common import dtype as mstype
import mindspore.dataset.transforms.c_transforms as c_transforms
import mindspore.dataset.transforms.py_transforms as py_transforms
import mindspore.dataset.vision.c_transforms as c_vision
from . import py_transforms_util as util
from .py_transforms_util import Implementation, FuncWrapper
from .validators import check_fill_value, check_slice_option, check_slice_op, check_one_hot_op, check_compose_call, \
    check_mask_op_new, check_pad_end, check_concat_type, check_random_transform_ops, check_plugin, check_type_cast
from ..core.datatypes import mstype_to_detype, nptype_to_detype
from ..vision.py_transforms_util import is_pil


class TensorOperation:
    """
    Base class Tensor Ops
    """

    def __init__(self):
        super().__init__()
        self.implementation = None
        self.callable_op_ = None

    def __call__(self, *input_tensor_list):
        """
        Call method.
        """
        # Check if Python implementation of op, or PIL input
        if (self.implementation == Implementation.PY) or \
                (len(input_tensor_list) == 1 and is_pil(input_tensor_list[0]) and getattr(self, '_execute_py', None)):
            return self._execute_py(*input_tensor_list)

        tensor_row = []
        for tensor in input_tensor_list:
            try:
                tensor_row.append(cde.Tensor(np.asarray(tensor)))
            except (RuntimeError, TypeError):
                raise TypeError("Invalid user input. Got {}: {}, cannot be converted into tensor." \
                                .format(type(tensor), tensor))
        if not hasattr(self, 'callable_op_') or self.callable_op_ is None:
            self.callable_op_ = cde.Execute(self.parse())
        output_tensor_list = self.callable_op_(tensor_row)
        output_numpy_list = [x.as_array() for x in output_tensor_list]
        return output_numpy_list[0] if len(output_numpy_list) == 1 else tuple(output_numpy_list)

    @staticmethod
    def parse():
        """parse function - not yet implemented"""
        raise NotImplementedError("TensorOperation has to implement parse() method.")


class PyTensorOperation:
    """
    Base Python Tensor Operations class
    """

    def __init__(self):
        self.transforms = []
        self.output_type = None

    def __call__(self, img):
        """
        Call method.

        Args:
            img (PIL Image): Image to be augmented.

        Returns:
            PIL Image, augmented image.
        """
        return self._execute_py(img)

    @classmethod
    def from_json(cls, json_string):
        """
        Base from_json for Python tensor operations class
        """
        json_obj = json.loads(json_string)
        new_op = cls.__new__(cls)
        new_op.__dict__ = json_obj
        if "transforms" in json_obj.keys():
            # operations which have transforms as input, need to call _from_json() for each transform to deseriallize
            transforms = []
            for json_op in json_obj["transforms"]:
                transforms.append(getattr(
                    sys.modules.get(json_op.get("python_module")),
                    json_op["tensor_op_name"]).from_json(json.dumps(json_op["tensor_op_params"])))
            new_op.transforms = transforms
        if "output_type" in json_obj.keys():
            output_type = np.dtype(json_obj["output_type"])
            new_op.output_type = output_type
        return new_op

    def to_json(self):
        """
        Base to_json for Python tensor operations class
        """
        json_obj = {}
        json_trans = {}
        if "transforms" in self.__dict__.keys():
            # operations which have transforms as input, need to call _to_json() for each transform to serialize
            json_list = []
            for transform in self.transforms:
                json_list.append(json.loads(transform.to_json()))
            json_trans["transforms"] = json_list
            self.__dict__.pop("transforms")
        if "output_type" in self.__dict__.keys():
            json_trans["output_type"] = np.dtype(
                self.__dict__["output_type"]).name
            self.__dict__.pop("output_type")
        json_obj["tensor_op_params"] = self.__dict__
        # append transforms to the tensor_op_params of the operation
        json_obj.get("tensor_op_params").update(json_trans)
        json_obj["tensor_op_name"] = self.__class__.__name__
        json_obj["python_module"] = self.__class__.__module__
        return json.dumps(json_obj)


class CompoundOperation(TensorOperation, PyTensorOperation, ABC):
    """
    Compound Tensor Operations class
    """

    def __init__(self, transforms):
        super(CompoundOperation, self).__init__()
        self.transforms = []
        trans_with_imple = []
        for op in transforms:
            if callable(op) and not hasattr(op, "implementation") and \
                    not isinstance(op, c_transforms.TensorOperation) and \
                    not isinstance(op, py_transforms.PyTensorOperation) and \
                    not isinstance(op, c_vision.ImageTensorOperation):
                op = util.FuncWrapper(op)
            if hasattr(op, "implementation"):
                if op.implementation is not None:
                    trans_with_imple.append(op)
            else:
                raise RuntimeError("Mixing old legacy c/py_transforms and new unified transforms is not allowed.")
            self.transforms.append(op)

        if all([t.implementation == Implementation.PY for t in self.transforms]):
            self.implementation = Implementation.PY
        elif all([t.implementation is not None for t in self.transforms]):
            self.implementation = Implementation.C
        elif not trans_with_imple:
            self.implementation = None
        elif all([t.implementation == Implementation.PY for t in trans_with_imple]):
            self.implementation = Implementation.PY
        elif all([t.implementation == Implementation.C for t in trans_with_imple]):
            self.implementation = Implementation.C

    @staticmethod
    def parse():
        """parse function - not yet implemented"""
        raise NotImplementedError("CompoundOperation has to implement parse() method.")

    def parse_transforms(self):
        operations = []
        for op in self.transforms:
            if op and getattr(op, 'parse', None):
                operations.append(op.parse())
            else:
                operations.append(op)
        return operations


def not_random(function):
    """
    Specify the function as "not random", i.e., it produces deterministic result.
    A Python function can only be cached after it is specified as "not random".
    """
    function.random = False
    return function


[docs]class Compose(CompoundOperation): """ Compose a list of transforms into a single transform. .. Note:: Compose takes a list of transformations in `mindspore.dataset.transforms` / `mindspore.dataset.vision` and user-defined Python callable objects to combine as single data augmentation. For user-defined Python callable objects, the return value is required to be type numpy.ndarray. Args: transforms (list): List of transformations to be applied. Raises: TypeError: If `transforms` is not of type list. ValueError: If `transforms` is empty. TypeError: If elements of `transforms` are neither Python callable objects nor data processing operations in transforms.py. Supported Platforms: ``CPU`` Examples: >>> import mindspore.dataset as ds >>> import mindspore.dataset.transforms as transforms >>> import mindspore.dataset.vision as vision >>> >>> # create a dataset that reads all files in dataset_dir with 8 threads >>> image_folder_dataset_dir = "/path/to/image_folder_dataset_directory" >>> image_folder_dataset = ds.ImageFolderDataset(image_folder_dataset_dir, num_parallel_workers=8) >>> >>> # create a list of transformations to be applied to the image data >>> transform = transforms.Compose([vision.Decode(to_pil=True), ... vision.RandomHorizontalFlip(0.5), ... vision.ToTensor(), ... vision.Normalize((0.491, 0.482, 0.447), (0.247, 0.243, 0.262), is_hwc=False), ... vision.RandomErasing()]) >>> # apply the transform to the dataset through dataset.map function >>> image_folder_dataset = image_folder_dataset.map(operations=transform, input_columns=["image"]) >>> >>> # Compose is also be invoked implicitly, by just passing in a list of ops >>> # the above example then becomes: >>> transforms_list = [vision.Decode(to_pil=True), ... vision.RandomHorizontalFlip(0.5), ... vision.ToTensor(), ... vision.Normalize((0.491, 0.482, 0.447), (0.247, 0.243, 0.262), is_hwc=False), ... vision.RandomErasing()] >>> >>> # apply the transform to the dataset through dataset.map() >>> image_folder_dataset = image_folder_dataset.map(operations=transforms_list, input_columns=["image"]) >>> >>> # Certain C++ and Python ops can be combined, but not all of them >>> # An example of combined operations >>> arr = [0, 1] >>> dataset = ds.NumpySlicesDataset(arr, column_names=["cols"], shuffle=False) >>> transformed_list = [transforms.OneHot(2), ... transforms.Mask(transforms.Relational.EQ, 1)] >>> dataset = dataset.map(operations=transformed_list, input_columns=["cols"]) >>> >>> # Here is an example of mixing vision ops >>> import numpy as np >>> op_list=[vision.Decode(), ... vision.Resize((224, 244)), ... vision.ToPIL(), ... np.array, # need to convert PIL image to a NumPy array to pass it to C++ operation ... vision.Resize((24, 24))] >>> image_folder_dataset = image_folder_dataset.map(operations=op_list, input_columns=["image"]) """ @check_random_transform_ops def __init__(self, transforms): super().__init__(transforms) self.transforms = Compose.decompose(self.transforms) if all(hasattr(transform, "random") and not transform.random for transform in self.transforms): self.random = False # pylint: disable=missing-docstring @staticmethod def decompose(operations): # Remove all compose operation from the given list of operations. # # Args: # operations (list): list of transforms. # # Returns: # list of operations without compose operations. new_operations = [] for op in operations: if isinstance(op, Compose): new_operations.extend(Compose.decompose(op.transforms)) else: new_operations.append(op) return new_operations # pylint: disable=missing-docstring @staticmethod def reduce(operations): # Wraps adjacent Python operations in a Compose to allow mixing of Python and C++ operations. # # Args: # operations (list): list of tensor operations. # # Returns: # list, the reduced list of operations. new_ops, start_ind, end_ind = [], 0, 0 for i, op in enumerate(operations): if op.implementation == Implementation.C and not isinstance(op, FuncWrapper): # reset counts if start_ind != end_ind: if end_ind == start_ind + 1: composed_op = operations[start_ind] else: composed_op = Compose(operations[start_ind:end_ind]) composed_op.implementation = Implementation.PY new_ops.append(composed_op) new_ops.append(op) start_ind, end_ind = i + 1, i + 1 else: end_ind += 1 # do additional check in case the last operation is a Python operation if start_ind != end_ind: if end_ind == start_ind + 1: composed_op = operations[start_ind] else: composed_op = Compose(operations[start_ind:end_ind]) composed_op.implementation = Implementation.PY new_ops.append(composed_op) return new_ops def parse(self): operations = self.parse_transforms() return cde.ComposeOperation(operations) @check_compose_call def _execute_py(self, *args): """ Execute method. Returns: lambda function, Lambda function that takes in an args to apply transformations on. """ return util.compose(self.transforms, *args)
[docs]class Concatenate(TensorOperation): """ Tensor operation that concatenates all columns into a single tensor, only 1D tenspr is supported. Args: axis (int, optional): Concatenate the tensors along given axis. Default: ``0``. prepend (numpy.ndarray, optional): NumPy array to be prepended to the already concatenated tensors. Default: ``None``. append (numpy.ndarray, optional): NumPy array to be appended to the already concatenated tensors. Default: ``None``. Raises: TypeError: If `axis` is not of type int. TypeError: If `prepend` is not of type numpy.ndarray. TypeError: If `append` is not of type numpy.ndarray. Supported Platforms: ``CPU`` Examples: >>> import mindspore.dataset as ds >>> import mindspore.dataset.transforms as transforms >>> import numpy as np >>> >>> # concatenate string >>> prepend_tensor = np.array(["dw", "df"], dtype='S') >>> append_tensor = np.array(["dwsdf", "df"], dtype='S') >>> concatenate_op = transforms.Concatenate(0, prepend_tensor, append_tensor) >>> data = [["This","is","a","string"]] >>> dataset = ds.NumpySlicesDataset(data) >>> dataset = dataset.map(operations=concatenate_op) """ @check_concat_type def __init__(self, axis=0, prepend=None, append=None): super().__init__() self.axis = axis self.prepend = cde.Tensor(np.array(prepend)) if prepend is not None else prepend self.append = cde.Tensor(np.array(append)) if append is not None else append self.implementation = Implementation.C def parse(self): return cde.ConcatenateOperation(self.axis, self.prepend, self.append)
[docs]class Duplicate(TensorOperation): """ Duplicate the input tensor to output, only support transform one column each time. Raises: RuntimeError: If given tensor has two columns. Supported Platforms: ``CPU`` Examples: >>> import mindspore.dataset as ds >>> import mindspore.dataset.transforms as transforms >>> # Data before >>> # | x | >>> # +---------+ >>> # | [1,2,3] | >>> # +---------+ >>> data = [[1,2,3]] >>> numpy_slices_dataset = ds.NumpySlicesDataset(data, ["x"]) >>> numpy_slices_dataset = numpy_slices_dataset.map(operations=transforms.Duplicate(), ... input_columns=["x"], ... output_columns=["x", "y"]) >>> # Data after >>> # | x | y | >>> # +---------+---------+ >>> # | [1,2,3] | [1,2,3] | >>> # +---------+---------+ """ def __init__(self): super().__init__() self.implementation = Implementation.C def parse(self): return cde.DuplicateOperation()
[docs]class Fill(TensorOperation): """ Tensor operation to fill all elements in the tensor with the specified value. The output tensor will have the same shape and type as the input tensor. Args: fill_value (Union[str, bytes, int, float, bool]): scalar value to fill the tensor with. Raises: TypeError: If `fill_value` is not of type str, float, bool, int or bytes. Supported Platforms: ``CPU`` Examples: >>> import mindspore.dataset as ds >>> import mindspore.dataset.transforms as transforms >>> import numpy as np >>> >>> # generate a 1D integer numpy array from 0 to 4 >>> def generator_1d(): ... for i in range(5): ... yield (np.array([i]),) >>> generator_dataset = ds.GeneratorDataset(generator_1d, column_names="col1") >>> # [[0], [1], [2], [3], [4]] >>> fill_op = transforms.Fill(3) >>> generator_dataset = generator_dataset.map(operations=fill_op) >>> # [[3], [3], [3], [3], [3]] """ @check_fill_value def __init__(self, fill_value): super().__init__() self.fill_value = cde.Tensor(np.array(fill_value)) self.implementation = Implementation.C def parse(self): return cde.FillOperation(self.fill_value)
[docs]class Mask(TensorOperation): r""" Mask content of the input tensor with the given predicate. Any element of the tensor that matches the predicate will be evaluated to True, otherwise False. Args: operator (Relational): relational operators, it can be ``Relational.EQ``, ``Relational.NE``, ``Relational.LT``, ``Relational.GT``, ``Relational.LE``, ``Relational.GE``, take ``Relational.EQ`` as example, EQ refers to equal. constant (Union[str, int, float, bool]): Constant to be compared to. dtype (mindspore.dtype, optional): Type of the generated mask. Default: ``mstype.bool\_``. Raises: TypeError: `operator` is not of type Relational. TypeError: `constant` is not of type string int, float or bool. TypeError: `dtype` is not of type mindspore.dtype. Supported Platforms: ``CPU`` Examples: >>> import mindspore.dataset as ds >>> import mindspore.dataset.transforms as transforms >>> from mindspore.dataset.transforms import Relational >>> # Data before >>> # | col | >>> # +---------+ >>> # | [1,2,3] | >>> # +---------+ >>> data = [[1, 2, 3]] >>> numpy_slices_dataset = ds.NumpySlicesDataset(data, ["col"]) >>> numpy_slices_dataset = numpy_slices_dataset.map(operations=transforms.Mask(Relational.EQ, 2)) >>> # Data after >>> # | col | >>> # +--------------------+ >>> # | [False,True,False] | >>> # +--------------------+ """ @check_mask_op_new def __init__(self, operator, constant, dtype=mstype.bool_): super().__init__() self.operator = operator self.dtype = mstype_to_detype(dtype) self.constant = cde.Tensor(np.array(constant)) self.implementation = Implementation.C def parse(self): return cde.MaskOperation(DE_C_RELATIONAL.get(self.operator), self.constant, self.dtype)
[docs]class OneHot(TensorOperation): """ Tensor operation to apply one hot encoding. Args: num_classes (int): Number of classes of objects in dataset. It should be larger than the largest label number in the dataset. smoothing_rate (float, optional): Adjustable hyperparameter for label smoothing level. Default: ``0.0``, means no smoothing is applied. Raises: TypeError: `num_classes` is not of type int. TypeError: `smoothing_rate` is not of type float or int. ValueError: `smoothing_rate` is not in range [0.0, 1.0]. RuntimeError: Input tensor is not of type int. RuntimeError: Input tensor is not a 1-D tensor. Supported Platforms: ``CPU`` Examples: >>> import mindspore.dataset as ds >>> import mindspore.dataset.transforms as transforms >>> >>> mnist_dataset_dir = "/path/to/mnist_dataset_directory" >>> mnist_dataset = ds.MnistDataset(dataset_dir=mnist_dataset_dir) >>> >>> # Assume that dataset has 10 classes, thus the label ranges from 0 to 9 >>> onehot_op = transforms.OneHot(num_classes=10) >>> mnist_dataset = mnist_dataset.map(operations=onehot_op, input_columns=["label"]) """ @check_one_hot_op def __init__(self, num_classes, smoothing_rate=0.0): super().__init__() self.num_classes = num_classes self.random = False self.smoothing_rate = smoothing_rate def parse(self): return cde.OneHotOperation(self.num_classes, self.smoothing_rate)
[docs]class PadEnd(TensorOperation): """ Pad input tensor according to pad_shape, input tensor needs to have same rank. Args: pad_shape (list(int)): List of integers representing the shape needed. Dimensions that set to ``None`` will not be padded (i.e., original dim will be used). Shorter dimensions will truncate the values. pad_value (Union[str, bytes, int, float, bool], optional): Value used to pad. Default: ``None``. Default to ``0`` in case of tensors of Numbers, or empty string in case of tensors of strings. Raises: TypeError: If `pad_shape` is not of type list. TypeError: If `pad_value` is not of type str, float, bool, int or bytes. TypeError: If elements of `pad_shape` is not of type int. ValueError: If elements of `pad_shape` is not of positive. Supported Platforms: ``CPU`` Examples: >>> import mindspore.dataset as ds >>> import mindspore.dataset.transforms as transforms >>> # Data before >>> # | col | >>> # +---------+ >>> # | [1,2,3] | >>> # +---------| >>> data = [[1, 2, 3]] >>> numpy_slices_dataset = ds.NumpySlicesDataset(data, ["col"]) >>> numpy_slices_dataset = numpy_slices_dataset.map(operations=transforms.PadEnd(pad_shape=[4], ... pad_value=10)) >>> # Data after >>> # | col | >>> # +------------+ >>> # | [1,2,3,10] | >>> # +------------| """ @check_pad_end def __init__(self, pad_shape, pad_value=None): super().__init__() self.pad_shape = cde.TensorShape(pad_shape) self.pad_value = cde.Tensor(np.array(pad_value)) if pad_value is not None else pad_value self.implementation = Implementation.C def parse(self): return cde.PadEndOperation(self.pad_shape, self.pad_value)
class Plugin(TensorOperation): """ Plugin support for MindData. Use this class to dynamically load a .so file (shared library) and execute its symbols. Args: lib_path (str): Path to .so file which is compiled to support MindData plugin. func_name (str): Name of the function to load from the .so file. user_args (str, optional): Serialized args to pass to the plugin. Only needed if "func_name" requires one. Raises: TypeError: If `lib_path` is not of type string. TypeError: If `func_name` is not of type string. TypeError: If `user_args` is not of type string. Supported Platforms: ``CPU`` Examples: >>> import mindspore.dataset as ds >>> import mindspore.dataset.transforms as transforms >>> >>> plugin = transforms.Plugin("pluginlib.so", "PluginDecode") >>> image_folder_dataset = ds.ImageFolderDataset("/path/to/image_folder_dataset_directory") >>> image_folder_dataset = image_folder_dataset.map(operations=plugin) """ @check_plugin def __init__(self, lib_path, func_name, user_args=None): super().__init__() self.lib_path = lib_path self.func_name = func_name self.user_args = str() if (user_args is None) else user_args self.implementation = Implementation.C def parse(self): return cde.PluginOperation(self.lib_path, self.func_name, self.user_args)
[docs]class RandomApply(CompoundOperation): """ Randomly perform a series of transforms with a given probability. Args: transforms (list): List of transformations to be applied. prob (float, optional): The probability to apply the transformation list. Default: ``0.5``. Raises: TypeError: If `transforms` is not of type list. ValueError: If `transforms` is empty. TypeError: If elements of `transforms` are neither Python callable objects nor data processing operations in transforms.py. TypeError: If `prob` is not of type float. ValueError: If `prob` is not in range [0.0, 1.0]. Supported Platforms: ``CPU`` Examples: >>> import mindspore.dataset as ds >>> import mindspore.dataset.transforms as transforms >>> import mindspore.dataset.vision as vision >>> from mindspore.dataset.transforms import Compose >>> >>> transforms_list = [vision.RandomHorizontalFlip(0.5), ... vision.Normalize((0.491, 0.482, 0.447), (0.247, 0.243, 0.262)), ... vision.RandomErasing()] >>> composed_transform = Compose([vision.Decode(to_pil=True), ... transforms.RandomApply(transforms_list, prob=0.6), ... vision.ToTensor()]) >>> >>> image_folder_dataset = ds.ImageFolderDataset("/path/to/image_folder_dataset_directory") >>> image_folder_dataset = image_folder_dataset.map(operations=composed_transform, input_columns=["image"]) """ @check_random_transform_ops def __init__(self, transforms, prob=0.5): super().__init__(transforms) self.prob = prob def parse(self): operations = self.parse_transforms() return cde.RandomApplyOperation(self.prob, operations) def _execute_py(self, img): """ Execute method. Args: img (PIL image): Image to be randomly applied a list transformations. Returns: img (PIL image), Transformed image. """ return util.random_apply(img, self.transforms, self.prob)
[docs]class RandomChoice(CompoundOperation): """ Randomly select one transform from a list of transforms to perform operation. Args: transforms (list): List of transformations to be chosen from to apply. Raises: TypeError: If `transforms` is not of type list. ValueError: If `transforms` is empty. TypeError: If elements of `transforms` are neither Python callable objects nor data processing operations in transforms.py. Supported Platforms: ``CPU`` Examples: >>> import mindspore.dataset as ds >>> import mindspore.dataset.transforms as transforms >>> import mindspore.dataset.vision as vision >>> from mindspore.dataset.transforms import Compose >>> >>> transforms_list = [vision.RandomHorizontalFlip(0.5), ... vision.Normalize((0.491, 0.482, 0.447), (0.247, 0.243, 0.262)), ... vision.RandomErasing()] >>> composed_transform = Compose([vision.Decode(), ... transforms.RandomChoice(transforms_list), ... vision.ToTensor()]) >>> >>> image_folder_dataset = ds.ImageFolderDataset("/path/to/image_folder_dataset_directory") >>> image_folder_dataset = image_folder_dataset.map(operations=composed_transform, input_columns=["image"]) """ @check_random_transform_ops def __init__(self, transforms): super().__init__(transforms) def parse(self): operations = self.parse_transforms() return cde.RandomChoiceOperation(operations) def _execute_py(self, img): """ Execute method. Args: img (PIL image): Image to be applied transformation. Returns: img (PIL image), Transformed image. """ return util.random_choice(img, self.transforms)
[docs]class RandomOrder(PyTensorOperation): """ Perform a series of transforms to the input image in a random order. Args: transforms (list): List of the transformations to apply. Raises: TypeError: If `transforms` is not of type list. TypeError: If elements of `transforms` are neither Python callable objects nor data processing operations in mindspore.dataset.transforms.transforms. ValueError: If `transforms` is empty. Supported Platforms: ``CPU`` Examples: >>> import mindspore.dataset as ds >>> import mindspore.dataset.vision as vision >>> import mindspore.dataset.transforms as transforms >>> from mindspore.dataset.transforms import Compose >>> >>> transforms_list = [vision.RandomHorizontalFlip(0.5), ... vision.Normalize((0.491, 0.482, 0.447), (0.247, 0.243, 0.262)), ... vision.RandomErasing()] >>> composed_transform = Compose([vision.Decode(to_pil=False), ... transforms.RandomOrder(transforms_list), ... vision.ToTensor()]) >>> >>> image_folder_dataset = ds.ImageFolderDataset("/path/to/image_folder_dataset_directory") >>> image_folder_dataset = image_folder_dataset.map(operations=composed_transform, input_columns=["image"]) """ @check_random_transform_ops def __init__(self, transforms): super().__init__() self.transforms = transforms self.implementation = Implementation.PY def _execute_py(self, img): """ Execute method. Args: img (PIL image): Image to apply transformations in a random order. Returns: img (PIL image), Transformed image. """ return util.random_order(img, self.transforms)
[docs]class Relational(IntEnum): """ Relationship operator. Possible enumeration values are: ``Relational.EQ``, ``Relational.NE``, ``Relational.GT``, ``Relational.GE``, ``Relational.LT``, ``Relational.LE``. - Relational.EQ: refers to Equality. - Relational.NE: refers not equal, or Inequality. - Relational.GT: refers to Greater than. - Relational.GE: refers to Greater than or equal to. - Relational.LT: refers to Less than. - Relational.LE: refers to Less than or equal to. """ EQ = 0 NE = 1 GT = 2 GE = 3 LT = 4 LE = 5
DE_C_RELATIONAL = {Relational.EQ: cde.RelationalOp.EQ, Relational.NE: cde.RelationalOp.NE, Relational.GT: cde.RelationalOp.GT, Relational.GE: cde.RelationalOp.GE, Relational.LT: cde.RelationalOp.LT, Relational.LE: cde.RelationalOp.LE} class _SliceOption(cde.SliceOption): """ Internal class SliceOption to be used with SliceOperation Args: _SliceOption(Union[int, list(int), slice, None, Ellipsis, bool, _SliceOption]): 1. :py:obj:`int`: Slice this index only along the dimension. Negative index is supported. 2. :py:obj:`list(int)`: Slice these indices along the dimension. Negative indices are supported. 3. :py:obj:`slice`: Slice the generated indices from the slice object along the dimension. 4. :py:obj:`None`: Slice the whole dimension. Similar to :py:obj:`:` in Python indexing. 5. :py:obj:`Ellipsis`: Slice the whole dimension. Similar to :py:obj:`:` in Python indexing. 6. :py:obj:`boolean`: Slice the whole dimension. Similar to :py:obj:`:` in Python indexing. """ @check_slice_option def __init__(self, slice_option): if isinstance(slice_option, int) and not isinstance(slice_option, bool): slice_option = [slice_option] elif slice_option is Ellipsis: slice_option = True elif slice_option is None: slice_option = True super().__init__(slice_option)
[docs]class Slice(TensorOperation): """ Slice operation to extract a tensor out using the given n slices. The functionality of Slice is similar to NumPy's indexing feature (Currently only rank-1 tensors are supported). Args: slices (Union[int, list[int], slice, None, Ellipsis]): Maximum `n` number of arguments to slice a tensor of rank `n` . One object in slices can be one of: 1. :py:obj:`int`: Slice this index only along the first dimension. Negative index is supported. 2. :py:obj:`list(int)`: Slice these indices along the first dimension. Negative indices are supported. 3. :py:obj:`slice`: Slice the generated indices from the `slice <https://docs.python.org/3.7/library/functions.html?highlight=slice#slice>`_ object along the first dimension. Similar to start:stop:step. 4. :py:obj:`None`: Slice the whole dimension. Similar to :py:obj:`[:]` in Python indexing. 5. :py:obj:`Ellipsis`: Slice the whole dimension, same result with `None` . Raises: TypeError: If `slices` is not of type int, list[int], :py:obj:`slice` , :py:obj:`None` or :py:obj:`Ellipsis` . Supported Platforms: ``CPU`` Examples: >>> import mindspore.dataset as ds >>> import mindspore.dataset.transforms as transforms >>> # Data before >>> # | col | >>> # +---------+ >>> # | [1,2,3] | >>> # +---------| >>> data = [[1, 2, 3]] >>> numpy_slices_dataset = ds.NumpySlicesDataset(data, ["col"]) >>> # slice indices 1 and 2 only >>> numpy_slices_dataset = numpy_slices_dataset.map(operations=transforms.Slice(slice(1,3))) >>> # Data after >>> # | col | >>> # +---------+ >>> # | [2,3] | >>> # +---------| """ @check_slice_op def __init__(self, *slices): super().__init__() slice_input_ = list(slices) slice_input_ = [_SliceOption(slice_dim) for slice_dim in slice_input_] self.slice_input_ = slice_input_ self.implementation = Implementation.C def parse(self): return cde.SliceOperation(self.slice_input_)
[docs]class TypeCast(TensorOperation): """ Tensor operation to cast to a given MindSpore data type or NumPy data type. Note: This operation supports running on Ascend or GPU platforms by Offload. Args: data_type (Union[mindspore.dtype, numpy.dtype]): mindspore.dtype or numpy.dtype (e.g. `numpy.float32`) to be cast to. Raises: TypeError: If `data_type` is not of MindSpore data type bool, int, float, string or type :class:`numpy.dtype` . Supported Platforms: ``Ascend`` ``GPU`` ``CPU`` Examples: >>> import mindspore.dataset as ds >>> import mindspore.dataset.transforms as transforms >>> import numpy as np >>> from mindspore import dtype as mstype >>> >>> # Generate 1d int numpy array from 0 - 63 >>> def generator_1d(): ... for i in range(64): ... yield (np.array([i]),) >>> >>> dataset = ds.GeneratorDataset(generator_1d, column_names='col') >>> type_cast_op = transforms.TypeCast(mstype.int32) >>> dataset = dataset.map(operations=type_cast_op) """ @check_type_cast def __init__(self, data_type): super().__init__() if isinstance(data_type, typing.Type): data_type = mstype_to_detype(data_type) else: data_type = nptype_to_detype(data_type) self.data_type = str(data_type) self.implementation = Implementation.C def parse(self): return cde.TypeCastOperation(self.data_type)
[docs]class Unique(TensorOperation): """ Perform the unique operation on the input tensor, only support transform one column each time. Return 3 tensor: unique output tensor, index tensor, count tensor. - Output tensor contains all the unique elements of the input tensor in the same order that they occur in the input tensor. - Index tensor that contains the index of each element of the input tensor in the unique output tensor. - Count tensor that contains the count of each element of the output tensor in the input tensor. Note: Call batch op before calling this function. Raises: RuntimeError: If given Tensor has two columns. Supported Platforms: ``CPU`` Examples: >>> import mindspore.dataset as ds >>> import mindspore.dataset.transforms as transforms >>> # Data before >>> # | x | >>> # +--------------------+ >>> # | [[0,1,2], [1,2,3]] | >>> # +--------------------+ >>> data = [[[0,1,2], [1,2,3]]] >>> dataset = ds.NumpySlicesDataset(data, ["x"]) >>> dataset = dataset.map(operations=transforms.Unique(), ... input_columns=["x"], ... output_columns=["x", "y", "z"]) >>> # Data after >>> # | x | y |z | >>> # +---------+-----------------+---------+ >>> # | [0,1,2,3] | [0,1,2,1,2,3] | [1,2,2,1] >>> # +---------+-----------------+---------+ """ def __init__(self): super().__init__() self.implementation = Implementation.C def parse(self): return cde.UniqueOperation()