# Differences with torchtext.datasets.IWSLT2016

[![View Source On Gitee](https://mindspore-website.obs.cn-north-4.myhuaweicloud.com/website-images/master/resource/_static/logo_source_en.svg)](https://gitee.com/mindspore/docs/blob/master/docs/mindspore/source_en/note/api_mapping/pytorch_diff/IWSLT2016.md)

## torchtext.datasets.IWSLT2016

```python
class torchtext.datasets.IWSLT2016(
    root: str = '.data',
    split: Union[List[str], str] = ('train', 'valid', 'test'),
    language_pair: Sequence =('de', 'en'),
    valid_set: str ='tst2013',
    test_set: str ='tst2014')
```

For more information, see [torchtext.datasets.IWSLT2016](https://pytorch.org/text/0.9.0/datasets.html#iwslt2016).

## mindspore.dataset.IWSLT2016Dataset

```python
class mindspore.dataset.IWSLT2016Dataset(
    dataset_dir,
    usage=None,
    language_pair=None,
    valid_set=None,
    test_set=None,
    num_samples=None,
    num_parallel_workers=None,
    shuffle=Shuffle.GLOBAL,
    num_shards=None,
    shard_id=None,
    cache=None)
```

For more information, see [mindspore.dataset.IWSLT2016Dataset](https://mindspore.cn/docs/en/master/api_python/dataset/mindspore.dataset.IWSLT2016Dataset.html#mindspore.dataset.IWSLT2016Dataset).

## Differences

PyTorch: Read the IWSLT2016 dataset.

MindSpore: Read the IWSLT2016 dataset. Downloading dataset from web is not supported.

| Categories | Subcategories |PyTorch | MindSpore | Difference |
| --- | ---   | ---   | ---        |---  |
|Parameter | Parameter1 | root    | dataset_dir    | - |
|     | Parameter2 | split      | usage    |- |
|     | Parameter3 | language_pair      | language_pair    |- |
|     | Parameter4 | valid_set      | valid_set    |- |
|     | Parameter5 | test_set      | test_set    |- |
|     | Parameter6 | -    | num_samples | The number of images to be included in the dataset |
|     | Parameter7 | -    | num_parallel_workers | Number of worker threads to read the data |
|     | Parameter8 | -    | shuffle  | Whether to perform shuffle on the dataset |
|     | Parameter9 | -    | num_shards | Number of shards that the dataset will be divided into |
|     | Parameter10 | -    | shard_id | The shard ID within num_shards |
|     | Parameter11 | -    | cache | Use tensor caching service to speed up dataset processing |

## Code Example

```python
# PyTorch
import torchtext.datasets as datasets

root = "/path/to/dataset_root/"
train_iter, valid_iter, test_iter = datasets.IWSLT2016(root, split=('train', 'valid', 'test'))
data = next(iter(train_iter))

# MindSpore
import mindspore.dataset as ds

# Download IWSLT2016 dataset files, unzip into the following structure
# .
# └── /path/to/dataset_directory/
#      ├── subeval_files
#                └── texts
#                    ├── ar
#                    │    └── en
#                    │        └── ar-en
#                    ├── cs
#                    │    └── en
#                    │        └── cs-en
#                    ├── de
#                    │    └── en
#                    │        └── de-en
#                    │            ├── IWSLT16.TED.dev2010.de-en.de.xml
#                    │            ├── train.tags.de-en.de
#                    │            ├── ...
#                    ├── en
#                    │    ├── ar
#                    │    │   └── en-ar
#                    │    ├── cs
#                    │    │   └── en-cs
#                    │    ├── de
#                    │    │   └── en-de
#                    │    └── fr
#                    │        └── en-fr
#                    └── fr
#                        └── en
#                            └── fr-en2
root = "/path/to/dataset_directory/"
dataset = ds.IWSLT2016Dataset(root, usage='all')
data = next(iter(dataset))
```