| """ Dataset Factory |
| |
| Hacked together by / Copyright 2021, Ross Wightman |
| """ |
| import os |
| from typing import Optional |
|
|
| from torchvision.datasets import CIFAR100, CIFAR10, MNIST, KMNIST, FashionMNIST, ImageFolder |
| try: |
| from torchvision.datasets import Places365 |
| has_places365 = True |
| except ImportError: |
| has_places365 = False |
| try: |
| from torchvision.datasets import INaturalist |
| has_inaturalist = True |
| except ImportError: |
| has_inaturalist = False |
| try: |
| from torchvision.datasets import QMNIST |
| has_qmnist = True |
| except ImportError: |
| has_qmnist = False |
| try: |
| from torchvision.datasets import ImageNet |
| has_imagenet = True |
| except ImportError: |
| has_imagenet = False |
|
|
| from .dataset import IterableImageDataset, ImageDataset |
|
|
| _TORCH_BASIC_DS = dict( |
| cifar10=CIFAR10, |
| cifar100=CIFAR100, |
| mnist=MNIST, |
| kmnist=KMNIST, |
| fashion_mnist=FashionMNIST, |
| ) |
| _TRAIN_SYNONYM = dict(train=None, training=None) |
| _EVAL_SYNONYM = dict(val=None, valid=None, validation=None, eval=None, evaluation=None) |
|
|
|
|
| def _search_split(root, split): |
| |
| split_name = split.split('[')[0] |
| try_root = os.path.join(root, split_name) |
| if os.path.exists(try_root): |
| return try_root |
|
|
| def _try(syn): |
| for s in syn: |
| try_root = os.path.join(root, s) |
| if os.path.exists(try_root): |
| return try_root |
| return root |
| if split_name in _TRAIN_SYNONYM: |
| root = _try(_TRAIN_SYNONYM) |
| elif split_name in _EVAL_SYNONYM: |
| root = _try(_EVAL_SYNONYM) |
| return root |
|
|
|
|
| def create_dataset( |
| name: str, |
| root: Optional[str] = None, |
| split: str = 'validation', |
| search_split: bool = True, |
| class_map: dict = None, |
| load_bytes: bool = False, |
| is_training: bool = False, |
| download: bool = False, |
| batch_size: int = 1, |
| num_samples: Optional[int] = None, |
| seed: int = 42, |
| repeats: int = 0, |
| input_img_mode: str = 'RGB', |
| trust_remote_code: bool = False, |
| **kwargs, |
| ): |
| """ Dataset factory method |
| |
| In parentheses after each arg are the type of dataset supported for each arg, one of: |
| * Folder - default, timm folder (or tar) based ImageDataset |
| * Torch - torchvision based datasets |
| * HFDS - Hugging Face Datasets |
| * HFIDS - Hugging Face Datasets Iterable (streaming mode, with IterableDataset) |
| * TFDS - Tensorflow-datasets wrapper in IterabeDataset interface via IterableImageDataset |
| * WDS - Webdataset |
| * All - any of the above |
| |
| Args: |
| name: Dataset name, empty is okay for folder based datasets |
| root: Root folder of dataset (All) |
| split: Dataset split (All) |
| search_split: Search for split specific child fold from root so one can specify |
| `imagenet/` instead of `/imagenet/val`, etc on cmd line / config. (Folder, Torch) |
| class_map: Specify class -> index mapping via text file or dict (Folder) |
| load_bytes: Load data, return images as undecoded bytes (Folder) |
| download: Download dataset if not present and supported (HFIDS, TFDS, Torch) |
| is_training: Create dataset in train mode, this is different from the split. |
| For Iterable / TDFS it enables shuffle, ignored for other datasets. (TFDS, WDS, HFIDS) |
| batch_size: Batch size hint for iterable datasets (TFDS, WDS, HFIDS) |
| seed: Seed for iterable datasets (TFDS, WDS, HFIDS) |
| repeats: Dataset repeats per iteration i.e. epoch (TFDS, WDS, HFIDS) |
| input_img_mode: Input image color conversion mode e.g. 'RGB', 'L' (folder, TFDS, WDS, HFDS, HFIDS) |
| trust_remote_code: Trust remote code in Hugging Face Datasets if True (HFDS, HFIDS) |
| **kwargs: Other args to pass through to underlying Dataset and/or Reader classes |
| |
| Returns: |
| Dataset object |
| """ |
| kwargs = {k: v for k, v in kwargs.items() if v is not None} |
| name = name.lower() |
| if name.startswith('torch/'): |
| name = name.split('/', 2)[-1] |
| torch_kwargs = dict(root=root, download=download, **kwargs) |
| if name in _TORCH_BASIC_DS: |
| ds_class = _TORCH_BASIC_DS[name] |
| use_train = split in _TRAIN_SYNONYM |
| ds = ds_class(train=use_train, **torch_kwargs) |
| elif name == 'inaturalist' or name == 'inat': |
| assert has_inaturalist, 'Please update to PyTorch 1.10, torchvision 0.11+ for Inaturalist' |
| target_type = 'full' |
| split_split = split.split('/') |
| if len(split_split) > 1: |
| target_type = split_split[0].split('_') |
| if len(target_type) == 1: |
| target_type = target_type[0] |
| split = split_split[-1] |
| if split in _TRAIN_SYNONYM: |
| split = '2021_train' |
| elif split in _EVAL_SYNONYM: |
| split = '2021_valid' |
| ds = INaturalist(version=split, target_type=target_type, **torch_kwargs) |
| elif name == 'places365': |
| assert has_places365, 'Please update to a newer PyTorch and torchvision for Places365 dataset.' |
| if split in _TRAIN_SYNONYM: |
| split = 'train-standard' |
| elif split in _EVAL_SYNONYM: |
| split = 'val' |
| ds = Places365(split=split, **torch_kwargs) |
| elif name == 'qmnist': |
| assert has_qmnist, 'Please update to a newer PyTorch and torchvision for QMNIST dataset.' |
| use_train = split in _TRAIN_SYNONYM |
| ds = QMNIST(train=use_train, **torch_kwargs) |
| elif name == 'imagenet': |
| assert has_imagenet, 'Please update to a newer PyTorch and torchvision for ImageNet dataset.' |
| if split in _EVAL_SYNONYM: |
| split = 'val' |
| ds = ImageNet(split=split, **torch_kwargs) |
| elif name == 'image_folder' or name == 'folder': |
| |
| if search_split and os.path.isdir(root): |
| |
| root = _search_split(root, split) |
| ds = ImageFolder(root, **kwargs) |
| else: |
| assert False, f"Unknown torchvision dataset {name}" |
| elif name.startswith('hfds/'): |
| |
| |
| ds = ImageDataset( |
| root, |
| reader=name, |
| split=split, |
| class_map=class_map, |
| input_img_mode=input_img_mode, |
| trust_remote_code=trust_remote_code, |
| **kwargs, |
| ) |
| elif name.startswith('hfids/'): |
| ds = IterableImageDataset( |
| root, |
| reader=name, |
| split=split, |
| class_map=class_map, |
| is_training=is_training, |
| download=download, |
| batch_size=batch_size, |
| num_samples=num_samples, |
| repeats=repeats, |
| seed=seed, |
| input_img_mode=input_img_mode, |
| trust_remote_code=trust_remote_code, |
| **kwargs, |
| ) |
| elif name.startswith('tfds/'): |
| ds = IterableImageDataset( |
| root, |
| reader=name, |
| split=split, |
| class_map=class_map, |
| is_training=is_training, |
| download=download, |
| batch_size=batch_size, |
| num_samples=num_samples, |
| repeats=repeats, |
| seed=seed, |
| input_img_mode=input_img_mode, |
| **kwargs |
| ) |
| elif name.startswith('wds/'): |
| ds = IterableImageDataset( |
| root, |
| reader=name, |
| split=split, |
| class_map=class_map, |
| is_training=is_training, |
| batch_size=batch_size, |
| num_samples=num_samples, |
| repeats=repeats, |
| seed=seed, |
| input_img_mode=input_img_mode, |
| **kwargs |
| ) |
| else: |
| |
| if search_split and os.path.isdir(root): |
| |
| root = _search_split(root, split) |
| ds = ImageDataset( |
| root, |
| reader=name, |
| class_map=class_map, |
| load_bytes=load_bytes, |
| input_img_mode=input_img_mode, |
| **kwargs, |
| ) |
| return ds |
|
|