| | """ Quick n Simple Image Folder, Tarfile based DataSet |
| | |
| | Hacked together by / Copyright 2019, Ross Wightman |
| | """ |
| | import io |
| | import logging |
| | from typing import Optional |
| |
|
| | import torch |
| | import torch.utils.data as data |
| | from PIL import Image |
| |
|
| | from .readers import create_reader |
| |
|
| | _logger = logging.getLogger(__name__) |
| |
|
| |
|
| | _ERROR_RETRY = 50 |
| |
|
| |
|
| | class ImageDataset(data.Dataset): |
| |
|
| | def __init__( |
| | self, |
| | root, |
| | reader=None, |
| | split='train', |
| | class_map=None, |
| | load_bytes=False, |
| | input_img_mode='RGB', |
| | transform=None, |
| | target_transform=None, |
| | **kwargs, |
| | ): |
| | if reader is None or isinstance(reader, str): |
| | reader = create_reader( |
| | reader or '', |
| | root=root, |
| | split=split, |
| | class_map=class_map, |
| | **kwargs, |
| | ) |
| | self.reader = reader |
| | self.load_bytes = load_bytes |
| | self.input_img_mode = input_img_mode |
| | self.transform = transform |
| | self.target_transform = target_transform |
| | self._consecutive_errors = 0 |
| |
|
| | def __getitem__(self, index): |
| | img, target = self.reader[index] |
| |
|
| | try: |
| | img = img.read() if self.load_bytes else Image.open(img) |
| | except Exception as e: |
| | _logger.warning(f'Skipped sample (index {index}, file {self.reader.filename(index)}). {str(e)}') |
| | self._consecutive_errors += 1 |
| | if self._consecutive_errors < _ERROR_RETRY: |
| | return self.__getitem__((index + 1) % len(self.reader)) |
| | else: |
| | raise e |
| | self._consecutive_errors = 0 |
| |
|
| | if self.input_img_mode and not self.load_bytes: |
| | img = img.convert(self.input_img_mode) |
| | if self.transform is not None: |
| | img = self.transform(img) |
| |
|
| | if target is None: |
| | target = -1 |
| | elif self.target_transform is not None: |
| | target = self.target_transform(target) |
| |
|
| | return img, target |
| |
|
| | def __len__(self): |
| | return len(self.reader) |
| |
|
| | def filename(self, index, basename=False, absolute=False): |
| | return self.reader.filename(index, basename, absolute) |
| |
|
| | def filenames(self, basename=False, absolute=False): |
| | return self.reader.filenames(basename, absolute) |
| |
|
| |
|
| | class IterableImageDataset(data.IterableDataset): |
| |
|
| | def __init__( |
| | self, |
| | root, |
| | reader=None, |
| | split='train', |
| | class_map=None, |
| | is_training=False, |
| | batch_size=1, |
| | num_samples=None, |
| | seed=42, |
| | repeats=0, |
| | download=False, |
| | input_img_mode='RGB', |
| | input_key=None, |
| | target_key=None, |
| | transform=None, |
| | target_transform=None, |
| | max_steps=None, |
| | **kwargs, |
| | ): |
| | assert reader is not None |
| | if isinstance(reader, str): |
| | self.reader = create_reader( |
| | reader, |
| | root=root, |
| | split=split, |
| | class_map=class_map, |
| | is_training=is_training, |
| | batch_size=batch_size, |
| | num_samples=num_samples, |
| | seed=seed, |
| | repeats=repeats, |
| | download=download, |
| | input_img_mode=input_img_mode, |
| | input_key=input_key, |
| | target_key=target_key, |
| | max_steps=max_steps, |
| | **kwargs, |
| | ) |
| | else: |
| | self.reader = reader |
| | self.transform = transform |
| | self.target_transform = target_transform |
| | self._consecutive_errors = 0 |
| |
|
| | def __iter__(self): |
| | for img, target in self.reader: |
| | if self.transform is not None: |
| | img = self.transform(img) |
| | if self.target_transform is not None: |
| | target = self.target_transform(target) |
| | yield img, target |
| |
|
| | def __len__(self): |
| | if hasattr(self.reader, '__len__'): |
| | return len(self.reader) |
| | else: |
| | return 0 |
| |
|
| | def set_epoch(self, count): |
| | |
| | if hasattr(self.reader, 'set_epoch'): |
| | self.reader.set_epoch(count) |
| |
|
| | def set_loader_cfg( |
| | self, |
| | num_workers: Optional[int] = None, |
| | ): |
| | |
| | if hasattr(self.reader, 'set_loader_cfg'): |
| | self.reader.set_loader_cfg(num_workers=num_workers) |
| |
|
| | def filename(self, index, basename=False, absolute=False): |
| | assert False, 'Filename lookup by index not supported, use filenames().' |
| |
|
| | def filenames(self, basename=False, absolute=False): |
| | return self.reader.filenames(basename, absolute) |
| |
|
| |
|
| | class AugMixDataset(torch.utils.data.Dataset): |
| | """Dataset wrapper to perform AugMix or other clean/augmentation mixes""" |
| |
|
| | def __init__(self, dataset, num_splits=2): |
| | self.augmentation = None |
| | self.normalize = None |
| | self.dataset = dataset |
| | if self.dataset.transform is not None: |
| | self._set_transforms(self.dataset.transform) |
| | self.num_splits = num_splits |
| |
|
| | def _set_transforms(self, x): |
| | assert isinstance(x, (list, tuple)) and len(x) == 3, 'Expecting a tuple/list of 3 transforms' |
| | self.dataset.transform = x[0] |
| | self.augmentation = x[1] |
| | self.normalize = x[2] |
| |
|
| | @property |
| | def transform(self): |
| | return self.dataset.transform |
| |
|
| | @transform.setter |
| | def transform(self, x): |
| | self._set_transforms(x) |
| |
|
| | def _normalize(self, x): |
| | return x if self.normalize is None else self.normalize(x) |
| |
|
| | def __getitem__(self, i): |
| | x, y = self.dataset[i] |
| | x_list = [self._normalize(x)] |
| | |
| | for _ in range(self.num_splits - 1): |
| | x_list.append(self._normalize(self.augmentation(x))) |
| | return tuple(x_list), y |
| |
|
| | def __len__(self): |
| | return len(self.dataset) |
| |
|