| |
| |
| |
| |
|
|
|
|
| import logging |
|
|
| import numpy as np |
| import os |
|
|
| import torch |
|
|
| from torchvision import datasets, transforms |
|
|
| from timm.data import create_transform |
| from timm.data.constants import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD |
| import PIL |
|
|
| from fairseq.data import FairseqDataset |
| from .mae_image_dataset import caching_loader |
|
|
|
|
| logger = logging.getLogger(__name__) |
|
|
|
|
| def build_transform(is_train, input_size, color_jitter, aa, reprob, remode, recount): |
| mean = IMAGENET_DEFAULT_MEAN |
| std = IMAGENET_DEFAULT_STD |
| |
| if is_train: |
| |
| transform = create_transform( |
| input_size=input_size, |
| is_training=True, |
| color_jitter=color_jitter, |
| auto_augment=aa, |
| interpolation="bicubic", |
| re_prob=reprob, |
| re_mode=remode, |
| re_count=recount, |
| mean=mean, |
| std=std, |
| ) |
| return transform |
|
|
| |
| t = [] |
| if input_size <= 224: |
| crop_pct = 224 / 256 |
| else: |
| crop_pct = 1.0 |
| size = int(input_size / crop_pct) |
| t.append( |
| transforms.Resize( |
| size, interpolation=PIL.Image.BICUBIC |
| ), |
| ) |
| t.append(transforms.CenterCrop(input_size)) |
|
|
| t.append(transforms.ToTensor()) |
| t.append(transforms.Normalize(mean, std)) |
| return transforms.Compose(t) |
|
|
|
|
| class MaeFinetuningImageDataset(FairseqDataset): |
| def __init__( |
| self, |
| root: str, |
| split: str, |
| is_train: bool, |
| input_size, |
| color_jitter=None, |
| aa="rand-m9-mstd0.5-inc1", |
| reprob=0.25, |
| remode="pixel", |
| recount=1, |
| local_cache_path=None, |
| shuffle=True, |
| ): |
| FairseqDataset.__init__(self) |
|
|
| self.shuffle = shuffle |
|
|
| transform = build_transform( |
| is_train, input_size, color_jitter, aa, reprob, remode, recount |
| ) |
|
|
| path = os.path.join(root, split) |
| loader = caching_loader(local_cache_path, datasets.folder.default_loader) |
|
|
| self.dataset = datasets.ImageFolder(path, loader=loader, transform=transform) |
|
|
| logger.info(f"loaded {len(self.dataset)} examples") |
|
|
| def __getitem__(self, index): |
| img, label = self.dataset[index] |
| return {"id": index, "img": img, "label": label} |
|
|
| def __len__(self): |
| return len(self.dataset) |
|
|
| def collater(self, samples): |
| if len(samples) == 0: |
| return {} |
|
|
| collated_img = torch.stack([s["img"] for s in samples], dim=0) |
|
|
| res = { |
| "id": torch.LongTensor([s["id"] for s in samples]), |
| "net_input": { |
| "imgs": collated_img, |
| }, |
| } |
|
|
| if "label" in samples[0]: |
| res["net_input"]["labels"] = torch.LongTensor([s["label"] for s in samples]) |
|
|
| return res |
|
|
| def num_tokens(self, index): |
| return 1 |
|
|
| def size(self, index): |
| return 1 |
|
|
| def ordered_indices(self): |
| """Return an ordered list of indices. Batches will be constructed based |
| on this order.""" |
| if self.shuffle: |
| order = [np.random.permutation(len(self))] |
| else: |
| order = [np.arange(len(self))] |
|
|
| return order[0] |
|
|