| | import argparse |
| | import glob |
| | import os |
| | from PIL import Image |
| |
|
| | import cv2 |
| | import math |
| | import numpy as np |
| | import os |
| | import os.path as osp |
| | import random |
| | import time |
| | import torch |
| | from tqdm import tqdm |
| |
|
| | from basicsr.data.degradations import circular_lowpass_kernel, random_mixed_kernels |
| | from basicsr.data.transforms import augment |
| | from basicsr.utils import FileClient, get_root_logger, imfrombytes, img2tensor |
| | from basicsr.utils.registry import DATASET_REGISTRY |
| | from torch.utils import data as data |
| | from torchvision.transforms.functional import center_crop |
| | import torchvision.transforms as T |
| | from torchvision.utils import save_image |
| |
|
| | from basicsr.data.degradations import random_add_gaussian_noise_pt, random_add_poisson_noise_pt |
| | from basicsr.data.transforms import paired_random_crop |
| | from basicsr.utils import DiffJPEG, USMSharp |
| | from basicsr.utils.img_process_util import filter2D |
| | from basicsr.utils.registry import MODEL_REGISTRY |
| | from collections import OrderedDict |
| | from torch.nn import functional as F |
| |
|
| | cfg = { |
| | |
| | "name": "DF2K+OST", |
| | "type": "RealESRGANDataset", |
| | "dataroot_gt": "/home/CORP/hsiang.chen/Project/Datasets/IR/SuperResolution", |
| | "meta_train": [ |
| | "DIV2K/metas/DIV2K_train_HR.list", |
| | "Flickr2K/metas/Flickr2K.list", |
| | |
| | ], |
| | "meta_test": ["DIV2K/metas/DIV2K_valid_HR.list"], |
| |
|
| | |
| | "resize_prob": [0.2, 0.7, 0.1], |
| | "resize_range": [0.15, 1.5], |
| | "gaussian_noise_prob": 0.5, |
| | "noise_range": [1, 30], |
| | "poisson_scale_range": [0.05, 3], |
| | "gray_noise_prob": 0.4, |
| | "jpeg_range": [30, 95], |
| |
|
| | "blur_kernel_size": 21, |
| | "kernel_list": ['iso', 'aniso', 'generalized_iso', 'generalized_aniso', 'plateau_iso', 'plateau_aniso'], |
| | "kernel_prob": [0.45, 0.25, 0.12, 0.03, 0.12, 0.03], |
| | "sinc_prob": 0.1, |
| | "blur_sigma": [0.2, 3], |
| | "betag_range": [0.5, 4], |
| | "betap_range": [1, 2], |
| |
|
| | |
| | "second_blur_prob": 0.8, |
| | "resize_prob2": [0.3, 0.4, 0.3], |
| | "resize_range2": [0.3, 1.2], |
| | "gaussian_noise_prob2": 0.5, |
| | "noise_range2": [1, 25], |
| | "poisson_scale_range2": [0.05, 2.5], |
| | "gray_noise_prob2": 0.4, |
| | "jpeg_range2": [30, 95], |
| |
|
| | "blur_kernel_size2": 21, |
| | "kernel_list2": ['iso', 'aniso', 'generalized_iso', 'generalized_aniso', 'plateau_iso', 'plateau_aniso'], |
| | "kernel_prob2": [0.45, 0.25, 0.12, 0.03, 0.12, 0.03], |
| | "sinc_prob2": 0.1, |
| | "blur_sigma2": [0.2, 1.5], |
| | "betag_range2": [0.5, 4], |
| | "betap_range2": [1, 2], |
| |
|
| | "final_sinc_prob": 0.8, |
| |
|
| | "gt_size": 512, |
| | "keep_ratio": True, |
| | "use_hflip": True, |
| | "use_rot": False, |
| |
|
| | |
| | "use_shuffle": True, |
| | "num_worker_per_gpu": 5, |
| | "batch_size_per_gpu": 12, |
| | "dataset_enlarge_ratio": 1, |
| | "prefetch_mode": None, |
| | } |
| |
|
| | def set_seed(seed=42): |
| | random.seed(seed) |
| | np.random.seed(seed) |
| | torch.manual_seed(seed) |
| | torch.cuda.manual_seed(seed) |
| | torch.cuda.manual_seed_all(seed) |
| |
|
| | torch.backends.cudnn.deterministic = True |
| | torch.backends.cudnn.benchmark = False |
| |
|
| | @DATASET_REGISTRY.register() |
| | class NoiseDataset(data.Dataset): |
| | """Dataset used for Denoise model: |
| | synthetic Gaussian and Poisson noise dataset. |
| | """ |
| | def __init__(self, opt, train=True, level=None): |
| | super(NoiseDataset, self).__init__() |
| | self.opt = opt |
| |
|
| | |
| | self.data_rt = opt['dataroot_gt'] |
| |
|
| | |
| | self.train = train |
| | if self.train: |
| | self.metas = opt['meta_train'] |
| | else: |
| | self.metas = opt['meta_test'] |
| |
|
| | self.paths = [] |
| | for meta in self.metas: |
| | with open(os.path.join(self.data_rt, meta)) as fin: |
| | self.paths += [line.strip().split(' ')[1] for line in fin] |
| |
|
| | |
| |
|
| | self.device = torch.cuda.current_device() |
| | self.jpeger = DiffJPEG(differentiable=False).to(self.device) |
| | self.usm_sharpener = USMSharp().to(self.device) |
| | self.resize = opt['gt_size'] |
| | self.keep_ratio = opt['keep_ratio'] |
| |
|
| | |
| | self.crop = T.RandomCrop((self.resize, self.resize)) |
| | self.flip = T.RandomHorizontalFlip() |
| | self.transform = T.Compose( |
| | [ |
| | |
| | T.ToTensor(), |
| | ] |
| | ) |
| |
|
| | |
| | self.sigma = [0.0588, 0.0784, 0.098, 0.1451, 0.1961] |
| | if level: |
| | self.level = [level] |
| | else: |
| | self.level = [1,2,3,4,5] |
| |
|
| | def __getitem__(self, index): |
| | |
| | gt_path = self.paths[index] |
| | img_gt = Image.open(gt_path).convert("RGB") |
| |
|
| | |
| | |
| | h, w = img_gt.height, img_gt.width |
| | if self.keep_ratio: |
| | ratio = self.resize / min(h, w) |
| | h_new, w_new = round(h * ratio * 1.2), round(w * ratio * 1.2) |
| | img_gt = img_gt.resize((w_new, h_new), resample=Image.LANCZOS) |
| | else: |
| | img_gt = img_gt.resize((self.resize, self.resize), resample=Image.LANCZOS) |
| | |
| | img_gt = self.crop(img_gt) |
| | |
| | if self.train: |
| | img_gt = self.flip(img_gt) |
| | |
| | img_gt = self.transform(img_gt).to(torch.float32) |
| |
|
| | |
| | |
| | peak = 255 |
| | lam = torch.clamp(img_gt, 0, 1) * peak |
| | counts = torch.poisson(lam) |
| | img_poisson = torch.clamp(counts / float(peak), 0.0, 1.0) |
| | |
| | |
| | level = random.choice(self.level) |
| | noise = torch.randn(size=img_poisson.size()) |
| | img_poisson_gaussian = torch.clamp(img_poisson + self.sigma[level-1] * noise, 0., 1.) |
| |
|
| | return img_poisson_gaussian, img_gt, gt_path |
| |
|
| | def __len__(self): |
| | return len(self.paths) |
| |
|
| | def poisson_gaussian_sampler(): |
| | """ |
| | It is now used for DF2K dataset (DIV2K + Flickr 2K) |
| | """ |
| | parser = argparse.ArgumentParser() |
| | parser.add_argument('--level', type=int, default=None, help='train: one to many') |
| | args = parser.parse_args() |
| |
|
| | if args.level: |
| | level = args.level |
| | else: |
| | level = [1,3,5] |
| |
|
| | |
| | for number in level: |
| | print("="*100) |
| | print(f"Generate Noise Level {number}...") |
| |
|
| | dataset = NoiseDataset(cfg, train=True, level=number) |
| | data_dl = data.DataLoader( |
| | dataset, |
| | batch_size = 1 |
| | ) |
| | print("Train Data:", dataset.data_rt, len(data_dl)) |
| | meta_info = {} |
| | for sample in tqdm(data_dl): |
| | lq, hq, path = sample |
| | |
| | file_name = os.path.basename(path[0]) |
| | gt_folder = os.path.dirname(path[0]) |
| | if "DIV2K_train_HR" in gt_folder or "DIV2K_valid_HR" in gt_folder: |
| | hq_folder = gt_folder.replace("HR", f"pair/Noise_L{number}/HQ") |
| | lq_folder = gt_folder.replace("HR", f"pair/Noise_L{number}/LQ") |
| | else: |
| | hq_folder = os.path.join(gt_folder.replace("images", f"images_pair/Noise_L{number}"), "HQ/") |
| | lq_folder = os.path.join(gt_folder.replace("images", f"images_pair/Noise_L{number}"), "LQ/") |
| |
|
| | os.makedirs(hq_folder, exist_ok=True) |
| | os.makedirs(lq_folder, exist_ok=True) |
| |
|
| | hq_path = os.path.join(hq_folder, file_name) |
| | lq_path = os.path.join(lq_folder, file_name) |
| |
|
| | save_image(hq[0], hq_path) |
| | save_image(lq[0], lq_path) |
| |
|
| | dset = os.path.relpath(gt_folder, dataset.data_rt).split("/")[0] |
| | if dset not in meta_info: |
| | meta_info[dset] = [(lq_path, hq_path)] |
| | else: |
| | meta_info[dset].append((lq_path, hq_path)) |
| |
|
| | for dset, dlist in meta_info.items(): |
| | with open(os.path.join(dataset.data_rt,'{}/metas/{}_train_Noise_L{}.list'.format(dset, dset, number)), 'w') as fp: |
| | for item in dlist: |
| | fp.write('{} {} {}\n'.format(item[0], item[1], None)) |
| | print(os.path.join(dataset.data_rt,'{}/metas/{}_train_Noise_L{}.list'.format(dset, dset, number)), len(dlist)) |
| |
|
| | |
| | dataset = NoiseDataset(cfg, train=False) |
| | data_dl = data.DataLoader( |
| | dataset, |
| | batch_size = 1 |
| | ) |
| | print("Test Data:", dataset.data_rt, len(data_dl)) |
| | print("="*100) |
| | print(f"Generate Testing Noise...") |
| |
|
| | meta_info = {} |
| | for sample in tqdm(data_dl): |
| | lq, hq, path = sample |
| | |
| | file_name = os.path.basename(path[0]) |
| | gt_folder = os.path.dirname(path[0]) |
| | if "DIV2K_train_HR" in gt_folder or "DIV2K_valid_HR" in gt_folder: |
| | hq_folder = gt_folder.replace("HR", f"pair/Noise/HQ") |
| | lq_folder = gt_folder.replace("HR", f"pair/Noise/LQ") |
| | else: |
| | hq_folder = os.path.join(gt_folder.replace("images", f"images_pair/Noise"), "HQ/") |
| | lq_folder = os.path.join(gt_folder.replace("images", f"images_pair/Noise"), "LQ/") |
| |
|
| | os.makedirs(hq_folder, exist_ok=True) |
| | os.makedirs(lq_folder, exist_ok=True) |
| |
|
| | hq_path = os.path.join(hq_folder, file_name) |
| | lq_path = os.path.join(lq_folder, file_name) |
| |
|
| | save_image(hq[0], hq_path) |
| | save_image(lq[0], lq_path) |
| |
|
| | dset = os.path.relpath(gt_folder, dataset.data_rt).split("/")[0] |
| | if dset not in meta_info: |
| | meta_info[dset] = [(lq_path, hq_path)] |
| | else: |
| | meta_info[dset].append((lq_path, hq_path)) |
| |
|
| | for dset, dlist in meta_info.items(): |
| | with open(os.path.join(dataset.data_rt,'{}/metas/{}_valid_Noise.list'.format(dset, dset)), 'w') as fp: |
| | for item in dlist: |
| | fp.write('{} {} {}\n'.format(item[0], item[1], None)) |
| | print(os.path.join(dataset.data_rt,'{}/metas/{}_valid_Noise.list'.format(dset, dset)), len(dlist)) |
| |
|
| | if __name__ == '__main__': |
| | set_seed(1229) |
| | |
| | poisson_gaussian_sampler() |
| |
|
| | |