|
|
import argparse |
|
|
import glob |
|
|
import os |
|
|
from PIL import Image |
|
|
|
|
|
import cv2 |
|
|
import math |
|
|
import numpy as np |
|
|
import os |
|
|
import os.path as osp |
|
|
import random |
|
|
import time |
|
|
import torch |
|
|
from tqdm import tqdm |
|
|
|
|
|
from basicsr.data.degradations import circular_lowpass_kernel, random_mixed_kernels |
|
|
from basicsr.data.transforms import augment |
|
|
from basicsr.utils import FileClient, get_root_logger, imfrombytes, img2tensor |
|
|
from basicsr.utils.registry import DATASET_REGISTRY |
|
|
from torch.utils import data as data |
|
|
from torchvision.transforms.functional import center_crop |
|
|
import torchvision.transforms as T |
|
|
from torchvision.utils import save_image |
|
|
|
|
|
from basicsr.data.degradations import random_add_gaussian_noise_pt, random_add_poisson_noise_pt |
|
|
from basicsr.data.transforms import paired_random_crop |
|
|
from basicsr.utils import DiffJPEG, USMSharp |
|
|
from basicsr.utils.img_process_util import filter2D |
|
|
from basicsr.utils.registry import MODEL_REGISTRY |
|
|
from collections import OrderedDict |
|
|
from torch.nn import functional as F |
|
|
|
|
|
cfg = { |
|
|
|
|
|
"name": "DF2K+OST", |
|
|
"type": "RealESRGANDataset", |
|
|
"dataroot_gt": "/home/CORP/hsiang.chen/Project/Datasets/IR/SuperResolution", |
|
|
"meta_train": [ |
|
|
"DIV2K/metas/DIV2K_train_HR.list", |
|
|
"Flickr2K/metas/Flickr2K.list", |
|
|
|
|
|
], |
|
|
"meta_test": ["DIV2K/metas/DIV2K_valid_HR.list"], |
|
|
|
|
|
|
|
|
"resize_prob": [0.2, 0.7, 0.1], |
|
|
"resize_range": [0.15, 1.5], |
|
|
"gaussian_noise_prob": 0.5, |
|
|
"noise_range": [1, 30], |
|
|
"poisson_scale_range": [0.05, 3], |
|
|
"gray_noise_prob": 0.4, |
|
|
"jpeg_range": [30, 95], |
|
|
|
|
|
"blur_kernel_size": 21, |
|
|
"kernel_list": ['iso', 'aniso', 'generalized_iso', 'generalized_aniso', 'plateau_iso', 'plateau_aniso'], |
|
|
"kernel_prob": [0.45, 0.25, 0.12, 0.03, 0.12, 0.03], |
|
|
"sinc_prob": 0.1, |
|
|
"blur_sigma": [0.2, 3], |
|
|
"betag_range": [0.5, 4], |
|
|
"betap_range": [1, 2], |
|
|
|
|
|
|
|
|
"second_blur_prob": 0.8, |
|
|
"resize_prob2": [0.3, 0.4, 0.3], |
|
|
"resize_range2": [0.3, 1.2], |
|
|
"gaussian_noise_prob2": 0.5, |
|
|
"noise_range2": [1, 25], |
|
|
"poisson_scale_range2": [0.05, 2.5], |
|
|
"gray_noise_prob2": 0.4, |
|
|
"jpeg_range2": [30, 95], |
|
|
|
|
|
"blur_kernel_size2": 21, |
|
|
"kernel_list2": ['iso', 'aniso', 'generalized_iso', 'generalized_aniso', 'plateau_iso', 'plateau_aniso'], |
|
|
"kernel_prob2": [0.45, 0.25, 0.12, 0.03, 0.12, 0.03], |
|
|
"sinc_prob2": 0.1, |
|
|
"blur_sigma2": [0.2, 1.5], |
|
|
"betag_range2": [0.5, 4], |
|
|
"betap_range2": [1, 2], |
|
|
|
|
|
"final_sinc_prob": 0.8, |
|
|
|
|
|
"gt_size": 512, |
|
|
"keep_ratio": True, |
|
|
"use_hflip": True, |
|
|
"use_rot": False, |
|
|
|
|
|
|
|
|
"use_shuffle": True, |
|
|
"num_worker_per_gpu": 5, |
|
|
"batch_size_per_gpu": 12, |
|
|
"dataset_enlarge_ratio": 1, |
|
|
"prefetch_mode": None, |
|
|
} |
|
|
|
|
|
def set_seed(seed=42): |
|
|
random.seed(seed) |
|
|
np.random.seed(seed) |
|
|
torch.manual_seed(seed) |
|
|
torch.cuda.manual_seed(seed) |
|
|
torch.cuda.manual_seed_all(seed) |
|
|
|
|
|
torch.backends.cudnn.deterministic = True |
|
|
torch.backends.cudnn.benchmark = False |
|
|
|
|
|
@DATASET_REGISTRY.register() |
|
|
class NoiseDataset(data.Dataset): |
|
|
"""Dataset used for Denoise model: |
|
|
synthetic Gaussian and Poisson noise dataset. |
|
|
""" |
|
|
def __init__(self, opt, train=True, level=None): |
|
|
super(NoiseDataset, self).__init__() |
|
|
self.opt = opt |
|
|
|
|
|
|
|
|
self.data_rt = opt['dataroot_gt'] |
|
|
|
|
|
|
|
|
self.train = train |
|
|
if self.train: |
|
|
self.metas = opt['meta_train'] |
|
|
else: |
|
|
self.metas = opt['meta_test'] |
|
|
|
|
|
self.paths = [] |
|
|
for meta in self.metas: |
|
|
with open(os.path.join(self.data_rt, meta)) as fin: |
|
|
self.paths += [line.strip().split(' ')[1] for line in fin] |
|
|
|
|
|
|
|
|
|
|
|
self.device = torch.cuda.current_device() |
|
|
self.jpeger = DiffJPEG(differentiable=False).to(self.device) |
|
|
self.usm_sharpener = USMSharp().to(self.device) |
|
|
self.resize = opt['gt_size'] |
|
|
self.keep_ratio = opt['keep_ratio'] |
|
|
|
|
|
|
|
|
self.crop = T.RandomCrop((self.resize, self.resize)) |
|
|
self.flip = T.RandomHorizontalFlip() |
|
|
self.transform = T.Compose( |
|
|
[ |
|
|
|
|
|
T.ToTensor(), |
|
|
] |
|
|
) |
|
|
|
|
|
|
|
|
self.sigma = [0.0588, 0.0784, 0.098, 0.1451, 0.1961] |
|
|
if level: |
|
|
self.level = [level] |
|
|
else: |
|
|
self.level = [1,2,3,4,5] |
|
|
|
|
|
def __getitem__(self, index): |
|
|
|
|
|
gt_path = self.paths[index] |
|
|
img_gt = Image.open(gt_path).convert("RGB") |
|
|
|
|
|
|
|
|
|
|
|
h, w = img_gt.height, img_gt.width |
|
|
if self.keep_ratio: |
|
|
ratio = self.resize / min(h, w) |
|
|
h_new, w_new = round(h * ratio * 1.2), round(w * ratio * 1.2) |
|
|
img_gt = img_gt.resize((w_new, h_new), resample=Image.LANCZOS) |
|
|
else: |
|
|
img_gt = img_gt.resize((self.resize, self.resize), resample=Image.LANCZOS) |
|
|
|
|
|
img_gt = self.crop(img_gt) |
|
|
|
|
|
if self.train: |
|
|
img_gt = self.flip(img_gt) |
|
|
|
|
|
img_gt = self.transform(img_gt).to(torch.float32) |
|
|
|
|
|
|
|
|
|
|
|
peak = 255 |
|
|
lam = torch.clamp(img_gt, 0, 1) * peak |
|
|
counts = torch.poisson(lam) |
|
|
img_poisson = torch.clamp(counts / float(peak), 0.0, 1.0) |
|
|
|
|
|
|
|
|
level = random.choice(self.level) |
|
|
noise = torch.randn(size=img_poisson.size()) |
|
|
img_poisson_gaussian = torch.clamp(img_poisson + self.sigma[level-1] * noise, 0., 1.) |
|
|
|
|
|
return img_poisson_gaussian, img_gt, gt_path |
|
|
|
|
|
def __len__(self): |
|
|
return len(self.paths) |
|
|
|
|
|
def poisson_gaussian_sampler(): |
|
|
""" |
|
|
It is now used for DF2K dataset (DIV2K + Flickr 2K) |
|
|
""" |
|
|
parser = argparse.ArgumentParser() |
|
|
parser.add_argument('--level', type=int, default=None, help='train: one to many') |
|
|
args = parser.parse_args() |
|
|
|
|
|
if args.level: |
|
|
level = args.level |
|
|
else: |
|
|
level = [1,3,5] |
|
|
|
|
|
|
|
|
for number in level: |
|
|
print("="*100) |
|
|
print(f"Generate Noise Level {number}...") |
|
|
|
|
|
dataset = NoiseDataset(cfg, train=True, level=number) |
|
|
data_dl = data.DataLoader( |
|
|
dataset, |
|
|
batch_size = 1 |
|
|
) |
|
|
print("Train Data:", dataset.data_rt, len(data_dl)) |
|
|
meta_info = {} |
|
|
for sample in tqdm(data_dl): |
|
|
lq, hq, path = sample |
|
|
|
|
|
file_name = os.path.basename(path[0]) |
|
|
gt_folder = os.path.dirname(path[0]) |
|
|
if "DIV2K_train_HR" in gt_folder or "DIV2K_valid_HR" in gt_folder: |
|
|
hq_folder = gt_folder.replace("HR", f"pair/Noise_L{number}/HQ") |
|
|
lq_folder = gt_folder.replace("HR", f"pair/Noise_L{number}/LQ") |
|
|
else: |
|
|
hq_folder = os.path.join(gt_folder.replace("images", f"images_pair/Noise_L{number}"), "HQ/") |
|
|
lq_folder = os.path.join(gt_folder.replace("images", f"images_pair/Noise_L{number}"), "LQ/") |
|
|
|
|
|
os.makedirs(hq_folder, exist_ok=True) |
|
|
os.makedirs(lq_folder, exist_ok=True) |
|
|
|
|
|
hq_path = os.path.join(hq_folder, file_name) |
|
|
lq_path = os.path.join(lq_folder, file_name) |
|
|
|
|
|
save_image(hq[0], hq_path) |
|
|
save_image(lq[0], lq_path) |
|
|
|
|
|
dset = os.path.relpath(gt_folder, dataset.data_rt).split("/")[0] |
|
|
if dset not in meta_info: |
|
|
meta_info[dset] = [(lq_path, hq_path)] |
|
|
else: |
|
|
meta_info[dset].append((lq_path, hq_path)) |
|
|
|
|
|
for dset, dlist in meta_info.items(): |
|
|
with open(os.path.join(dataset.data_rt,'{}/metas/{}_train_Noise_L{}.list'.format(dset, dset, number)), 'w') as fp: |
|
|
for item in dlist: |
|
|
fp.write('{} {} {}\n'.format(item[0], item[1], None)) |
|
|
print(os.path.join(dataset.data_rt,'{}/metas/{}_train_Noise_L{}.list'.format(dset, dset, number)), len(dlist)) |
|
|
|
|
|
|
|
|
dataset = NoiseDataset(cfg, train=False) |
|
|
data_dl = data.DataLoader( |
|
|
dataset, |
|
|
batch_size = 1 |
|
|
) |
|
|
print("Test Data:", dataset.data_rt, len(data_dl)) |
|
|
print("="*100) |
|
|
print(f"Generate Testing Noise...") |
|
|
|
|
|
meta_info = {} |
|
|
for sample in tqdm(data_dl): |
|
|
lq, hq, path = sample |
|
|
|
|
|
file_name = os.path.basename(path[0]) |
|
|
gt_folder = os.path.dirname(path[0]) |
|
|
if "DIV2K_train_HR" in gt_folder or "DIV2K_valid_HR" in gt_folder: |
|
|
hq_folder = gt_folder.replace("HR", f"pair/Noise/HQ") |
|
|
lq_folder = gt_folder.replace("HR", f"pair/Noise/LQ") |
|
|
else: |
|
|
hq_folder = os.path.join(gt_folder.replace("images", f"images_pair/Noise"), "HQ/") |
|
|
lq_folder = os.path.join(gt_folder.replace("images", f"images_pair/Noise"), "LQ/") |
|
|
|
|
|
os.makedirs(hq_folder, exist_ok=True) |
|
|
os.makedirs(lq_folder, exist_ok=True) |
|
|
|
|
|
hq_path = os.path.join(hq_folder, file_name) |
|
|
lq_path = os.path.join(lq_folder, file_name) |
|
|
|
|
|
save_image(hq[0], hq_path) |
|
|
save_image(lq[0], lq_path) |
|
|
|
|
|
dset = os.path.relpath(gt_folder, dataset.data_rt).split("/")[0] |
|
|
if dset not in meta_info: |
|
|
meta_info[dset] = [(lq_path, hq_path)] |
|
|
else: |
|
|
meta_info[dset].append((lq_path, hq_path)) |
|
|
|
|
|
for dset, dlist in meta_info.items(): |
|
|
with open(os.path.join(dataset.data_rt,'{}/metas/{}_valid_Noise.list'.format(dset, dset)), 'w') as fp: |
|
|
for item in dlist: |
|
|
fp.write('{} {} {}\n'.format(item[0], item[1], None)) |
|
|
print(os.path.join(dataset.data_rt,'{}/metas/{}_valid_Noise.list'.format(dset, dset)), len(dlist)) |
|
|
|
|
|
if __name__ == '__main__': |
|
|
set_seed(1229) |
|
|
|
|
|
poisson_gaussian_sampler() |
|
|
|
|
|
|