AaronCIH commited on
Commit
31f842e
·
verified ·
1 Parent(s): efff475

Delete generate_noise.py

Browse files
Files changed (1) hide show
  1. generate_noise.py +0 -308
generate_noise.py DELETED
@@ -1,308 +0,0 @@
1
- import argparse
2
- import glob
3
- import os
4
- from PIL import Image
5
-
6
- import cv2
7
- import math
8
- import numpy as np
9
- import os
10
- import os.path as osp
11
- import random
12
- import time
13
- import torch
14
- from tqdm import tqdm
15
-
16
- from basicsr.data.degradations import circular_lowpass_kernel, random_mixed_kernels
17
- from basicsr.data.transforms import augment
18
- from basicsr.utils import FileClient, get_root_logger, imfrombytes, img2tensor
19
- from basicsr.utils.registry import DATASET_REGISTRY
20
- from torch.utils import data as data
21
- from torchvision.transforms.functional import center_crop
22
- import torchvision.transforms as T
23
- from torchvision.utils import save_image
24
-
25
- from basicsr.data.degradations import random_add_gaussian_noise_pt, random_add_poisson_noise_pt
26
- from basicsr.data.transforms import paired_random_crop
27
- from basicsr.utils import DiffJPEG, USMSharp
28
- from basicsr.utils.img_process_util import filter2D
29
- from basicsr.utils.registry import MODEL_REGISTRY
30
- from collections import OrderedDict
31
- from torch.nn import functional as F
32
-
33
- cfg = {
34
- # dataset info.
35
- "name": "DF2K+OST",
36
- "type": "RealESRGANDataset",
37
- "dataroot_gt": "/home/CORP/hsiang.chen/Project/Datasets/IR/Denoise",
38
- "meta_train": [
39
- "BSD68/metas/BSD68.list",
40
- "BSD400/metas/BSD400.list",
41
- "CBSD68/metas/CBSD68.list",
42
- "Kodak/metas/Kodak.list",
43
- "McMaster/metas/McMaster.list",
44
- "Set12/metas/Set12.list",
45
- "Urban100/metas/Urban100.list",
46
- "WaterlooED/metas/WaterlooED.list",
47
- ],
48
- # "meta_test": ["DIV2K/metas/DIV2K_valid_HR.list"],
49
-
50
- # the first degradation process
51
- "resize_prob": [0.2, 0.7, 0.1], # up, down, keep
52
- "resize_range": [0.15, 1.5],
53
- "gaussian_noise_prob": 0.5,
54
- "noise_range": [1, 30],
55
- "poisson_scale_range": [0.05, 3],
56
- "gray_noise_prob": 0.4,
57
- "jpeg_range": [30, 95],
58
-
59
- "blur_kernel_size": 21,
60
- "kernel_list": ['iso', 'aniso', 'generalized_iso', 'generalized_aniso', 'plateau_iso', 'plateau_aniso'],
61
- "kernel_prob": [0.45, 0.25, 0.12, 0.03, 0.12, 0.03],
62
- "sinc_prob": 0.1,
63
- "blur_sigma": [0.2, 3],
64
- "betag_range": [0.5, 4],
65
- "betap_range": [1, 2],
66
-
67
- # the second degradation process
68
- "second_blur_prob": 0.8,
69
- "resize_prob2": [0.3, 0.4, 0.3], # up, down, keep
70
- "resize_range2": [0.3, 1.2],
71
- "gaussian_noise_prob2": 0.5,
72
- "noise_range2": [1, 25],
73
- "poisson_scale_range2": [0.05, 2.5],
74
- "gray_noise_prob2": 0.4,
75
- "jpeg_range2": [30, 95],
76
-
77
- "blur_kernel_size2": 21,
78
- "kernel_list2": ['iso', 'aniso', 'generalized_iso', 'generalized_aniso', 'plateau_iso', 'plateau_aniso'],
79
- "kernel_prob2": [0.45, 0.25, 0.12, 0.03, 0.12, 0.03],
80
- "sinc_prob2": 0.1,
81
- "blur_sigma2": [0.2, 1.5],
82
- "betag_range2": [0.5, 4],
83
- "betap_range2": [1, 2],
84
-
85
- "final_sinc_prob": 0.8,
86
-
87
- "gt_size": 512,
88
- "keep_ratio": True,
89
- "use_hflip": True,
90
- "use_rot": False,
91
-
92
- # data loader
93
- "use_shuffle": True,
94
- "num_worker_per_gpu": 5,
95
- "batch_size_per_gpu": 12,
96
- "dataset_enlarge_ratio": 1,
97
- "prefetch_mode": None,
98
- }
99
-
100
- def set_seed(seed=42):
101
- random.seed(seed)
102
- np.random.seed(seed)
103
- torch.manual_seed(seed)
104
- torch.cuda.manual_seed(seed)
105
- torch.cuda.manual_seed_all(seed)
106
-
107
- torch.backends.cudnn.deterministic = True
108
- torch.backends.cudnn.benchmark = False
109
-
110
- @DATASET_REGISTRY.register()
111
- class NoiseDataset(data.Dataset):
112
- """Dataset used for Denoise model:
113
- synthetic Gaussian and Poisson noise dataset.
114
- """
115
- def __init__(self, opt, train=True, level=None):
116
- super(NoiseDataset, self).__init__()
117
- self.opt = opt
118
-
119
- # kernel define
120
- self.data_rt = opt['dataroot_gt']
121
-
122
- # dataload
123
- self.train = train
124
- if self.train:
125
- self.metas = opt['meta_train']
126
- else:
127
- self.metas = opt['meta_test']
128
-
129
- self.paths = []
130
- for meta in self.metas:
131
- with open(os.path.join(self.data_rt, meta)) as fin:
132
- self.paths += [line.strip().split(' ')[1] for line in fin]
133
-
134
- # hyperparameter
135
-
136
- self.device = torch.cuda.current_device()
137
- self.jpeger = DiffJPEG(differentiable=False).to(self.device) # simulate JPEG compression artifacts
138
- self.usm_sharpener = USMSharp().to(self.device) # do usm sharpening
139
- self.resize = opt['gt_size']
140
- self.keep_ratio = opt['keep_ratio']
141
-
142
- # function
143
- self.crop = T.RandomCrop((self.resize, self.resize))
144
- self.flip = T.RandomHorizontalFlip()
145
- self.transform = T.Compose(
146
- [
147
- # T.ToDtype(torch.float32, scale=True), # only support for torch 2.++
148
- T.ToTensor(),
149
- ]
150
- )
151
-
152
- # noise
153
- self.sigma = [0.0588, 0.0784, 0.098, 0.1451, 0.1961] # 5 levels: 15, 20, 25, 37, 50
154
- if level:
155
- self.level = [level]
156
- else:
157
- self.level = [1,2,3,4,5]
158
-
159
- def __getitem__(self, index):
160
- # -------------------------------- Load gt images -------------------------------- #
161
- gt_path = self.paths[index]
162
- img_gt = Image.open(gt_path).convert("RGB")
163
-
164
- # -------------------------------- Image Process --------------------------------
165
- # resize
166
- h, w = img_gt.height, img_gt.width
167
- if self.keep_ratio:
168
- ratio = self.resize / min(h, w)
169
- h_new, w_new = round(h * ratio * 1.2), round(w * ratio * 1.2)
170
- img_gt = img_gt.resize((w_new, h_new), resample=Image.LANCZOS)
171
- else:
172
- img_gt = img_gt.resize((self.resize, self.resize), resample=Image.LANCZOS)
173
- # crop and
174
- img_gt = self.crop(img_gt)
175
- # flip (only for train)
176
- if self.train:
177
- img_gt = self.flip(img_gt)
178
- # transform to tensor
179
- img_gt = self.transform(img_gt).to(torch.float32)
180
-
181
- # -------------------------------- Generate Noise --------------------------------
182
- # Poisson Noise
183
- peak = 255
184
- lam = torch.clamp(img_gt, 0, 1) * peak
185
- counts = torch.poisson(lam)
186
- img_poisson = torch.clamp(counts / float(peak), 0.0, 1.0)
187
-
188
- # Gaussian Noise
189
- level = random.choice(self.level)
190
- noise = torch.randn(size=img_poisson.size())
191
- img_poisson_gaussian = torch.clamp(img_poisson + self.sigma[level-1] * noise, 0., 1.)
192
-
193
- return img_poisson_gaussian, img_gt, gt_path
194
-
195
- def __len__(self):
196
- return len(self.paths)
197
-
198
- def poisson_gaussian_sampler():
199
- """
200
- It is now used for DF2K dataset (DIV2K + Flickr 2K)
201
- """
202
- parser = argparse.ArgumentParser()
203
- parser.add_argument('--level', type=int, default=None, help='train: one to many')
204
- args = parser.parse_args()
205
-
206
- if args.level:
207
- level = args.level
208
- else:
209
- level = [1,3,5]
210
-
211
- # generate training dataset
212
- for number in level:
213
- print("="*100)
214
- print(f"Generate Noise Level {number}...")
215
-
216
- dataset = NoiseDataset(cfg, train=True, level=number)
217
- data_dl = data.DataLoader(
218
- dataset,
219
- batch_size = 1
220
- )
221
- print("Train Data:", dataset.data_rt, len(data_dl))
222
- meta_info = {}
223
- for sample in tqdm(data_dl):
224
- lq, hq, path = sample
225
- # /home/CORP/hsiang.chen/Project/Datasets/IR/SuperResolution/DIV2K/DIV2K_train_HR/0098.png
226
- file_name = os.path.basename(path[0])
227
- gt_folder = os.path.dirname(path[0]) # /home/CORP/hsiang.chen/Project/Datasets/IR/SuperResolution/DIV2K/DIV2K_train_HR/
228
- if "DIV2K_train_HR" in gt_folder or "DIV2K_valid_HR" in gt_folder:
229
- hq_folder = gt_folder.replace("HR", f"pair/Noise_L{number}/HQ")
230
- lq_folder = gt_folder.replace("HR", f"pair/Noise_L{number}/LQ")
231
- elif "Denoise" in gt_folder:
232
- hq_folder = os.path.join(gt_folder.replace("image", f"image_pair/Noise_L{number}"), "HQ/")
233
- lq_folder = os.path.join(gt_folder.replace("image", f"image_pair/Noise_L{number}"), "LQ/")
234
- else:
235
- hq_folder = os.path.join(gt_folder.replace("images", f"images_pair/Noise_L{number}"), "HQ/")
236
- lq_folder = os.path.join(gt_folder.replace("images", f"images_pair/Noise_L{number}"), "LQ/")
237
-
238
- os.makedirs(hq_folder, exist_ok=True)
239
- os.makedirs(lq_folder, exist_ok=True)
240
-
241
- hq_path = os.path.join(hq_folder, file_name)
242
- lq_path = os.path.join(lq_folder, file_name)
243
-
244
- save_image(hq[0], hq_path)
245
- save_image(lq[0], lq_path)
246
-
247
- dset = os.path.relpath(gt_folder, dataset.data_rt).split("/")[0]
248
- if dset not in meta_info:
249
- meta_info[dset] = [(lq_path, hq_path)]
250
- else:
251
- meta_info[dset].append((lq_path, hq_path))
252
-
253
- for dset, dlist in meta_info.items():
254
- with open(os.path.join(dataset.data_rt,'{}/metas/{}_train_Noise_L{}.list'.format(dset, dset, number)), 'w') as fp:
255
- for item in dlist:
256
- fp.write('{} {} {}\n'.format(item[0], item[1], None))
257
- print(os.path.join(dataset.data_rt,'{}/metas/{}_train_Noise_L{}.list'.format(dset, dset, number)), len(dlist))
258
-
259
- # generate testing dataset
260
- # dataset = NoiseDataset(cfg, train=False)
261
- # data_dl = data.DataLoader(
262
- # dataset,
263
- # batch_size = 1
264
- # )
265
- # print("Test Data:", dataset.data_rt, len(data_dl))
266
- # print("="*100)
267
- # print(f"Generate Testing Noise...")
268
-
269
- # meta_info = {}
270
- # for sample in tqdm(data_dl):
271
- # lq, hq, path = sample
272
- # # /home/CORP/hsiang.chen/Project/Datasets/IR/SuperResolution/DIV2K/DIV2K_train_HR/0098.png
273
- # file_name = os.path.basename(path[0])
274
- # gt_folder = os.path.dirname(path[0]) # /home/CORP/hsiang.chen/Project/Datasets/IR/SuperResolution/DIV2K/DIV2K_train_HR/
275
- # if "DIV2K_train_HR" in gt_folder or "DIV2K_valid_HR" in gt_folder:
276
- # hq_folder = gt_folder.replace("HR", f"pair/Noise/HQ")
277
- # lq_folder = gt_folder.replace("HR", f"pair/Noise/LQ")
278
- # else:
279
- # hq_folder = os.path.join(gt_folder.replace("images", f"images_pair/Noise"), "HQ/")
280
- # lq_folder = os.path.join(gt_folder.replace("images", f"images_pair/Noise"), "LQ/")
281
-
282
- # os.makedirs(hq_folder, exist_ok=True)
283
- # os.makedirs(lq_folder, exist_ok=True)
284
-
285
- # hq_path = os.path.join(hq_folder, file_name)
286
- # lq_path = os.path.join(lq_folder, file_name)
287
-
288
- # save_image(hq[0], hq_path)
289
- # save_image(lq[0], lq_path)
290
-
291
- # dset = os.path.relpath(gt_folder, dataset.data_rt).split("/")[0]
292
- # if dset not in meta_info:
293
- # meta_info[dset] = [(lq_path, hq_path)]
294
- # else:
295
- # meta_info[dset].append((lq_path, hq_path))
296
-
297
- # for dset, dlist in meta_info.items():
298
- # with open(os.path.join(dataset.data_rt,'{}/metas/{}_valid_Noise.list'.format(dset, dset)), 'w') as fp:
299
- # for item in dlist:
300
- # fp.write('{} {} {}\n'.format(item[0], item[1], None))
301
- # print(os.path.join(dataset.data_rt,'{}/metas/{}_valid_Noise.list'.format(dset, dset)), len(dlist))
302
-
303
- if __name__ == '__main__':
304
- set_seed(1229)
305
- # poisson_gaussian for data generation
306
- poisson_gaussian_sampler()
307
-
308
- # python 3_generate_noise.py