AaronCIH commited on
Commit
efff475
·
verified ·
1 Parent(s): 2e23197

Delete generate_lowresolution.py

Browse files
Files changed (1) hide show
  1. generate_lowresolution.py +0 -537
generate_lowresolution.py DELETED
@@ -1,537 +0,0 @@
1
- import argparse
2
- import glob
3
- import os
4
- from PIL import Image
5
-
6
- import cv2
7
- import math
8
- import numpy as np
9
- import os
10
- import os.path as osp
11
- import random
12
- import time
13
- import torch
14
- from tqdm import tqdm
15
-
16
- from basicsr.data.degradations import circular_lowpass_kernel, random_mixed_kernels
17
- from basicsr.data.transforms import augment
18
- from basicsr.utils import FileClient, get_root_logger, imfrombytes, img2tensor
19
- from basicsr.utils.registry import DATASET_REGISTRY
20
- from torch.utils import data as data
21
- from torchvision.transforms.functional import center_crop
22
- import torchvision.transforms as T
23
- from torchvision.utils import save_image
24
-
25
- from basicsr.data.degradations import random_add_gaussian_noise_pt, random_add_poisson_noise_pt
26
- from basicsr.data.transforms import paired_random_crop
27
- from basicsr.utils import DiffJPEG, USMSharp
28
- from basicsr.utils.img_process_util import filter2D
29
- from basicsr.utils.registry import MODEL_REGISTRY
30
- from collections import OrderedDict
31
- from torch.nn import functional as F
32
-
33
- cfg = {
34
- # dataset info.
35
- "name": "DF2K+OST",
36
- "type": "RealESRGANDataset",
37
- "dataroot_gt": "/home/CORP/hsiang.chen/Project/Datasets/IR/SuperResolution",
38
- "meta_train": [
39
- "DIV2K/metas/DIV2K_train_HR.list",
40
- "Flickr2K/metas/Flickr2K.list",
41
- "OST/metas/OST.list",
42
- ],
43
- "meta_test": ["DIV2K/metas/DIV2K_valid_HR.list"],
44
-
45
- # the first degradation process
46
- "resize_prob": [0.2, 0.7, 0.1], # up, down, keep
47
- "resize_range": [0.15, 1.5],
48
- "gaussian_noise_prob": 0.5,
49
- "noise_range": [1, 30],
50
- "poisson_scale_range": [0.05, 3],
51
- "gray_noise_prob": 0.4,
52
- "jpeg_range": [30, 95],
53
-
54
- "blur_kernel_size": 21,
55
- "kernel_list": ['iso', 'aniso', 'generalized_iso', 'generalized_aniso', 'plateau_iso', 'plateau_aniso'],
56
- "kernel_prob": [0.45, 0.25, 0.12, 0.03, 0.12, 0.03],
57
- "sinc_prob": 0.1,
58
- "blur_sigma": [0.2, 3],
59
- "betag_range": [0.5, 4],
60
- "betap_range": [1, 2],
61
-
62
- # the second degradation process
63
- "second_blur_prob": 0.8,
64
- "resize_prob2": [0.3, 0.4, 0.3], # up, down, keep
65
- "resize_range2": [0.3, 1.2],
66
- "gaussian_noise_prob2": 0.5,
67
- "noise_range2": [1, 25],
68
- "poisson_scale_range2": [0.05, 2.5],
69
- "gray_noise_prob2": 0.4,
70
- "jpeg_range2": [30, 95],
71
-
72
- "blur_kernel_size2": 21,
73
- "kernel_list2": ['iso', 'aniso', 'generalized_iso', 'generalized_aniso', 'plateau_iso', 'plateau_aniso'],
74
- "kernel_prob2": [0.45, 0.25, 0.12, 0.03, 0.12, 0.03],
75
- "sinc_prob2": 0.1,
76
- "blur_sigma2": [0.2, 1.5],
77
- "betag_range2": [0.5, 4],
78
- "betap_range2": [1, 2],
79
-
80
- "final_sinc_prob": 0.8,
81
-
82
- "gt_size": 512,
83
- "keep_ratio": True,
84
- "use_hflip": True,
85
- "use_rot": False,
86
-
87
- # data loader
88
- "use_shuffle": True,
89
- "num_worker_per_gpu": 5,
90
- "batch_size_per_gpu": 12,
91
- "dataset_enlarge_ratio": 1,
92
- "prefetch_mode": None,
93
- }
94
-
95
- def set_seed(seed=42):
96
- random.seed(seed)
97
- np.random.seed(seed)
98
- torch.manual_seed(seed)
99
- torch.cuda.manual_seed(seed)
100
- torch.cuda.manual_seed_all(seed)
101
-
102
- torch.backends.cudnn.deterministic = True
103
- torch.backends.cudnn.benchmark = False
104
-
105
- @DATASET_REGISTRY.register()
106
- class RealESRGANDataset(data.Dataset):
107
- """Dataset used for Real-ESRGAN model:
108
- Real-ESRGAN: Training Real-World Blind Super-Resolution with Pure Synthetic Data.
109
-
110
- It loads gt (Ground-Truth) images, and augments them.
111
- It also generates blur kernels and sinc kernels for generating low-quality images.
112
- Note that the low-quality images are processed in tensors on GPUS for faster processing.
113
-
114
- Args:
115
- opt (dict): Config for train datasets. It contains the following keys:
116
- dataroot_gt (str): Data root path for gt.
117
- meta_info (str): Path for meta information file.
118
- io_backend (dict): IO backend type and other kwarg.
119
- use_hflip (bool): Use horizontal flips.
120
- use_rot (bool): Use rotation (use vertical flip and transposing h and w for implementation).
121
- Please see more options in the codes.
122
- """
123
-
124
- def __init__(self, opt, train=True):
125
- super(RealESRGANDataset, self).__init__()
126
- self.opt = opt
127
- self.file_client = None
128
-
129
- # kernel define
130
- self.data_rt = opt['dataroot_gt']
131
-
132
- # dataload
133
- self.train = train
134
- if self.train:
135
- self.metas = opt['meta_train']
136
- else:
137
- self.metas = opt['meta_test']
138
-
139
- self.paths = []
140
- for meta in self.metas:
141
- with open(os.path.join(self.data_rt, meta)) as fin:
142
- self.paths += [line.strip().split(' ')[1] for line in fin]
143
-
144
- # Hyperparameter of Degradation
145
- # blur settings for the first degradation
146
- self.blur_kernel_size = opt['blur_kernel_size']
147
- self.kernel_list = opt['kernel_list']
148
- self.kernel_prob = opt['kernel_prob'] # a list for each kernel probability
149
- self.blur_sigma = opt['blur_sigma']
150
- self.betag_range = opt['betag_range'] # betag used in generalized Gaussian blur kernels
151
- self.betap_range = opt['betap_range'] # betap used in plateau blur kernels
152
- self.sinc_prob = opt['sinc_prob'] # the probability for sinc filters
153
-
154
- # blur settings for the second degradation
155
- self.blur_kernel_size2 = opt['blur_kernel_size2']
156
- self.kernel_list2 = opt['kernel_list2']
157
- self.kernel_prob2 = opt['kernel_prob2']
158
- self.blur_sigma2 = opt['blur_sigma2']
159
- self.betag_range2 = opt['betag_range2']
160
- self.betap_range2 = opt['betap_range2']
161
- self.sinc_prob2 = opt['sinc_prob2']
162
-
163
- # a final sinc filter
164
- self.final_sinc_prob = opt['final_sinc_prob']
165
-
166
- self.kernel_range = [2 * v + 1 for v in range(3, 11)] # kernel size ranges from 7 to 21
167
- # TODO: kernel range is now hard-coded, should be in the configure file
168
- self.pulse_tensor = torch.zeros(21, 21).float() # convolving with pulse tensor brings no blurry effect
169
- self.pulse_tensor[10, 10] = 1
170
-
171
- self.device = torch.cuda.current_device()
172
- self.jpeger = DiffJPEG(differentiable=False).to(self.device) # simulate JPEG compression artifacts
173
- self.usm_sharpener = USMSharp().to(self.device) # do usm sharpening
174
- self.resize = opt['gt_size']
175
- self.keep_ratio = opt['keep_ratio']
176
-
177
- # function
178
- self.crop = T.RandomCrop((self.resize, self.resize))
179
- self.flip = T.RandomHorizontalFlip()
180
- self.transform = T.Compose(
181
- [
182
- # T.ToDtype(torch.float32, scale=True), # only support for torch 2.++
183
- T.ToTensor(),
184
- ]
185
- )
186
-
187
- def __getitem__(self, index):
188
- # -------------------------------- Load gt images -------------------------------- #
189
- gt_path = self.paths[index]
190
- img_gt = Image.open(gt_path).convert("RGB")
191
-
192
- # -------------------------------- Image Process --------------------------------
193
- # resize
194
- h, w = img_gt.height, img_gt.width
195
- if self.keep_ratio:
196
- ratio = self.resize / min(h, w)
197
- h_new, w_new = round(h * ratio * 1.2), round(w * ratio * 1.2)
198
- img_gt = img_gt.resize((w_new, h_new), resample=Image.LANCZOS)
199
- else:
200
- img_gt = img_gt.resize((self.resize, self.resize), resample=Image.LANCZOS)
201
- # crop and
202
- img_gt = self.crop(img_gt)
203
- # flip (only for train)
204
- if self.train:
205
- img_gt = self.flip(img_gt)
206
- # transform to tensor
207
- img_gt = self.transform(img_gt).to(torch.float32)
208
-
209
- # -------------------------------- Generate Kernels --------------------------------
210
- kernel, kernel2, sinc_kernel = self.generate_kernel()
211
-
212
- # ------------------------- Generate Low Resolutino Sample -------------------------
213
- lq, hq = self.generate_lr({
214
- "gt": img_gt.unsqueeze(0),
215
- "kernel1": kernel,
216
- "kernel2": kernel2,
217
- "sinc_kernel": sinc_kernel,
218
- })
219
-
220
- return lq, hq, gt_path
221
-
222
- def generate_kernel(self, ):
223
- # ------------------------ Generate kernels (used in the first degradation) ------------------------ #
224
- kernel_size = random.choice(self.kernel_range)
225
- if np.random.uniform() < self.opt['sinc_prob']:
226
- # this sinc filter setting is for kernels ranging from [7, 21]
227
- if kernel_size < 13:
228
- omega_c = np.random.uniform(np.pi / 3, np.pi)
229
- else:
230
- omega_c = np.random.uniform(np.pi / 5, np.pi)
231
- kernel = circular_lowpass_kernel(omega_c, kernel_size, pad_to=False)
232
- else:
233
- kernel = random_mixed_kernels(
234
- self.kernel_list,
235
- self.kernel_prob,
236
- kernel_size,
237
- self.blur_sigma,
238
- self.blur_sigma, [-math.pi, math.pi],
239
- self.betag_range,
240
- self.betap_range,
241
- noise_range=None)
242
- # pad kernel
243
- pad_size = (21 - kernel_size) // 2
244
- kernel = np.pad(kernel, ((pad_size, pad_size), (pad_size, pad_size)))
245
- kernel = torch.FloatTensor(kernel)
246
-
247
- # ------------------------ Generate kernels (used in the second degradation) ------------------------ #
248
- kernel_size = random.choice(self.kernel_range)
249
- if np.random.uniform() < self.opt['sinc_prob2']:
250
- if kernel_size < 13:
251
- omega_c = np.random.uniform(np.pi / 3, np.pi)
252
- else:
253
- omega_c = np.random.uniform(np.pi / 5, np.pi)
254
- kernel2 = circular_lowpass_kernel(omega_c, kernel_size, pad_to=False)
255
- else:
256
- kernel2 = random_mixed_kernels(
257
- self.kernel_list2,
258
- self.kernel_prob2,
259
- kernel_size,
260
- self.blur_sigma2,
261
- self.blur_sigma2, [-math.pi, math.pi],
262
- self.betag_range2,
263
- self.betap_range2,
264
- noise_range=None)
265
-
266
- # pad kernel
267
- pad_size = (21 - kernel_size) // 2
268
- kernel2 = np.pad(kernel2, ((pad_size, pad_size), (pad_size, pad_size)))
269
- kernel2 = torch.FloatTensor(kernel2)
270
-
271
- # ------------------------------------- the final sinc kernel ------------------------------------- #
272
- if np.random.uniform() < self.opt['final_sinc_prob']:
273
- kernel_size = random.choice(self.kernel_range)
274
- omega_c = np.random.uniform(np.pi / 3, np.pi)
275
- sinc_kernel = circular_lowpass_kernel(omega_c, kernel_size, pad_to=21)
276
- sinc_kernel = torch.FloatTensor(sinc_kernel)
277
- else:
278
- sinc_kernel = self.pulse_tensor
279
- return kernel, kernel2, sinc_kernel
280
-
281
- def generate_lr(self, data):
282
- """Accept data from dataloader, and then add two-order degradations to obtain LQ images.
283
- """
284
- # training data synthesis
285
- self.gt = data['gt'].to(self.device)
286
- self.gt_usm = self.usm_sharpener(self.gt)
287
-
288
- self.kernel1 = data['kernel1'].to(self.device)
289
- self.kernel2 = data['kernel2'].to(self.device)
290
- self.sinc_kernel = data['sinc_kernel'].to(self.device)
291
-
292
- ori_h, ori_w = self.gt.size()[2:4]
293
-
294
- # ----------------------- The first degradation process ----------------------- #
295
- # blur
296
- out = filter2D(self.gt_usm, self.kernel1)
297
- # random resize
298
- updown_type = random.choices(['up', 'down', 'keep'], self.opt['resize_prob'])[0]
299
- if updown_type == 'up':
300
- scale = np.random.uniform(1, self.opt['resize_range'][1])
301
- elif updown_type == 'down':
302
- scale = np.random.uniform(self.opt['resize_range'][0], 1)
303
- else:
304
- scale = 1
305
- mode = random.choice(['area', 'bilinear', 'bicubic'])
306
- out = F.interpolate(out, scale_factor=scale, mode=mode)
307
- # add noise
308
- gray_noise_prob = self.opt['gray_noise_prob']
309
- if np.random.uniform() < self.opt['gaussian_noise_prob']:
310
- out = random_add_gaussian_noise_pt(
311
- out, sigma_range=self.opt['noise_range'], clip=True, rounds=False, gray_prob=gray_noise_prob)
312
- else:
313
- out = random_add_poisson_noise_pt(
314
- out,
315
- scale_range=self.opt['poisson_scale_range'],
316
- gray_prob=gray_noise_prob,
317
- clip=True,
318
- rounds=False)
319
- # JPEG compression
320
- jpeg_p = out.new_zeros(out.size(0)).uniform_(*self.opt['jpeg_range'])
321
- out = torch.clamp(out, 0, 1) # clamp to [0, 1], otherwise JPEGer will result in unpleasant artifacts
322
- out = self.jpeger(out, quality=jpeg_p)
323
-
324
- # ----------------------- The second degradation process ----------------------- #
325
- # blur
326
- if np.random.uniform() < self.opt['second_blur_prob']:
327
- out = filter2D(out, self.kernel2)
328
- # random resize
329
- updown_type = random.choices(['up', 'down', 'keep'], self.opt['resize_prob2'])[0]
330
- if updown_type == 'up':
331
- scale = np.random.uniform(1, self.opt['resize_range2'][1])
332
- elif updown_type == 'down':
333
- scale = np.random.uniform(self.opt['resize_range2'][0], 1)
334
- else:
335
- scale = 1
336
- mode = random.choice(['area', 'bilinear', 'bicubic'])
337
- out = F.interpolate(
338
- out, size=(int(ori_h * scale), int(ori_w * scale)), mode=mode)
339
- # add noise
340
- gray_noise_prob = self.opt['gray_noise_prob2']
341
- if np.random.uniform() < self.opt['gaussian_noise_prob2']:
342
- out = random_add_gaussian_noise_pt(
343
- out, sigma_range=self.opt['noise_range2'], clip=True, rounds=False, gray_prob=gray_noise_prob)
344
- else:
345
- out = random_add_poisson_noise_pt(
346
- out,
347
- scale_range=self.opt['poisson_scale_range2'],
348
- gray_prob=gray_noise_prob,
349
- clip=True,
350
- rounds=False)
351
-
352
- # JPEG compression + the final sinc filter
353
- # We also need to resize images to desired sizes. We group [resize back + sinc filter] together
354
- # as one operation.
355
- # We consider two orders:
356
- # 1. [resize back + sinc filter] + JPEG compression
357
- # 2. JPEG compression + [resize back + sinc filter]
358
- # Empirically, we find other combinations (sinc + JPEG + Resize) will introduce twisted lines.
359
- if np.random.uniform() < 0.5:
360
- # resize back + the final sinc filter
361
- mode = random.choice(['area', 'bilinear', 'bicubic'])
362
- out = F.interpolate(out, size=(ori_h, ori_w), mode=mode)
363
- out = filter2D(out, self.sinc_kernel)
364
- # JPEG compression
365
- jpeg_p = out.new_zeros(out.size(0)).uniform_(*self.opt['jpeg_range2'])
366
- out = torch.clamp(out, 0, 1)
367
- out = self.jpeger(out, quality=jpeg_p)
368
- else:
369
- # JPEG compression
370
- jpeg_p = out.new_zeros(out.size(0)).uniform_(*self.opt['jpeg_range2'])
371
- out = torch.clamp(out, 0, 1)
372
- out = self.jpeger(out, quality=jpeg_p)
373
- # resize back + the final sinc filter
374
- mode = random.choice(['area', 'bilinear', 'bicubic'])
375
- out = F.interpolate(out, size=(ori_h, ori_w), mode=mode)
376
- out = filter2D(out, self.sinc_kernel)
377
-
378
- # clamp and round
379
- lq = torch.clamp((out * 255.0).round(), 0, 255) / 255.
380
- lq = lq.contiguous() # for the warning: grad and param do not obey the gradient layout contract
381
-
382
- hq = self.usm_sharpener(self.gt)
383
- return lq[0], hq[0]
384
-
385
- def __len__(self):
386
- return len(self.paths)
387
-
388
- def real_esrgan_sampler():
389
- """
390
- Generate multi-scale versions for GT images with LANCZOS resampling.
391
- It is now used for DF2K dataset (DIV2K + Flickr 2K)
392
- """
393
- parser = argparse.ArgumentParser()
394
- parser.add_argument('--num_samples', type=int, default=3, help='train: one to many')
395
- args = parser.parse_args()
396
-
397
- # generate training dataset
398
- dataset = RealESRGANDataset(cfg, train=True)
399
- data_dl = data.DataLoader(
400
- dataset,
401
- batch_size = 1
402
- )
403
- print("Train Data:", dataset.data_rt, len(data_dl))
404
- for number in range(args.num_samples):
405
- print("="*100)
406
- print(f"Generate round {number}...")
407
-
408
- meta_info = {}
409
- for sample in tqdm(data_dl):
410
- lq, hq, path = sample
411
- # /home/CORP/hsiang.chen/Project/Datasets/IR/SuperResolution/DIV2K/DIV2K_train_HR/0098.png
412
- file_name = os.path.basename(path[0])
413
- gt_folder = os.path.dirname(path[0]) # /home/CORP/hsiang.chen/Project/Datasets/IR/SuperResolution/DIV2K/DIV2K_train_HR/
414
- if "DIV2K_train_HR" in gt_folder or "DIV2K_valid_HR" in gt_folder:
415
- hq_folder = gt_folder.replace("HR", f"pair/SR{number+1}/HR")
416
- lq_folder = gt_folder.replace("HR", f"pair/SR{number+1}/LR")
417
- else:
418
- hq_folder = os.path.join(gt_folder.replace("images", f"images_pair/SR{number+1}"), "HR/")
419
- lq_folder = os.path.join(gt_folder.replace("images", f"images_pair/SR{number+1}"), "LR/")
420
-
421
- os.makedirs(hq_folder, exist_ok=True)
422
- os.makedirs(lq_folder, exist_ok=True)
423
-
424
- hq_path = os.path.join(hq_folder, file_name)
425
- lq_path = os.path.join(lq_folder, file_name)
426
-
427
- save_image(hq[0], hq_path)
428
- save_image(lq[0], lq_path)
429
-
430
- dset = os.path.relpath(gt_folder, dataset.data_rt).split("/")[0]
431
- if dset not in meta_info:
432
- meta_info[dset] = [(lq_path, hq_path)]
433
- else:
434
- meta_info[dset].append((lq_path, hq_path))
435
-
436
- for dset, dlist in meta_info.items():
437
- with open(os.path.join(dataset.data_rt,'{}/metas/{}_train_SR{}.list'.format(dset, dset, number+1)), 'w') as fp:
438
- for item in dlist:
439
- fp.write('{} {} {}\n'.format(item[0], item[1], None))
440
- print(os.path.join(dataset.data_rt,'{}/metas/{}_train_SR{}.list'.format(dset, dset, number+1)), len(dlist))
441
-
442
-
443
- # generate testing dataset
444
- dataset = RealESRGANDataset(cfg, train=False)
445
- data_dl = data.DataLoader(
446
- dataset,
447
- batch_size = 1
448
- )
449
- print("Test Data:", dataset.data_rt, len(data_dl))
450
- print("="*100)
451
- print(f"Generate round {number}...")
452
-
453
- meta_info = {}
454
- for sample in tqdm(data_dl):
455
- lq, hq, path = sample
456
- # /home/CORP/hsiang.chen/Project/Datasets/IR/SuperResolution/DIV2K/DIV2K_train_HR/0098.png
457
- file_name = os.path.basename(path[0])
458
- gt_folder = os.path.dirname(path[0]) # /home/CORP/hsiang.chen/Project/Datasets/IR/SuperResolution/DIV2K/DIV2K_train_HR/
459
- if "DIV2K_train_HR" in gt_folder or "DIV2K_valid_HR" in gt_folder:
460
- hq_folder = gt_folder.replace("HR", f"pair/SR/HR")
461
- lq_folder = gt_folder.replace("HR", f"pair/SR/LR")
462
- else:
463
- hq_folder = os.path.join(gt_folder.replace("images", f"images_pair/SR"), "HR/")
464
- lq_folder = os.path.join(gt_folder.replace("images", f"images_pair/SR"), "LR/")
465
-
466
- os.makedirs(hq_folder, exist_ok=True)
467
- os.makedirs(lq_folder, exist_ok=True)
468
-
469
- hq_path = os.path.join(hq_folder, file_name)
470
- lq_path = os.path.join(lq_folder, file_name)
471
-
472
- save_image(hq[0], hq_path)
473
- save_image(lq[0], lq_path)
474
-
475
- dset = os.path.relpath(gt_folder, dataset.data_rt).split("/")[0]
476
- if dset not in meta_info:
477
- meta_info[dset] = [(lq_path, hq_path)]
478
- else:
479
- meta_info[dset].append((lq_path, hq_path))
480
-
481
- for dset, dlist in meta_info.items():
482
- with open(os.path.join(dataset.data_rt,'{}/metas/{}_valid_SR.list'.format(dset, dset)), 'w') as fp:
483
- for item in dlist:
484
- fp.write('{} {} {}\n'.format(item[0], item[1], None))
485
- print(os.path.join(dataset.data_rt,'{}/metas/{}_valid_SR.list'.format(dset, dset)), len(dlist))
486
-
487
- def simple_multiscale():
488
- """
489
- Generate multi-scale versions for GT images with LANCZOS resampling.
490
- It is now used for DF2K dataset (DIV2K + Flickr 2K)
491
- """
492
- parser = argparse.ArgumentParser()
493
- parser.add_argument('--input', type=str, default='DIV2K/DIV2K_train_HR', help='Input folder')
494
- parser.add_argument('--output', type=str, default='DIV2K/DIV2K_train_multiscale', help='Output folder')
495
- args = parser.parse_args()
496
- os.makedirs(args.output, exist_ok=True)
497
-
498
- # For DF2K, we consider the following three scales,
499
- # and the smallest image whose shortest edge is 400
500
- scale_list = [0.75, 0.5, 1 / 3]
501
- shortest_edge = 400
502
-
503
- path_list = sorted(glob.glob(os.path.join(args.input, '*')))
504
- for path in path_list:
505
- basename = os.path.splitext(os.path.basename(path))[0]
506
-
507
- img = Image.open(path)
508
- width, height = img.size
509
- for idx, scale in enumerate(scale_list):
510
- print(f'\t{scale:.2f}')
511
- rlt = img.resize((int(width * scale), int(height * scale)), resample=Image.LANCZOS)
512
- rlt = rlt.resize((width, height), resample=Image.NEAREST)
513
- rlt.save(os.path.join(args.output, f'{basename}T{idx}.png'))
514
-
515
- # save the smallest image which the shortest edge is 400
516
- if width < height:
517
- ratio = height / width
518
- width = shortest_edge
519
- height = int(width * ratio)
520
- else:
521
- ratio = width / height
522
- height = shortest_edge
523
- width = int(height * ratio)
524
- rlt = img.resize((int(width), int(height)), resample=Image.LANCZOS)
525
- rlt = rlt.resize(img.size, resample=Image.NEAREST)
526
- rlt.save(os.path.join(args.output, f'{basename}T{idx+1}.png'))
527
-
528
-
529
- if __name__ == '__main__':
530
- set_seed(1229)
531
- # simple version
532
- # simple_multiscale()
533
-
534
- # Real-ESRGAN for data generation
535
- real_esrgan_sampler()
536
-
537
- # python 2_generate_lowresolution.py --num_samples 3