simon-donike commited on
Commit
5dfd515
·
verified ·
1 Parent(s): 353e1d4

Upload Configs

Browse files
Files changed (2) hide show
  1. config_RGB-NIR.yaml +63 -0
  2. config_SWIR.yaml +61 -0
config_RGB-NIR.yaml ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Data:
2
+ train_batch_size: 12
3
+ val_batch_size: 8
4
+ num_workers: 6
5
+ prefetch_factor: 2
6
+ dataset_type: SISR_WW
7
+ normalization: normalise_10k
8
+ Model:
9
+ in_bands: 4
10
+ continue_training: false
11
+ load_checkpoint: false
12
+ Training:
13
+ gpus:
14
+ - 0
15
+ - 1
16
+ - 2
17
+ - 3
18
+ max_epochs: 9999
19
+ val_check_interval: 1.0
20
+ limit_val_batches: 250
21
+ pretrain_g_only: true
22
+ g_pretrain_steps: 15000
23
+ adv_loss_ramp_steps: 2500
24
+ label_smoothing: true
25
+ EMA:
26
+ enabled: true
27
+ decay: 0.999
28
+ update_after_step: 0
29
+ use_num_updates: true
30
+ Losses:
31
+ adv_loss_beta: 0.001
32
+ adv_loss_schedule: cosine
33
+ l1_weight: 1.0
34
+ sam_weight: 0.05
35
+ perceptual_weight: 0.1
36
+ perceptual_metric: vgg
37
+ tv_weight: 0.0
38
+ max_val: 1.0
39
+ ssim_win: 11
40
+ Generator:
41
+ model_type: SRResNet
42
+ large_kernel_size: 9
43
+ small_kernel_size: 3
44
+ n_channels: 96
45
+ n_blocks: 16
46
+ scaling_factor: 4
47
+ Discriminator:
48
+ model_type: patchgan
49
+ n_blocks: 3
50
+ Optimizers:
51
+ optim_g_lr: 0.0001
52
+ optim_d_lr: 0.0001
53
+ Schedulers:
54
+ g_warmup_steps: 5000
55
+ g_warmup_type: cosine
56
+ metric: val_metrics/l1
57
+ patience_g: 100
58
+ patience_d: 100
59
+ factor_g: 0.5
60
+ factor_d: 0.5
61
+ verbose: true
62
+ Logging:
63
+ num_val_images: 5
config_SWIR.yaml ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Data:
2
+ train_batch_size: 24
3
+ val_batch_size: 8
4
+ num_workers: 8
5
+ prefetch_factor: 4
6
+ dataset_type: S2_6b
7
+ normalization: normalise_10k
8
+ Model:
9
+ in_bands: 6
10
+ continue_training: false
11
+ load_checkpoint: false
12
+ Training:
13
+ gpus:
14
+ - 2
15
+ - 3
16
+ max_epochs: 9999
17
+ val_check_interval: 1.0
18
+ limit_val_batches: 250
19
+ pretrain_g_only: true
20
+ g_pretrain_steps: 15000
21
+ adv_loss_ramp_steps: 2500
22
+ label_smoothing: true
23
+ EMA:
24
+ enabled: false
25
+ decay: 0.999
26
+ update_after_step: 0
27
+ use_num_updates: true
28
+ Losses:
29
+ adv_loss_beta: 0.001
30
+ adv_loss_schedule: cosine
31
+ l1_weight: 1.0
32
+ sam_weight: 0.05
33
+ perceptual_weight: 0.1
34
+ perceptual_metric: vgg
35
+ tv_weight: 0.0
36
+ max_val: 1.0
37
+ ssim_win: 11
38
+ Generator:
39
+ model_type: rcab
40
+ large_kernel_size: 9
41
+ small_kernel_size: 3
42
+ n_channels: 96
43
+ n_blocks: 32
44
+ scaling_factor: 8
45
+ Discriminator:
46
+ model_type: standard
47
+ n_blocks: 8
48
+ Optimizers:
49
+ optim_g_lr: 0.0001
50
+ optim_d_lr: 0.0001
51
+ Schedulers:
52
+ g_warmup_steps: 2500
53
+ g_warmup_type: cosine
54
+ metric: val_metrics/l1
55
+ patience_g: 50
56
+ patience_d: 50
57
+ factor_g: 0.5
58
+ factor_d: 0.5
59
+ verbose: true
60
+ Logging:
61
+ num_val_images: 5