valentinsingularity commited on
Commit
a6b0dee
·
verified ·
1 Parent(s): 7450196

Upload 3 files

Browse files
configs/c_absorption_L2_d500_s150_is30_r60_ws5_q03_nq256.yaml ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ general:
2
+ name: c_absorption_L2_d500_s150_is30_r60_ws5_q03_nq256
3
+ root_dir: null
4
+
5
+ dset:
6
+ prior_sampler:
7
+ cls: SubpriorParametricSampler
8
+ kwargs:
9
+ param_ranges:
10
+ thicknesses: [1., 500.]
11
+ roughnesses: [0., 60.]
12
+ slds: [0., 150.]
13
+ islds: [0., 30.]
14
+ bound_width_ranges:
15
+ thicknesses: [1.0e-2, 500.]
16
+ roughnesses: [1.0e-2, 60.]
17
+ slds: [1.0e-2, 5.]
18
+ islds: [1.0e-2, 5.]
19
+ model_name: model_with_absorption
20
+ max_num_layers: 2
21
+ constrained_roughness: true
22
+ constrained_isld: true
23
+ max_thickness_share: 0.5
24
+ max_sld_share: 0.2
25
+ logdist: false
26
+ scale_params_by_ranges: false
27
+ scaled_range: [-1., 1.]
28
+ device: 'cuda'
29
+
30
+ q_generator:
31
+ cls: ConstantQ
32
+ kwargs:
33
+ q: [0.02, 0.3, 256]
34
+ device: 'cuda'
35
+
36
+ intensity_noise:
37
+ cls: BasicExpIntensityNoise
38
+ kwargs:
39
+ relative_errors: [0.0, 0.2]
40
+ abs_errors: 0.0
41
+ consistent_rel_err: false
42
+ logdist: false
43
+ apply_shift: false
44
+ shift_range: [-0.05, 0.05]
45
+ apply_scaling: false
46
+ scale_range: [-0.01, 0.01]
47
+
48
+ curves_scaler:
49
+ cls: LogAffineCurvesScaler
50
+ kwargs:
51
+ weight: 0.2
52
+ bias: 1.0
53
+ eps: 1.0e-10
54
+
55
+ model:
56
+ network:
57
+ cls: NetworkWithPriorsConvEmb
58
+ pretrained_name: null
59
+ device: 'cuda'
60
+ kwargs:
61
+ in_channels: 1
62
+ hidden_channels: [32, 64, 128, 256, 512]
63
+ dim_embedding: 128
64
+ dim_avpool: 1
65
+ embedding_net_activation: 'gelu'
66
+ use_batch_norm: true
67
+ dim_out: 11
68
+ layer_width: 1024
69
+ num_blocks: 6
70
+ repeats_per_block: 2
71
+ mlp_activation: 'gelu'
72
+ dropout_rate: 0.0
73
+ pretrained_embedding_net: null
74
+
75
+ training:
76
+ num_iterations: 50000
77
+ batch_size: 4096
78
+ lr: 1.0e-4
79
+ grad_accumulation_steps: 1
80
+ clip_grad_norm_max: True
81
+ train_with_q_input: False
82
+ update_tqdm_freq: 1
83
+ optimizer: AdamW
84
+ trainer_kwargs:
85
+ optim_kwargs:
86
+ betas: [0.9, 0.999]
87
+ weight_decay: 0.0005
88
+ callbacks:
89
+ save_best_model:
90
+ enable: true
91
+ freq: 100
92
+ lr_scheduler:
93
+ cls: StepLR
94
+ kwargs:
95
+ step_size: 2500
96
+ gamma: 0.9
97
+ logger:
98
+ use_neptune: false
configs/c_absorption_L5_d200_s150_is30_r60_ws5_q03_nq256.yaml ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ general:
2
+ name: c_absorption_L5_d200_s150_is30_r60_ws5_q03_nq256
3
+ root_dir: null
4
+
5
+ dset:
6
+ prior_sampler:
7
+ cls: SubpriorParametricSampler
8
+ kwargs:
9
+ param_ranges:
10
+ thicknesses: [1., 200.]
11
+ roughnesses: [0., 60.]
12
+ slds: [0., 150.]
13
+ islds: [0., 30.]
14
+ bound_width_ranges:
15
+ thicknesses: [1.0e-2, 200.]
16
+ roughnesses: [1.0e-2, 60.]
17
+ slds: [1.0e-2, 5.]
18
+ islds: [1.0e-2, 5.]
19
+ model_name: model_with_absorption
20
+ max_num_layers: 5
21
+ constrained_roughness: true
22
+ constrained_isld: true
23
+ max_thickness_share: 0.5
24
+ max_sld_share: 0.2
25
+ logdist: false
26
+ scale_params_by_ranges: false
27
+ scaled_range: [-1., 1.]
28
+ device: 'cuda'
29
+
30
+ q_generator:
31
+ cls: ConstantQ
32
+ kwargs:
33
+ q: [0.02, 0.3, 256]
34
+ device: 'cuda'
35
+
36
+ intensity_noise:
37
+ cls: BasicExpIntensityNoise
38
+ kwargs:
39
+ relative_errors: [0.0, 0.2]
40
+ abs_errors: 0.0
41
+ consistent_rel_err: false
42
+ logdist: false
43
+ apply_shift: false
44
+ shift_range: [-0.05, 0.05]
45
+ apply_scaling: false
46
+ scale_range: [-0.01, 0.01]
47
+
48
+ curves_scaler:
49
+ cls: LogAffineCurvesScaler
50
+ kwargs:
51
+ weight: 0.2
52
+ bias: 1.0
53
+ eps: 1.0e-10
54
+
55
+ model:
56
+ network:
57
+ cls: NetworkWithPriorsConvEmb
58
+ pretrained_name: null
59
+ device: 'cuda'
60
+ kwargs:
61
+ in_channels: 1
62
+ hidden_channels: [32, 64, 128, 256, 512]
63
+ dim_embedding: 256
64
+ dim_avpool: 1
65
+ embedding_net_activation: 'gelu'
66
+ use_batch_norm: true
67
+ dim_out: 23
68
+ layer_width: 1024
69
+ num_blocks: 8
70
+ repeats_per_block: 2
71
+ mlp_activation: 'gelu'
72
+ dropout_rate: 0.0
73
+ pretrained_embedding_net: null
74
+
75
+ training:
76
+ num_iterations: 50000
77
+ batch_size: 4096
78
+ lr: 1.0e-4
79
+ grad_accumulation_steps: 1
80
+ clip_grad_norm_max: True
81
+ train_with_q_input: False
82
+ update_tqdm_freq: 1
83
+ optimizer: AdamW
84
+ trainer_kwargs:
85
+ optim_kwargs:
86
+ betas: [0.9, 0.999]
87
+ weight_decay: 0.0005
88
+ callbacks:
89
+ save_best_model:
90
+ enable: true
91
+ freq: 100
92
+ lr_scheduler:
93
+ cls: StepLR
94
+ kwargs:
95
+ step_size: 2500
96
+ gamma: 0.9
97
+ logger:
98
+ use_neptune: false
configs/c_repeating_multilayer_trained1.yaml ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ general:
2
+ name: c_repeating_multilayer_trained1
3
+ root_dir: null
4
+
5
+ dset:
6
+ prior_sampler:
7
+ cls: SubpriorParametricSampler
8
+ kwargs:
9
+ param_ranges:
10
+ d_full_rel: [0, 25]
11
+ rel_sigmas: [0, 5]
12
+ dr_sigmoid_rel_pos: [-10, 10]
13
+ dr_sigmoid_rel_width: [0, 20]
14
+ d_block1_rel: [0.01, 0.99]
15
+ d_block: [10, 20]
16
+ s_block_rel: [0., 0.3]
17
+ r_block: [0., 20.]
18
+ dr: [-10., 10.]
19
+ d3_rel: [0, 1]
20
+ s3_rel: [0, 1]
21
+ r3: [0., 25]
22
+ d_sio2: [0, 10]
23
+ s_sio2: [0, 10]
24
+ s_si: [0., 10]
25
+ r_sio2: [17., 19.]
26
+ r_si: [19., 21.]
27
+ bound_width_ranges:
28
+ d_full_rel: [0.1, 25]
29
+ rel_sigmas: [0.1, 5]
30
+ dr_sigmoid_rel_pos: [0.1, 20]
31
+ dr_sigmoid_rel_width: [0.1, 20]
32
+ d_block1_rel: [0.01, 1.0]
33
+ d_block: [0.1, 10.]
34
+ s_block_rel: [0.1, 0.3]
35
+ r_block: [0.1, 5.]
36
+ dr: [0.1, 5.]
37
+ d3_rel: [0.01, 1]
38
+ s3_rel: [0.01, 1]
39
+ r3: [0.01, 25]
40
+ d_sio2: [0.01, 10]
41
+ s_sio2: [0.01, 10]
42
+ s_si: [0.01, 10]
43
+ r_sio2: [0.01, 2]
44
+ r_si: [0.01, 2]
45
+ model_name: repeating_multilayer_v3
46
+ max_num_layers: 30
47
+ logdist: false
48
+ scale_params_by_ranges: false
49
+ scaled_range: [-1., 1.]
50
+ device: 'cuda'
51
+
52
+ q_generator:
53
+ cls: ConstantQ
54
+ kwargs:
55
+ q: [0.02, 0.5, 256]
56
+ device: 'cuda'
57
+
58
+ intensity_noise:
59
+ cls: BasicExpIntensityNoise
60
+ kwargs:
61
+ relative_errors: [0.0, 0.2]
62
+ abs_errors: 0.0
63
+ consistent_rel_err: true
64
+ logdist: false
65
+ apply_shift: true
66
+ shift_range: [-0.3, 0.3]
67
+ apply_scaling: true
68
+ scale_range: [-0.02, 0.02]
69
+
70
+ curves_scaler:
71
+ cls: LogAffineCurvesScaler
72
+ kwargs:
73
+ weight: 0.2
74
+ bias: 1.0
75
+ eps: 1.0e-10
76
+
77
+ model:
78
+ network:
79
+ cls: NetworkWithPriorsConvEmb
80
+ pretrained_name: null
81
+ device: 'cuda'
82
+ kwargs:
83
+ in_channels: 1
84
+ hidden_channels: [32, 64, 128, 256, 512]
85
+ dim_embedding: 128
86
+ dim_avpool: 1
87
+ embedding_net_activation: 'gelu'
88
+ use_batch_norm: true
89
+ dim_out: 17
90
+ layer_width: 512
91
+ num_blocks: 6
92
+ repeats_per_block: 2
93
+ mlp_activation: 'gelu'
94
+ dropout_rate: 0.0
95
+ pretrained_embedding_net: null
96
+
97
+ training:
98
+ num_iterations: 10000
99
+ batch_size: 4096
100
+ lr: 1.0e-4
101
+ grad_accumulation_steps: 1
102
+ clip_grad_norm_max: null
103
+ train_with_q_input: False
104
+ update_tqdm_freq: 1
105
+ optimizer: AdamW
106
+ trainer_kwargs:
107
+ optim_kwargs:
108
+ betas: [0.9, 0.999]
109
+ weight_decay: 0.0005
110
+ callbacks:
111
+ save_best_model:
112
+ enable: true
113
+ freq: 500
114
+ lr_scheduler:
115
+ cls: StepLR
116
+ kwargs:
117
+ step_size: 500
118
+ gamma: 0.5
119
+ logger:
120
+ use_neptune: false