Nayoung Kim commited on
Commit
4c4307a
·
verified ·
1 Parent(s): 8675601

Add files using upload-large-folder tool

Browse files
SHA256SUMS ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ 2f9345c143cd7576472a0b88815504e4540447e5fcf2c4e927f62f431beeb10c ./gen/seq_module_unconditional/seq_module_unconditional_ckpt.tar.gz
2
+ 938f13187c5e76df0c44c959ae2c6c44a764c2d8c673a2ac104e5cfa3cd05199 ./gen/seq_module_unconditional/config.yaml
3
+ 944de5df74846043142ab9e16c135eb8a2ccfdc10b761bde513be423dffb056b ./gen/sp_module/sp_module_ckpt.tar.gz
4
+ 7c0c2b66f020a6bfc3e7fd01bead5b1e7cc679c77a5871fc8a5e5a092e1d12e9 ./gen/sp_module/config.yaml
5
+ d537cc4cc0238f66f9872c19304aed98cfb1d5c9c2f9c72bfd82b2f2a1e6474c ./gen/seq_module_conditional/seq_module_conditional_ckpt.tar.gz
6
+ a0b1f4ef43dbb617d06edca6124a9755def9f9163853f951d017a87a3986fc16 ./gen/seq_module_conditional/config.yaml
7
+ cf376f406334d3a86c7dc07dfc942df5af16b3a69adc781ce0322c4104851321 ./csp/sp_module_ckpt.tar.gz
8
+ a839101b25a878268b70a564593d098d723bdb43647674b79d5bbd6728b5ff24 ./csp/config.yaml
csp/config.yaml ADDED
@@ -0,0 +1,220 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ preprocess:
2
+ task: csp
3
+ num_cpus: 60
4
+ lmdb_dir: ${paths.data_dir}/lmdb
5
+ split_dir: ${paths.data_dir}/splits
6
+ metal_dir: ${paths.data_dir}/metals
7
+ seq_dir: ${paths.data_dir}/seqs
8
+ filter:
9
+ num_cpus: 16
10
+ max_bbs: 20
11
+ max_atoms: 200
12
+ max_cps: 20
13
+ prop_list:
14
+ - working_capacity_vacuum_swing [mmol/g]
15
+ - working_capacity_temperature_swing [mmol/g]
16
+ mof_matcher:
17
+ optimizer:
18
+ steps: 200
19
+ popsize: 20
20
+ maxiter: 20
21
+ tolerance:
22
+ ltol: 0.2
23
+ stol: 0.5
24
+ angle_tol: 5.0
25
+ mof_checker:
26
+ use_matched_coords: true
27
+ data:
28
+ train_sample_limit: null
29
+ val_sample_limit: null
30
+ dataset_prefix: MetalOxo_final
31
+ lmdb_dir: ${paths.data_dir}/lmdb/${experiment.task}
32
+ split_dir: ${paths.data_dir}/splits
33
+ rot_symmetry: true
34
+ corrupt_trans: ${experiment.training.corrupt_trans}
35
+ corrupt_rots: ${experiment.training.corrupt_rots}
36
+ corrupt_lattice: ${experiment.training.corrupt_lattice}
37
+ corrupt_torsions: ${experiment.training.corrupt_torsions}
38
+ t_sample_dist: uniform
39
+ t_uniform:
40
+ t_min: 0.0
41
+ t_max: 1.0
42
+ t_logit_normal:
43
+ mean: -0.5
44
+ std: 1.0
45
+ lattice:
46
+ lognormal:
47
+ loc: ${lattice.lognormal.loc}
48
+ scale: ${lattice.lognormal.scale}
49
+ uniform:
50
+ low: 60.0
51
+ high: 120.0
52
+ eps: 0.1
53
+ loader:
54
+ sampler: dynamic
55
+ overfit:
56
+ num_samples: 100000
57
+ dynamic:
58
+ max_num_atoms: 1500
59
+ max_batch_size: null
60
+ num_workers: 32
61
+ prefetch_factor: 10
62
+ batch_size:
63
+ train: 128
64
+ valid: 64
65
+ predict: 20
66
+ experiment:
67
+ task: csp
68
+ project: mofflow_2
69
+ name: 146M-csp
70
+ debug: false
71
+ seed: 123
72
+ num_devices: 8
73
+ warm_start: null
74
+ warm_start_cfg_override: true
75
+ visualize_freq: 1000
76
+ use_ema: false
77
+ ema:
78
+ decay: 0.999
79
+ apply_ema_every_n_steps: 1
80
+ start_step: 0
81
+ evaluate_ema_weights_instead: false
82
+ training:
83
+ corrupt_trans: true
84
+ corrupt_rots: true
85
+ corrupt_lattice: true
86
+ corrupt_torsions: true
87
+ translation_loss_weight: 3.0
88
+ rotation_loss_weight: 1.0
89
+ lattice_loss_weight: 1.0
90
+ torsion_angle_loss_weight: 1.0
91
+ torsion_norm_loss_weight: 0.005
92
+ t_normalize_clip: 0.9
93
+ optimizer:
94
+ lr: 1.0e-05
95
+ betas:
96
+ - 0.9
97
+ - 0.98
98
+ eps: 1.0e-08
99
+ weight_decay: 0.0
100
+ lr_scheduler_type: null
101
+ lr_scheduler:
102
+ linear_warmup:
103
+ warmup_epochs: 10
104
+ min_lr: 1.0e-07
105
+ reduce_on_plateau:
106
+ factor: 0.6
107
+ patience: 10
108
+ min_lr: 1.0e-07
109
+ trainer:
110
+ overfit_batches: 0
111
+ min_epochs: 10
112
+ max_epochs: 200
113
+ accelerator: gpu
114
+ log_every_n_steps: 1
115
+ deterministic: false
116
+ strategy: ddp
117
+ val_check_interval: 1.0
118
+ check_val_every_n_epoch: 1
119
+ accumulate_grad_batches: 1
120
+ gradient_clip_val: 10.0
121
+ checkpointer:
122
+ dirpath: ${paths.ckpt_dir}
123
+ filename: epoch_{epoch}-step_{step}-loss_{train/loss:.4f}
124
+ auto_insert_metric_name: false
125
+ save_last: true
126
+ save_top_k: 10
127
+ monitor: step
128
+ mode: max
129
+ model:
130
+ init_mode: 1in5out
131
+ corrupt_trans: ${experiment.training.corrupt_trans}
132
+ corrupt_rots: ${experiment.training.corrupt_rots}
133
+ corrupt_lattice: ${experiment.training.corrupt_lattice}
134
+ corrupt_torsions: ${experiment.training.corrupt_torsions}
135
+ initialize_cfg:
136
+ max_atom_types: 100
137
+ atom_type_dim: 64
138
+ atom_feats_dim: 39
139
+ time_embed_dim: 64
140
+ coords_feats_dim: 64
141
+ node_embed_dim: ${model.node_embed_dim}
142
+ add_pos_embed: true
143
+ node_embed_dim: 1024
144
+ interaction_cfg:
145
+ num_layers: 10
146
+ max_radius: 50
147
+ max_neighbors: 130
148
+ dist_embed_dim: 128
149
+ trans_encoder_cfg:
150
+ ln_mode: pre
151
+ init_mode: ${model.init_mode}
152
+ dropout: 0.0
153
+ ffn_bias: false
154
+ norm: rmsnorm
155
+ attention:
156
+ node_embed_dim: ${model.node_embed_dim}
157
+ dist_embed_dim: ${model.interaction_cfg.dist_embed_dim}
158
+ no_heads: 16
159
+ encoder:
160
+ ffn_dim: ${eval:'4 * ${model.node_embed_dim}'}
161
+ pool_cfg:
162
+ node_embed_dim: ${model.node_embed_dim}
163
+ max_radius: .inf
164
+ max_neighbors: 1000
165
+ gaussian:
166
+ max_radius: 20
167
+ dist_embed_dim: 128
168
+ no_heads: 16
169
+ output_model_cfg:
170
+ torsion:
171
+ normalize: false
172
+ node_embed_dim: ${model.node_embed_dim}
173
+ max_radius: 5
174
+ max_neighbors: 24
175
+ gaussian:
176
+ max_radius: 5
177
+ dist_embed_dim: 128
178
+ no_heads: 16
179
+ rotation:
180
+ node_embed_dim: ${model.node_embed_dim}
181
+ use_svd: false
182
+ paths:
183
+ root_dir: ${oc.env:PROJECT_ROOT}
184
+ log_dir: ${paths.root_dir}/logs/${experiment.project}/${experiment.name}
185
+ logger_dir: ${paths.root_dir}/logs
186
+ ckpt_dir: ${paths.log_dir}/ckpt
187
+ inference_dir: ${paths.log_dir}/inference
188
+ data_dir: ${paths.root_dir}/data
189
+ lattice:
190
+ lognormal:
191
+ loc:
192
+ - 2.5542190074920654
193
+ - 2.754688262939453
194
+ - 2.9632623195648193
195
+ scale:
196
+ - 0.37387514114379883
197
+ - 0.30108675360679626
198
+ - 0.31256887316703796
199
+ logger: wandb
200
+ wandb:
201
+ name: ${experiment.name}
202
+ project: ${experiment.project}
203
+ save_dir: ${paths.logger_dir}
204
+ wandb_watch:
205
+ log: all
206
+ log_freq: 500
207
+ tensorboard:
208
+ name: tensorboard
209
+ version: ${experiment.project}/${experiment.name}
210
+ save_dir: ${paths.logger_dir}
211
+ log_graph: false
212
+ default_hp_metric: true
213
+ callbacks:
214
+ tensorboard:
215
+ log_freq: 1
216
+ log_grad_hist: false
217
+ log_weight_hist: false
218
+ log_grad_norm: true
219
+ log_weight_norm: true
220
+ log_norm_input: true
csp/sp_module_ckpt.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf376f406334d3a86c7dc07dfc942df5af16b3a69adc781ce0322c4104851321
3
+ size 1582384815
gen/seq_module_conditional/config.yaml ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ data:
2
+ train_sample_limit: null
3
+ val_sample_limit: null
4
+ seq_dir: ${paths.data_dir}/seqs
5
+ dataset_prefix: mof_sequence
6
+ vocab_path: ${data.seq_dir}/vocab.json
7
+ max_len: null
8
+ loader:
9
+ num_workers: 16
10
+ prefetch_factor: 10
11
+ max_tokens: 8000
12
+ max_batch_size: null
13
+ model:
14
+ conditional: true
15
+ max_seq_len: 2048
16
+ attention:
17
+ dim: 1024
18
+ depth: 6
19
+ heads: 8
20
+ rotary_pos_emb: true
21
+ attn_flash: true
22
+ use_scalenorm: true
23
+ experiment:
24
+ project: mofflow_2
25
+ name: 63M-seq_cond
26
+ debug: false
27
+ seed: 123
28
+ num_devices: 4
29
+ warm_start: null
30
+ warm_start_cfg_override: true
31
+ sample_seq_freq: 500
32
+ use_ema: false
33
+ ema:
34
+ decay: 0.999
35
+ apply_ema_every_n_steps: 1
36
+ start_step: 0
37
+ evaluate_ema_weights_instead: true
38
+ optimizer:
39
+ lr: 0.0003
40
+ betas:
41
+ - 0.9
42
+ - 0.999
43
+ eps: 1.0e-08
44
+ weight_decay: 0.0
45
+ lr_scheduler_type: null
46
+ lr_scheduler:
47
+ linear_warmup:
48
+ warmup_epochs: 10
49
+ min_lr: 1.0e-07
50
+ reduce_on_plateau:
51
+ factor: 0.6
52
+ patience: 10
53
+ min_lr: 1.0e-06
54
+ trainer:
55
+ overfit_batches: 0
56
+ min_epochs: 10
57
+ max_epochs: 20
58
+ accelerator: gpu
59
+ log_every_n_steps: 1
60
+ deterministic: false
61
+ strategy: ddp
62
+ val_check_interval: 1.0
63
+ check_val_every_n_epoch: 1
64
+ accumulate_grad_batches: 1
65
+ gradient_clip_val: 1.0
66
+ checkpointer:
67
+ dirpath: ${paths.ckpt_dir}
68
+ monitor: step
69
+ save_last: true
70
+ save_top_k: -1
71
+ filename: epoch_{epoch}-step_{step}-loss_{valid/loss:.4f}
72
+ auto_insert_metric_name: false
73
+ mode: max
74
+ logger: wandb
75
+ wandb:
76
+ name: ${experiment.name}
77
+ project: ${experiment.project}
78
+ save_dir: ${paths.logger_dir}
79
+ wandb_watch:
80
+ log: all
81
+ log_freq: 500
82
+ tensorboard:
83
+ name: tensorboard
84
+ version: ${experiment.project}/${experiment.name}
85
+ save_dir: ${paths.logger_dir}
86
+ log_graph: false
87
+ default_hp_metric: true
88
+ callbacks:
89
+ tensorboard:
90
+ log_freq: ${experiment.trainer.log_every_n_steps}
91
+ log_grad_hist: false
92
+ log_weight_hist: false
93
+ log_grad_norm: true
94
+ log_weight_norm: true
95
+ log_norm_input: true
96
+ paths:
97
+ root_dir: ${oc.env:PROJECT_ROOT}
98
+ log_dir: ${paths.root_dir}/logs/${experiment.project}/${experiment.name}
99
+ logger_dir: ${paths.root_dir}/logs
100
+ ckpt_dir: ${paths.log_dir}/ckpt
101
+ inference_dir: ${paths.log_dir}/inference
102
+ data_dir: ${paths.root_dir}/data
gen/seq_module_conditional/seq_module_conditional_ckpt.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d537cc4cc0238f66f9872c19304aed98cfb1d5c9c2f9c72bfd82b2f2a1e6474c
3
+ size 829792726
gen/seq_module_unconditional/config.yaml ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ data:
2
+ train_sample_limit: null
3
+ val_sample_limit: null
4
+ seq_dir: ${paths.data_dir}/seqs
5
+ dataset_prefix: mof_sequence
6
+ vocab_path: ${data.seq_dir}/vocab.json
7
+ max_len: null
8
+ loader:
9
+ num_workers: 16
10
+ prefetch_factor: 10
11
+ max_tokens: 8000
12
+ max_batch_size: null
13
+ model:
14
+ conditional: false
15
+ max_seq_len: 2048
16
+ attention:
17
+ dim: 1024
18
+ depth: 6
19
+ heads: 8
20
+ rotary_pos_emb: true
21
+ attn_flash: true
22
+ use_scalenorm: true
23
+ experiment:
24
+ project: mofflow_2
25
+ name: 63M-seq
26
+ debug: false
27
+ seed: 123
28
+ num_devices: 8
29
+ warm_start: null
30
+ warm_start_cfg_override: true
31
+ sample_seq_freq: 500
32
+ use_ema: false
33
+ ema:
34
+ decay: 0.999
35
+ apply_ema_every_n_steps: 1
36
+ start_step: 0
37
+ evaluate_ema_weights_instead: true
38
+ optimizer:
39
+ lr: 0.0003
40
+ betas:
41
+ - 0.9
42
+ - 0.999
43
+ eps: 1.0e-08
44
+ weight_decay: 0.0
45
+ lr_scheduler_type: null
46
+ lr_scheduler:
47
+ linear_warmup:
48
+ warmup_epochs: 10
49
+ min_lr: 1.0e-07
50
+ reduce_on_plateau:
51
+ factor: 0.6
52
+ patience: 10
53
+ min_lr: 1.0e-06
54
+ trainer:
55
+ overfit_batches: 0
56
+ min_epochs: 10
57
+ max_epochs: 20
58
+ accelerator: gpu
59
+ log_every_n_steps: 1
60
+ deterministic: false
61
+ strategy: ddp
62
+ val_check_interval: 1.0
63
+ check_val_every_n_epoch: 1
64
+ accumulate_grad_batches: 1
65
+ gradient_clip_val: 1.0
66
+ checkpointer:
67
+ dirpath: ${paths.ckpt_dir}
68
+ monitor: step
69
+ save_last: true
70
+ save_top_k: 20
71
+ filename: epoch_{epoch}-step_{step}-loss_{train/loss:.4f}
72
+ auto_insert_metric_name: false
73
+ mode: max
74
+ logger: wandb
75
+ wandb:
76
+ name: ${experiment.name}
77
+ project: ${experiment.project}
78
+ save_dir: ${paths.logger_dir}
79
+ wandb_watch:
80
+ log: all
81
+ log_freq: 500
82
+ tensorboard:
83
+ name: tensorboard
84
+ version: ${experiment.project}/${experiment.name}
85
+ save_dir: ${paths.logger_dir}
86
+ log_graph: false
87
+ default_hp_metric: true
88
+ callbacks:
89
+ tensorboard:
90
+ log_freq: ${experiment.trainer.log_every_n_steps}
91
+ log_grad_hist: false
92
+ log_weight_hist: false
93
+ log_grad_norm: true
94
+ log_weight_norm: true
95
+ log_norm_input: true
96
+ paths:
97
+ root_dir: ${oc.env:PROJECT_ROOT}
98
+ log_dir: ${paths.root_dir}/logs/${experiment.project}/${experiment.name}
99
+ logger_dir: ${paths.root_dir}/logs
100
+ ckpt_dir: ${paths.log_dir}/ckpt
101
+ inference_dir: ${paths.log_dir}/inference
102
+ data_dir: ${paths.root_dir}/data
gen/seq_module_unconditional/seq_module_unconditional_ckpt.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f9345c143cd7576472a0b88815504e4540447e5fcf2c4e927f62f431beeb10c
3
+ size 690673315
gen/sp_module/config.yaml ADDED
@@ -0,0 +1,220 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ preprocess:
2
+ task: csp
3
+ num_cpus: 60
4
+ lmdb_dir: ${paths.data_dir}/lmdb
5
+ split_dir: ${paths.data_dir}/splits
6
+ metal_dir: ${paths.data_dir}/metals
7
+ seq_dir: ${paths.data_dir}/seqs
8
+ filter:
9
+ num_cpus: 16
10
+ max_bbs: 20
11
+ max_atoms: 200
12
+ max_cps: 20
13
+ prop_list:
14
+ - working_capacity_vacuum_swing [mmol/g]
15
+ - working_capacity_temperature_swing [mmol/g]
16
+ mof_matcher:
17
+ optimizer:
18
+ steps: 200
19
+ popsize: 20
20
+ maxiter: 20
21
+ tolerance:
22
+ ltol: 0.2
23
+ stol: 0.5
24
+ angle_tol: 5.0
25
+ mof_checker:
26
+ use_matched_coords: true
27
+ data:
28
+ train_sample_limit: null
29
+ val_sample_limit: null
30
+ dataset_prefix: MetalOxo_final
31
+ lmdb_dir: ${paths.data_dir}/lmdb/${experiment.task}
32
+ split_dir: ${paths.data_dir}/splits
33
+ rot_symmetry: true
34
+ corrupt_trans: ${experiment.training.corrupt_trans}
35
+ corrupt_rots: ${experiment.training.corrupt_rots}
36
+ corrupt_lattice: ${experiment.training.corrupt_lattice}
37
+ corrupt_torsions: ${experiment.training.corrupt_torsions}
38
+ t_sample_dist: uniform
39
+ t_uniform:
40
+ t_min: 0.0
41
+ t_max: 1.0
42
+ t_logit_normal:
43
+ mean: -0.5
44
+ std: 1.0
45
+ lattice:
46
+ lognormal:
47
+ loc: ${lattice.lognormal.loc}
48
+ scale: ${lattice.lognormal.scale}
49
+ uniform:
50
+ low: 60.0
51
+ high: 120.0
52
+ eps: 0.1
53
+ loader:
54
+ sampler: dynamic
55
+ overfit:
56
+ num_samples: 100000
57
+ dynamic:
58
+ max_num_atoms: 1500
59
+ max_batch_size: null
60
+ num_workers: 32
61
+ prefetch_factor: 10
62
+ batch_size:
63
+ train: 128
64
+ valid: 64
65
+ predict: 20
66
+ experiment:
67
+ task: gen
68
+ project: mofflow_2
69
+ name: 146M-gen
70
+ debug: false
71
+ seed: 123
72
+ num_devices: 8
73
+ warm_start: null
74
+ warm_start_cfg_override: true
75
+ visualize_freq: 1000
76
+ use_ema: false
77
+ ema:
78
+ decay: 0.999
79
+ apply_ema_every_n_steps: 1
80
+ start_step: 0
81
+ evaluate_ema_weights_instead: false
82
+ training:
83
+ corrupt_trans: true
84
+ corrupt_rots: true
85
+ corrupt_lattice: true
86
+ corrupt_torsions: true
87
+ translation_loss_weight: 3.0
88
+ rotation_loss_weight: 1.0
89
+ lattice_loss_weight: 1.0
90
+ torsion_angle_loss_weight: 1.0
91
+ torsion_norm_loss_weight: 0.005
92
+ t_normalize_clip: 0.9
93
+ optimizer:
94
+ lr: 1.0e-05
95
+ betas:
96
+ - 0.9
97
+ - 0.98
98
+ eps: 1.0e-08
99
+ weight_decay: 0.0
100
+ lr_scheduler_type: null
101
+ lr_scheduler:
102
+ linear_warmup:
103
+ warmup_epochs: 10
104
+ min_lr: 1.0e-07
105
+ reduce_on_plateau:
106
+ factor: 0.6
107
+ patience: 10
108
+ min_lr: 1.0e-07
109
+ trainer:
110
+ overfit_batches: 0
111
+ min_epochs: 10
112
+ max_epochs: 200
113
+ accelerator: gpu
114
+ log_every_n_steps: 1
115
+ deterministic: false
116
+ strategy: ddp
117
+ val_check_interval: 1.0
118
+ check_val_every_n_epoch: 1
119
+ accumulate_grad_batches: 1
120
+ gradient_clip_val: 10.0
121
+ checkpointer:
122
+ dirpath: ${paths.ckpt_dir}
123
+ filename: epoch_{epoch}-step_{step}-loss_{train/loss:.4f}
124
+ auto_insert_metric_name: false
125
+ save_last: true
126
+ save_top_k: 10
127
+ monitor: step
128
+ mode: max
129
+ model:
130
+ init_mode: 1in5out
131
+ corrupt_trans: ${experiment.training.corrupt_trans}
132
+ corrupt_rots: ${experiment.training.corrupt_rots}
133
+ corrupt_lattice: ${experiment.training.corrupt_lattice}
134
+ corrupt_torsions: ${experiment.training.corrupt_torsions}
135
+ initialize_cfg:
136
+ max_atom_types: 100
137
+ atom_type_dim: 64
138
+ atom_feats_dim: 39
139
+ time_embed_dim: 64
140
+ coords_feats_dim: 64
141
+ node_embed_dim: ${model.node_embed_dim}
142
+ add_pos_embed: true
143
+ node_embed_dim: 1024
144
+ interaction_cfg:
145
+ num_layers: 10
146
+ max_radius: 50
147
+ max_neighbors: 130
148
+ dist_embed_dim: 128
149
+ trans_encoder_cfg:
150
+ ln_mode: pre
151
+ init_mode: ${model.init_mode}
152
+ dropout: 0.0
153
+ ffn_bias: false
154
+ norm: rmsnorm
155
+ attention:
156
+ node_embed_dim: ${model.node_embed_dim}
157
+ dist_embed_dim: ${model.interaction_cfg.dist_embed_dim}
158
+ no_heads: 16
159
+ encoder:
160
+ ffn_dim: ${eval:'4 * ${model.node_embed_dim}'}
161
+ pool_cfg:
162
+ node_embed_dim: ${model.node_embed_dim}
163
+ max_radius: .inf
164
+ max_neighbors: 1000
165
+ gaussian:
166
+ max_radius: 20
167
+ dist_embed_dim: 128
168
+ no_heads: 16
169
+ output_model_cfg:
170
+ torsion:
171
+ normalize: false
172
+ node_embed_dim: ${model.node_embed_dim}
173
+ max_radius: 5
174
+ max_neighbors: 24
175
+ gaussian:
176
+ max_radius: 5
177
+ dist_embed_dim: 128
178
+ no_heads: 16
179
+ rotation:
180
+ node_embed_dim: ${model.node_embed_dim}
181
+ use_svd: false
182
+ paths:
183
+ root_dir: ${oc.env:PROJECT_ROOT}
184
+ log_dir: ${paths.root_dir}/logs/${experiment.project}/${experiment.name}
185
+ logger_dir: ${paths.root_dir}/logs
186
+ ckpt_dir: ${paths.log_dir}/ckpt
187
+ inference_dir: ${paths.log_dir}/inference
188
+ data_dir: ${paths.root_dir}/data
189
+ lattice:
190
+ lognormal:
191
+ loc:
192
+ - 2.5542190074920654
193
+ - 2.754688262939453
194
+ - 2.9632623195648193
195
+ scale:
196
+ - 0.37387514114379883
197
+ - 0.30108675360679626
198
+ - 0.31256887316703796
199
+ logger: wandb
200
+ wandb:
201
+ name: ${experiment.name}
202
+ project: ${experiment.project}
203
+ save_dir: ${paths.logger_dir}
204
+ wandb_watch:
205
+ log: all
206
+ log_freq: 500
207
+ tensorboard:
208
+ name: tensorboard
209
+ version: ${experiment.project}/${experiment.name}
210
+ save_dir: ${paths.logger_dir}
211
+ log_graph: false
212
+ default_hp_metric: true
213
+ callbacks:
214
+ tensorboard:
215
+ log_freq: 1
216
+ log_grad_hist: false
217
+ log_weight_hist: false
218
+ log_grad_norm: true
219
+ log_weight_norm: true
220
+ log_norm_input: true
gen/sp_module/sp_module_ckpt.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:944de5df74846043142ab9e16c135eb8a2ccfdc10b761bde513be423dffb056b
3
+ size 1582456820