PIEthonista commited on
Commit
371fe69
Β·
1 Parent(s): dff2967
Files changed (27) hide show
  1. {03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume}/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A.yaml +8 -4
  2. 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/args_135_iter_224808.pickle β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/args_105_iter_175218.pickle +2 -2
  3. 03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A/generative_model_135_iter_224808.npy β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/generative_model_105_iter_175218.npy +1 -1
  4. {03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume}/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A.yaml +6 -2
  5. 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume/args_75_iter_125628.pickle β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/args_80_iter_133893.pickle +2 -2
  6. 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume/generative_model_75_iter_125628.npy β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/generative_model_80_iter_133893.npy +1 -1
  7. {03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x_resume}/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A.yaml +10 -2
  8. 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/args_75_iter_125628.pickle β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x_resume/args_80_iter_133893.pickle +2 -2
  9. 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/generative_model_75_iter_125628.npy β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x_resume/generative_model_80_iter_133893.npy +1 -1
  10. {03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume}/03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A.yaml +6 -2
  11. 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/args_75_iter_125628.pickle β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/args_80_iter_133893.pickle +2 -2
  12. 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/generative_model_75_iter_125628.npy β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/generative_model_80_iter_133893.npy +1 -1
  13. 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/args_75_iter_125628.pickle +0 -3
  14. {03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume}/03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A.yaml +6 -2
  15. 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/args_80_iter_133893.pickle +3 -0
  16. 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/generative_model_75_iter_125628.npy β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/generative_model_80_iter_133893.npy +1 -1
  17. 03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x_resume/03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A.yaml +393 -0
  18. 03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x_resume/args_4_iter_47490.pickle +3 -0
  19. 03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x_resume/generative_model_4_iter_47490.npy +3 -0
  20. 03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_2x_resume/args_5_iter_56988.pickle +0 -3
  21. {03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_2x_resume β†’ 03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x_resume}/03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A.yaml +12 -4
  22. 03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x_resume/args_7_iter_75984.pickle +3 -0
  23. 03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_2x_resume/generative_model_5_iter_56988.npy β†’ 03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x_resume/generative_model_7_iter_75984.npy +1 -1
  24. 03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A/args_135_iter_224808.pickle +0 -3
  25. {03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A β†’ 03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume}/03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A.yaml +12 -6
  26. 03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume/args_105_iter_175218.pickle +3 -0
  27. 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/generative_model_135_iter_224808.npy β†’ 03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume/generative_model_105_iter_175218.npy +1 -1
{03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume}/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A.yaml RENAMED
@@ -2,7 +2,8 @@ proj_name: Control-GeoLDM
2
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
3
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x
4
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x
5
- exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x
 
6
 
7
 
8
  # ========================================================================================================== Training Mode (ldm/vae/both)
@@ -296,9 +297,12 @@ n_report_steps: 50
296
  # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume
297
  # resume_model_ckpt: generative_model_84_iter_140505.npy
298
  # resume_optim_ckpt: optim_84_iter_140505.npy
299
- resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x_resume
300
- resume_model_ckpt: generative_model_76_iter_127281.npy
301
- resume_optim_ckpt: optim_76_iter_127281.npy
 
 
 
302
 
303
  save_model: true
304
 
 
2
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
3
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x
4
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x
5
+ # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x
6
+ exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x
7
 
8
 
9
  # ========================================================================================================== Training Mode (ldm/vae/both)
 
297
  # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume
298
  # resume_model_ckpt: generative_model_84_iter_140505.npy
299
  # resume_optim_ckpt: optim_84_iter_140505.npy
300
+ # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x_resume
301
+ # resume_model_ckpt: generative_model_76_iter_127281.npy
302
+ # resume_optim_ckpt: optim_76_iter_127281.npy
303
+ resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume
304
+ resume_model_ckpt: generative_model_135_iter_224808.npy
305
+ resume_optim_ckpt: optim_135_iter_224808.npy
306
 
307
  save_model: true
308
 
03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/args_135_iter_224808.pickle β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/args_105_iter_175218.pickle RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:81c69beda6fc5c42a0523fcedb22afeed8eb170c216b15215bdd705627b6e317
3
- size 5710
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccb58d6c3a8d1fd23393dd6b13c2fa6ffc88c781cf579fe05b2951ce40e72d22
3
+ size 6104
03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A/generative_model_135_iter_224808.npy β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/generative_model_105_iter_175218.npy RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:74803916b1fd755f3b6fa346885792b031824d4dc6ed5fded39d04580f358397
3
  size 53576312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49fab49636a334f20d5f3b2024f8f9c72944a37c84a4cfe0db08dc643a402668
3
  size 53576312
{03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume}/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A.yaml RENAMED
@@ -2,7 +2,8 @@ proj_name: Control-GeoLDM
2
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
3
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x
4
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x
5
- exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x
 
6
 
7
 
8
  # ========================================================================================================== Training Mode (ldm/vae/both)
@@ -296,7 +297,10 @@ n_report_steps: 50
296
  # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume
297
  # resume_model_ckpt: generative_model_84_iter_140505.npy
298
  # resume_optim_ckpt: optim_84_iter_140505.npy
299
- resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x_resume
 
 
 
300
  resume_model_ckpt: generative_model_75_iter_125628.npy
301
  resume_optim_ckpt: optim_75_iter_125628.npy
302
 
 
2
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
3
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x
4
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x
5
+ # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x
6
+ exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x
7
 
8
 
9
  # ========================================================================================================== Training Mode (ldm/vae/both)
 
297
  # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume
298
  # resume_model_ckpt: generative_model_84_iter_140505.npy
299
  # resume_optim_ckpt: optim_84_iter_140505.npy
300
+ # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x_resume
301
+ # resume_model_ckpt: generative_model_75_iter_125628.npy
302
+ # resume_optim_ckpt: optim_75_iter_125628.npy
303
+ resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume
304
  resume_model_ckpt: generative_model_75_iter_125628.npy
305
  resume_optim_ckpt: optim_75_iter_125628.npy
306
 
03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume/args_75_iter_125628.pickle β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/args_80_iter_133893.pickle RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87d55dd84aac61dbcb37450676b8999cab749933af7ee14565ccc0349aeae494
3
- size 5700
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:beb58b277f6d8b49845e43ef7907fd9b0040451eca253f616b9a81ace3ee8cdb
3
+ size 6102
03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume/generative_model_75_iter_125628.npy β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/generative_model_80_iter_133893.npy RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c234296d787f7121cf40b9f958a3e4714a899c122a37de7d3a61fbb9555f8653
3
  size 53575942
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa3fa622427b66d2418340bfda4d3d936e58d337c1484931ade587f2a25dab57
3
  size 53575942
{03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x_resume}/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A.yaml RENAMED
@@ -1,6 +1,8 @@
1
  proj_name: Control-GeoLDM
2
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
3
- exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x
 
 
4
 
5
 
6
  # ========================================================================================================== Training Mode (ldm/vae/both)
@@ -288,7 +290,13 @@ n_report_steps: 50
288
 
289
  # ========================================================================================================== Saving & Resuming
290
  # resume: null
291
- resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
 
 
 
 
 
 
292
  resume_model_ckpt: generative_model_80_iter_133893.npy
293
  resume_optim_ckpt: optim_80_iter_133893.npy
294
 
 
1
  proj_name: Control-GeoLDM
2
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
3
+ # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x
4
+ # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x
5
+ exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x
6
 
7
 
8
  # ========================================================================================================== Training Mode (ldm/vae/both)
 
290
 
291
  # ========================================================================================================== Saving & Resuming
292
  # resume: null
293
+ # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
294
+ # resume_model_ckpt: generative_model_80_iter_133893.npy
295
+ # resume_optim_ckpt: optim_80_iter_133893.npy
296
+ # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume
297
+ # resume_model_ckpt: generative_model_75_iter_125628.npy
298
+ # resume_optim_ckpt: optim_75_iter_125628.npy
299
+ resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x_resume
300
  resume_model_ckpt: generative_model_80_iter_133893.npy
301
  resume_optim_ckpt: optim_80_iter_133893.npy
302
 
03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/args_75_iter_125628.pickle β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x_resume/args_80_iter_133893.pickle RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d8916a0dbab4f3c448ffd3c169d7f4b6cb5fc1164a4c853d4b061027f4bed00c
3
- size 5710
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:070f13b65383f59392c07e9e2f8507dfd5a0f81943b0a10e56afd0769938a1be
3
+ size 6102
03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/generative_model_75_iter_125628.npy β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.9__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x_resume/generative_model_80_iter_133893.npy RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:710e0ed2b42e05e78713e9918ed196a14e1ef86b6aa5f7be7e47268c779eb7af
3
  size 53575942
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:817da2f283474984da0dc6be696e6a7c004b1429a3905f8789602cf757b7af98
3
  size 53575942
{03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume}/03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A.yaml RENAMED
@@ -2,7 +2,8 @@ proj_name: Control-GeoLDM
2
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
3
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x
4
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x
5
- exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x
 
6
 
7
 
8
  # ========================================================================================================== Training Mode (ldm/vae/both)
@@ -296,7 +297,10 @@ n_report_steps: 50
296
  # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume
297
  # resume_model_ckpt: generative_model_82_iter_137199.npy
298
  # resume_optim_ckpt: optim_82_iter_137199.npy
299
- resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x_resume
 
 
 
300
  resume_model_ckpt: generative_model_75_iter_125628.npy
301
  resume_optim_ckpt: optim_75_iter_125628.npy
302
 
 
2
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
3
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x
4
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x
5
+ # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x
6
+ exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x
7
 
8
 
9
  # ========================================================================================================== Training Mode (ldm/vae/both)
 
297
  # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume
298
  # resume_model_ckpt: generative_model_82_iter_137199.npy
299
  # resume_optim_ckpt: optim_82_iter_137199.npy
300
+ # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x_resume
301
+ # resume_model_ckpt: generative_model_75_iter_125628.npy
302
+ # resume_optim_ckpt: optim_75_iter_125628.npy
303
+ resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume
304
  resume_model_ckpt: generative_model_75_iter_125628.npy
305
  resume_optim_ckpt: optim_75_iter_125628.npy
306
 
03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/args_75_iter_125628.pickle β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/args_80_iter_133893.pickle RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a37e79ffff388edaaf86eed139cf88521b7087099e25135cda14c0a17747d81
3
- size 5706
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:993b81d0cba58de4c4ed5bb46b3a56fb667b35b5c7a8001de6302a93e23624a3
3
+ size 6098
03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/generative_model_75_iter_125628.npy β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/generative_model_80_iter_133893.npy RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b78e4f16f584b7bb3d619c2ab445bfa0fd55162787588c4214324553e605905b
3
  size 53575942
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07f2f056909e379f4ac6a5fd3eeb4adb6c2f36edde69ad087f1087cf25a574fb
3
  size 53575942
03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/args_75_iter_125628.pickle DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1a9ea77f2e066fb9ef3f6f492394c18fd1687822cc66172907da6e2f67d15eec
3
- size 5706
 
 
 
 
{03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume}/03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A.yaml RENAMED
@@ -2,7 +2,8 @@ proj_name: Control-GeoLDM
2
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
3
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x
4
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x
5
- exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x
 
6
 
7
 
8
  # ========================================================================================================== Training Mode (ldm/vae/both)
@@ -296,7 +297,10 @@ n_report_steps: 50
296
  # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume
297
  # resume_model_ckpt: generative_model_89_iter_148770.npy
298
  # resume_optim_ckpt: optim_89_iter_148770.npy
299
- resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x_resume
 
 
 
300
  resume_model_ckpt: generative_model_75_iter_125628.npy
301
  resume_optim_ckpt: optim_75_iter_125628.npy
302
 
 
2
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
3
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x
4
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x
5
+ # exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x
6
+ exp_name: 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x
7
 
8
 
9
  # ========================================================================================================== Training Mode (ldm/vae/both)
 
297
  # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume
298
  # resume_model_ckpt: generative_model_89_iter_148770.npy
299
  # resume_optim_ckpt: optim_89_iter_148770.npy
300
+ # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x_resume
301
+ # resume_model_ckpt: generative_model_75_iter_125628.npy
302
+ # resume_optim_ckpt: optim_75_iter_125628.npy
303
+ resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume
304
  resume_model_ckpt: generative_model_75_iter_125628.npy
305
  resume_optim_ckpt: optim_75_iter_125628.npy
306
 
03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/args_80_iter_133893.pickle ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa55000301c933844199dc9af8431d90b3dd10844e541953a1768d2f7d93e38a
3
+ size 6098
03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/generative_model_75_iter_125628.npy β†’ 03_latent2_nf256_ds1k_fusBSum_CA_conditionBlocks34_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_4x_resume/generative_model_80_iter_133893.npy RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ce0e20a07221263b534873efc0116dc2b946a992668f37f25cdadb261aca2040
3
  size 53575942
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a028c94486b0e1e2f21f1ee8ca84a15ecaee523cea30e65b2506dd766d9bef52
3
  size 53575942
03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x_resume/03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A.yaml ADDED
@@ -0,0 +1,393 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ proj_name: Control-GeoLDM
2
+ # exp_name: 03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A
3
+ # exp_name: 03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_1x
4
+ # exp_name: 03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_2x
5
+ # exp_name: 03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x
6
+ exp_name: 03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_4x
7
+
8
+
9
+ # ========================================================================================================== Training Mode (ldm/vae/both)
10
+ # Train second stage LatentDiffusionModel model
11
+ train_diffusion: true
12
+
13
+ # training mode: VAE | LDM | ControlNet
14
+ training_mode: ControlNet
15
+ loss_analysis: false
16
+
17
+ # Specify ligand & pocket VAE weights path, set to null for random initialisation
18
+ # set checkpoint (ckpt) to null to automatically select best
19
+ ligand_ae_path: null # use GEOM LDM's VAE
20
+ ligand_ae_ckpt: null # use GEOM LDM's VAE
21
+ pocket_ae_path: outputs_selected/vae_pockets/AMP__01_VAE_vaenorm_True10__float32__latent2_nf256_epoch100_bs4_lr1e-5_InvClassFreq_Smooth0.25_XH_x30_h15_NoEMA__20240623__10A__PKT_Only_2x_resume
22
+ pocket_ae_ckpt: generative_model_3_iter_76064.npy
23
+
24
+ # Specify LDM weights path, set to null for random initialisation
25
+ ldm_path: outputs_selected/ldm/ori_drugs_latent2
26
+ ldm_ckpt: generative_model_ema.npy
27
+
28
+ # Zero out all weights of fusion blocks instead of randomly instantiated
29
+ zero_fusion_block_weights: false
30
+
31
+
32
+ # Train 1st stage AutoEncoder model (no effect if train_diffusion=False)
33
+ trainable_ligand_ae_encoder: false
34
+ trainable_ligand_ae_decoder: false
35
+ trainable_pocket_ae_encoder: false
36
+
37
+ # Train 2nd stage LDM model
38
+ trainable_ldm: false
39
+
40
+ # Train 3rd stage ControlNet
41
+ trainable_controlnet: true
42
+ trainable_fusion_blocks: true
43
+
44
+
45
+ # can contain multiple: homo | onehot | lumo | num_atoms | etc
46
+ conditioning: []
47
+
48
+ # include atom charge, according to periodic table
49
+ include_charges: false # true for qm9
50
+
51
+ # only works for ldm, not for VAE
52
+ condition_time: true
53
+
54
+ # Time Noisy, t/2, adopted from [https://arxiv.org/abs/2405.06659]
55
+ time_noisy: false
56
+
57
+ vis_activations: false
58
+ vis_activations_batch_samples: 5
59
+ vis_activations_batch_size: 1
60
+ vis_activations_specific_ylim: [0, 40]
61
+
62
+ random_seed: 42
63
+
64
+
65
+ # ========================================================================================================== Dataset
66
+
67
+ # pre-computed dataset stats
68
+ # dataset: d_20240623_CrossDocked_LG_PKT__10A__LIGAND
69
+ dataset: d_20241115_GEOM_LDM_CrossDocked_LG_PKT_MMseq2_split__10A__LIGAND
70
+
71
+ # pre-computed training dataset
72
+ data_file: ./data/d_20241115_CrossDocked_LG_PKT_MMseq2_split/d_20241115_CrossDocked_LG_PKT_MMseq2_split__10.0A.npz
73
+ # data_file: ./data/d_20240623_CrossDocked_LG_PKT/d_20240623_CrossDocked_LG_PKT__10.0A_split_811.npz
74
+ # data_file: ./data/d_20240623_CrossDocked_LG_PKT/d_20240623_CrossDocked_LG_PKT__10.0A__subset_0.01_split_811.npz
75
+ data_splitted: true
76
+
77
+ # Quick Vina 2.1
78
+ compute_qvina: true
79
+ qvina_search_size: 20 # search size (all 3 axes) in Angstroms around ligand center
80
+ qvina_exhaustiveness: 16
81
+ qvina_seed: 42
82
+ qvina_cleanup_files: false # cleanup tmp pdb, pdbqt files
83
+ qvina_save_csv: true # save results in csv
84
+ pocket_pdb_dir: ./data/d_20241115_CrossDocked_LG_PKT_MMseq2_split/test_val_paired_files/val_pocket
85
+ match_raw_file_by_id: true
86
+ mgltools_env_name: mgltools-python2 # for pdb -> pdbqt conversion
87
+
88
+ ligand_add_H: false # add hydrogens via: [mgltools] prepare_ligand4.py -l .. -o .. -A hydrogens
89
+ pocket_add_H: false # add hydrogens via: [mgltools] prepare_receptor4.py -r .. -o .. -A checkhydrogens
90
+ pocket_remove_nonstd_resi: false # remove any pocket residues not in this list:
91
+ # ['CYS','ILE','SER','VAL','GLN','LYS','ASN',
92
+ # 'PRO','THR','PHE','ALA','HIS','GLY','ASP',
93
+ # 'LEU', 'ARG', 'TRP', 'GLU', 'TYR','MET',
94
+ # 'HID', 'HSP', 'HIE', 'HIP', 'CYX', 'CSS']
95
+
96
+
97
+ # set to null if you're running this dataset for the first time.
98
+ # Script will generate a random permutation to shuffle the dataset.
99
+ # Please set the path to the DATASET_permutation.npy file after it is generated.
100
+ # permutation_file_path: ./data/d_20240623_CrossDocked_LG_PKT/d_20240623_CrossDocked_LG_PKT__10.0A_LG100_PKT600_permutation.npy
101
+ permutation_file_path: null
102
+
103
+ # what data to load for VAE training: ligand | pocket | all
104
+ vae_data_mode: ligand
105
+
106
+ # When set to an integer value, QM9 will only contain molecules of that amount of atoms, default null
107
+ filter_n_atoms: null
108
+
109
+ # Only use molecules below this size. Int, default null ~!geom
110
+ filter_molecule_size: 100
111
+ filter_pocket_size: 600 # refer EDA
112
+
113
+ # Organize data by size to reduce average memory usage. ~!geom
114
+ sequential: false
115
+
116
+ # Number of worker for the dataloader
117
+ num_workers: 60 # match cpu count
118
+
119
+ # use data augmentation (i.e. random rotation of x atom coordinates)
120
+ data_augmentation: false
121
+
122
+ # remove hydrogen atoms
123
+ remove_h: false
124
+
125
+
126
+
127
+
128
+ # ========================================================================================================== Training Params
129
+ start_epoch: 0
130
+ test_epochs: 2 # 4
131
+
132
+
133
+ n_epochs: 1000 # 3000 takes 20 epoches on paper (bs:32), hence 80 epochs for bs:8
134
+ batch_size: 10 # 14
135
+ lr: 1.0e-4
136
+
137
+ # weight of KL term in ELBO, default 0.01
138
+ kl_weight: 0.01
139
+
140
+ # ode_regularization weightage, default 1e-3
141
+ ode_regularization: 0.001
142
+ # brute_force: false
143
+ # actnorm: true
144
+ break_train_epoch: false
145
+
146
+ # Data Parallel for multi GPU support
147
+ dp: true
148
+ clip_grad: true
149
+
150
+ # Amount of EMA decay, 0 means off. A reasonable value is 0.999.
151
+ ema_decay: 0 # 0.99
152
+
153
+ # add noise to x before encoding, data augmenting
154
+ augment_noise: 0
155
+
156
+ # Number of samples to compute the stability, default 500
157
+ n_stability_samples: 90 # 98, 50
158
+ n_stability_samples_batch_size: 10 # 7, 14
159
+
160
+ # Dataset partition where pocket samples will be drawn from for analyzing
161
+ # generated ligands' stability: train | test | val
162
+ n_stability_eval_split: val
163
+
164
+
165
+ # disables CUDA training
166
+ no_cuda: false
167
+
168
+ # hutch | exact
169
+ trace: hutch
170
+
171
+ # verbose logging
172
+ verbose: false
173
+
174
+ dtype: torch.float32
175
+
176
+ # enable mixed precision training (fp32, fp16)
177
+ mixed_precision_training: true
178
+ mixed_precision_autocast_dtype: torch.bfloat16
179
+
180
+ # use model checkpointing during training to reduce GPU memory usage
181
+ use_checkpointing: true
182
+
183
+ # sqrt: checkpointing is done on the sqrt(block_num)'th Equivariant block of each EGNN for most optimal perf
184
+ # all: checkpointing is done on all Equivariant blocks. Not optimal but helps if input size is too large
185
+ checkpointing_mode: sqrt
186
+
187
+ # splits tensors into managable chunks and performs forward propagation without breaking GPU memory limit
188
+ forward_tensor_chunk_size: 50000
189
+
190
+
191
+
192
+
193
+
194
+
195
+
196
+
197
+ # ========================================================================================================== LDM
198
+ # our_dynamics | schnet | simple_dynamics | kernel_dynamics | egnn_dynamics | gnn_dynamics
199
+ model: egnn_dynamics
200
+
201
+ probabilistic_model: diffusion
202
+
203
+ # Training complexity is O(1) (unaffected), but sampling complexity is O(steps), default 500
204
+ diffusion_steps: 1000
205
+
206
+ # learned, cosine, polynomial_<power>
207
+ diffusion_noise_schedule: polynomial_2
208
+
209
+ # default 1e-5
210
+ diffusion_noise_precision: 1.0e-05 # ~!fp16
211
+
212
+ # vlb | l2
213
+ diffusion_loss_type: l2
214
+
215
+ # number of latent features, default 4
216
+ latent_nf: 2
217
+
218
+ # VAE configs below will be overriden by the original VAE's configs (.pkl file)
219
+ # normalize factors for [x, h_cat/categorical/one-hot, h_int/integer/charges]
220
+ normalize_factors: [1, 4, 10]
221
+
222
+ # vae_normalize_x: true
223
+ # vae_normalize_method: scale # scale | linear
224
+ # vae_normalize_factors: [10, 1, 1]
225
+
226
+ # reweight_class_loss: null # "inv_class_freq"
227
+ # smoothing_factor: 1.0 # [0.1 - 1.0) 1.0 is essentially disabling
228
+
229
+ # error_x_weight: null # error_x custom weighting
230
+ # error_h_weight: null
231
+
232
+
233
+ # ========================================================================================================== Network Architecture
234
+
235
+ # number of layers of EquivariantBlock to use in VAE's Encoder
236
+ encoder_n_layers: 1
237
+
238
+ # number of layers of EquivariantBlock to use in LDM and VAE's Decoder
239
+ n_layers: 4
240
+
241
+ # number of GCL Blocks to use in each EquivariantBlock
242
+ inv_sublayers: 1
243
+
244
+ # model's internal operating number of features
245
+ nf: 256
246
+
247
+ # use tanh in the coord_mlp
248
+ tanh: true
249
+
250
+ # use attention in the EGNN
251
+ attention: true
252
+
253
+ # diff/(|diff| + norm_constant)
254
+ norm_constant: 1
255
+
256
+ # whether using or not the sin embedding
257
+ sin_embedding: false
258
+
259
+ # uniform | variational | argmax_variational | deterministic
260
+ dequantization: argmax_variational
261
+
262
+ # Normalize the sum aggregation of EGNN
263
+ normalization_factor: 1
264
+
265
+ # EGNN aggregation method: sum | mean
266
+ aggregation_method: sum
267
+
268
+
269
+ # Fusion Block specific settings
270
+ fusion_weights: [0.5, 0.5, 0.5, 0.5] # [0.25, 0.5, 0.75, 1]
271
+ # Condition fusion method:
272
+ # - scaled_sum : (h1_i,x1_i) = (h1_i,x1_i) + w_i * (f_h1_i,f_x1_i)
273
+ # - balanced_sum : (h1_i,x1_i) = [(1 - w_i) * (h1_i,x1_i)] + [w_i * (f_h1_i,f_x1_i)]
274
+ # - replace : (h1_i,x1_i) = (f_h1_i,f_x1_i)
275
+ fusion_mode: balanced_sum
276
+
277
+ # Initial Noise Injection / Feedback Mechanism
278
+ noise_injection_weights: [0.5, 0.5] # pkt = w[0]*lg + w[1]*pkt
279
+ noise_injection_aggregation_method: mean # mean | sum
280
+ noise_injection_normalization_factor: 1 # aggregation normalization factor
281
+
282
+
283
+
284
+
285
+ # ========================================================================================================== Logging
286
+ # Can be used to visualize multiple times per epoch, default 1e8
287
+ visualize_sample_chain: true
288
+ visualize_every_batch: 20000
289
+ visualize_sample_chain_epochs: 2 # for 1% testing dataset, others set to 1
290
+ n_report_steps: 50
291
+
292
+
293
+
294
+
295
+ # ========================================================================================================== Saving & Resuming
296
+ # resume: null
297
+ # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A
298
+ # resume_model_ckpt: generative_model_6_iter_66486.npy
299
+ # resume_optim_ckpt: optim_6_iter_66486.npy
300
+ # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_1x_resume
301
+ # resume_model_ckpt: generative_model_6_iter_66486.npy
302
+ # resume_optim_ckpt: optim_6_iter_66486.npy
303
+ # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_2x_resume
304
+ # resume_model_ckpt: generative_model_4_iter_47490.npy
305
+ # resume_optim_ckpt: optim_4_iter_47490.npy
306
+ resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x_resume
307
+ resume_model_ckpt: generative_model_4_iter_47490.npy
308
+ resume_optim_ckpt: optim_4_iter_47490.npy
309
+
310
+ save_model: true
311
+
312
+
313
+
314
+ # ========================================================================================================== Wandb
315
+ # disable wandb
316
+ no_wandb: false
317
+ wandb_usr: gohyixian456
318
+ # True = wandb online -- False = wandb offline
319
+ online: true
320
+
321
+
322
+
323
+
324
+ pocket_vae:
325
+ dataset: d_20240623_CrossDocked_LG_PKT__10A__LIGAND+POCKET
326
+ vae_data_mode: pocket
327
+ remove_h: false
328
+ ca_only: false
329
+
330
+ # can contain multiple: homo | onehot | lumo | num_atoms | etc
331
+ conditioning: []
332
+
333
+ # egnn_dynamics
334
+ model: egnn_dynamics
335
+
336
+ # include atom charge, according to periodic table
337
+ include_charges: false
338
+
339
+ # Amount of EMA decay, 0 means off. A reasonable value is 0.999.
340
+ ema_decay: 0
341
+
342
+ # weight of KL term in ELBO, default 0.01
343
+ kl_weight: 0.01
344
+
345
+ # number of latent features, default 4 (have to match ligand VAE & LDM's latent_nf)
346
+ latent_nf: 2
347
+
348
+ # number of layers of EquivariantBlock to use in VAE's Encoder
349
+ encoder_n_layers: 1
350
+
351
+ # number of layers of EquivariantBlock to use in VAE's Decoder
352
+ n_layers: 4
353
+
354
+ # number of GCL Blocks to use in each EquivariantBlock
355
+ inv_sublayers: 1
356
+
357
+ # model's internal operating number of features
358
+ nf: 256
359
+
360
+ # use tanh in the coord_mlp
361
+ tanh: true
362
+
363
+ # use attention in the EGNN
364
+ attention: true
365
+
366
+ # diff/(|diff| + norm_constant)
367
+ norm_constant: 1
368
+
369
+ # whether using or not the sin embedding
370
+ sin_embedding: false
371
+
372
+ # uniform | variational | argmax_variational | deterministic
373
+ dequantization: argmax_variational
374
+
375
+ # Normalize the sum aggregation of EGNN
376
+ normalization_factor: 1
377
+
378
+ # EGNN aggregation method: sum | mean
379
+ aggregation_method: sum
380
+
381
+ # normalize factors for [x, h_cat/categorical/one-hot, h_int/integer/charges]
382
+ normalize_factors: [1, 4, 10]
383
+
384
+ vae_normalize_x: true
385
+ vae_normalize_method: scale # scale | linear
386
+ vae_normalize_factors: [10, 1, 1]
387
+
388
+ reweight_class_loss: "inv_class_freq"
389
+ reweight_coords_loss: "inv_class_freq"
390
+ smoothing_factor: 0.25 # [0.1 - 1.0) 1.0 is essentially disabling
391
+
392
+ error_x_weight: 30
393
+ error_h_weight: 15
03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x_resume/args_4_iter_47490.pickle ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30f955bbb5ed8bff26aae10897ab7be087e8219bc469c5232ca266e9c16cb6bb
3
+ size 5260
03_latent2_nf256_ds1k_fusBSum_conditionAll_0.5__GEOM_LDM_base__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x_resume/generative_model_4_iter_47490.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34eed4038e7d9946a9b02478cb8460eb8b09d8810499a95c058fe3197772af02
3
+ size 53567010
03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_2x_resume/args_5_iter_56988.pickle DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e265c6c80c375510446ea18e8c7c918bcbbbf7c4d7cae1e379f9aa8b27c7979d
3
- size 5422
 
 
 
 
{03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_2x_resume β†’ 03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x_resume}/03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A.yaml RENAMED
@@ -1,7 +1,9 @@
1
  proj_name: Control-GeoLDM
2
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A
3
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_1x
4
- exp_name: 03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_2x
 
 
5
 
6
 
7
  # ========================================================================================================== Training Mode (ldm/vae/both)
@@ -291,9 +293,15 @@ n_report_steps: 50
291
  # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A
292
  # resume_model_ckpt: generative_model_5_iter_56988.npy
293
  # resume_optim_ckpt: optim_5_iter_56988.npy
294
- resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_1x_resume
295
- resume_model_ckpt: generative_model_6_iter_66486.npy
296
- resume_optim_ckpt: optim_6_iter_66486.npy
 
 
 
 
 
 
297
 
298
  save_model: true
299
 
 
1
  proj_name: Control-GeoLDM
2
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A
3
  # exp_name: 03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_1x
4
+ # exp_name: 03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_2x
5
+ # exp_name: 03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x
6
+ exp_name: 03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_4x
7
 
8
 
9
  # ========================================================================================================== Training Mode (ldm/vae/both)
 
293
  # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A
294
  # resume_model_ckpt: generative_model_5_iter_56988.npy
295
  # resume_optim_ckpt: optim_5_iter_56988.npy
296
+ # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_1x_resume
297
+ # resume_model_ckpt: generative_model_6_iter_66486.npy
298
+ # resume_optim_ckpt: optim_6_iter_66486.npy
299
+ # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_2x_resume
300
+ # resume_model_ckpt: generative_model_5_iter_56988.npy
301
+ # resume_optim_ckpt: optim_5_iter_56988.npy
302
+ resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x_resume
303
+ resume_model_ckpt: generative_model_7_iter_75984.npy
304
+ resume_optim_ckpt: optim_7_iter_75984.npy
305
 
306
  save_model: true
307
 
03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x_resume/args_7_iter_75984.pickle ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f37e26821ddc295ae1a7fe672171d04256b98e46124fd00ce2b50ed62080f8c
3
+ size 5811
03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_2x_resume/generative_model_5_iter_56988.npy β†’ 03_latent2_nf256_ds1k_fusBSum_conditionBlocks34_0.1__epoch1k_bs10_lr1e-4_NoEMA__20241115__10A_3x_resume/generative_model_7_iter_75984.npy RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bfe6dc165698fecbafb4948e34f69a4e0c1d62c79be4a22dda8b69baeeb69b3
3
  size 53550562
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a85b31835c3fc18a14c66857fa94861030b1213a8ee706acbabd1426a8af6be7
3
  size 53550562
03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A/args_135_iter_224808.pickle DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a53ec1e5f606b364dd6579772c91d961b0576c88991fffe029868968a346e71
3
- size 5430
 
 
 
 
{03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A β†’ 03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume}/03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A.yaml RENAMED
@@ -1,5 +1,7 @@
1
  proj_name: Control-GeoLDM
2
- exp_name: 03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
 
 
3
 
4
 
5
  # ========================================================================================================== Training Mode (ldm/vae/both)
@@ -31,7 +33,8 @@ trainable_ligand_ae_decoder: false
31
  trainable_pocket_ae_encoder: false
32
 
33
  # Train 2nd stage LDM model
34
- trainable_ldm: false
 
35
 
36
  # Train 3rd stage ControlNet
37
  trainable_controlnet: true
@@ -286,10 +289,13 @@ n_report_steps: 50
286
 
287
 
288
  # ========================================================================================================== Saving & Resuming
289
- # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.5__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
290
- # resume_model_ckpt: generative_model_8_iter_14049.npy
291
- # resume_optim_ckpt: optim_8_iter_14049.npy
292
- resume: null
 
 
 
293
 
294
  save_model: true
295
 
 
1
  proj_name: Control-GeoLDM
2
+ # exp_name: 03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
3
+ # exp_name: 03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x
4
+ exp_name: FT_03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_2x
5
 
6
 
7
  # ========================================================================================================== Training Mode (ldm/vae/both)
 
33
  trainable_pocket_ae_encoder: false
34
 
35
  # Train 2nd stage LDM model
36
+ # trainable_ldm: false
37
+ trainable_ldm: true
38
 
39
  # Train 3rd stage ControlNet
40
  trainable_controlnet: true
 
289
 
290
 
291
  # ========================================================================================================== Saving & Resuming
292
+ # resume: null
293
+ # resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A
294
+ # resume_model_ckpt: generative_model_135_iter_224808.npy
295
+ # resume_optim_ckpt: optim_135_iter_224808.npy
296
+ resume: outputs_selected/controlnet/03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume
297
+ resume_model_ckpt: generative_model_105_iter_175218.npy
298
+ resume_optim_ckpt: optim_105_iter_175218.npy
299
 
300
  save_model: true
301
 
03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume/args_105_iter_175218.pickle ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10cec7a99405a19a6a249e986c91ee129549a82d332e4f13e98dccbccfc0650b
3
+ size 6061
03_latent2_nf256_ds1k_fusBSum_CA_conditionAll_0.1__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_3x_resume/generative_model_135_iter_224808.npy β†’ 03_latent2_nf256_ds1k_fusReplace_CA__epoch1k_bs60_lr1e-4_NoEMA__20241203__10A_1x_resume/generative_model_105_iter_175218.npy RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:041b33cb54d1628bf5516aa1ae250dd8c7f631835463dc86bc2ce00becebbe0d
3
  size 53576312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cec230630488ff56b5ba8a1539efac5f6fdaddd43c61646acafb3063e80e1195
3
  size 53576312