| [06-18 08:14:01] (SR/zrk_utils/arg_util.py, line 196)=> [tf32] [precis] torch.get_float32_matmul_precision(): high |
| [06-18 08:14:01] (SR/zrk_utils/arg_util.py, line 197)=> [tf32] [ conv ] torch.backends.cudnn.allow_tf32: True |
| [06-18 08:14:01] (SR/zrk_utils/arg_util.py, line 198)=> [tf32] [matmul] torch.backends.cuda.matmul.allow_tf32: True |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| ======================================================= RESTART [06-18 08:16:44] ======================================================= |
| [06-18 08:16:44] (SR/zrk_utils/arg_util.py, line 196)=> [tf32] [precis] torch.get_float32_matmul_precision(): high |
| [06-18 08:16:44] (SR/zrk_utils/arg_util.py, line 197)=> [tf32] [ conv ] torch.backends.cudnn.allow_tf32: True |
| [06-18 08:16:44] (SR/zrk_utils/arg_util.py, line 198)=> [tf32] [matmul] torch.backends.cuda.matmul.allow_tf32: True |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| ======================================================= RESTART [06-18 08:18:55] ======================================================= |
| [06-18 08:18:55] (SR/zrk_utils/arg_util.py, line 196)=> [tf32] [precis] torch.get_float32_matmul_precision(): high |
| [06-18 08:18:55] (SR/zrk_utils/arg_util.py, line 197)=> [tf32] [ conv ] torch.backends.cudnn.allow_tf32: True |
| [06-18 08:18:55] (SR/zrk_utils/arg_util.py, line 198)=> [tf32] [matmul] torch.backends.cuda.matmul.allow_tf32: True |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| ======================================================= RESTART [06-18 08:22:08] ======================================================= |
| [06-18 08:22:08] (SR/zrk_utils/arg_util.py, line 196)=> [tf32] [precis] torch.get_float32_matmul_precision(): high |
| [06-18 08:22:08] (SR/zrk_utils/arg_util.py, line 197)=> [tf32] [ conv ] torch.backends.cudnn.allow_tf32: True |
| [06-18 08:22:08] (SR/zrk_utils/arg_util.py, line 198)=> [tf32] [matmul] torch.backends.cuda.matmul.allow_tf32: True |
| [06-18 08:22:10] (/projects/VARSR/train.py, line 85)=> global bs=8, local bs=4 |
| [06-18 08:22:10] (/projects/VARSR/train.py, line 86)=> initial args: |
| { |
| vae_model_path : checkpoints/VQVAE.pth |
| var_pretrain_path : checkpoints/checkpoints_VAR/var_d24.pth |
| wandb_flag : True |
| exp_name : VARDahazing0616 |
| vfast : 0 |
| tfast : 0 |
| depth : 24 |
| ini : -1 |
| hd : 0.02 |
| aln : 0.5 |
| alng : 0.0001 |
| fp16 : 1 |
| tblr : 5e-05 |
| tlr : 5e-05 |
| twd : 0.05 |
| twde : 0.05 |
| tclip : 2.0 |
| ls : 0.0 |
| bs : 8 |
| batch_size : 4 |
| glb_batch_size : 8 |
| ac : 1 |
| ep : 5 |
| wp : 0.1 |
| wp0 : 0.005 |
| wpe : 0.01 |
| sche : lin0 |
| opt : adamw |
| afuse : True |
| saln : False |
| anorm : True |
| fuse : True |
| pn : 1_2_3_4_6_9_13_18_24_32 |
| patch_size : 16 |
| patch_nums : (1, 2, 3, 4, 6, 9, 13, 18, 24, 32) |
| resos : (16, 32, 48, 64, 96, 144, 208, 288, 384, 512) |
| data_load_reso : 512 |
| mid_reso : 1.125 |
| hflip : False |
| workers : 0 |
| dataset_opt_path : options/VARSR_Dehazing.yml |
| val_freq : 5000 |
| wandb_entity : benzlxs |
| project : VARSR |
| label_B : 1 |
| pg : 0.0 |
| pg0 : 4 |
| pgwp : 0.016666666666666666 |
| cmd : --depth=24 --bs=4 --ep=5 --fp16=1 --tblr=5e-5 --alng=1e-4 --wpe=0.01 --wandb_flag=True --fuse=1 --exp_name=VARDahazing0616 --opt=adamw |
| branch : master |
| commit_id : 80bd1f0d832951506610a3e82320a8322a6f1a7a |
| commit_msg : Update README.md |
| acc_mean : None |
| acc_tail : None |
| L_mean : None |
| L_tail : None |
| vacc_mean : None |
| vacc_tail : None |
| vL_mean : None |
| vL_tail : None |
| grad_norm : None |
| cur_lr : None |
| cur_wd : None |
| cur_it : |
| cur_ep : |
| remain_time : |
| finish_time : |
| local_out_dir_path : results/VARDahazing0616 |
| tb_log_dir_path : results/VARDahazing0616/tb-VARd24__pn1_2_3_4_6_9_13_18_24_32__b8ep5adamlr5e-05wd0.05 |
| log_txt_path : results/VARDahazing0616/log.txt |
| last_ckpt_path : results/VARDahazing0616/ar-ckpt-last.pth |
| tf32 : True |
| seed : None |
| same_seed_for_all_ranks: 0 |
| local_debug : False |
| dbg_nan : False |
| } |
|
|
| [06-18 08:22:10] (/projects/VARSR/train.py, line 97)=> [build dahazing paired dataset] ... |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| ======================================================= RESTART [06-18 08:37:53] ======================================================= |
| [06-18 08:37:53] (SR/zrk_utils/arg_util.py, line 196)=> [tf32] [precis] torch.get_float32_matmul_precision(): high |
| [06-18 08:37:53] (SR/zrk_utils/arg_util.py, line 197)=> [tf32] [ conv ] torch.backends.cudnn.allow_tf32: True |
| [06-18 08:37:53] (SR/zrk_utils/arg_util.py, line 198)=> [tf32] [matmul] torch.backends.cuda.matmul.allow_tf32: True |
| [06-18 08:37:55] (/projects/VARSR/train.py, line 85)=> global bs=8, local bs=4 |
| [06-18 08:37:55] (/projects/VARSR/train.py, line 86)=> initial args: |
| { |
| vae_model_path : checkpoints/VQVAE.pth |
| var_pretrain_path : checkpoints/checkpoints_VAR/var_d24.pth |
| wandb_flag : True |
| exp_name : VARDahazing0616 |
| vfast : 0 |
| tfast : 0 |
| depth : 24 |
| ini : -1 |
| hd : 0.02 |
| aln : 0.5 |
| alng : 0.0001 |
| fp16 : 1 |
| tblr : 5e-05 |
| tlr : 5e-05 |
| twd : 0.05 |
| twde : 0.05 |
| tclip : 2.0 |
| ls : 0.0 |
| bs : 8 |
| batch_size : 4 |
| glb_batch_size : 8 |
| ac : 1 |
| ep : 5 |
| wp : 0.1 |
| wp0 : 0.005 |
| wpe : 0.01 |
| sche : lin0 |
| opt : adamw |
| afuse : True |
| saln : False |
| anorm : True |
| fuse : True |
| pn : 1_2_3_4_6_9_13_18_24_32 |
| patch_size : 16 |
| patch_nums : (1, 2, 3, 4, 6, 9, 13, 18, 24, 32) |
| resos : (16, 32, 48, 64, 96, 144, 208, 288, 384, 512) |
| data_load_reso : 512 |
| mid_reso : 1.125 |
| hflip : False |
| workers : 0 |
| dataset_opt_path : options/VARSR_Dehazing.yml |
| val_freq : 5000 |
| wandb_entity : benzlxs |
| project : VARSR |
| label_B : 1 |
| pg : 0.0 |
| pg0 : 4 |
| pgwp : 0.016666666666666666 |
| cmd : --depth=24 --bs=4 --ep=5 --fp16=1 --tblr=5e-5 --alng=1e-4 --wpe=0.01 --wandb_flag=True --fuse=1 --exp_name=VARDahazing0616 --opt=adamw |
| branch : master |
| commit_id : 80bd1f0d832951506610a3e82320a8322a6f1a7a |
| commit_msg : Update README.md |
| acc_mean : None |
| acc_tail : None |
| L_mean : None |
| L_tail : None |
| vacc_mean : None |
| vacc_tail : None |
| vL_mean : None |
| vL_tail : None |
| grad_norm : None |
| cur_lr : None |
| cur_wd : None |
| cur_it : |
| cur_ep : |
| remain_time : |
| finish_time : |
| local_out_dir_path : results/VARDahazing0616 |
| tb_log_dir_path : results/VARDahazing0616/tb-VARd24__pn1_2_3_4_6_9_13_18_24_32__b8ep5adamlr5e-05wd0.05 |
| log_txt_path : results/VARDahazing0616/log.txt |
| last_ckpt_path : results/VARDahazing0616/ar-ckpt-last.pth |
| tf32 : True |
| seed : None |
| same_seed_for_all_ranks: 0 |
| local_debug : False |
| dbg_nan : False |
| } |
|
|
| [06-18 08:37:55] (/projects/VARSR/train.py, line 97)=> [build dahazing paired dataset] ... |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| ======================================================= RESTART [06-18 08:39:37] ======================================================= |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| ======================================================= RESTART [06-18 08:44:29] ======================================================= |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| ======================================================= RESTART [06-18 08:49:35] ======================================================= |
| [06-18 08:44:29] (SR/zrk_utils/arg_util.py, line 196)=> [tf32] [precis] torch.get_float32_matmul_precision(): high |
| [06-18 08:44:29] (SR/zrk_utils/arg_util.py, line 197)=> [tf32] [ conv ] torch.backends.cudnn.allow_tf32: True |
| [06-18 08:44:29] (SR/zrk_utils/arg_util.py, line 198)=> [tf32] [matmul] torch.backends.cuda.matmul.allow_tf32: True |
| [06-18 08:44:31] (train.py , line 85)=> global bs=4, local bs=4 |
| [06-18 08:44:31] (train.py , line 86)=> initial args: |
| { |
| vae_model_path : checkpoints/VQVAE.pth |
| var_pretrain_path : checkpoints/checkpoints_VAR/var_d24.pth |
| wandb_flag : True |
| exp_name : VARDahazing0616 |
| vfast : 0 |
| tfast : 0 |
| depth : 24 |
| ini : -1 |
| hd : 0.02 |
| aln : 0.5 |
| alng : 0.0001 |
| fp16 : 1 |
| tblr : 5e-05 |
| tlr : 5e-05 |
| twd : 0.05 |
| twde : 0.05 |
| tclip : 2.0 |
| ls : 0.0 |
| bs : 4 |
| batch_size : 4 |
| glb_batch_size : 4 |
| ac : 1 |
| ep : 5 |
| wp : 0.1 |
| wp0 : 0.005 |
| wpe : 0.01 |
| sche : lin0 |
| opt : adamw |
| afuse : True |
| saln : False |
| anorm : True |
| fuse : True |
| pn : 1_2_3_4_6_9_13_18_24_32 |
| patch_size : 16 |
| patch_nums : (1, 2, 3, 4, 6, 9, 13, 18, 24, 32) |
| resos : (16, 32, 48, 64, 96, 144, 208, 288, 384, 512) |
| data_load_reso : 512 |
| mid_reso : 1.125 |
| hflip : False |
| workers : 0 |
| dataset_opt_path : options/VARSR_Dehazing.yml |
| val_freq : 5000 |
| wandb_entity : benzlxs |
| project : VARSR |
| label_B : 1 |
| pg : 0.0 |
| pg0 : 4 |
| pgwp : 0.016666666666666666 |
| cmd : --depth=24 --bs=4 --ep=5 --fp16=1 --tblr=5e-5 --alng=1e-4 --wpe=0.01 --wandb_flag=True --fuse=1 --exp_name=VARDahazing0616 --opt=adamw |
| branch : master |
| commit_id : 80bd1f0d832951506610a3e82320a8322a6f1a7a |
| commit_msg : Update README.md |
| acc_mean : None |
| acc_tail : None |
| L_mean : None |
| L_tail : None |
| vacc_mean : None |
| vacc_tail : None |
| vL_mean : None |
| vL_tail : None |
| grad_norm : None |
| cur_lr : None |
| cur_wd : None |
| cur_it : |
| cur_ep : |
| remain_time : |
| finish_time : |
| local_out_dir_path : results/VARDahazing0616 |
| tb_log_dir_path : results/VARDahazing0616/tb-VARd24__pn1_2_3_4_6_9_13_18_24_32__b4ep5adamlr5e-05wd0.05 |
| log_txt_path : results/VARDahazing0616/log.txt |
| last_ckpt_path : results/VARDahazing0616/ar-ckpt-last.pth |
| tf32 : True |
| seed : None |
| same_seed_for_all_ranks: 0 |
| local_debug : False |
| dbg_nan : False |
| } |
|
|
| [06-18 08:44:32] (train.py , line 97)=> [build dahazing paired dataset] ... |
|
|
| [06-18 08:49:35] (SR/zrk_utils/arg_util.py, line 196)=> [tf32] [precis] torch.get_float32_matmul_precision(): high |
| [06-18 08:49:35] (SR/zrk_utils/arg_util.py, line 197)=> [tf32] [ conv ] torch.backends.cudnn.allow_tf32: True |
| [06-18 08:49:35] (SR/zrk_utils/arg_util.py, line 198)=> [tf32] [matmul] torch.backends.cuda.matmul.allow_tf32: True |
| [06-18 08:49:37] (train.py , line 85)=> global bs=4, local bs=4 |
| [06-18 08:49:37] (train.py , line 86)=> initial args: |
| { |
| vae_model_path : checkpoints/VQVAE.pth |
| var_pretrain_path : checkpoints/checkpoints_VAR/var_d24.pth |
| wandb_flag : True |
| exp_name : VARDahazing0616 |
| vfast : 0 |
| tfast : 0 |
| depth : 24 |
| ini : -1 |
| hd : 0.02 |
| aln : 0.5 |
| alng : 0.0001 |
| fp16 : 1 |
| tblr : 5e-05 |
| tlr : 5e-05 |
| twd : 0.05 |
| twde : 0.05 |
| tclip : 2.0 |
| ls : 0.0 |
| bs : 4 |
| batch_size : 4 |
| glb_batch_size : 4 |
| ac : 1 |
| ep : 5 |
| wp : 0.1 |
| wp0 : 0.005 |
| wpe : 0.01 |
| sche : lin0 |
| opt : adamw |
| afuse : True |
| saln : False |
| anorm : True |
| fuse : True |
| pn : 1_2_3_4_6_9_13_18_24_32 |
| patch_size : 16 |
| patch_nums : (1, 2, 3, 4, 6, 9, 13, 18, 24, 32) |
| resos : (16, 32, 48, 64, 96, 144, 208, 288, 384, 512) |
| data_load_reso : 512 |
| mid_reso : 1.125 |
| hflip : False |
| workers : 0 |
| dataset_opt_path : options/VARSR_Dehazing.yml |
| val_freq : 5000 |
| wandb_entity : benzlxs |
| project : VARSR |
| label_B : 1 |
| pg : 0.0 |
| pg0 : 4 |
| pgwp : 0.016666666666666666 |
| cmd : --depth=24 --bs=4 --ep=5 --fp16=1 --tblr=5e-5 --alng=1e-4 --wpe=0.01 --wandb_flag=True --fuse=1 --exp_name=VARDahazing0616 --opt=adamw |
| branch : master |
| commit_id : 80bd1f0d832951506610a3e82320a8322a6f1a7a |
| commit_msg : Update README.md |
| acc_mean : None |
| acc_tail : None |
| L_mean : None |
| L_tail : None |
| vacc_mean : None |
| vacc_tail : None |
| vL_mean : None |
| vL_tail : None |
| grad_norm : None |
| cur_lr : None |
| cur_wd : None |
| cur_it : |
| cur_ep : |
| remain_time : |
| finish_time : |
| local_out_dir_path : results/VARDahazing0616 |
| tb_log_dir_path : results/VARDahazing0616/tb-VARd24__pn1_2_3_4_6_9_13_18_24_32__b4ep5adamlr5e-05wd0.05 |
| log_txt_path : results/VARDahazing0616/log.txt |
| last_ckpt_path : results/VARDahazing0616/ar-ckpt-last.pth |
| tf32 : True |
| seed : None |
| same_seed_for_all_ranks: 0 |
| local_debug : False |
| dbg_nan : False |
| } |
|
|
| [06-18 08:49:38] (train.py , line 97)=> [build dahazing paired dataset] ... |
|
|
| [06-18 08:49:35] (SR/zrk_utils/arg_util.py, line 196)=> [tf32] [precis] torch.get_float32_matmul_precision(): high |
| [06-18 08:49:35] (SR/zrk_utils/arg_util.py, line 197)=> [tf32] [ conv ] torch.backends.cudnn.allow_tf32: True |
| [06-18 08:49:35] (SR/zrk_utils/arg_util.py, line 198)=> [tf32] [matmul] torch.backends.cuda.matmul.allow_tf32: True |
| [06-18 08:49:37] (train.py , line 85)=> global bs=4, local bs=4 |
| [06-18 08:49:37] (train.py , line 86)=> initial args: |
| { |
| vae_model_path : checkpoints/VQVAE.pth |
| var_pretrain_path : checkpoints/checkpoints_VAR/var_d24.pth |
| wandb_flag : True |
| exp_name : VARDahazing0616 |
| vfast : 0 |
| tfast : 0 |
| depth : 24 |
| ini : -1 |
| hd : 0.02 |
| aln : 0.5 |
| alng : 0.0001 |
| fp16 : 1 |
| tblr : 5e-05 |
| tlr : 5e-05 |
| twd : 0.05 |
| twde : 0.05 |
| tclip : 2.0 |
| ls : 0.0 |
| bs : 4 |
| batch_size : 4 |
| glb_batch_size : 4 |
| ac : 1 |
| ep : 5 |
| wp : 0.1 |
| wp0 : 0.005 |
| wpe : 0.01 |
| sche : lin0 |
| opt : adamw |
| afuse : True |
| saln : False |
| anorm : True |
| fuse : True |
| pn : 1_2_3_4_6_9_13_18_24_32 |
| patch_size : 16 |
| patch_nums : (1, 2, 3, 4, 6, 9, 13, 18, 24, 32) |
| resos : (16, 32, 48, 64, 96, 144, 208, 288, 384, 512) |
| data_load_reso : 512 |
| mid_reso : 1.125 |
| hflip : False |
| workers : 0 |
| dataset_opt_path : options/VARSR_Dehazing.yml |
| val_freq : 5000 |
| wandb_entity : benzlxs |
| project : VARSR |
| label_B : 1 |
| pg : 0.0 |
| pg0 : 4 |
| pgwp : 0.016666666666666666 |
| cmd : --depth=24 --bs=4 --ep=5 --fp16=1 --tblr=5e-5 --alng=1e-4 --wpe=0.01 --wandb_flag=True --fuse=1 --exp_name=VARDahazing0616 --opt=adamw |
| branch : master |
| commit_id : 80bd1f0d832951506610a3e82320a8322a6f1a7a |
| commit_msg : Update README.md |
| acc_mean : None |
| acc_tail : None |
| L_mean : None |
| L_tail : None |
| vacc_mean : None |
| vacc_tail : None |
| vL_mean : None |
| vL_tail : None |
| grad_norm : None |
| cur_lr : None |
| cur_wd : None |
| cur_it : |
| cur_ep : |
| remain_time : |
| finish_time : |
| local_out_dir_path : results/VARDahazing0616 |
| tb_log_dir_path : results/VARDahazing0616/tb-VARd24__pn1_2_3_4_6_9_13_18_24_32__b4ep5adamlr5e-05wd0.05 |
| log_txt_path : results/VARDahazing0616/log.txt |
| last_ckpt_path : results/VARDahazing0616/ar-ckpt-last.pth |
| tf32 : True |
| seed : None |
| same_seed_for_all_ranks: 0 |
| local_debug : False |
| dbg_nan : False |
| } |
|
|
| [06-18 08:49:38] (train.py , line 97)=> [build dahazing paired dataset] ... |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| ======================================================= RESTART [06-18 13:47:22] ======================================================= |
| [06-18 13:47:22] (SR/zrk_utils/arg_util.py, line 196)=> [tf32] [precis] torch.get_float32_matmul_precision(): high |
| [06-18 13:47:22] (SR/zrk_utils/arg_util.py, line 197)=> [tf32] [ conv ] torch.backends.cudnn.allow_tf32: True |
| [06-18 13:47:22] (SR/zrk_utils/arg_util.py, line 198)=> [tf32] [matmul] torch.backends.cuda.matmul.allow_tf32: True |
| [06-18 13:47:23] (/projects/VARSR/train.py, line 85)=> global bs=8, local bs=4 |
| [06-18 13:47:23] (/projects/VARSR/train.py, line 86)=> initial args: |
| { |
| vae_model_path : checkpoints/VQVAE.pth |
| var_pretrain_path : checkpoints/checkpoints_VAR/var_d24.pth |
| wandb_flag : True |
| exp_name : VARDahazing0616 |
| vfast : 0 |
| tfast : 0 |
| depth : 24 |
| ini : -1 |
| hd : 0.02 |
| aln : 0.5 |
| alng : 0.0001 |
| fp16 : 1 |
| tblr : 5e-05 |
| tlr : 5e-05 |
| twd : 0.05 |
| twde : 0.05 |
| tclip : 2.0 |
| ls : 0.0 |
| bs : 8 |
| batch_size : 4 |
| glb_batch_size : 8 |
| ac : 1 |
| ep : 5 |
| wp : 0.1 |
| wp0 : 0.005 |
| wpe : 0.01 |
| sche : lin0 |
| opt : adamw |
| afuse : True |
| saln : False |
| anorm : True |
| fuse : True |
| pn : 1_2_3_4_6_9_13_18_24_32 |
| patch_size : 16 |
| patch_nums : (1, 2, 3, 4, 6, 9, 13, 18, 24, 32) |
| resos : (16, 32, 48, 64, 96, 144, 208, 288, 384, 512) |
| data_load_reso : 512 |
| mid_reso : 1.125 |
| hflip : False |
| workers : 0 |
| dataset_opt_path : options/VARSR_Dehazing.yml |
| val_freq : 5000 |
| wandb_entity : benzlxs |
| project : VARSR |
| label_B : 1 |
| pg : 0.0 |
| pg0 : 4 |
| pgwp : 0.016666666666666666 |
| cmd : --depth=24 --bs=4 --ep=5 --fp16=1 --tblr=5e-5 --alng=1e-4 --wpe=0.01 --wandb_flag=True --fuse=1 --exp_name=VARDahazing0616 --opt=adamw |
| branch : master |
| commit_id : 80bd1f0d832951506610a3e82320a8322a6f1a7a |
| commit_msg : Update README.md |
| acc_mean : None |
| acc_tail : None |
| L_mean : None |
| L_tail : None |
| vacc_mean : None |
| vacc_tail : None |
| vL_mean : None |
| vL_tail : None |
| grad_norm : None |
| cur_lr : None |
| cur_wd : None |
| cur_it : |
| cur_ep : |
| remain_time : |
| finish_time : |
| local_out_dir_path : results/VARDahazing0616 |
| tb_log_dir_path : results/VARDahazing0616/tb-VARd24__pn1_2_3_4_6_9_13_18_24_32__b8ep5adamlr5e-05wd0.05 |
| log_txt_path : results/VARDahazing0616/log.txt |
| last_ckpt_path : results/VARDahazing0616/ar-ckpt-last.pth |
| tf32 : True |
| seed : None |
| same_seed_for_all_ranks: 0 |
| local_debug : False |
| dbg_nan : False |
| } |
|
|
| [06-18 13:47:24] (/projects/VARSR/train.py, line 97)=> [build dahazing paired dataset] ... |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| ======================================================= RESTART [06-18 13:50:35] ======================================================= |
| [06-18 13:50:35] (SR/zrk_utils/arg_util.py, line 196)=> [tf32] [precis] torch.get_float32_matmul_precision(): high |
| [06-18 13:50:35] (SR/zrk_utils/arg_util.py, line 197)=> [tf32] [ conv ] torch.backends.cudnn.allow_tf32: True |
| [06-18 13:50:35] (SR/zrk_utils/arg_util.py, line 198)=> [tf32] [matmul] torch.backends.cuda.matmul.allow_tf32: True |
| [06-18 13:50:36] (/projects/VARSR/train.py, line 85)=> global bs=8, local bs=4 |
| [06-18 13:50:36] (/projects/VARSR/train.py, line 86)=> initial args: |
| { |
| vae_model_path : checkpoints/VQVAE.pth |
| var_pretrain_path : checkpoints/checkpoints_VAR/var_d24.pth |
| wandb_flag : True |
| exp_name : VARDahazing0616 |
| vfast : 0 |
| tfast : 0 |
| depth : 24 |
| ini : -1 |
| hd : 0.02 |
| aln : 0.5 |
| alng : 0.0001 |
| fp16 : 1 |
| tblr : 5e-05 |
| tlr : 5e-05 |
| twd : 0.05 |
| twde : 0.05 |
| tclip : 2.0 |
| ls : 0.0 |
| bs : 8 |
| batch_size : 4 |
| glb_batch_size : 8 |
| ac : 1 |
| ep : 5 |
| wp : 0.1 |
| wp0 : 0.005 |
| wpe : 0.01 |
| sche : lin0 |
| opt : adamw |
| afuse : True |
| saln : False |
| anorm : True |
| fuse : True |
| pn : 1_2_3_4_6_9_13_18_24_32 |
| patch_size : 16 |
| patch_nums : (1, 2, 3, 4, 6, 9, 13, 18, 24, 32) |
| resos : (16, 32, 48, 64, 96, 144, 208, 288, 384, 512) |
| data_load_reso : 512 |
| mid_reso : 1.125 |
| hflip : False |
| workers : 0 |
| dataset_opt_path : options/VARSR_Dehazing.yml |
| val_freq : 5000 |
| wandb_entity : benzlxs |
| project : VARSR |
| label_B : 1 |
| pg : 0.0 |
| pg0 : 4 |
| pgwp : 0.016666666666666666 |
| cmd : --depth=24 --bs=4 --ep=5 --fp16=1 --tblr=5e-5 --alng=1e-4 --wpe=0.01 --wandb_flag=True --fuse=1 --exp_name=VARDahazing0616 --opt=adamw |
| branch : master |
| commit_id : 80bd1f0d832951506610a3e82320a8322a6f1a7a |
| commit_msg : Update README.md |
| acc_mean : None |
| acc_tail : None |
| L_mean : None |
| L_tail : None |
| vacc_mean : None |
| vacc_tail : None |
| vL_mean : None |
| vL_tail : None |
| grad_norm : None |
| cur_lr : None |
| cur_wd : None |
| cur_it : |
| cur_ep : |
| remain_time : |
| finish_time : |
| local_out_dir_path : results/VARDahazing0616 |
| tb_log_dir_path : results/VARDahazing0616/tb-VARd24__pn1_2_3_4_6_9_13_18_24_32__b8ep5adamlr5e-05wd0.05 |
| log_txt_path : results/VARDahazing0616/log.txt |
| last_ckpt_path : results/VARDahazing0616/ar-ckpt-last.pth |
| tf32 : True |
| seed : None |
| same_seed_for_all_ranks: 0 |
| local_debug : False |
| dbg_nan : False |
| } |
|
|
| [06-18 13:50:37] (/projects/VARSR/train.py, line 97)=> [build dahazing paired dataset] ... |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| ======================================================= RESTART [06-18 18:41:00] ======================================================= |
| [06-18 18:41:00] (SR/zrk_utils/arg_util.py, line 196)=> [tf32] [precis] torch.get_float32_matmul_precision(): high |
| [06-18 18:41:00] (SR/zrk_utils/arg_util.py, line 197)=> [tf32] [ conv ] torch.backends.cudnn.allow_tf32: True |
| [06-18 18:41:00] (SR/zrk_utils/arg_util.py, line 198)=> [tf32] [matmul] torch.backends.cuda.matmul.allow_tf32: True |
| [06-18 18:41:01] (/projects/VARSR/train.py, line 85)=> global bs=8, local bs=4 |
| [06-18 18:41:01] (/projects/VARSR/train.py, line 86)=> initial args: |
| { |
| vae_model_path : checkpoints/VQVAE.pth |
| var_pretrain_path : checkpoints/checkpoints_VAR/var_d24.pth |
| wandb_flag : True |
| exp_name : VARDahazing0616 |
| vfast : 0 |
| tfast : 0 |
| depth : 24 |
| ini : -1 |
| hd : 0.02 |
| aln : 0.5 |
| alng : 0.0001 |
| fp16 : 1 |
| tblr : 5e-05 |
| tlr : 5e-05 |
| twd : 0.05 |
| twde : 0.05 |
| tclip : 2.0 |
| ls : 0.0 |
| bs : 8 |
| batch_size : 4 |
| glb_batch_size : 8 |
| ac : 1 |
| ep : 5 |
| wp : 0.1 |
| wp0 : 0.005 |
| wpe : 0.01 |
| sche : lin0 |
| opt : adamw |
| afuse : True |
| saln : False |
| anorm : True |
| fuse : True |
| pn : 1_2_3_4_6_9_13_18_24_32 |
| patch_size : 16 |
| patch_nums : (1, 2, 3, 4, 6, 9, 13, 18, 24, 32) |
| resos : (16, 32, 48, 64, 96, 144, 208, 288, 384, 512) |
| data_load_reso : 512 |
| mid_reso : 1.125 |
| hflip : False |
| workers : 0 |
| dataset_opt_path : options/VARSR_Dehazing.yml |
| val_freq : 5000 |
| wandb_entity : benzlxs |
| project : VARSR |
| label_B : 1 |
| pg : 0.0 |
| pg0 : 4 |
| pgwp : 0.016666666666666666 |
| cmd : --depth=24 --bs=4 --ep=5 --fp16=1 --tblr=5e-5 --alng=1e-4 --wpe=0.01 --wandb_flag=True --fuse=1 --exp_name=VARDahazing0616 --opt=adamw |
| branch : master |
| commit_id : 80bd1f0d832951506610a3e82320a8322a6f1a7a |
| commit_msg : Update README.md |
| acc_mean : None |
| acc_tail : None |
| L_mean : None |
| L_tail : None |
| vacc_mean : None |
| vacc_tail : None |
| vL_mean : None |
| vL_tail : None |
| grad_norm : None |
| cur_lr : None |
| cur_wd : None |
| cur_it : |
| cur_ep : |
| remain_time : |
| finish_time : |
| local_out_dir_path : results/VARDahazing0616 |
| tb_log_dir_path : results/VARDahazing0616/tb-VARd24__pn1_2_3_4_6_9_13_18_24_32__b8ep5adamlr5e-05wd0.05 |
| log_txt_path : results/VARDahazing0616/log.txt |
| last_ckpt_path : results/VARDahazing0616/ar-ckpt-last.pth |
| tf32 : True |
| seed : None |
| same_seed_for_all_ranks: 0 |
| local_debug : False |
| dbg_nan : False |
| } |
|
|
| [06-18 18:41:02] (/projects/VARSR/train.py, line 97)=> [build dahazing paired dataset] ... |
|
|
| [06-18 18:41:18] (/VARSR/zrk_utils/data.py, line 22)=> [Dataset] len(dataset)=18130 |
| [06-18 18:41:18] (/VARSR/zrk_utils/data.py, line 22)=> [Dataset] len(dataset)=1135 |
| [06-18 18:41:18] (/projects/VARSR/train.py, line 146)=> [auto_resume] no ckpt found @ results/VARDahazing0616/ar-ckpt*.pth |
| [06-18 18:41:18] (/projects/VARSR/train.py, line 146)=> [auto_resume quit] |
| [06-18 18:41:18] (/projects/VARSR/train.py, line 147)=> [dataloader multi processing] ... [dataloader multi processing](*) finished! (0.00s) |
| [06-18 18:41:18] (/projects/VARSR/train.py, line 153)=> [dataloader] gbs=8, lbs=4, iters_train=2267, types(tr, va)=('PairedImageDataset_Dehazing', 'PairedImageDataset_Dehazing') |
| [06-18 18:41:18] (R/models/var4dehazing.py, line 183)=> [VAR config ] embed_dim=1536, num_heads=24, depth=24, mlp_ratio=4.0 |
| [drop ratios ] drop_rate=0.0, attn_drop_rate=0.0, drop_path_rate=0.1 (tensor([0.0000, 0.0043, 0.0087, 0.0130, 0.0174, 0.0217, 0.0261, 0.0304, 0.0348, |
| 0.0391, 0.0435, 0.0478, 0.0522, 0.0565, 0.0609, 0.0652, 0.0696, 0.0739, |
| 0.0783, 0.0826, 0.0870, 0.0913, 0.0957, 0.1000])) |
|
|
| [06-18 18:41:18] (R/models/var4dehazing.py, line 207)=> torch.Size([1, 1, 3264, 3264]) |
| [06-18 18:41:19] (R/models/var4dehazing.py, line 504)=> [init_weights] VAR_dehazing with init_std=0.0147314 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| ======================================================= RESTART [06-18 20:21:22] ======================================================= |
| [06-18 20:21:22] (SR/zrk_utils/arg_util.py, line 196)=> [tf32] [precis] torch.get_float32_matmul_precision(): high |
| [06-18 20:21:22] (SR/zrk_utils/arg_util.py, line 197)=> [tf32] [ conv ] torch.backends.cudnn.allow_tf32: True |
| [06-18 20:21:22] (SR/zrk_utils/arg_util.py, line 198)=> [tf32] [matmul] torch.backends.cuda.matmul.allow_tf32: True |
| [06-18 20:21:23] (/projects/VARSR/train.py, line 85)=> global bs=8, local bs=4 |
| [06-18 20:21:23] (/projects/VARSR/train.py, line 86)=> initial args: |
| { |
| vae_model_path : checkpoints/VQVAE.pth |
| var_pretrain_path : checkpoints/checkpoints_VAR/var_d24.pth |
| wandb_flag : True |
| exp_name : VARDahazing0616 |
| vfast : 0 |
| tfast : 0 |
| depth : 24 |
| ini : -1 |
| hd : 0.02 |
| aln : 0.5 |
| alng : 0.0001 |
| fp16 : 1 |
| tblr : 5e-05 |
| tlr : 5e-05 |
| twd : 0.05 |
| twde : 0.05 |
| tclip : 2.0 |
| ls : 0.0 |
| bs : 8 |
| batch_size : 4 |
| glb_batch_size : 8 |
| ac : 1 |
| ep : 5 |
| wp : 0.1 |
| wp0 : 0.005 |
| wpe : 0.01 |
| sche : lin0 |
| opt : adamw |
| afuse : True |
| saln : False |
| anorm : True |
| fuse : True |
| pn : 1_2_3_4_6_9_13_18_24_32 |
| patch_size : 16 |
| patch_nums : (1, 2, 3, 4, 6, 9, 13, 18, 24, 32) |
| resos : (16, 32, 48, 64, 96, 144, 208, 288, 384, 512) |
| data_load_reso : 512 |
| mid_reso : 1.125 |
| hflip : False |
| workers : 0 |
| dataset_opt_path : options/VARSR_Dehazing.yml |
| val_freq : 5000 |
| wandb_entity : benzlxs |
| project : VARSR |
| label_B : 1 |
| pg : 0.0 |
| pg0 : 4 |
| pgwp : 0.016666666666666666 |
| cmd : --depth=24 --bs=4 --ep=5 --fp16=1 --tblr=5e-5 --alng=1e-4 --wpe=0.01 --wandb_flag=True --fuse=1 --exp_name=VARDahazing0616 --opt=adamw |
| branch : master |
| commit_id : 80bd1f0d832951506610a3e82320a8322a6f1a7a |
| commit_msg : Update README.md |
| acc_mean : None |
| acc_tail : None |
| L_mean : None |
| L_tail : None |
| vacc_mean : None |
| vacc_tail : None |
| vL_mean : None |
| vL_tail : None |
| grad_norm : None |
| cur_lr : None |
| cur_wd : None |
| cur_it : |
| cur_ep : |
| remain_time : |
| finish_time : |
| local_out_dir_path : results/VARDahazing0616 |
| tb_log_dir_path : results/VARDahazing0616/tb-VARd24__pn1_2_3_4_6_9_13_18_24_32__b8ep5adamlr5e-05wd0.05 |
| log_txt_path : results/VARDahazing0616/log.txt |
| last_ckpt_path : results/VARDahazing0616/ar-ckpt-last.pth |
| tf32 : True |
| seed : None |
| same_seed_for_all_ranks: 0 |
| local_debug : False |
| dbg_nan : False |
| } |
|
|
| [06-18 20:21:23] (/projects/VARSR/train.py, line 97)=> [build dahazing paired dataset] ... |
|
|
| [06-18 20:21:31] (/VARSR/zrk_utils/data.py, line 22)=> [Dataset] len(dataset)=18130 |
| [06-18 20:21:31] (/VARSR/zrk_utils/data.py, line 22)=> [Dataset] len(dataset)=1135 |
| [06-18 20:21:31] (/projects/VARSR/train.py, line 146)=> [auto_resume] no ckpt found @ results/VARDahazing0616/ar-ckpt*.pth |
| [06-18 20:21:31] (/projects/VARSR/train.py, line 146)=> [auto_resume quit] |
| [06-18 20:21:31] (/projects/VARSR/train.py, line 147)=> [dataloader multi processing] ... [dataloader multi processing](*) finished! (0.00s) |
| [06-18 20:21:31] (/projects/VARSR/train.py, line 153)=> [dataloader] gbs=8, lbs=4, iters_train=2267, types(tr, va)=('PairedImageDataset_Dehazing', 'PairedImageDataset_Dehazing') |
| [06-18 20:21:32] (R/models/var4dehazing.py, line 183)=> [VAR config ] embed_dim=1536, num_heads=24, depth=24, mlp_ratio=4.0 |
| [drop ratios ] drop_rate=0.0, attn_drop_rate=0.0, drop_path_rate=0.1 (tensor([0.0000, 0.0043, 0.0087, 0.0130, 0.0174, 0.0217, 0.0261, 0.0304, 0.0348, |
| 0.0391, 0.0435, 0.0478, 0.0522, 0.0565, 0.0609, 0.0652, 0.0696, 0.0739, |
| 0.0783, 0.0826, 0.0870, 0.0913, 0.0957, 0.1000])) |
|
|
| [06-18 20:21:32] (R/models/var4dehazing.py, line 207)=> torch.Size([1, 1, 3264, 3264]) |
| [06-18 20:21:33] (R/models/var4dehazing.py, line 504)=> [init_weights] VAR_dehazing with init_std=0.0147314 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| ======================================================= RESTART [06-18 20:22:42] ======================================================= |
| [06-18 20:22:42] (SR/zrk_utils/arg_util.py, line 196)=> [tf32] [precis] torch.get_float32_matmul_precision(): high |
| [06-18 20:22:42] (SR/zrk_utils/arg_util.py, line 197)=> [tf32] [ conv ] torch.backends.cudnn.allow_tf32: True |
| [06-18 20:22:42] (SR/zrk_utils/arg_util.py, line 198)=> [tf32] [matmul] torch.backends.cuda.matmul.allow_tf32: True |
| [06-18 20:22:44] (/projects/VARSR/train.py, line 85)=> global bs=8, local bs=4 |
| [06-18 20:22:44] (/projects/VARSR/train.py, line 86)=> initial args: |
| { |
| vae_model_path : checkpoints/VQVAE.pth |
| var_pretrain_path : checkpoints/checkpoints_VAR/var_d24.pth |
| wandb_flag : True |
| exp_name : VARDahazing0616 |
| vfast : 0 |
| tfast : 0 |
| depth : 24 |
| ini : -1 |
| hd : 0.02 |
| aln : 0.5 |
| alng : 0.0001 |
| fp16 : 1 |
| tblr : 5e-05 |
| tlr : 5e-05 |
| twd : 0.05 |
| twde : 0.05 |
| tclip : 2.0 |
| ls : 0.0 |
| bs : 8 |
| batch_size : 4 |
| glb_batch_size : 8 |
| ac : 1 |
| ep : 5 |
| wp : 0.1 |
| wp0 : 0.005 |
| wpe : 0.01 |
| sche : lin0 |
| opt : adamw |
| afuse : True |
| saln : False |
| anorm : True |
| fuse : True |
| pn : 1_2_3_4_6_9_13_18_24_32 |
| patch_size : 16 |
| patch_nums : (1, 2, 3, 4, 6, 9, 13, 18, 24, 32) |
| resos : (16, 32, 48, 64, 96, 144, 208, 288, 384, 512) |
| data_load_reso : 512 |
| mid_reso : 1.125 |
| hflip : False |
| workers : 0 |
| dataset_opt_path : options/VARSR_Dehazing.yml |
| val_freq : 5000 |
| wandb_entity : benzlxs |
| project : VARSR |
| label_B : 1 |
| pg : 0.0 |
| pg0 : 4 |
| pgwp : 0.016666666666666666 |
| cmd : --depth=24 --bs=4 --ep=5 --fp16=1 --tblr=5e-5 --alng=1e-4 --wpe=0.01 --wandb_flag=True --fuse=1 --exp_name=VARDahazing0616 --opt=adamw |
| branch : master |
| commit_id : 80bd1f0d832951506610a3e82320a8322a6f1a7a |
| commit_msg : Update README.md |
| acc_mean : None |
| acc_tail : None |
| L_mean : None |
| L_tail : None |
| vacc_mean : None |
| vacc_tail : None |
| vL_mean : None |
| vL_tail : None |
| grad_norm : None |
| cur_lr : None |
| cur_wd : None |
| cur_it : |
| cur_ep : |
| remain_time : |
| finish_time : |
| local_out_dir_path : results/VARDahazing0616 |
| tb_log_dir_path : results/VARDahazing0616/tb-VARd24__pn1_2_3_4_6_9_13_18_24_32__b8ep5adamlr5e-05wd0.05 |
| log_txt_path : results/VARDahazing0616/log.txt |
| last_ckpt_path : results/VARDahazing0616/ar-ckpt-last.pth |
| tf32 : True |
| seed : None |
| same_seed_for_all_ranks: 0 |
| local_debug : False |
| dbg_nan : False |
| } |
|
|
| [06-18 20:22:44] (/projects/VARSR/train.py, line 97)=> [build dahazing paired dataset] ... |
|
|
| [06-18 20:23:08] (/VARSR/zrk_utils/data.py, line 22)=> [Dataset] len(dataset)=18130 |
| [06-18 20:23:08] (/VARSR/zrk_utils/data.py, line 22)=> [Dataset] len(dataset)=1135 |
| [06-18 20:23:08] (/projects/VARSR/train.py, line 146)=> [auto_resume] no ckpt found @ results/VARDahazing0616/ar-ckpt*.pth |
| [06-18 20:23:08] (/projects/VARSR/train.py, line 146)=> [auto_resume quit] |
| [06-18 20:23:08] (/projects/VARSR/train.py, line 147)=> [dataloader multi processing] ... [dataloader multi processing](*) finished! (0.00s) |
| [06-18 20:23:08] (/projects/VARSR/train.py, line 153)=> [dataloader] gbs=8, lbs=4, iters_train=2267, types(tr, va)=('PairedImageDataset_Dehazing', 'PairedImageDataset_Dehazing') |
| [06-18 20:23:09] (R/models/var4dehazing.py, line 183)=> [VAR config ] embed_dim=1536, num_heads=24, depth=24, mlp_ratio=4.0 |
| [drop ratios ] drop_rate=0.0, attn_drop_rate=0.0, drop_path_rate=0.1 (tensor([0.0000, 0.0043, 0.0087, 0.0130, 0.0174, 0.0217, 0.0261, 0.0304, 0.0348, |
| 0.0391, 0.0435, 0.0478, 0.0522, 0.0565, 0.0609, 0.0652, 0.0696, 0.0739, |
| 0.0783, 0.0826, 0.0870, 0.0913, 0.0957, 0.1000])) |
|
|
| [06-18 20:23:09] (R/models/var4dehazing.py, line 207)=> torch.Size([1, 1, 3264, 3264]) |
| [06-18 20:23:10] (R/models/var4dehazing.py, line 504)=> [init_weights] VAR_dehazing with init_std=0.0147314 |
| [06-18 20:23:13] (/projects/VARSR/train.py, line 53)=> pos_start |
| [06-18 20:23:13] (/projects/VARSR/train.py, line 53)=> lvl_1L |
| [06-18 20:23:13] (/projects/VARSR/train.py, line 53)=> attn_bias_for_masking |
| [06-18 20:23:14] (/projects/VARSR/train.py, line 57)=> [VARTrainer.load_state_dict] missing: ['pos_start', 'lvl_1L', 'attn_bias_for_masking', 'con_embedding.conv_in.weight', 'con_embedding.conv_in.bias', 'con_embedding.blocks.0.weight', 'con_embedding.blocks.0.bias', 'con_embedding.blocks.1.weight', 'con_embedding.blocks.1.bias', 'con_embedding.blocks.2.weight', 'con_embedding.blocks.2.bias', 'con_embedding.blocks.3.weight', 'con_embedding.blocks.3.bias', 'con_embedding.blocks.4.weight', 'con_embedding.blocks.4.bias', 'con_embedding.blocks.5.weight', 'con_embedding.blocks.5.bias', 'con_embedding.blocks.6.weight', 'con_embedding.blocks.6.bias', 'con_embedding.blocks.7.weight', 'con_embedding.blocks.7.bias', 'con_embedding.conv_out.weight', 'con_embedding.conv_out.bias', 'decoder_norm.weight', 'decoder_norm.bias', 'diffloss.net.time_embed.mlp.0.weight', 'diffloss.net.time_embed.mlp.0.bias', 'diffloss.net.time_embed.mlp.2.weight', 'diffloss.net.time_embed.mlp.2.bias', 'diffloss.net.cond_embed.weight', 'diffloss.net.cond_embed.bias', 'diffloss.net.input_proj.weight', 'diffloss.net.input_proj.bias', 'diffloss.net.res_blocks.0.in_ln.weight', 'diffloss.net.res_blocks.0.in_ln.bias', 'diffloss.net.res_blocks.0.mlp.0.weight', 'diffloss.net.res_blocks.0.mlp.0.bias', 'diffloss.net.res_blocks.0.mlp.2.weight', 'diffloss.net.res_blocks.0.mlp.2.bias', 'diffloss.net.res_blocks.0.adaLN_modulation.1.weight', 'diffloss.net.res_blocks.0.adaLN_modulation.1.bias', 'diffloss.net.res_blocks.1.in_ln.weight', 'diffloss.net.res_blocks.1.in_ln.bias', 'diffloss.net.res_blocks.1.mlp.0.weight', 'diffloss.net.res_blocks.1.mlp.0.bias', 'diffloss.net.res_blocks.1.mlp.2.weight', 'diffloss.net.res_blocks.1.mlp.2.bias', 'diffloss.net.res_blocks.1.adaLN_modulation.1.weight', 'diffloss.net.res_blocks.1.adaLN_modulation.1.bias', 'diffloss.net.res_blocks.2.in_ln.weight', 'diffloss.net.res_blocks.2.in_ln.bias', 'diffloss.net.res_blocks.2.mlp.0.weight', 'diffloss.net.res_blocks.2.mlp.0.bias', 'diffloss.net.res_blocks.2.mlp.2.weight', 'diffloss.net.res_blocks.2.mlp.2.bias', 'diffloss.net.res_blocks.2.adaLN_modulation.1.weight', 'diffloss.net.res_blocks.2.adaLN_modulation.1.bias', 'diffloss.net.res_blocks.3.in_ln.weight', 'diffloss.net.res_blocks.3.in_ln.bias', 'diffloss.net.res_blocks.3.mlp.0.weight', 'diffloss.net.res_blocks.3.mlp.0.bias', 'diffloss.net.res_blocks.3.mlp.2.weight', 'diffloss.net.res_blocks.3.mlp.2.bias', 'diffloss.net.res_blocks.3.adaLN_modulation.1.weight', 'diffloss.net.res_blocks.3.adaLN_modulation.1.bias', 'diffloss.net.res_blocks.4.in_ln.weight', 'diffloss.net.res_blocks.4.in_ln.bias', 'diffloss.net.res_blocks.4.mlp.0.weight', 'diffloss.net.res_blocks.4.mlp.0.bias', 'diffloss.net.res_blocks.4.mlp.2.weight', 'diffloss.net.res_blocks.4.mlp.2.bias', 'diffloss.net.res_blocks.4.adaLN_modulation.1.weight', 'diffloss.net.res_blocks.4.adaLN_modulation.1.bias', 'diffloss.net.res_blocks.5.in_ln.weight', 'diffloss.net.res_blocks.5.in_ln.bias', 'diffloss.net.res_blocks.5.mlp.0.weight', 'diffloss.net.res_blocks.5.mlp.0.bias', 'diffloss.net.res_blocks.5.mlp.2.weight', 'diffloss.net.res_blocks.5.mlp.2.bias', 'diffloss.net.res_blocks.5.adaLN_modulation.1.weight', 'diffloss.net.res_blocks.5.adaLN_modulation.1.bias', 'diffloss.net.final_layer.linear.weight', 'diffloss.net.final_layer.linear.bias', 'diffloss.net.final_layer.adaLN_modulation.1.weight', 'diffloss.net.final_layer.adaLN_modulation.1.bias'] |
| [06-18 20:23:14] (/projects/VARSR/train.py, line 58)=> [VARTrainer.load_state_dict] unexpected: ['pos_1LC'] |
| [06-18 20:23:14] (/projects/VARSR/train.py, line 193)=> [INIT] VAR model = VAR_dehazing( |
| drop_path_rate=0.1 |
| (con_embedding): ControlNetConditioningEmbedding( |
| (conv_in): Conv2d(3, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) |
| (blocks): ModuleList( |
| (0): Conv2d(32, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) |
| (1): Conv2d(32, 128, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1)) |
| (2): Conv2d(128, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) |
| (3): Conv2d(128, 256, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1)) |
| (4): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) |
| (5): Conv2d(256, 512, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1)) |
| (6): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) |
| (7): Conv2d(512, 1536, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1)) |
| ) |
| (conv_out): Conv2d(1536, 1536, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) |
| ) |
| (word_embed): Linear(in_features=32, out_features=1536, bias=True) |
| (class_emb): Embedding(2, 1536) |
| (lvl_embed): Embedding(10, 1536) |
| (shared_ada_lin): Identity() |
| (blocks): ModuleList( |
| (0): AdaLNSelfAttn_RoPE( |
| shared_aln=False |
| (drop_path): Identity() |
| (attn): SelfAttention_RoPE( |
| using_flash=False, using_xform=False, attn_l2_norm=True |
| (mat_qkv): Linear(in_features=1536, out_features=4608, bias=False) |
| (proj): Linear(in_features=1536, out_features=1536, bias=True) |
| (proj_drop): Identity() |
| ) |
| (ffn): FFN( |
| fused_mlp_func=False |
| (fc1): Linear(in_features=1536, out_features=6144, bias=True) |
| (act): GELU(approximate='tanh') |
| (fc2): Linear(in_features=6144, out_features=1536, bias=True) |
| (drop): Identity() |
| ) |
| (ln_wo_grad): LayerNorm((1536,), eps=1e-06, elementwise_affine=False) |
| (ada_lin): Sequential( |
| (0): SiLU() |
| (1): Linear(in_features=1536, out_features=9216, bias=True) |
| ) |
| ) |
| (1-23): 23 x AdaLNSelfAttn_RoPE( |
| shared_aln=False |
| (drop_path): DropPath((drop_prob=...)) |
| (attn): SelfAttention_RoPE( |
| using_flash=False, using_xform=False, attn_l2_norm=True |
| (mat_qkv): Linear(in_features=1536, out_features=4608, bias=False) |
| (proj): Linear(in_features=1536, out_features=1536, bias=True) |
| (proj_drop): Identity() |
| ) |
| (ffn): FFN( |
| fused_mlp_func=False |
| (fc1): Linear(in_features=1536, out_features=6144, bias=True) |
| (act): GELU(approximate='tanh') |
| (fc2): Linear(in_features=6144, out_features=1536, bias=True) |
| (drop): Identity() |
| ) |
| (ln_wo_grad): LayerNorm((1536,), eps=1e-06, elementwise_affine=False) |
| (ada_lin): Sequential( |
| (0): SiLU() |
| (1): Linear(in_features=1536, out_features=9216, bias=True) |
| ) |
| ) |
| ) |
| (head_nm): AdaLNBeforeHead( |
| (ln_wo_grad): LayerNorm((1536,), eps=1e-06, elementwise_affine=False) |
| (ada_lin): Sequential( |
| (0): SiLU() |
| (1): Linear(in_features=1536, out_features=3072, bias=True) |
| ) |
| ) |
| (head): Linear(in_features=1536, out_features=4096, bias=True) |
| (decoder_norm): LayerNorm((1536,), eps=1e-06, elementwise_affine=True) |
| (diffloss): DiffLoss( |
| (net): SimpleMLPAdaLN( |
| (time_embed): TimestepEmbedder( |
| (mlp): Sequential( |
| (0): Linear(in_features=256, out_features=1024, bias=True) |
| (1): SiLU() |
| (2): Linear(in_features=1024, out_features=1024, bias=True) |
| ) |
| ) |
| (cond_embed): Linear(in_features=1536, out_features=1024, bias=True) |
| (input_proj): Linear(in_features=32, out_features=1024, bias=True) |
| (res_blocks): ModuleList( |
| (0-5): 6 x ResBlock( |
| (in_ln): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) |
| (mlp): Sequential( |
| (0): Linear(in_features=1024, out_features=1024, bias=True) |
| (1): SiLU() |
| (2): Linear(in_features=1024, out_features=1024, bias=True) |
| ) |
| (adaLN_modulation): Sequential( |
| (0): SiLU() |
| (1): Linear(in_features=1024, out_features=3072, bias=True) |
| ) |
| ) |
| ) |
| (final_layer): FinalLayer( |
| (norm_final): LayerNorm((1024,), eps=1e-06, elementwise_affine=False) |
| (linear): Linear(in_features=1024, out_features=64, bias=True) |
| (adaLN_modulation): Sequential( |
| (0): SiLU() |
| (1): Linear(in_features=1024, out_features=2048, bias=True) |
| ) |
| ) |
| ) |
| ) |
| ) |
|
|
|
|
| [06-18 20:23:14] (/projects/VARSR/train.py, line 195)=> [INIT][#para] VAE=108.95, VAE.enc=44.11, VAE.dec=64.65, VAE.quant=0.17 |
| [06-18 20:23:14] (/projects/VARSR/train.py, line 196)=> [INIT][#para] VAR=1101.92 |
|
|
|
|
| [06-18 20:23:14] (ARSR/utils/lr_control.py, line 99)=> [get_param_groups] param_groups = |
| { 'D': { 'lr_sc': 1.0, |
| 'params': "('con_embedding.conv_in.weight, con_embedding.blocks.0.weight, con_embedding.blocks.1.weight, con_embedding.blocks.2.weight, con_embedding.blocks.3.weight, con_embedding.blocks.4.weight, '\n" |
| " 'con_embedding.blocks.5.weight, con_embedding.blocks.6.weight, con_embedding.blocks.7.weight, con_embedding.conv_out.weight, word_embed.weight, class_emb.weight, blocks.0.attn.mat_qkv.weight, '\n" |
| " 'blocks.0.attn.proj.weight, blocks.0.ffn.fc1.weight, blocks.0.ffn.fc2.weight, blocks.0.ada_lin.1.weight, blocks.1.attn.mat_qkv.weight, blocks.1.attn.proj.weight, blocks.1.ffn.fc1.weight, '\n" |
| " 'blocks.1.ffn.fc2.weight, blocks.1.ada_lin.1.weight, blocks.2.attn.mat_qkv.weight, blocks.2.attn.proj.weight, blocks.2.ffn.fc1.weight, blocks.2.ffn.fc2.weight, blocks.2.ada_lin.1.weight, '\n" |
| " 'blocks.3.attn.mat_qkv.weight, blocks.3.attn.proj.weight, blocks.3.ffn.fc1.weight, blocks.3.ffn.fc2.weight, blocks.3.ada_lin.1.weight, blocks.4.attn.mat_qkv.weight, blocks.4.attn.proj.weight, '\n" |
| " 'blocks.4.ffn.fc1.weight, blocks.4.ffn.fc2.weight, blocks.4.ada_lin.1.weight, blocks.5.attn.mat_qkv.weight, blocks.5.attn.proj.weight, blocks.5.ffn.fc1.weight, blocks.5.ffn.fc2.weight, '\n" |
| " 'blocks.5.ada_lin.1.weight, blocks.6.attn.mat_qkv.weight, blocks.6.attn.proj.weight, blocks.6.ffn.fc1.weight, blocks.6.ffn.fc2.weight, blocks.6.ada_lin.1.weight, blocks.7.attn.mat_qkv.weight, '\n" |
| " 'blocks.7.attn.proj.weight, blocks.7.ffn.fc1.weight, blocks.7.ffn.fc2.weight, blocks.7.ada_lin.1.weight, blocks.8.attn.mat_qkv.weight, blocks.8.attn.proj.weight, blocks.8.ffn.fc1.weight, '\n" |
| " 'blocks.8.ffn.fc2.weight, blocks.8.ada_lin.1.weight, blocks.9.attn.mat_qkv.weight, blocks.9.attn.proj.weight, blocks.9.ffn.fc1.weight, blocks.9.ffn.fc2.weight, blocks.9.ada_lin.1.weight, '\n" |
| " 'blocks.10.attn.mat_qkv.weight, blocks.10.attn.proj.weight, blocks.10.ffn.fc1.weight, blocks.10.ffn.fc2.weight, blocks.10.ada_lin.1.weight, blocks.11.attn.mat_qkv.weight, '\n" |
| " 'blocks.11.attn.proj.weight, blocks.11.ffn.fc1.weight, blocks.11.ffn.fc2.weight, blocks.11.ada_lin.1.weight, blocks.12.attn.mat_qkv.weight, blocks.12.attn.proj.weight, blocks.12.ffn.fc1.weight, '\n" |
| " 'blocks.12.ffn.fc2.weight, blocks.12.ada_lin.1.weight, blocks.13.attn.mat_qkv.weight, blocks.13.attn.proj.weight, blocks.13.ffn.fc1.weight, blocks.13.ffn.fc2.weight, blocks.13.ada_lin.1.weight, '\n" |
| " 'blocks.14.attn.mat_qkv.weight, blocks.14.attn.proj.weight, blocks.14.ffn.fc1.weight, blocks.14.ffn.fc2.weight, blocks.14.ada_lin.1.weight, blocks.15.attn.mat_qkv.weight, '\n" |
| " 'blocks.15.attn.proj.weight, blocks.15.ffn.fc1.weight, blocks.15.ffn.fc2.weight, blocks.15.ada_lin.1.weight, blocks.16.attn.mat_qkv.weight, blocks.16.attn.proj.weight, blocks.16.ffn.fc1.weight, '\n" |
| " 'blocks.16.ffn.fc2.weight, blocks.16.ada_lin.1.weight, blocks.17.attn.mat_qkv.weight, blocks.17.attn.proj.weight, blocks.17.ffn.fc1.weight, blocks.17.ffn.fc2.weight, blocks.17.ada_lin.1.weight, '\n" |
| " 'blocks.18.attn.mat_qkv.weight, blocks.18.attn.proj.weight, blocks.18.ffn.fc1.weight, blocks.18.ffn.fc2.weight, blocks.18.ada_lin.1.weight, blocks.19.attn.mat_qkv.weight, '\n" |
| " 'blocks.19.attn.proj.weight, blocks.19.ffn.fc1.weight, blocks.19.ffn.fc2.weight, blocks.19.ada_lin.1.weight, blocks.20.attn.mat_qkv.weight, blocks.20.attn.proj.weight, blocks.20.ffn.fc1.weight, '\n" |
| " 'blocks.20.ffn.fc2.weight, blocks.20.ada_lin.1.weight, blocks.21.attn.mat_qkv.weight, blocks.21.attn.proj.weight, blocks.21.ffn.fc1.weight, blocks.21.ffn.fc2.weight, blocks.21.ada_lin.1.weight, '\n" |
| " 'blocks.22.attn.mat_qkv.weight, blocks.22.attn.proj.weight, blocks.22.ffn.fc1.weight, blocks.22.ffn.fc2.weight, blocks.22.ada_lin.1.weight, blocks.23.attn.mat_qkv.weight, '\n" |
| " 'blocks.23.attn.proj.weight, blocks.23.ffn.fc1.weight, blocks.23.ffn.fc2.weight, blocks.23.ada_lin.1.weight, head_nm.ada_lin.1.weight, head.weight')", |
| 'wd_sc': 1.0}, |
| 'ND': { 'lr_sc': 1.0, |
| 'params': "('pos_start, con_embedding.conv_in.bias, con_embedding.blocks.0.bias, con_embedding.blocks.1.bias, con_embedding.blocks.2.bias, con_embedding.blocks.3.bias, con_embedding.blocks.4.bias, '\n" |
| " 'con_embedding.blocks.5.bias, con_embedding.blocks.6.bias, con_embedding.blocks.7.bias, con_embedding.conv_out.bias, word_embed.bias, lvl_embed.weight, blocks.0.attn.scale_mul_1H11, '\n" |
| " 'blocks.0.attn.q_bias, blocks.0.attn.v_bias, blocks.0.attn.proj.bias, blocks.0.ffn.fc1.bias, blocks.0.ffn.fc2.bias, blocks.0.ada_lin.1.bias, blocks.1.attn.scale_mul_1H11, blocks.1.attn.q_bias, '\n" |
| " 'blocks.1.attn.v_bias, blocks.1.attn.proj.bias, blocks.1.ffn.fc1.bias, blocks.1.ffn.fc2.bias, blocks.1.ada_lin.1.bias, blocks.2.attn.scale_mul_1H11, blocks.2.attn.q_bias, blocks.2.attn.v_bias, '\n" |
| " 'blocks.2.attn.proj.bias, blocks.2.ffn.fc1.bias, blocks.2.ffn.fc2.bias, blocks.2.ada_lin.1.bias, blocks.3.attn.scale_mul_1H11, blocks.3.attn.q_bias, blocks.3.attn.v_bias, blocks.3.attn.proj.bias, '\n" |
| " 'blocks.3.ffn.fc1.bias, blocks.3.ffn.fc2.bias, blocks.3.ada_lin.1.bias, blocks.4.attn.scale_mul_1H11, blocks.4.attn.q_bias, blocks.4.attn.v_bias, blocks.4.attn.proj.bias, blocks.4.ffn.fc1.bias, '\n" |
| " 'blocks.4.ffn.fc2.bias, blocks.4.ada_lin.1.bias, blocks.5.attn.scale_mul_1H11, blocks.5.attn.q_bias, blocks.5.attn.v_bias, blocks.5.attn.proj.bias, blocks.5.ffn.fc1.bias, blocks.5.ffn.fc2.bias, '\n" |
| " 'blocks.5.ada_lin.1.bias, blocks.6.attn.scale_mul_1H11, blocks.6.attn.q_bias, blocks.6.attn.v_bias, blocks.6.attn.proj.bias, blocks.6.ffn.fc1.bias, blocks.6.ffn.fc2.bias, blocks.6.ada_lin.1.bias, '\n" |
| " 'blocks.7.attn.scale_mul_1H11, blocks.7.attn.q_bias, blocks.7.attn.v_bias, blocks.7.attn.proj.bias, blocks.7.ffn.fc1.bias, blocks.7.ffn.fc2.bias, blocks.7.ada_lin.1.bias, '\n" |
| " 'blocks.8.attn.scale_mul_1H11, blocks.8.attn.q_bias, blocks.8.attn.v_bias, blocks.8.attn.proj.bias, blocks.8.ffn.fc1.bias, blocks.8.ffn.fc2.bias, blocks.8.ada_lin.1.bias, '\n" |
| " 'blocks.9.attn.scale_mul_1H11, blocks.9.attn.q_bias, blocks.9.attn.v_bias, blocks.9.attn.proj.bias, blocks.9.ffn.fc1.bias, blocks.9.ffn.fc2.bias, blocks.9.ada_lin.1.bias, '\n" |
| " 'blocks.10.attn.scale_mul_1H11, blocks.10.attn.q_bias, blocks.10.attn.v_bias, blocks.10.attn.proj.bias, blocks.10.ffn.fc1.bias, blocks.10.ffn.fc2.bias, blocks.10.ada_lin.1.bias, '\n" |
| " 'blocks.11.attn.scale_mul_1H11, blocks.11.attn.q_bias, blocks.11.attn.v_bias, blocks.11.attn.proj.bias, blocks.11.ffn.fc1.bias, blocks.11.ffn.fc2.bias, blocks.11.ada_lin.1.bias, '\n" |
| " 'blocks.12.attn.scale_mul_1H11, blocks.12.attn.q_bias, blocks.12.attn.v_bias, blocks.12.attn.proj.bias, blocks.12.ffn.fc1.bias, blocks.12.ffn.fc2.bias, blocks.12.ada_lin.1.bias, '\n" |
| " 'blocks.13.attn.scale_mul_1H11, blocks.13.attn.q_bias, blocks.13.attn.v_bias, blocks.13.attn.proj.bias, blocks.13.ffn.fc1.bias, blocks.13.ffn.fc2.bias, blocks.13.ada_lin.1.bias, '\n" |
| " 'blocks.14.attn.scale_mul_1H11, blocks.14.attn.q_bias, blocks.14.attn.v_bias, blocks.14.attn.proj.bias, blocks.14.ffn.fc1.bias, blocks.14.ffn.fc2.bias, blocks.14.ada_lin.1.bias, '\n" |
| " 'blocks.15.attn.scale_mul_1H11, blocks.15.attn.q_bias, blocks.15.attn.v_bias, blocks.15.attn.proj.bias, blocks.15.ffn.fc1.bias, blocks.15.ffn.fc2.bias, blocks.15.ada_lin.1.bias, '\n" |
| " 'blocks.16.attn.scale_mul_1H11, blocks.16.attn.q_bias, blocks.16.attn.v_bias, blocks.16.attn.proj.bias, blocks.16.ffn.fc1.bias, blocks.16.ffn.fc2.bias, blocks.16.ada_lin.1.bias, '\n" |
| " 'blocks.17.attn.scale_mul_1H11, blocks.17.attn.q_bias, blocks.17.attn.v_bias, blocks.17.attn.proj.bias, blocks.17.ffn.fc1.bias, blocks.17.ffn.fc2.bias, blocks.17.ada_lin.1.bias, '\n" |
| " 'blocks.18.attn.scale_mul_1H11, blocks.18.attn.q_bias, blocks.18.attn.v_bias, blocks.18.attn.proj.bias, blocks.18.ffn.fc1.bias, blocks.18.ffn.fc2.bias, blocks.18.ada_lin.1.bias, '\n" |
| " 'blocks.19.attn.scale_mul_1H11, blocks.19.attn.q_bias, blocks.19.attn.v_bias, blocks.19.attn.proj.bias, blocks.19.ffn.fc1.bias, blocks.19.ffn.fc2.bias, blocks.19.ada_lin.1.bias, '\n" |
| " 'blocks.20.attn.scale_mul_1H11, blocks.20.attn.q_bias, blocks.20.attn.v_bias, blocks.20.attn.proj.bias, blocks.20.ffn.fc1.bias, blocks.20.ffn.fc2.bias, blocks.20.ada_lin.1.bias, '\n" |
| " 'blocks.21.attn.scale_mul_1H11, blocks.21.attn.q_bias, blocks.21.attn.v_bias, blocks.21.attn.proj.bias, blocks.21.ffn.fc1.bias, blocks.21.ffn.fc2.bias, blocks.21.ada_lin.1.bias, '\n" |
| " 'blocks.22.attn.scale_mul_1H11, blocks.22.attn.q_bias, blocks.22.attn.v_bias, blocks.22.attn.proj.bias, blocks.22.ffn.fc1.bias, blocks.22.ffn.fc2.bias, blocks.22.ada_lin.1.bias, '\n" |
| " 'blocks.23.attn.scale_mul_1H11, blocks.23.attn.q_bias, blocks.23.attn.v_bias, blocks.23.attn.proj.bias, blocks.23.ffn.fc1.bias, blocks.23.ffn.fc2.bias, blocks.23.ada_lin.1.bias, '\n" |
| " 'head_nm.ada_lin.1.bias, head.bias, decoder_norm.weight, decoder_norm.bias, diffloss.net.time_embed.mlp.0.weight, diffloss.net.time_embed.mlp.0.bias, diffloss.net.time_embed.mlp.2.weight, '\n" |
| " 'diffloss.net.time_embed.mlp.2.bias, diffloss.net.cond_embed.weight, diffloss.net.cond_embed.bias, diffloss.net.input_proj.weight, diffloss.net.input_proj.bias, '\n" |
| " 'diffloss.net.res_blocks.0.in_ln.weight, diffloss.net.res_blocks.0.in_ln.bias, diffloss.net.res_blocks.0.mlp.0.weight, diffloss.net.res_blocks.0.mlp.0.bias, diffloss.net.res_blocks.0.mlp.2.weight, '\n" |
| " 'diffloss.net.res_blocks.0.mlp.2.bias, diffloss.net.res_blocks.0.adaLN_modulation.1.weight, diffloss.net.res_blocks.0.adaLN_modulation.1.bias, diffloss.net.res_blocks.1.in_ln.weight, '\n" |
| " 'diffloss.net.res_blocks.1.in_ln.bias, diffloss.net.res_blocks.1.mlp.0.weight, diffloss.net.res_blocks.1.mlp.0.bias, diffloss.net.res_blocks.1.mlp.2.weight, diffloss.net.res_blocks.1.mlp.2.bias, '\n" |
| " 'diffloss.net.res_blocks.1.adaLN_modulation.1.weight, diffloss.net.res_blocks.1.adaLN_modulation.1.bias, diffloss.net.res_blocks.2.in_ln.weight, diffloss.net.res_blocks.2.in_ln.bias, '\n" |
| " 'diffloss.net.res_blocks.2.mlp.0.weight, diffloss.net.res_blocks.2.mlp.0.bias, diffloss.net.res_blocks.2.mlp.2.weight, diffloss.net.res_blocks.2.mlp.2.bias, '\n" |
| " 'diffloss.net.res_blocks.2.adaLN_modulation.1.weight, diffloss.net.res_blocks.2.adaLN_modulation.1.bias, diffloss.net.res_blocks.3.in_ln.weight, diffloss.net.res_blocks.3.in_ln.bias, '\n" |
| " 'diffloss.net.res_blocks.3.mlp.0.weight, diffloss.net.res_blocks.3.mlp.0.bias, diffloss.net.res_blocks.3.mlp.2.weight, diffloss.net.res_blocks.3.mlp.2.bias, '\n" |
| " 'diffloss.net.res_blocks.3.adaLN_modulation.1.weight, diffloss.net.res_blocks.3.adaLN_modulation.1.bias, diffloss.net.res_blocks.4.in_ln.weight, diffloss.net.res_blocks.4.in_ln.bias, '\n" |
| " 'diffloss.net.res_blocks.4.mlp.0.weight, diffloss.net.res_blocks.4.mlp.0.bias, diffloss.net.res_blocks.4.mlp.2.weight, diffloss.net.res_blocks.4.mlp.2.bias, '\n" |
| " 'diffloss.net.res_blocks.4.adaLN_modulation.1.weight, diffloss.net.res_blocks.4.adaLN_modulation.1.bias, diffloss.net.res_blocks.5.in_ln.weight, diffloss.net.res_blocks.5.in_ln.bias, '\n" |
| " 'diffloss.net.res_blocks.5.mlp.0.weight, diffloss.net.res_blocks.5.mlp.0.bias, diffloss.net.res_blocks.5.mlp.2.weight, diffloss.net.res_blocks.5.mlp.2.bias, '\n" |
| " 'diffloss.net.res_blocks.5.adaLN_modulation.1.weight, diffloss.net.res_blocks.5.adaLN_modulation.1.bias, diffloss.net.final_layer.linear.weight, diffloss.net.final_layer.linear.bias, '\n" |
| " 'diffloss.net.final_layer.adaLN_modulation.1.weight, diffloss.net.final_layer.adaLN_modulation.1.bias')", |
| 'wd_sc': 0.0}} |
|
|
| [06-18 20:23:14] (ARSR/utils/lr_control.py, line 104)=> [get_param_groups][rank0] type(model).__name__='VAR_dehazing' count=379, numel=1101916448 |
| [06-18 20:23:14] (ARSR/utils/lr_control.py, line 105)=> |
| [06-18 20:23:14] (/projects/VARSR/train.py, line 211)=> [INIT] optim=functools.partial(<class 'torch.optim.adamw.AdamW'>, betas=(0.9, 0.95), fused=True), opt_kw={'lr': 5e-05, 'weight_decay': 0} |
|
|
| [06-18 20:23:23] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 0/5] [ 0/2267] eta: 5:32:58 tlr: 2.5e-07 tnm: 29.13 Lm: 10.083 (10.083) Lt: 9.557 (9.557) Accm: 0.04 (0.04) Acct: 0.10 (0.10) time: 8.8129 data: 0.7686 |
| [06-18 20:58:02] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 0/5] [ 566/2267] eta: 1:44:24 tlr: 5e-05 tnm: 1.22 Lm: 8.801 (8.801) Lt: 8.717 (8.717) Accm: 0.56 (0.56) Acct: 0.13 (0.13) time: 3.6906 data: 0.8209 |
| [06-18 21:33:35] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 0/5] [1133/2267] eta: 1:10:20 tlr: 5e-05 tnm: 0.74 Lm: 7.528 (8.376) Lt: 7.878 (8.399) Accm: 0.74 (0.62) Acct: 0.17 (0.23) time: 3.8047 data: 0.9043 |
| [06-18 22:09:02] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 0/5] [1699/2267] eta: 0:35:20 tlr: 5e-05 tnm: 0.84 Lm: 7.523 (8.126) Lt: 7.820 (8.226) Accm: 0.90 (0.79) Acct: 0.29 (0.34) time: 3.6712 data: 0.7746 |
| [06-18 22:44:37] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 0/5] [2266/2267] eta: 0:00:03 tlr: 5e-05 tnm: 0.84 Lm: 7.518 (7.916) Lt: 7.761 (8.090) Accm: 1.07 (1.05) Acct: 0.42 (0.51) time: 3.7091 data: 0.8078 |
| [06-18 22:44:37] (/VARSR/zrk_utils/misc.py, line 344)=> [Ep]: [ 0/5] Total time: 2:21:22 (3.742 s / it) |
| [06-18 22:44:37] (/projects/VARSR/train.py, line 423)=> [ep0] (training ) Lm: 7.935 (7.935), Lt: 8.120 (8.120), Acc m&t: 1.00 0.47, Remain: 9:21:32, Finish: 2025-06-19 10:06 |
| [06-18 22:44:41] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 1/5] [ 0/2267] eta: 2:24:26 tlr: 5e-05 tnm: 0.90 Lm: 7.230 (7.230) Lt: 7.532 (7.532) Accm: 1.63 (1.63) Acct: 1.10 (1.10) time: 3.8231 data: 0.6894 |
| [06-18 23:19:33] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 1/5] [ 566/2267] eta: 1:44:48 tlr: 5e-05 tnm: 0.79 Lm: 7.185 (7.185) Lt: 7.553 (7.553) Accm: 1.72 (1.72) Acct: 0.94 (0.94) time: 3.7238 data: 0.7989 |
| [06-18 23:54:36] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 1/5] [1133/2267] eta: 1:09:59 tlr: 5e-05 tnm: 0.83 Lm: 7.139 (7.137) Lt: 7.532 (7.532) Accm: 1.81 (1.85) Acct: 1.00 (0.96) time: 3.6393 data: 0.7187 |
| [06-19 00:29:21] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 1/5] [1699/2267] eta: 0:34:59 tlr: 4.7e-05 tnm: 1.01 Lm: 7.091 (7.109) Lt: 7.512 (7.509) Accm: 1.95 (2.01) Acct: 1.05 (1.02) time: 3.6697 data: 0.7540 |
| [06-19 01:04:29] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 1/5] [2266/2267] eta: 0:00:03 tlr: 4.4e-05 tnm: 1.05 Lm: 7.042 (7.083) Lt: 7.491 (7.495) Accm: 2.10 (2.04) Acct: 1.00 (1.02) time: 3.6405 data: 0.7603 |
| [06-19 01:04:29] (/VARSR/zrk_utils/misc.py, line 344)=> [Ep]: [ 1/5] Total time: 2:19:52 (3.702 s / it) |
| [06-19 01:04:29] (/projects/VARSR/train.py, line 423)=> [ep1] (training ) Lm: 7.085 (7.085), Lt: 7.508 (7.508), Acc m&t: 1.97 0.95, Remain: 6:52:48, Finish: 2025-06-19 09:57 |
| [06-19 01:04:33] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 2/5] [ 0/2267] eta: 2:14:38 tlr: 4.4e-05 tnm: 0.98 Lm: 6.934 (6.934) Lt: 7.404 (7.404) Accm: 2.19 (2.19) Acct: 0.93 (0.93) time: 3.5634 data: 0.6454 |
| [06-19 01:39:45] (/projects/VARSR/train.py, line 363)=> [*] [ep2] (val 1135) Lm: 6.8316, Lt: 7.2320, Acc m&t: 2.54 1.54, Val cost: 419.09s |
| [06-19 01:39:45] (/projects/VARSR/train.py, line 367)=> [saving ckpt] ... [saving ckpt](*) finished! @ results/VARDahazing0616/ar-ckpt-best.pth |
| [06-19 01:46:22] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 2/5] [ 566/2267] eta: 2:05:39 tlr: 4e-05 tnm: 1.15 Lm: 6.759 (6.759) Lt: 7.316 (7.316) Accm: 2.84 (2.84) Acct: 1.14 (1.14) time: 3.7289 data: 0.8361 |
| [06-19 02:21:54] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 2/5] [1133/2267] eta: 1:17:25 tlr: 3.7e-05 tnm: 1.30 Lm: 6.585 (6.674) Lt: 7.242 (7.291) Accm: 3.36 (3.01) Acct: 1.27 (1.18) time: 3.7628 data: 0.8396 |
| [06-19 02:56:58] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 2/5] [1699/2267] eta: 0:37:34 tlr: 3.3e-05 tnm: 1.62 Lm: 6.544 (6.568) Lt: 7.235 (7.220) Accm: 3.43 (3.46) Acct: 1.31 (1.38) time: 3.6585 data: 0.7778 |
| [06-19 03:31:56] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 2/5] [2266/2267] eta: 0:00:03 tlr: 2.9e-05 tnm: 1.49 Lm: 6.502 (6.522) Lt: 7.227 (7.185) Accm: 3.49 (3.61) Acct: 1.34 (1.42) time: 3.6719 data: 0.7520 |
| [06-19 03:31:56] (/VARSR/zrk_utils/misc.py, line 344)=> [Ep]: [ 2/5] Total time: 2:27:26 (3.902 s / it) |
| [06-19 03:31:56] (/projects/VARSR/train.py, line 423)=> [ep2] (training ) Lm: 6.601 (6.601), Lt: 7.239 (7.239), Acc m&t: 3.31 1.28, Remain: 4:37:44, Finish: 2025-06-19 10:09 |
| [06-19 03:32:00] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 3/5] [ 0/2267] eta: 2:16:07 tlr: 2.9e-05 tnm: 1.68 Lm: 6.134 (6.134) Lt: 6.922 (6.922) Accm: 4.75 (4.75) Acct: 1.66 (1.66) time: 3.6028 data: 0.6484 |
| [06-19 04:06:19] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 3/5] [ 566/2267] eta: 1:43:09 tlr: 2.6e-05 tnm: 1.59 Lm: 5.999 (5.999) Lt: 6.788 (6.788) Accm: 5.33 (5.33) Acct: 2.10 (2.10) time: 3.6572 data: 0.7595 |
| [06-19 04:40:59] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 3/5] [1133/2267] eta: 1:09:03 tlr: 2.2e-05 tnm: 1.92 Lm: 5.877 (5.959) Lt: 6.744 (6.773) Accm: 5.60 (5.42) Acct: 2.29 (2.16) time: 3.8565 data: 0.7627 |
| [06-19 05:15:50] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 3/5] [1699/2267] eta: 0:34:42 tlr: 1.9e-05 tnm: 2.37 Lm: 5.871 (5.888) Lt: 6.699 (6.699) Accm: 5.76 (5.81) Acct: 2.42 (2.34) time: 3.6987 data: 0.7211 |
| [06-19 05:50:59] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 3/5] [2266/2267] eta: 0:00:03 tlr: 1.5e-05 tnm: 2.20 Lm: 5.865 (5.869) Lt: 6.655 (6.673) Accm: 5.92 (5.93) Acct: 2.54 (2.41) time: 3.7162 data: 0.8073 |
| [06-19 05:50:59] (/VARSR/zrk_utils/misc.py, line 344)=> [Ep]: [ 3/5] Total time: 2:19:02 (3.680 s / it) |
| [06-19 05:50:59] (/projects/VARSR/train.py, line 423)=> [ep3] (training ) Lm: 5.924 (5.924), Lt: 6.683 (6.683), Acc m&t: 5.82 2.52, Remain: 2:19:57, Finish: 2025-06-19 10:10 |
| [06-19 05:51:02] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 4/5] [ 0/2267] eta: 2:12:57 tlr: 1.5e-05 tnm: 1.97 Lm: 5.730 (5.730) Lt: 6.450 (6.450) Accm: 5.97 (5.97) Acct: 3.03 (3.03) time: 3.5190 data: 0.6160 |
| [06-19 06:25:29] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 4/5] [ 566/2267] eta: 1:43:30 tlr: 1.1e-05 tnm: 1.92 Lm: 5.768 (5.768) Lt: 6.413 (6.413) Accm: 5.73 (5.73) Acct: 3.09 (3.09) time: 3.6240 data: 0.6995 |
| [06-19 06:54:19] (/projects/VARSR/train.py, line 363)=> [*] [ep4] (val 1135) Lm: 6.1785, Lt: 6.6731, Acc m&t: 4.82 2.99, Val cost: 359.43s |
| [06-19 06:54:19] (/projects/VARSR/train.py, line 367)=> [saving ckpt] ... [saving ckpt](*) finished! @ results/VARDahazing0616/ar-ckpt-best.pth |
| [06-19 07:07:19] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 4/5] [1133/2267] eta: 1:16:19 tlr: 7.7e-06 tnm: 2.26 Lm: 5.730 (5.611) Lt: 6.377 (6.308) Accm: 5.97 (6.54) Acct: 3.15 (3.39) time: 3.6319 data: 0.7192 |
| [06-19 07:42:13] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 4/5] [1699/2267] eta: 0:37:09 tlr: 4.1e-06 tnm: 2.13 Lm: 5.603 (5.577) Lt: 6.250 (6.262) Accm: 6.48 (6.65) Acct: 3.48 (3.50) time: 3.7271 data: 0.8105 |
| [06-19 08:17:44] (/VARSR/zrk_utils/misc.py, line 322)=> [Ep]: [ 4/5] [2266/2267] eta: 0:00:03 tlr: 5e-07 tnm: 2.01 Lm: 5.477 (5.518) Lt: 6.123 (6.212) Accm: 6.99 (7.06) Acct: 3.81 (3.72) time: 3.6919 data: 0.7889 |
| [06-19 08:17:44] (/VARSR/zrk_utils/misc.py, line 344)=> [Ep]: [ 4/5] Total time: 2:26:44 (3.884 s / it) |
| [06-19 08:17:44] (/projects/VARSR/train.py, line 423)=> [ep4] (training ) Lm: 5.504 (5.504), Lt: 6.209 (6.209), Acc m&t: 7.24 3.73, Remain: 0:00:55, Finish: 2025-06-19 10:18 |
| [06-19 08:17:44] (/projects/VARSR/train.py, line 427)=> |
|
|
|
|
| [06-19 08:17:44] (/projects/VARSR/train.py, line 428)=> [*] [PT finished] Total cost: 11.9h, Lm: 5.504 (5.503819465637207), Lt: 6.209 (6.209009504318237) |
| [06-19 08:17:44] (/projects/VARSR/train.py, line 429)=> |
|
|
|
|
| [06-19 08:17:50] (/projects/VARSR/train.py, line 436)=> final args: |
|
|
| { |
| vae_model_path : checkpoints/VQVAE.pth |
| var_pretrain_path : checkpoints/checkpoints_VAR/var_d24.pth |
| wandb_flag : True |
| exp_name : VARDahazing0616 |
| vfast : 0 |
| tfast : 0 |
| depth : 24 |
| ini : -1 |
| hd : 0.02 |
| aln : 0.5 |
| alng : 0.0001 |
| fp16 : 1 |
| tblr : 5e-05 |
| tlr : 5e-05 |
| twd : 0.05 |
| twde : 0.05 |
| tclip : 2.0 |
| ls : 0.0 |
| bs : 8 |
| batch_size : 4 |
| glb_batch_size : 8 |
| ac : 1 |
| ep : 5 |
| wp : 0.1 |
| wp0 : 0.005 |
| wpe : 0.01 |
| sche : lin0 |
| opt : adamw |
| afuse : True |
| saln : False |
| anorm : True |
| fuse : True |
| pn : 1_2_3_4_6_9_13_18_24_32 |
| patch_size : 16 |
| patch_nums : (1, 2, 3, 4, 6, 9, 13, 18, 24, 32) |
| resos : (16, 32, 48, 64, 96, 144, 208, 288, 384, 512) |
| data_load_reso : 512 |
| mid_reso : 1.125 |
| hflip : False |
| workers : 0 |
| dataset_opt_path : options/VARSR_Dehazing.yml |
| val_freq : 5000 |
| wandb_entity : benzlxs |
| project : VARSR |
| label_B : 1 |
| pg : 0.0 |
| pg0 : 4 |
| pgwp : 0.016666666666666666 |
| cmd : --depth=24 --bs=4 --ep=5 --fp16=1 --tblr=5e-5 --alng=1e-4 --wpe=0.01 --wandb_flag=True --fuse=1 --exp_name=VARDahazing0616 --opt=adamw |
| branch : master |
| commit_id : 80bd1f0d832951506610a3e82320a8322a6f1a7a |
| commit_msg : Update README.md |
| acc_mean : 7.236607149243355 |
| acc_tail : 3.73291015625 |
| L_mean : 5.503819465637207 |
| L_tail : 6.209009504318237 |
| vacc_mean : 4.817337989807129 |
| vacc_tail : 2.9948203563690186 |
| vL_mean : 6.178485870361328 |
| vL_tail : 6.6730804443359375 |
| grad_norm : 2.0592718839645388 |
| cur_lr : 5.000000000000001e-07 |
| cur_wd : 0.05 |
| cur_it : 2267/2267 |
| cur_ep : 5/5 |
| remain_time : - |
| finish_time : 2025-06-19 10:16 |
| local_out_dir_path : results/VARDahazing0616 |
| tb_log_dir_path : results/VARDahazing0616/tb-VARd24__pn1_2_3_4_6_9_13_18_24_32__b8ep5adamlr5e-05wd0.05 |
| log_txt_path : results/VARDahazing0616/log.txt |
| last_ckpt_path : results/VARDahazing0616/ar-ckpt-last.pth |
| tf32 : True |
| seed : None |
| same_seed_for_all_ranks: 0 |
| local_debug : False |
| dbg_nan : False |
| } |
|
|
|
|