Spaces:
Paused
Paused
| import torch | |
| def keys_in_state_dict(ckpt, device='cpu'): | |
| if device=="cpu": | |
| a = torch.load(ckpt, map_location=torch.device('cpu'))["state_dict"] | |
| else: | |
| a = torch.load(ckpt)["state_dict"] | |
| print("keys_in_state_dict", a.keys()) | |
| def check_ckpt_diff(ckpt_a, ckpt_b, key_include=None, key_exclude=None, device='cpu', verbose=True): | |
| if device=="cpu": | |
| a = torch.load(ckpt_a, map_location=torch.device('cpu'))["state_dict"] | |
| b = torch.load(ckpt_b, map_location=torch.device('cpu'))["state_dict"] | |
| else: | |
| a = torch.load(ckpt_a)["state_dict"] | |
| b = torch.load(ckpt_b)["state_dict"] | |
| a_sum = 0 | |
| b_sum = 0 | |
| difference_count = 0 | |
| for k in a.keys(): | |
| if key_include is not None and key_include not in k: | |
| continue | |
| if key_exclude is not None and key_exclude in k: | |
| continue | |
| if k in b.keys(): | |
| a_sum += torch.sum(a[k]) | |
| b_sum += torch.sum(b[k]) | |
| if verbose: | |
| if torch.sum(a[k]) != torch.sum(b[k]): | |
| print(f"key {k} is different") | |
| difference_count += 1 | |
| print("a_sum: ", a_sum) | |
| print("b_sum: ", b_sum) | |
| print("diff: ", a_sum - b_sum) | |
| if verbose: | |
| print("difference_count: ", difference_count) | |
| return bool(a_sum - b_sum) | |
| # Transformer no freeze: | |
| # check_ckpt_diff("/fsx/clap_logs/2022_09_11-19_37_08-model_PANN-14-lr_0.001-b_160-j_4-p_fp32/checkpoints/epoch_10.pt", "/fsx/clap_logs/2022_09_11-19_37_08-model_PANN-14-lr_0.001-b_160-j_4-p_fp32/checkpoints/epoch_100.pt", "text_branch.resblocks") | |
| check_ckpt_diff("/fsx/clap_logs/2022_09_29-23_42_40-model_PANN-14-lr_0.001-b_160-j_4-p_fp32/checkpoints/epoch_1.pt", | |
| "/fsx/clap_logs/2022_09_29-23_42_40-model_PANN-14-lr_0.001-b_160-j_4-p_fp32/checkpoints/epoch_2.pt", | |
| "text_branch.resblocks") | |
| # key module.text_branch.resblocks.0.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.0.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.0.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.0.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.0.ln_1.weight is different | |
| # key module.text_branch.resblocks.0.ln_1.bias is different | |
| # key module.text_branch.resblocks.0.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.0.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.0.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.0.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.0.ln_2.weight is different | |
| # key module.text_branch.resblocks.0.ln_2.bias is different | |
| # key module.text_branch.resblocks.1.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.1.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.1.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.1.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.1.ln_1.weight is different | |
| # key module.text_branch.resblocks.1.ln_1.bias is different | |
| # key module.text_branch.resblocks.1.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.1.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.1.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.1.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.1.ln_2.weight is different | |
| # key module.text_branch.resblocks.1.ln_2.bias is different | |
| # key module.text_branch.resblocks.2.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.2.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.2.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.2.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.2.ln_1.weight is different | |
| # key module.text_branch.resblocks.2.ln_1.bias is different | |
| # key module.text_branch.resblocks.2.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.2.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.2.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.2.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.2.ln_2.weight is different | |
| # key module.text_branch.resblocks.2.ln_2.bias is different | |
| # key module.text_branch.resblocks.3.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.3.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.3.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.3.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.3.ln_1.weight is different | |
| # key module.text_branch.resblocks.3.ln_1.bias is different | |
| # key module.text_branch.resblocks.3.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.3.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.3.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.3.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.3.ln_2.weight is different | |
| # key module.text_branch.resblocks.3.ln_2.bias is different | |
| # key module.text_branch.resblocks.4.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.4.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.4.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.4.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.4.ln_1.weight is different | |
| # key module.text_branch.resblocks.4.ln_1.bias is different | |
| # key module.text_branch.resblocks.4.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.4.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.4.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.4.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.4.ln_2.weight is different | |
| # key module.text_branch.resblocks.4.ln_2.bias is different | |
| # key module.text_branch.resblocks.5.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.5.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.5.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.5.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.5.ln_1.weight is different | |
| # key module.text_branch.resblocks.5.ln_1.bias is different | |
| # key module.text_branch.resblocks.5.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.5.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.5.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.5.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.5.ln_2.weight is different | |
| # key module.text_branch.resblocks.5.ln_2.bias is different | |
| # key module.text_branch.resblocks.6.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.6.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.6.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.6.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.6.ln_1.weight is different | |
| # key module.text_branch.resblocks.6.ln_1.bias is different | |
| # key module.text_branch.resblocks.6.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.6.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.6.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.6.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.6.ln_2.weight is different | |
| # key module.text_branch.resblocks.6.ln_2.bias is different | |
| # key module.text_branch.resblocks.7.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.7.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.7.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.7.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.7.ln_1.weight is different | |
| # key module.text_branch.resblocks.7.ln_1.bias is different | |
| # key module.text_branch.resblocks.7.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.7.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.7.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.7.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.7.ln_2.weight is different | |
| # key module.text_branch.resblocks.7.ln_2.bias is different | |
| # key module.text_branch.resblocks.8.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.8.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.8.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.8.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.8.ln_1.weight is different | |
| # key module.text_branch.resblocks.8.ln_1.bias is different | |
| # key module.text_branch.resblocks.8.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.8.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.8.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.8.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.8.ln_2.weight is different | |
| # key module.text_branch.resblocks.8.ln_2.bias is different | |
| # key module.text_branch.resblocks.9.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.9.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.9.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.9.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.9.ln_1.weight is different | |
| # key module.text_branch.resblocks.9.ln_1.bias is different | |
| # key module.text_branch.resblocks.9.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.9.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.9.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.9.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.9.ln_2.weight is different | |
| # key module.text_branch.resblocks.9.ln_2.bias is different | |
| # key module.text_branch.resblocks.10.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.10.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.10.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.10.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.10.ln_1.weight is different | |
| # key module.text_branch.resblocks.10.ln_1.bias is different | |
| # key module.text_branch.resblocks.10.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.10.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.10.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.10.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.10.ln_2.weight is different | |
| # key module.text_branch.resblocks.10.ln_2.bias is different | |
| # key module.text_branch.resblocks.11.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.11.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.11.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.11.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.11.ln_1.weight is different | |
| # key module.text_branch.resblocks.11.ln_1.bias is different | |
| # key module.text_branch.resblocks.11.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.11.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.11.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.11.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.11.ln_2.weight is different | |
| # key module.text_branch.resblocks.11.ln_2.bias is different | |
| # a_sum: tensor(12113.6445) | |
| # b_sum: tensor(9883.4424) | |
| # diff: tensor(2230.2021) | |
| # True | |
| # Transformer freeze: | |
| # check_ckpt_diff("/fsx/clap_logs/2022_09_16-18_55_10-model_PANN-14-lr_0.001-b_160-j_4-p_fp32/checkpoints/epoch_10.pt", "/fsx/clap_logs/2022_09_16-18_55_10-model_PANN-14-lr_0.001-b_160-j_4-p_fp32/checkpoints/epoch_100.pt", "text_branch.resblocks") | |
| # key module.text_branch.resblocks.0.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.0.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.0.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.0.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.0.ln_1.weight is different | |
| # key module.text_branch.resblocks.0.ln_1.bias is different | |
| # key module.text_branch.resblocks.0.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.0.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.0.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.0.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.0.ln_2.weight is different | |
| # key module.text_branch.resblocks.0.ln_2.bias is different | |
| # key module.text_branch.resblocks.1.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.1.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.1.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.1.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.1.ln_1.weight is different | |
| # key module.text_branch.resblocks.1.ln_1.bias is different | |
| # key module.text_branch.resblocks.1.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.1.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.1.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.1.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.1.ln_2.weight is different | |
| # key module.text_branch.resblocks.1.ln_2.bias is different | |
| # key module.text_branch.resblocks.2.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.2.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.2.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.2.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.2.ln_1.weight is different | |
| # key module.text_branch.resblocks.2.ln_1.bias is different | |
| # key module.text_branch.resblocks.2.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.2.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.2.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.2.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.2.ln_2.weight is different | |
| # key module.text_branch.resblocks.2.ln_2.bias is different | |
| # key module.text_branch.resblocks.3.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.3.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.3.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.3.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.3.ln_1.weight is different | |
| # key module.text_branch.resblocks.3.ln_1.bias is different | |
| # key module.text_branch.resblocks.3.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.3.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.3.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.3.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.3.ln_2.weight is different | |
| # key module.text_branch.resblocks.3.ln_2.bias is different | |
| # key module.text_branch.resblocks.4.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.4.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.4.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.4.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.4.ln_1.weight is different | |
| # key module.text_branch.resblocks.4.ln_1.bias is different | |
| # key module.text_branch.resblocks.4.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.4.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.4.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.4.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.4.ln_2.weight is different | |
| # key module.text_branch.resblocks.4.ln_2.bias is different | |
| # key module.text_branch.resblocks.5.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.5.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.5.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.5.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.5.ln_1.weight is different | |
| # key module.text_branch.resblocks.5.ln_1.bias is different | |
| # key module.text_branch.resblocks.5.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.5.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.5.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.5.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.5.ln_2.weight is different | |
| # key module.text_branch.resblocks.5.ln_2.bias is different | |
| # key module.text_branch.resblocks.6.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.6.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.6.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.6.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.6.ln_1.weight is different | |
| # key module.text_branch.resblocks.6.ln_1.bias is different | |
| # key module.text_branch.resblocks.6.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.6.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.6.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.6.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.6.ln_2.weight is different | |
| # key module.text_branch.resblocks.6.ln_2.bias is different | |
| # key module.text_branch.resblocks.7.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.7.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.7.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.7.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.7.ln_1.weight is different | |
| # key module.text_branch.resblocks.7.ln_1.bias is different | |
| # key module.text_branch.resblocks.7.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.7.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.7.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.7.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.7.ln_2.weight is different | |
| # key module.text_branch.resblocks.7.ln_2.bias is different | |
| # key module.text_branch.resblocks.8.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.8.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.8.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.8.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.8.ln_1.weight is different | |
| # key module.text_branch.resblocks.8.ln_1.bias is different | |
| # key module.text_branch.resblocks.8.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.8.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.8.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.8.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.8.ln_2.weight is different | |
| # key module.text_branch.resblocks.8.ln_2.bias is different | |
| # key module.text_branch.resblocks.9.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.9.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.9.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.9.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.9.ln_1.weight is different | |
| # key module.text_branch.resblocks.9.ln_1.bias is different | |
| # key module.text_branch.resblocks.9.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.9.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.9.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.9.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.9.ln_2.weight is different | |
| # key module.text_branch.resblocks.9.ln_2.bias is different | |
| # key module.text_branch.resblocks.10.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.10.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.10.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.10.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.10.ln_1.weight is different | |
| # key module.text_branch.resblocks.10.ln_1.bias is different | |
| # key module.text_branch.resblocks.10.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.10.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.10.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.10.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.10.ln_2.weight is different | |
| # key module.text_branch.resblocks.10.ln_2.bias is different | |
| # key module.text_branch.resblocks.11.attn.in_proj_weight is different | |
| # key module.text_branch.resblocks.11.attn.in_proj_bias is different | |
| # key module.text_branch.resblocks.11.attn.out_proj.weight is different | |
| # key module.text_branch.resblocks.11.attn.out_proj.bias is different | |
| # key module.text_branch.resblocks.11.ln_1.weight is different | |
| # key module.text_branch.resblocks.11.ln_1.bias is different | |
| # key module.text_branch.resblocks.11.mlp.c_fc.weight is different | |
| # key module.text_branch.resblocks.11.mlp.c_fc.bias is different | |
| # key module.text_branch.resblocks.11.mlp.c_proj.weight is different | |
| # key module.text_branch.resblocks.11.mlp.c_proj.bias is different | |
| # key module.text_branch.resblocks.11.ln_2.weight is different | |
| # key module.text_branch.resblocks.11.ln_2.bias is different | |
| # a_sum: tensor(12133.6348) | |
| # b_sum: tensor(10423.9521) | |
| # diff: tensor(1709.6826) | |
| # True | |
| # bert no freeze: | |
| # check_ckpt_diff("/fsx/clap_logs/2022_09_14-02_33_11-model_PANN-14-lr_0.0001-b_160-j_4-p_fp32/checkpoints/epoch_10.pt", "/fsx/clap_logs/2022_09_14-02_33_11-model_PANN-14-lr_0.0001-b_160-j_4-p_fp32/checkpoints/epoch_100.pt", "text_branch.encoder") | |
| # key module.text_branch.encoder.layer.0.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.0.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.0.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.0.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.0.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.0.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.0.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.0.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.0.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.0.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.0.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.0.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.0.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.0.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.0.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.0.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.1.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.1.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.1.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.1.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.1.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.1.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.1.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.1.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.1.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.1.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.1.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.1.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.1.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.1.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.1.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.1.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.2.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.2.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.2.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.2.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.2.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.2.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.2.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.2.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.2.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.2.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.2.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.2.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.2.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.2.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.2.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.2.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.3.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.3.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.3.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.3.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.3.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.3.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.3.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.3.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.3.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.3.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.3.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.3.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.3.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.3.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.3.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.3.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.4.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.4.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.4.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.4.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.4.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.4.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.4.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.4.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.4.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.4.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.4.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.4.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.4.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.4.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.4.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.4.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.5.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.5.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.5.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.5.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.5.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.5.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.5.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.5.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.5.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.5.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.5.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.5.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.5.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.5.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.5.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.5.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.6.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.6.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.6.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.6.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.6.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.6.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.6.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.6.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.6.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.6.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.6.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.6.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.6.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.6.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.6.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.6.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.7.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.7.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.7.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.7.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.7.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.7.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.7.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.7.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.7.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.7.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.7.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.7.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.7.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.7.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.7.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.7.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.8.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.8.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.8.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.8.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.8.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.8.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.8.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.8.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.8.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.8.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.8.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.8.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.8.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.8.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.8.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.8.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.9.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.9.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.9.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.9.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.9.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.9.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.9.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.9.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.9.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.9.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.9.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.9.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.9.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.9.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.9.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.9.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.10.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.10.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.10.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.10.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.10.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.10.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.10.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.10.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.10.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.10.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.10.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.10.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.10.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.10.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.10.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.10.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.11.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.11.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.11.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.11.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.11.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.11.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.11.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.11.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.11.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.11.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.11.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.11.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.11.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.11.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.11.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.11.output.LayerNorm.bias is different | |
| # a_sum: tensor(15185.1230) | |
| # b_sum: tensor(15576.5596) | |
| # diff: tensor(-391.4365) | |
| # True | |
| # bert freeze: | |
| # check_ckpt_diff("/fsx/clap_logs/2022_09_13-01_25_15-model_PANN-14-lr_0.0001-b_160-j_4-p_fp32/checkpoints/epoch_10.pt", "/fsx/clap_logs/2022_09_13-01_25_15-model_PANN-14-lr_0.0001-b_160-j_4-p_fp32/checkpoints/epoch_100.pt", "text_branch.encoder") | |
| # key module.text_branch.encoder.layer.0.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.0.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.0.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.0.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.0.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.0.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.0.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.0.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.0.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.0.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.0.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.0.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.0.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.0.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.0.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.0.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.1.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.1.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.1.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.1.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.1.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.1.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.1.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.1.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.1.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.1.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.1.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.1.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.1.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.1.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.1.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.1.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.2.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.2.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.2.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.2.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.2.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.2.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.2.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.2.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.2.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.2.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.2.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.2.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.2.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.2.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.2.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.2.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.3.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.3.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.3.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.3.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.3.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.3.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.3.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.3.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.3.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.3.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.3.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.3.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.3.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.3.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.3.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.3.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.4.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.4.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.4.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.4.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.4.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.4.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.4.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.4.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.4.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.4.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.4.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.4.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.4.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.4.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.4.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.4.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.5.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.5.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.5.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.5.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.5.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.5.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.5.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.5.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.5.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.5.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.5.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.5.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.5.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.5.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.5.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.5.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.6.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.6.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.6.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.6.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.6.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.6.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.6.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.6.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.6.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.6.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.6.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.6.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.6.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.6.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.6.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.6.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.7.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.7.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.7.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.7.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.7.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.7.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.7.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.7.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.7.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.7.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.7.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.7.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.7.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.7.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.7.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.7.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.8.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.8.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.8.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.8.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.8.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.8.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.8.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.8.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.8.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.8.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.8.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.8.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.8.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.8.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.8.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.8.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.9.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.9.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.9.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.9.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.9.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.9.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.9.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.9.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.9.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.9.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.9.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.9.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.9.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.9.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.9.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.9.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.10.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.10.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.10.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.10.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.10.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.10.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.10.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.10.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.10.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.10.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.10.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.10.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.10.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.10.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.10.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.10.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.11.attention.self.query.weight is different | |
| # key module.text_branch.encoder.layer.11.attention.self.query.bias is different | |
| # key module.text_branch.encoder.layer.11.attention.self.key.weight is different | |
| # key module.text_branch.encoder.layer.11.attention.self.key.bias is different | |
| # key module.text_branch.encoder.layer.11.attention.self.value.weight is different | |
| # key module.text_branch.encoder.layer.11.attention.self.value.bias is different | |
| # key module.text_branch.encoder.layer.11.attention.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.11.attention.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.11.attention.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.11.attention.output.LayerNorm.bias is different | |
| # key module.text_branch.encoder.layer.11.intermediate.dense.weight is different | |
| # key module.text_branch.encoder.layer.11.intermediate.dense.bias is different | |
| # key module.text_branch.encoder.layer.11.output.dense.weight is different | |
| # key module.text_branch.encoder.layer.11.output.dense.bias is different | |
| # key module.text_branch.encoder.layer.11.output.LayerNorm.weight is different | |
| # key module.text_branch.encoder.layer.11.output.LayerNorm.bias is different | |
| # a_sum: tensor(15078.6641) | |
| # b_sum: tensor(15540.0723) | |
| # diff: tensor(-461.4082) | |
| # True | |
| # linear_prob_text | |
| # check_ckpt_diff("/fsx/clap_logs/2022_09_15-02_05_29-linear_probemodel_PANN-14-lr_0.0001-b_512-j_4-p_fp32/checkpoints/pretrain_epoch_10_lp_epoch_50.pt", "/fsx/clap_logs/2022_09_15-02_05_29-linear_probemodel_PANN-14-lr_0.0001-b_512-j_4-p_fp32/checkpoints/pretrain_epoch_10_lp_epoch_100.pt", "text_branch.resblocks") | |
| # a_sum: tensor(12111.0244) | |
| # b_sum: tensor(12111.0244) | |
| # diff: tensor(0.) | |
| # linear_prob_audio | |
| # check_ckpt_diff("/fsx/clap_logs/2022_09_15-02_05_29-linear_probemodel_PANN-14-lr_0.0001-b_512-j_4-p_fp32/checkpoints/pretrain_epoch_10_lp_epoch_50.pt", "/fsx/clap_logs/2022_09_15-02_05_29-linear_probemodel_PANN-14-lr_0.0001-b_512-j_4-p_fp32/checkpoints/pretrain_epoch_10_lp_epoch_100.pt", "clap_model") | |
| # key clap_model.audio_branch.bn0.num_batches_tracked is different | |
| # key clap_model.audio_branch.conv_block1.bn1.running_mean is different | |
| # key clap_model.audio_branch.conv_block1.bn1.running_var is different | |
| # key clap_model.audio_branch.conv_block1.bn1.num_batches_tracked is different | |
| # key clap_model.audio_branch.conv_block1.bn2.running_mean is different | |
| # key clap_model.audio_branch.conv_block1.bn2.running_var is different | |
| # key clap_model.audio_branch.conv_block1.bn2.num_batches_tracked is different | |
| # key clap_model.audio_branch.conv_block2.bn1.running_mean is different | |
| # key clap_model.audio_branch.conv_block2.bn1.running_var is different | |
| # key clap_model.audio_branch.conv_block2.bn1.num_batches_tracked is different | |
| # key clap_model.audio_branch.conv_block2.bn2.running_mean is different | |
| # key clap_model.audio_branch.conv_block2.bn2.running_var is different | |
| # key clap_model.audio_branch.conv_block2.bn2.num_batches_tracked is different | |
| # key clap_model.audio_branch.conv_block3.bn1.running_mean is different | |
| # key clap_model.audio_branch.conv_block3.bn1.running_var is different | |
| # key clap_model.audio_branch.conv_block3.bn1.num_batches_tracked is different | |
| # key clap_model.audio_branch.conv_block3.bn2.running_mean is different | |
| # key clap_model.audio_branch.conv_block3.bn2.running_var is different | |
| # key clap_model.audio_branch.conv_block3.bn2.num_batches_tracked is different | |
| # key clap_model.audio_branch.conv_block4.bn1.running_mean is different | |
| # key clap_model.audio_branch.conv_block4.bn1.running_var is different | |
| # key clap_model.audio_branch.conv_block4.bn1.num_batches_tracked is different | |
| # key clap_model.audio_branch.conv_block4.bn2.running_mean is different | |
| # key clap_model.audio_branch.conv_block4.bn2.running_var is different | |
| # key clap_model.audio_branch.conv_block4.bn2.num_batches_tracked is different | |
| # key clap_model.audio_branch.conv_block5.bn1.running_mean is different | |
| # key clap_model.audio_branch.conv_block5.bn1.running_var is different | |
| # key clap_model.audio_branch.conv_block5.bn1.num_batches_tracked is different | |
| # key clap_model.audio_branch.conv_block5.bn2.running_mean is different | |
| # key clap_model.audio_branch.conv_block5.bn2.running_var is different | |
| # key clap_model.audio_branch.conv_block5.bn2.num_batches_tracked is different | |
| # key clap_model.audio_branch.conv_block6.bn1.running_mean is different | |
| # key clap_model.audio_branch.conv_block6.bn1.running_var is different | |
| # key clap_model.audio_branch.conv_block6.bn1.num_batches_tracked is different | |
| # key clap_model.audio_branch.conv_block6.bn2.running_mean is different | |
| # key clap_model.audio_branch.conv_block6.bn2.running_var is different | |
| # key clap_model.audio_branch.conv_block6.bn2.num_batches_tracked is different | |
| # a_sum: tensor(120061.5078) | |
| # b_sum: tensor(122656.0469) | |
| # diff: tensor(-2594.5391) | |
| # True | |