Training in progress, step 70000, checkpoint
Browse files
last-checkpoint/generation_config.json
CHANGED
|
@@ -1,5 +1,4 @@
|
|
| 1 |
{
|
| 2 |
-
"_from_model_config": true,
|
| 3 |
"bos_token_id": 0,
|
| 4 |
"do_sample": true,
|
| 5 |
"eos_token_id": 2,
|
|
|
|
| 1 |
{
|
|
|
|
| 2 |
"bos_token_id": 0,
|
| 3 |
"do_sample": true,
|
| 4 |
"eos_token_id": 2,
|
last-checkpoint/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 409608164
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c8156f2d05f008ab14e4f9ff4cf00464ab8dedc324048799bb7813f56f1455aa
|
| 3 |
size 409608164
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 814647162
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:22a9fe24e12b0fccc361f7f058d9a04ed40403345f911cf5e190c1b713dd223b
|
| 3 |
size 814647162
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7f1a78ce8fa051669a7cd4c6bca63f629c83f912ad6325f1a5a7d9a4e1fd04c0
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1192
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b232468396fd1b672ffe7d6d3dd6ae81216e79f635927559657b61a1ac09dbf5
|
| 3 |
size 1192
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
-
"best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_reduce-
|
| 4 |
-
"epoch": 6.
|
| 5 |
"eval_steps": 2500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -1043,6 +1043,118 @@
|
|
| 1043 |
"learning_rate": 5e-05,
|
| 1044 |
"loss": 0.0005,
|
| 1045 |
"step": 65000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1046 |
}
|
| 1047 |
],
|
| 1048 |
"logging_steps": 500,
|
|
@@ -1062,7 +1174,7 @@
|
|
| 1062 |
"attributes": {}
|
| 1063 |
}
|
| 1064 |
},
|
| 1065 |
-
"total_flos":
|
| 1066 |
"train_batch_size": 128,
|
| 1067 |
"trial_name": null,
|
| 1068 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.5933073654390935,
|
| 3 |
+
"best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_reduce-rand-smiles/checkpoint-70000",
|
| 4 |
+
"epoch": 6.904714933912015,
|
| 5 |
"eval_steps": 2500,
|
| 6 |
+
"global_step": 70000,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 1043 |
"learning_rate": 5e-05,
|
| 1044 |
"loss": 0.0005,
|
| 1045 |
"step": 65000
|
| 1046 |
+
},
|
| 1047 |
+
{
|
| 1048 |
+
"epoch": 6.460840402446242,
|
| 1049 |
+
"grad_norm": 0.007632725406438112,
|
| 1050 |
+
"learning_rate": 5e-05,
|
| 1051 |
+
"loss": 0.0006,
|
| 1052 |
+
"step": 65500
|
| 1053 |
+
},
|
| 1054 |
+
{
|
| 1055 |
+
"epoch": 6.510159794831328,
|
| 1056 |
+
"grad_norm": 0.021445443853735924,
|
| 1057 |
+
"learning_rate": 5e-05,
|
| 1058 |
+
"loss": 0.0006,
|
| 1059 |
+
"step": 66000
|
| 1060 |
+
},
|
| 1061 |
+
{
|
| 1062 |
+
"epoch": 6.5594791872164135,
|
| 1063 |
+
"grad_norm": 0.006442953832447529,
|
| 1064 |
+
"learning_rate": 5e-05,
|
| 1065 |
+
"loss": 0.0007,
|
| 1066 |
+
"step": 66500
|
| 1067 |
+
},
|
| 1068 |
+
{
|
| 1069 |
+
"epoch": 6.608798579601499,
|
| 1070 |
+
"grad_norm": 0.013008514419198036,
|
| 1071 |
+
"learning_rate": 5e-05,
|
| 1072 |
+
"loss": 0.0007,
|
| 1073 |
+
"step": 67000
|
| 1074 |
+
},
|
| 1075 |
+
{
|
| 1076 |
+
"epoch": 6.658117971986585,
|
| 1077 |
+
"grad_norm": 0.010181767866015434,
|
| 1078 |
+
"learning_rate": 5e-05,
|
| 1079 |
+
"loss": 0.0007,
|
| 1080 |
+
"step": 67500
|
| 1081 |
+
},
|
| 1082 |
+
{
|
| 1083 |
+
"epoch": 6.707437364371671,
|
| 1084 |
+
"grad_norm": 0.03756224736571312,
|
| 1085 |
+
"learning_rate": 5e-05,
|
| 1086 |
+
"loss": 0.0007,
|
| 1087 |
+
"step": 68000
|
| 1088 |
+
},
|
| 1089 |
+
{
|
| 1090 |
+
"epoch": 6.756756756756757,
|
| 1091 |
+
"grad_norm": 0.04266593977808952,
|
| 1092 |
+
"learning_rate": 5e-05,
|
| 1093 |
+
"loss": 0.0007,
|
| 1094 |
+
"step": 68500
|
| 1095 |
+
},
|
| 1096 |
+
{
|
| 1097 |
+
"epoch": 6.806076149141843,
|
| 1098 |
+
"grad_norm": 0.020077640190720558,
|
| 1099 |
+
"learning_rate": 5e-05,
|
| 1100 |
+
"loss": 0.0008,
|
| 1101 |
+
"step": 69000
|
| 1102 |
+
},
|
| 1103 |
+
{
|
| 1104 |
+
"epoch": 6.855395541526929,
|
| 1105 |
+
"grad_norm": 0.017746519297361374,
|
| 1106 |
+
"learning_rate": 5e-05,
|
| 1107 |
+
"loss": 0.0007,
|
| 1108 |
+
"step": 69500
|
| 1109 |
+
},
|
| 1110 |
+
{
|
| 1111 |
+
"epoch": 6.904714933912015,
|
| 1112 |
+
"grad_norm": 0.030025839805603027,
|
| 1113 |
+
"learning_rate": 5e-05,
|
| 1114 |
+
"loss": 0.0007,
|
| 1115 |
+
"step": 70000
|
| 1116 |
+
},
|
| 1117 |
+
{
|
| 1118 |
+
"epoch": 6.904714933912015,
|
| 1119 |
+
"eval_all_ligands_equal": 0.5933073654390935,
|
| 1120 |
+
"eval_e3_equal": 0.8271954674220963,
|
| 1121 |
+
"eval_e3_graph_edit_distance": Infinity,
|
| 1122 |
+
"eval_e3_graph_edit_distance_norm": Infinity,
|
| 1123 |
+
"eval_e3_has_attachment_point(s)": 0.9917669971671388,
|
| 1124 |
+
"eval_e3_heavy_atoms_difference": 0.4047450424929179,
|
| 1125 |
+
"eval_e3_heavy_atoms_difference_norm": 0.00843519659565475,
|
| 1126 |
+
"eval_e3_tanimoto_similarity": 0.0,
|
| 1127 |
+
"eval_e3_valid": 0.9917669971671388,
|
| 1128 |
+
"eval_has_all_attachment_points": 0.990350566572238,
|
| 1129 |
+
"eval_has_three_substructures": 0.9999114730878187,
|
| 1130 |
+
"eval_heavy_atoms_difference": 5.034082861189802,
|
| 1131 |
+
"eval_heavy_atoms_difference_norm": 0.06779925037570428,
|
| 1132 |
+
"eval_linker_equal": 0.84782223796034,
|
| 1133 |
+
"eval_linker_graph_edit_distance": 2.5672804532577904e+61,
|
| 1134 |
+
"eval_linker_graph_edit_distance_norm": Infinity,
|
| 1135 |
+
"eval_linker_has_attachment_point(s)": 0.9974327195467422,
|
| 1136 |
+
"eval_linker_heavy_atoms_difference": 0.17236189801699717,
|
| 1137 |
+
"eval_linker_heavy_atoms_difference_norm": 0.0012798111547145205,
|
| 1138 |
+
"eval_linker_tanimoto_similarity": 0.0,
|
| 1139 |
+
"eval_linker_valid": 0.9974327195467422,
|
| 1140 |
+
"eval_loss": 0.3657114505767822,
|
| 1141 |
+
"eval_num_fragments": 2.9999114730878187,
|
| 1142 |
+
"eval_poi_equal": 0.7852337110481586,
|
| 1143 |
+
"eval_poi_graph_edit_distance": Infinity,
|
| 1144 |
+
"eval_poi_graph_edit_distance_norm": Infinity,
|
| 1145 |
+
"eval_poi_has_attachment_point(s)": 0.9508675637393768,
|
| 1146 |
+
"eval_poi_heavy_atoms_difference": 1.4603399433427762,
|
| 1147 |
+
"eval_poi_heavy_atoms_difference_norm": 0.046673695753555136,
|
| 1148 |
+
"eval_poi_tanimoto_similarity": 0.0,
|
| 1149 |
+
"eval_poi_valid": 0.9508675637393768,
|
| 1150 |
+
"eval_reassembly": 0.6011862606232294,
|
| 1151 |
+
"eval_reassembly_nostereo": 0.6374822946175638,
|
| 1152 |
+
"eval_runtime": 2302.2232,
|
| 1153 |
+
"eval_samples_per_second": 4.907,
|
| 1154 |
+
"eval_steps_per_second": 0.077,
|
| 1155 |
+
"eval_tanimoto_similarity": 0.0,
|
| 1156 |
+
"eval_valid": 0.9407754957507082,
|
| 1157 |
+
"step": 70000
|
| 1158 |
}
|
| 1159 |
],
|
| 1160 |
"logging_steps": 500,
|
|
|
|
| 1174 |
"attributes": {}
|
| 1175 |
}
|
| 1176 |
},
|
| 1177 |
+
"total_flos": 6.110533685712476e+17,
|
| 1178 |
"train_batch_size": 128,
|
| 1179 |
"trial_name": null,
|
| 1180 |
"trial_params": null
|