| { | |
| "best_metric": 0.5964058073654391, | |
| "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-rand-smiles/checkpoint-95000", | |
| "epoch": 9.863878477017163, | |
| "eval_steps": 2500, | |
| "global_step": 100000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.049319392385085814, | |
| "grad_norm": 0.9180974364280701, | |
| "learning_rate": 3.57653791130186e-05, | |
| "loss": 2.3658, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.09863878477017163, | |
| "grad_norm": 0.5833754539489746, | |
| "learning_rate": 4.987090502485352e-05, | |
| "loss": 0.3343, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.14795817715525744, | |
| "grad_norm": 0.4139927327632904, | |
| "learning_rate": 4.9090579537757233e-05, | |
| "loss": 0.1224, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.19727756954034326, | |
| "grad_norm": 0.2650364339351654, | |
| "learning_rate": 4.762477336752997e-05, | |
| "loss": 0.0645, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.24659696192542907, | |
| "grad_norm": 0.1992732584476471, | |
| "learning_rate": 4.5515195009118114e-05, | |
| "loss": 0.0409, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.2959163543105149, | |
| "grad_norm": 0.2138502597808838, | |
| "learning_rate": 4.282187104726926e-05, | |
| "loss": 0.029, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.3452357466956007, | |
| "grad_norm": 0.17667913436889648, | |
| "learning_rate": 3.962143814181834e-05, | |
| "loss": 0.0219, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.3945551390806865, | |
| "grad_norm": 0.19663669168949127, | |
| "learning_rate": 3.600496238484107e-05, | |
| "loss": 0.0174, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.4438745314657723, | |
| "grad_norm": 0.18696776032447815, | |
| "learning_rate": 3.207534807835037e-05, | |
| "loss": 0.0139, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.49319392385085814, | |
| "grad_norm": 0.20042544603347778, | |
| "learning_rate": 2.7944409663963666e-05, | |
| "loss": 0.012, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.49319392385085814, | |
| "eval_all_ligands_equal": 0.4823831444759207, | |
| "eval_e3_equal": 0.7897485835694051, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9971671388101983, | |
| "eval_e3_heavy_atoms_difference": 0.1648371104815864, | |
| "eval_e3_heavy_atoms_difference_norm": -0.0036399760782591085, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9971671388101983, | |
| "eval_has_all_attachment_points": 0.9872521246458924, | |
| "eval_has_three_substructures": 0.9994688385269122, | |
| "eval_heavy_atoms_difference": 4.9585694050991505, | |
| "eval_heavy_atoms_difference_norm": 0.06500452298457288, | |
| "eval_linker_equal": 0.6945821529745042, | |
| "eval_linker_graph_edit_distance": 2.5672804532577904e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9974327195467422, | |
| "eval_linker_heavy_atoms_difference": 0.5344369688385269, | |
| "eval_linker_heavy_atoms_difference_norm": 0.010326204764079937, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9974327195467422, | |
| "eval_loss": 0.30552199482917786, | |
| "eval_num_fragments": 3.000531161473088, | |
| "eval_poi_equal": 0.7353930594900849, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9479461756373938, | |
| "eval_poi_heavy_atoms_difference": 1.4326310198300283, | |
| "eval_poi_heavy_atoms_difference_norm": 0.0421069121424993, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9479461756373938, | |
| "eval_reassembly": 0.4899079320113314, | |
| "eval_reassembly_nostereo": 0.5198300283286119, | |
| "eval_runtime": 2427.2767, | |
| "eval_samples_per_second": 4.654, | |
| "eval_steps_per_second": 0.073, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9440509915014165, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.542513316235944, | |
| "grad_norm": 0.23216110467910767, | |
| "learning_rate": 2.372969012074575e-05, | |
| "loss": 0.0101, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.5918327086210298, | |
| "grad_norm": 0.14501763880252838, | |
| "learning_rate": 1.9551116361503457e-05, | |
| "loss": 0.0089, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.6411521010061156, | |
| "grad_norm": 0.12931592762470245, | |
| "learning_rate": 1.55275867959039e-05, | |
| "loss": 0.0078, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.6904714933912014, | |
| "grad_norm": 0.1443043202161789, | |
| "learning_rate": 1.1773588158934163e-05, | |
| "loss": 0.0072, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.7397908857762873, | |
| "grad_norm": 0.13492755591869354, | |
| "learning_rate": 8.395937870495241e-06, | |
| "loss": 0.0064, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.7397908857762873, | |
| "eval_all_ligands_equal": 0.5282400849858357, | |
| "eval_e3_equal": 0.8016997167138811, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9944228045325779, | |
| "eval_e3_heavy_atoms_difference": 0.342864730878187, | |
| "eval_e3_heavy_atoms_difference_norm": 0.0053395654721611, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9944228045325779, | |
| "eval_has_all_attachment_points": 0.9873406515580736, | |
| "eval_has_three_substructures": 0.9984950424929179, | |
| "eval_heavy_atoms_difference": 4.519475920679887, | |
| "eval_heavy_atoms_difference_norm": 0.059262409138829596, | |
| "eval_linker_equal": 0.7532754957507082, | |
| "eval_linker_graph_edit_distance": 4.0722379603399434e+61, | |
| "eval_linker_graph_edit_distance_norm": 0.04522498623517566, | |
| "eval_linker_has_attachment_point(s)": 0.99592776203966, | |
| "eval_linker_heavy_atoms_difference": 0.2563739376770538, | |
| "eval_linker_heavy_atoms_difference_norm": -2.297886744053873e-06, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.99592776203966, | |
| "eval_loss": 0.3155761659145355, | |
| "eval_num_fragments": 3.000265580736544, | |
| "eval_poi_equal": 0.7551345609065155, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9531692634560907, | |
| "eval_poi_heavy_atoms_difference": 1.3395007082152974, | |
| "eval_poi_heavy_atoms_difference_norm": 0.03956240679400847, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9531692634560907, | |
| "eval_reassembly": 0.5359419263456091, | |
| "eval_reassembly_nostereo": 0.5688739376770539, | |
| "eval_runtime": 2393.2072, | |
| "eval_samples_per_second": 4.72, | |
| "eval_steps_per_second": 0.074, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9476805949008499, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.789110278161373, | |
| "grad_norm": 0.12762106955051422, | |
| "learning_rate": 5.490744620021837e-06, | |
| "loss": 0.0059, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.8384296705464589, | |
| "grad_norm": 0.12324267625808716, | |
| "learning_rate": 3.1406736605785858e-06, | |
| "loss": 0.0056, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.8877490629315447, | |
| "grad_norm": 0.10565518587827682, | |
| "learning_rate": 1.412594626586694e-06, | |
| "loss": 0.0053, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.9370684553166305, | |
| "grad_norm": 0.10529914498329163, | |
| "learning_rate": 3.5567880489165705e-07, | |
| "loss": 0.0052, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.9863878477017163, | |
| "grad_norm": 0.11241015791893005, | |
| "learning_rate": 0.0, | |
| "loss": 0.0051, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.9863878477017163, | |
| "eval_all_ligands_equal": 0.5389518413597734, | |
| "eval_e3_equal": 0.8036473087818697, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9966359773371105, | |
| "eval_e3_heavy_atoms_difference": 0.29762747875354106, | |
| "eval_e3_heavy_atoms_difference_norm": 0.0033101986712786185, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9966359773371105, | |
| "eval_has_all_attachment_points": 0.9905276203966006, | |
| "eval_has_three_substructures": 0.9992032577903682, | |
| "eval_heavy_atoms_difference": 4.104904390934844, | |
| "eval_heavy_atoms_difference_norm": 0.0535843946056828, | |
| "eval_linker_equal": 0.7668201133144475, | |
| "eval_linker_graph_edit_distance": 2.8328611898017e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9971671388101983, | |
| "eval_linker_heavy_atoms_difference": 0.3031161473087819, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0033820491296884894, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9971671388101983, | |
| "eval_loss": 0.31840986013412476, | |
| "eval_num_fragments": 3.000265580736544, | |
| "eval_poi_equal": 0.7619511331444759, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9589235127478754, | |
| "eval_poi_heavy_atoms_difference": 1.1856409348441925, | |
| "eval_poi_heavy_atoms_difference_norm": 0.035200135322798254, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9589235127478754, | |
| "eval_reassembly": 0.5459454674220963, | |
| "eval_reassembly_nostereo": 0.5798512747875354, | |
| "eval_runtime": 2383.5536, | |
| "eval_samples_per_second": 4.739, | |
| "eval_steps_per_second": 0.074, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9546742209631728, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 1.0357072400868022, | |
| "grad_norm": 0.1099235862493515, | |
| "learning_rate": 4.867483222184158e-05, | |
| "loss": 0.0148, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 1.085026632471888, | |
| "grad_norm": 0.09051310271024704, | |
| "learning_rate": 4.8545628694730624e-05, | |
| "loss": 0.0081, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 1.1343460248569737, | |
| "grad_norm": 0.11217254400253296, | |
| "learning_rate": 4.841060399978481e-05, | |
| "loss": 0.0067, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 1.1836654172420595, | |
| "grad_norm": 0.08800435066223145, | |
| "learning_rate": 4.826979151905655e-05, | |
| "loss": 0.0061, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 1.2329848096271454, | |
| "grad_norm": 0.08104352653026581, | |
| "learning_rate": 4.812322606550813e-05, | |
| "loss": 0.0058, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 1.2823042020122313, | |
| "grad_norm": 0.10292906314134598, | |
| "learning_rate": 4.797094387440491e-05, | |
| "loss": 0.0056, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 1.3316235943973171, | |
| "grad_norm": 0.07915246486663818, | |
| "learning_rate": 4.78129825943569e-05, | |
| "loss": 0.0051, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.3809429867824028, | |
| "grad_norm": 0.07057974487543106, | |
| "learning_rate": 4.7649381278011e-05, | |
| "loss": 0.0045, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 1.4302623791674887, | |
| "grad_norm": 0.07770856469869614, | |
| "learning_rate": 4.748018037239592e-05, | |
| "loss": 0.0043, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 1.4795817715525745, | |
| "grad_norm": 0.08110259473323822, | |
| "learning_rate": 4.73054217089226e-05, | |
| "loss": 0.0042, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 1.5289011639376602, | |
| "grad_norm": 0.06894150376319885, | |
| "learning_rate": 4.712514849304219e-05, | |
| "loss": 0.0039, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 1.578220556322746, | |
| "grad_norm": 0.057050593197345734, | |
| "learning_rate": 4.693940529356444e-05, | |
| "loss": 0.0053, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 1.627539948707832, | |
| "grad_norm": 0.04995543137192726, | |
| "learning_rate": 4.674823803163899e-05, | |
| "loss": 0.0034, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 1.6768593410929178, | |
| "grad_norm": 0.06972987949848175, | |
| "learning_rate": 4.655169396940229e-05, | |
| "loss": 0.0031, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 1.7261787334780037, | |
| "grad_norm": 0.05143571272492409, | |
| "learning_rate": 4.6349821698293025e-05, | |
| "loss": 0.0031, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 1.7754981258630895, | |
| "grad_norm": 0.06793902069330215, | |
| "learning_rate": 4.6142671127038905e-05, | |
| "loss": 0.0029, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 1.8248175182481752, | |
| "grad_norm": 0.05504274740815163, | |
| "learning_rate": 4.593029346931777e-05, | |
| "loss": 0.0029, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 1.874136910633261, | |
| "grad_norm": 0.07636507600545883, | |
| "learning_rate": 4.571274123109606e-05, | |
| "loss": 0.0028, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 1.9234563030183467, | |
| "grad_norm": 0.08288216590881348, | |
| "learning_rate": 4.549006819764779e-05, | |
| "loss": 0.0027, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 1.9727756954034326, | |
| "grad_norm": 0.07712626457214355, | |
| "learning_rate": 4.52623294202573e-05, | |
| "loss": 0.0027, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 2.0220950877885184, | |
| "grad_norm": 0.025654463097453117, | |
| "learning_rate": 4.502958120260894e-05, | |
| "loss": 0.0027, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 2.0714144801736043, | |
| "grad_norm": 0.05340616777539253, | |
| "learning_rate": 4.479188108686714e-05, | |
| "loss": 0.0021, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 2.12073387255869, | |
| "grad_norm": 0.043434690684080124, | |
| "learning_rate": 4.4549287839450324e-05, | |
| "loss": 0.0021, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 2.170053264943776, | |
| "grad_norm": 0.042540039867162704, | |
| "learning_rate": 4.4301861436502156e-05, | |
| "loss": 0.0021, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 2.219372657328862, | |
| "grad_norm": 0.047575388103723526, | |
| "learning_rate": 4.404966304906363e-05, | |
| "loss": 0.002, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 2.2686920497139473, | |
| "grad_norm": 0.06077829375863075, | |
| "learning_rate": 4.379275502794983e-05, | |
| "loss": 0.0022, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 2.318011442099033, | |
| "grad_norm": 0.03555487096309662, | |
| "learning_rate": 4.353120088833501e-05, | |
| "loss": 0.002, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 2.367330834484119, | |
| "grad_norm": 0.04414715990424156, | |
| "learning_rate": 4.326506529404972e-05, | |
| "loss": 0.0021, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 2.416650226869205, | |
| "grad_norm": 0.0375969260931015, | |
| "learning_rate": 4.2994959806435226e-05, | |
| "loss": 0.0022, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 2.465969619254291, | |
| "grad_norm": 0.054521363228559494, | |
| "learning_rate": 4.2719868638689734e-05, | |
| "loss": 0.0018, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 2.5152890116393767, | |
| "grad_norm": 0.026706155389547348, | |
| "learning_rate": 4.244039660132742e-05, | |
| "loss": 0.0016, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 2.5646084040244626, | |
| "grad_norm": 0.03463255986571312, | |
| "learning_rate": 4.2156612788007085e-05, | |
| "loss": 0.0016, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 2.6139277964095484, | |
| "grad_norm": 0.03429891914129257, | |
| "learning_rate": 4.186858735838457e-05, | |
| "loss": 0.0017, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 2.6632471887946343, | |
| "grad_norm": 0.06138679385185242, | |
| "learning_rate": 4.15763915207673e-05, | |
| "loss": 0.0017, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 2.7125665811797197, | |
| "grad_norm": 0.04362751543521881, | |
| "learning_rate": 4.1280097514509486e-05, | |
| "loss": 0.0016, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 2.7618859735648056, | |
| "grad_norm": 0.046160563826560974, | |
| "learning_rate": 4.097977859215244e-05, | |
| "loss": 0.0015, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 2.8112053659498915, | |
| "grad_norm": 0.0806930810213089, | |
| "learning_rate": 4.0675509001314444e-05, | |
| "loss": 0.0015, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 2.8605247583349773, | |
| "grad_norm": 0.04074354097247124, | |
| "learning_rate": 4.036798407361931e-05, | |
| "loss": 0.0015, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 2.909844150720063, | |
| "grad_norm": 0.02577074058353901, | |
| "learning_rate": 4.005604729882533e-05, | |
| "loss": 0.0014, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 2.959163543105149, | |
| "grad_norm": 0.02091473527252674, | |
| "learning_rate": 3.974038822893393e-05, | |
| "loss": 0.0014, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 3.008482935490235, | |
| "grad_norm": 0.02712525986135006, | |
| "learning_rate": 3.942108490409635e-05, | |
| "loss": 0.0014, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 3.0578023278753204, | |
| "grad_norm": 0.04912711679935455, | |
| "learning_rate": 3.9099514736586965e-05, | |
| "loss": 0.0024, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 3.1071217202604062, | |
| "grad_norm": 0.03684419021010399, | |
| "learning_rate": 3.877317438846952e-05, | |
| "loss": 0.0012, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 3.156441112645492, | |
| "grad_norm": 0.034598346799612045, | |
| "learning_rate": 3.8443428908946436e-05, | |
| "loss": 0.0011, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 3.205760505030578, | |
| "grad_norm": 0.023367149755358696, | |
| "learning_rate": 3.8110359820741256e-05, | |
| "loss": 0.0011, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 3.255079897415664, | |
| "grad_norm": 0.04803726077079773, | |
| "learning_rate": 3.777404946827081e-05, | |
| "loss": 0.0011, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 3.3043992898007497, | |
| "grad_norm": 0.02714533731341362, | |
| "learning_rate": 3.743458099728716e-05, | |
| "loss": 0.0011, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 3.3537186821858356, | |
| "grad_norm": 0.04370482265949249, | |
| "learning_rate": 3.7092038334321655e-05, | |
| "loss": 0.0011, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 3.4030380745709214, | |
| "grad_norm": 0.035204458981752396, | |
| "learning_rate": 3.674650616593575e-05, | |
| "loss": 0.0012, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 3.4523574669560073, | |
| "grad_norm": 0.013022173196077347, | |
| "learning_rate": 3.639806991778405e-05, | |
| "loss": 0.0012, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 3.501676859341093, | |
| "grad_norm": 0.027150381356477737, | |
| "learning_rate": 3.6046815733494614e-05, | |
| "loss": 0.0011, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 3.5509962517261786, | |
| "grad_norm": 0.034229960292577744, | |
| "learning_rate": 3.5692830453371675e-05, | |
| "loss": 0.001, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 3.6003156441112645, | |
| "grad_norm": 0.04263359308242798, | |
| "learning_rate": 3.533620159292621e-05, | |
| "loss": 0.001, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 3.6496350364963503, | |
| "grad_norm": 0.026501759886741638, | |
| "learning_rate": 3.497701732123957e-05, | |
| "loss": 0.0011, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 3.698954428881436, | |
| "grad_norm": 0.014317413792014122, | |
| "learning_rate": 3.4615366439165435e-05, | |
| "loss": 0.001, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 3.748273821266522, | |
| "grad_norm": 0.04115162417292595, | |
| "learning_rate": 3.425133835737574e-05, | |
| "loss": 0.001, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 3.797593213651608, | |
| "grad_norm": 0.01732572540640831, | |
| "learning_rate": 3.388502307425566e-05, | |
| "loss": 0.0011, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 3.8469126060366934, | |
| "grad_norm": 0.02579200640320778, | |
| "learning_rate": 3.3516511153653423e-05, | |
| "loss": 0.0009, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 3.8962319984217793, | |
| "grad_norm": 0.023984914645552635, | |
| "learning_rate": 3.314663697789466e-05, | |
| "loss": 0.0009, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 3.945551390806865, | |
| "grad_norm": 0.05176176130771637, | |
| "learning_rate": 3.2774009559667783e-05, | |
| "loss": 0.001, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 3.994870783191951, | |
| "grad_norm": 0.018639806658029556, | |
| "learning_rate": 3.239946017898311e-05, | |
| "loss": 0.0009, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 4.044190175577037, | |
| "grad_norm": 0.026381010189652443, | |
| "learning_rate": 3.202308143539852e-05, | |
| "loss": 0.0008, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 4.093509567962123, | |
| "grad_norm": 0.025508331134915352, | |
| "learning_rate": 3.164572428162694e-05, | |
| "loss": 0.0007, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 4.142828960347209, | |
| "grad_norm": 0.015003887936472893, | |
| "learning_rate": 3.1265969589091056e-05, | |
| "loss": 0.0008, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 4.1921483527322945, | |
| "grad_norm": 0.02331295795738697, | |
| "learning_rate": 3.0884665765669805e-05, | |
| "loss": 0.0008, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 4.24146774511738, | |
| "grad_norm": 0.01973959244787693, | |
| "learning_rate": 3.0501907080816822e-05, | |
| "loss": 0.0008, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 4.290787137502466, | |
| "grad_norm": 0.027150608599185944, | |
| "learning_rate": 3.0118557695949296e-05, | |
| "loss": 0.0008, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 4.340106529887552, | |
| "grad_norm": 0.02721910923719406, | |
| "learning_rate": 2.9733175947482144e-05, | |
| "loss": 0.0007, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 4.389425922272638, | |
| "grad_norm": 0.01915474608540535, | |
| "learning_rate": 2.9346624019533298e-05, | |
| "loss": 0.0008, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 4.438745314657724, | |
| "grad_norm": 0.01952786184847355, | |
| "learning_rate": 2.8958997479041255e-05, | |
| "loss": 0.0008, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 4.48806470704281, | |
| "grad_norm": 0.05904557555913925, | |
| "learning_rate": 2.857039215862017e-05, | |
| "loss": 0.0007, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 4.537384099427895, | |
| "grad_norm": 0.025458356365561485, | |
| "learning_rate": 2.8180904132867185e-05, | |
| "loss": 0.0007, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 4.5867034918129805, | |
| "grad_norm": 0.025726348161697388, | |
| "learning_rate": 2.7791410964224228e-05, | |
| "loss": 0.0009, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 4.636022884198066, | |
| "grad_norm": 0.01956965960562229, | |
| "learning_rate": 2.740123043483287e-05, | |
| "loss": 0.0007, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 4.685342276583152, | |
| "grad_norm": 0.038180191069841385, | |
| "learning_rate": 2.7009674984217738e-05, | |
| "loss": 0.0007, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 4.734661668968238, | |
| "grad_norm": 0.0360223725438118, | |
| "learning_rate": 2.661762268318811e-05, | |
| "loss": 0.0007, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 4.783981061353324, | |
| "grad_norm": 0.03571155667304993, | |
| "learning_rate": 2.6225170458535525e-05, | |
| "loss": 0.0007, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 4.83330045373841, | |
| "grad_norm": 0.012124253436923027, | |
| "learning_rate": 2.5832415335924354e-05, | |
| "loss": 0.0007, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 4.882619846123496, | |
| "grad_norm": 0.029337838292121887, | |
| "learning_rate": 2.5439454415904175e-05, | |
| "loss": 0.0006, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 4.931939238508582, | |
| "grad_norm": 0.04516094923019409, | |
| "learning_rate": 2.504638484990372e-05, | |
| "loss": 0.0006, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 4.9812586308936675, | |
| "grad_norm": 0.015657415613532066, | |
| "learning_rate": 2.4653303816212188e-05, | |
| "loss": 0.0006, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 5.030578023278753, | |
| "grad_norm": 0.009272584691643715, | |
| "learning_rate": 2.42603084959539e-05, | |
| "loss": 0.0006, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 5.079897415663839, | |
| "grad_norm": 0.016377432271838188, | |
| "learning_rate": 2.386749604906234e-05, | |
| "loss": 0.0005, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 5.129216808048925, | |
| "grad_norm": 0.02414529025554657, | |
| "learning_rate": 2.3475748311231863e-05, | |
| "loss": 0.0005, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 5.178536200434011, | |
| "grad_norm": 0.023295147344470024, | |
| "learning_rate": 2.3083592035157375e-05, | |
| "loss": 0.0005, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 5.227855592819097, | |
| "grad_norm": 0.020559698343276978, | |
| "learning_rate": 2.2691909551163278e-05, | |
| "loss": 0.0005, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 5.277174985204182, | |
| "grad_norm": 0.022504296153783798, | |
| "learning_rate": 2.2300797694611534e-05, | |
| "loss": 0.0005, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 5.326494377589268, | |
| "grad_norm": 0.006661570630967617, | |
| "learning_rate": 2.191035315978832e-05, | |
| "loss": 0.0005, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 5.375813769974354, | |
| "grad_norm": 0.011046066880226135, | |
| "learning_rate": 2.1520672475998373e-05, | |
| "loss": 0.0005, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 5.425133162359439, | |
| "grad_norm": 0.02395332045853138, | |
| "learning_rate": 2.1131851983700107e-05, | |
| "loss": 0.0005, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 5.474452554744525, | |
| "grad_norm": 0.006757025141268969, | |
| "learning_rate": 2.074398781068747e-05, | |
| "loss": 0.0005, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 5.523771947129611, | |
| "grad_norm": 0.01942395232617855, | |
| "learning_rate": 2.0357175848324307e-05, | |
| "loss": 0.0005, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 5.573091339514697, | |
| "grad_norm": 0.0312858484685421, | |
| "learning_rate": 1.997228184710801e-05, | |
| "loss": 0.0005, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 5.622410731899783, | |
| "grad_norm": 0.013257316313683987, | |
| "learning_rate": 1.958785833460854e-05, | |
| "loss": 0.0005, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 5.671730124284869, | |
| "grad_norm": 0.010021144524216652, | |
| "learning_rate": 1.9204772861767664e-05, | |
| "loss": 0.0005, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 5.721049516669955, | |
| "grad_norm": 0.010143189691007137, | |
| "learning_rate": 1.8823881951300727e-05, | |
| "loss": 0.0005, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 5.7703689090550405, | |
| "grad_norm": 0.028758643195033073, | |
| "learning_rate": 1.844375318521728e-05, | |
| "loss": 0.0005, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 5.819688301440126, | |
| "grad_norm": 0.01780054345726967, | |
| "learning_rate": 1.8065245315036443e-05, | |
| "loss": 0.0005, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 5.869007693825212, | |
| "grad_norm": 0.007259385660290718, | |
| "learning_rate": 1.768845191897044e-05, | |
| "loss": 0.0005, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 5.918327086210298, | |
| "grad_norm": 0.014262279495596886, | |
| "learning_rate": 1.7313466151363385e-05, | |
| "loss": 0.0004, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 5.967646478595384, | |
| "grad_norm": 0.011568223126232624, | |
| "learning_rate": 1.6941124932601034e-05, | |
| "loss": 0.0005, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 6.01696587098047, | |
| "grad_norm": 0.04736906662583351, | |
| "learning_rate": 1.6570027997553185e-05, | |
| "loss": 0.0004, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 6.066285263365556, | |
| "grad_norm": 0.01865927129983902, | |
| "learning_rate": 1.6201015198057866e-05, | |
| "loss": 0.0004, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 6.115604655750641, | |
| "grad_norm": 0.014169979840517044, | |
| "learning_rate": 1.5834177764868226e-05, | |
| "loss": 0.0004, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 6.164924048135727, | |
| "grad_norm": 0.009675228968262672, | |
| "learning_rate": 1.5469606390923247e-05, | |
| "loss": 0.0004, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 6.2142434405208125, | |
| "grad_norm": 0.00304154260084033, | |
| "learning_rate": 1.5107391208925827e-05, | |
| "loss": 0.0004, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 6.263562832905898, | |
| "grad_norm": 0.004374227486550808, | |
| "learning_rate": 1.474762176905931e-05, | |
| "loss": 0.0004, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 6.312882225290984, | |
| "grad_norm": 0.006118281278759241, | |
| "learning_rate": 1.439038701684809e-05, | |
| "loss": 0.0004, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 6.36220161767607, | |
| "grad_norm": 0.006160255055874586, | |
| "learning_rate": 1.4035775271167664e-05, | |
| "loss": 0.0004, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 6.411521010061156, | |
| "grad_norm": 0.009528923779726028, | |
| "learning_rate": 1.3684575241351546e-05, | |
| "loss": 0.0004, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 6.460840402446242, | |
| "grad_norm": 0.02038782648742199, | |
| "learning_rate": 1.3335466168024479e-05, | |
| "loss": 0.0004, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 6.510159794831328, | |
| "grad_norm": 0.005516901146620512, | |
| "learning_rate": 1.2989240908505564e-05, | |
| "loss": 0.0004, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 6.5594791872164135, | |
| "grad_norm": 0.0024911893997341394, | |
| "learning_rate": 1.2645985059801791e-05, | |
| "loss": 0.0004, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 6.608798579601499, | |
| "grad_norm": 0.021164007484912872, | |
| "learning_rate": 1.2306460783748302e-05, | |
| "loss": 0.0004, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 6.658117971986585, | |
| "grad_norm": 0.029552442952990532, | |
| "learning_rate": 1.196939123002433e-05, | |
| "loss": 0.0004, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 6.707437364371671, | |
| "grad_norm": 0.010598881170153618, | |
| "learning_rate": 1.1635543223774156e-05, | |
| "loss": 0.0004, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 6.756756756756757, | |
| "grad_norm": 0.00585036538541317, | |
| "learning_rate": 1.130499930198588e-05, | |
| "loss": 0.0004, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 6.806076149141843, | |
| "grad_norm": 0.011820879764854908, | |
| "learning_rate": 1.0977841184781343e-05, | |
| "loss": 0.0003, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 6.855395541526929, | |
| "grad_norm": 0.03137628361582756, | |
| "learning_rate": 1.0654793624981977e-05, | |
| "loss": 0.0003, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 6.904714933912015, | |
| "grad_norm": 0.013417651876807213, | |
| "learning_rate": 1.0334641736320572e-05, | |
| "loss": 0.0003, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 6.904714933912015, | |
| "eval_all_ligands_equal": 0.5924220963172805, | |
| "eval_e3_equal": 0.8257790368271954, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9915014164305949, | |
| "eval_e3_heavy_atoms_difference": 0.14606940509915015, | |
| "eval_e3_heavy_atoms_difference_norm": -0.002032805775238171, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9915014164305949, | |
| "eval_has_all_attachment_points": 0.9909702549575071, | |
| "eval_has_three_substructures": 0.9996458923512748, | |
| "eval_heavy_atoms_difference": 4.543112606232294, | |
| "eval_heavy_atoms_difference_norm": 0.06065633437693917, | |
| "eval_linker_equal": 0.8415368271954674, | |
| "eval_linker_graph_edit_distance": 2.2131728045325776e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9977868271954674, | |
| "eval_linker_heavy_atoms_difference": 0.16979461756373937, | |
| "eval_linker_heavy_atoms_difference_norm": -0.0016392134755984211, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9977868271954674, | |
| "eval_loss": 0.39393511414527893, | |
| "eval_num_fragments": 3.0003541076487252, | |
| "eval_poi_equal": 0.7886862606232294, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9554709631728046, | |
| "eval_poi_heavy_atoms_difference": 1.2542492917847026, | |
| "eval_poi_heavy_atoms_difference_norm": 0.034870780705027825, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9554709631728046, | |
| "eval_reassembly": 0.5998583569405099, | |
| "eval_reassembly_nostereo": 0.6323477337110481, | |
| "eval_runtime": 2303.8454, | |
| "eval_samples_per_second": 4.903, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9464412181303116, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 6.9540343262971, | |
| "grad_norm": 0.01884830743074417, | |
| "learning_rate": 1.001811555300422e-05, | |
| "loss": 0.0003, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 7.0033537186821855, | |
| "grad_norm": 0.007204363122582436, | |
| "learning_rate": 9.705293329559965e-06, | |
| "loss": 0.0004, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 7.052673111067271, | |
| "grad_norm": 0.009515893645584583, | |
| "learning_rate": 9.396866661960307e-06, | |
| "loss": 0.0003, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 7.101992503452357, | |
| "grad_norm": 0.01400610152631998, | |
| "learning_rate": 9.091675648786607e-06, | |
| "loss": 0.0003, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 7.151311895837443, | |
| "grad_norm": 0.010973923839628696, | |
| "learning_rate": 8.790417638516349e-06, | |
| "loss": 0.0003, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 7.151311895837443, | |
| "eval_all_ligands_equal": 0.5918909348441926, | |
| "eval_e3_equal": 0.8251593484419264, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9886685552407932, | |
| "eval_e3_heavy_atoms_difference": 0.3155099150141643, | |
| "eval_e3_heavy_atoms_difference_norm": 0.004674876264079149, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9886685552407932, | |
| "eval_has_all_attachment_points": 0.9888456090651558, | |
| "eval_has_three_substructures": 0.9991147308781869, | |
| "eval_heavy_atoms_difference": 4.57542492917847, | |
| "eval_heavy_atoms_difference_norm": 0.06105317584065338, | |
| "eval_linker_equal": 0.8426876770538244, | |
| "eval_linker_graph_edit_distance": 3.718130311614731e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9962818696883853, | |
| "eval_linker_heavy_atoms_difference": 0.26044617563739375, | |
| "eval_linker_heavy_atoms_difference_norm": 0.004330713276040957, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9962818696883853, | |
| "eval_loss": 0.39214199781417847, | |
| "eval_num_fragments": 3.0001770538243626, | |
| "eval_poi_equal": 0.7898371104815864, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9601628895184136, | |
| "eval_poi_heavy_atoms_difference": 1.192723087818697, | |
| "eval_poi_heavy_atoms_difference_norm": 0.035486140958842484, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9601628895184136, | |
| "eval_reassembly": 0.5992386685552408, | |
| "eval_reassembly_nostereo": 0.630842776203966, | |
| "eval_runtime": 2302.3208, | |
| "eval_samples_per_second": 4.906, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9479461756373938, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 7.200631288222529, | |
| "grad_norm": 0.003939173649996519, | |
| "learning_rate": 8.493167110937786e-06, | |
| "loss": 0.0003, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 7.249950680607615, | |
| "grad_norm": 0.006046623457223177, | |
| "learning_rate": 8.199997555072267e-06, | |
| "loss": 0.0003, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 7.299270072992701, | |
| "grad_norm": 0.008457995019853115, | |
| "learning_rate": 7.911555290403444e-06, | |
| "loss": 0.0003, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 7.348589465377787, | |
| "grad_norm": 0.0126716373488307, | |
| "learning_rate": 7.626755570937758e-06, | |
| "loss": 0.0003, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 7.397908857762872, | |
| "grad_norm": 0.0033882916904985905, | |
| "learning_rate": 7.346251025449738e-06, | |
| "loss": 0.0003, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 7.397908857762872, | |
| "eval_all_ligands_equal": 0.5922450424929179, | |
| "eval_e3_equal": 0.8269298866855525, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9876062322946175, | |
| "eval_e3_heavy_atoms_difference": 0.2445113314447592, | |
| "eval_e3_heavy_atoms_difference_norm": 0.0028633304101074964, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9876062322946175, | |
| "eval_has_all_attachment_points": 0.9893767705382436, | |
| "eval_has_three_substructures": 0.9987606232294618, | |
| "eval_heavy_atoms_difference": 4.683958923512748, | |
| "eval_heavy_atoms_difference_norm": 0.06305476454593091, | |
| "eval_linker_equal": 0.8430417847025495, | |
| "eval_linker_graph_edit_distance": 2.8328611898017e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9971671388101983, | |
| "eval_linker_heavy_atoms_difference": 0.24026203966005666, | |
| "eval_linker_heavy_atoms_difference_norm": 0.004181593662072319, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9971671388101983, | |
| "eval_loss": 0.39461877942085266, | |
| "eval_num_fragments": 2.9998229461756374, | |
| "eval_poi_equal": 0.7895715297450425, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9584808781869688, | |
| "eval_poi_heavy_atoms_difference": 1.2901912181303117, | |
| "eval_poi_heavy_atoms_difference_norm": 0.03941144643806612, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9584808781869688, | |
| "eval_reassembly": 0.5990616147308782, | |
| "eval_reassembly_nostereo": 0.6306657223796034, | |
| "eval_runtime": 2312.2916, | |
| "eval_samples_per_second": 4.885, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9463526912181303, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 7.447228250147958, | |
| "grad_norm": 0.01033720001578331, | |
| "learning_rate": 7.0701110028643575e-06, | |
| "loss": 0.0003, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 7.496547642533044, | |
| "grad_norm": 0.011238239705562592, | |
| "learning_rate": 6.798403773068926e-06, | |
| "loss": 0.0003, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 7.54586703491813, | |
| "grad_norm": 0.01063668355345726, | |
| "learning_rate": 6.531726389498183e-06, | |
| "loss": 0.0003, | |
| "step": 76500 | |
| }, | |
| { | |
| "epoch": 7.595186427303216, | |
| "grad_norm": 0.008092471398413181, | |
| "learning_rate": 6.269075957435105e-06, | |
| "loss": 0.0003, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 7.644505819688302, | |
| "grad_norm": 0.0076515693217515945, | |
| "learning_rate": 6.011056357445671e-06, | |
| "loss": 0.0003, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 7.644505819688302, | |
| "eval_all_ligands_equal": 0.5951664305949008, | |
| "eval_e3_equal": 0.8265757790368272, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9892882436260623, | |
| "eval_e3_heavy_atoms_difference": 0.2443342776203966, | |
| "eval_e3_heavy_atoms_difference_norm": 0.0019018359231366443, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9892882436260623, | |
| "eval_has_all_attachment_points": 0.9888456090651558, | |
| "eval_has_three_substructures": 0.9992917847025495, | |
| "eval_heavy_atoms_difference": 4.572237960339943, | |
| "eval_heavy_atoms_difference_norm": 0.06191163540242748, | |
| "eval_linker_equal": 0.8482648725212465, | |
| "eval_linker_graph_edit_distance": 2.301699716713881e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9976983002832861, | |
| "eval_linker_heavy_atoms_difference": 0.22016643059490085, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0030724834199448584, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9976983002832861, | |
| "eval_loss": 0.3914594054222107, | |
| "eval_num_fragments": 3.0003541076487252, | |
| "eval_poi_equal": 0.7890403682719547, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9580382436260623, | |
| "eval_poi_heavy_atoms_difference": 1.2062677053824362, | |
| "eval_poi_heavy_atoms_difference_norm": 0.03675441221373476, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9580382436260623, | |
| "eval_reassembly": 0.601628895184136, | |
| "eval_reassembly_nostereo": 0.6334100566572238, | |
| "eval_runtime": 2314.03, | |
| "eval_samples_per_second": 4.882, | |
| "eval_steps_per_second": 0.076, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9475035410764873, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 7.693825212073387, | |
| "grad_norm": 0.008383137173950672, | |
| "learning_rate": 5.757731379518924e-06, | |
| "loss": 0.0003, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 7.7431446044584735, | |
| "grad_norm": 0.006558096036314964, | |
| "learning_rate": 5.5096559996688554e-06, | |
| "loss": 0.0003, | |
| "step": 78500 | |
| }, | |
| { | |
| "epoch": 7.7924639968435585, | |
| "grad_norm": 0.012070530094206333, | |
| "learning_rate": 5.26589727980922e-06, | |
| "loss": 0.0003, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 7.841783389228644, | |
| "grad_norm": 0.005785902496427298, | |
| "learning_rate": 5.027017407123047e-06, | |
| "loss": 0.0003, | |
| "step": 79500 | |
| }, | |
| { | |
| "epoch": 7.89110278161373, | |
| "grad_norm": 0.010161773301661015, | |
| "learning_rate": 4.79307543969916e-06, | |
| "loss": 0.0003, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 7.89110278161373, | |
| "eval_all_ligands_equal": 0.5949893767705382, | |
| "eval_e3_equal": 0.8257790368271954, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9878718130311614, | |
| "eval_e3_heavy_atoms_difference": 0.10968484419263456, | |
| "eval_e3_heavy_atoms_difference_norm": -0.002931533414796491, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9878718130311614, | |
| "eval_has_all_attachment_points": 0.9905276203966006, | |
| "eval_has_three_substructures": 0.9989376770538244, | |
| "eval_heavy_atoms_difference": 4.5701133144475925, | |
| "eval_heavy_atoms_difference_norm": 0.06153108609413112, | |
| "eval_linker_equal": 0.8472025495750708, | |
| "eval_linker_graph_edit_distance": 2.6558073654390935e+61, | |
| "eval_linker_graph_edit_distance_norm": 0.02850188111677236, | |
| "eval_linker_has_attachment_point(s)": 0.9973441926345609, | |
| "eval_linker_heavy_atoms_difference": 0.1844900849858357, | |
| "eval_linker_heavy_atoms_difference_norm": -0.0007419158301506531, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9973441926345609, | |
| "eval_loss": 0.3949120342731476, | |
| "eval_num_fragments": 3.0003541076487252, | |
| "eval_poi_equal": 0.7887747875354107, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9589235127478754, | |
| "eval_poi_heavy_atoms_difference": 1.2103399433427762, | |
| "eval_poi_heavy_atoms_difference_norm": 0.034957290196438014, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9589235127478754, | |
| "eval_reassembly": 0.6022485835694051, | |
| "eval_reassembly_nostereo": 0.6329674220963173, | |
| "eval_runtime": 2318.4828, | |
| "eval_samples_per_second": 4.872, | |
| "eval_steps_per_second": 0.076, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9466182719546742, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 7.940422173998816, | |
| "grad_norm": 0.021738652139902115, | |
| "learning_rate": 4.564582083701799e-06, | |
| "loss": 0.0003, | |
| "step": 80500 | |
| }, | |
| { | |
| "epoch": 7.989741566383902, | |
| "grad_norm": 0.01311516109853983, | |
| "learning_rate": 4.340678043241161e-06, | |
| "loss": 0.0003, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 8.039060958768989, | |
| "grad_norm": 0.03375653550028801, | |
| "learning_rate": 4.121881591200602e-06, | |
| "loss": 0.0002, | |
| "step": 81500 | |
| }, | |
| { | |
| "epoch": 8.088380351154074, | |
| "grad_norm": 0.01043145451694727, | |
| "learning_rate": 3.908246820460176e-06, | |
| "loss": 0.0002, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 8.137699743539159, | |
| "grad_norm": 0.008790087886154652, | |
| "learning_rate": 3.6998265477814154e-06, | |
| "loss": 0.0002, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 8.137699743539159, | |
| "eval_all_ligands_equal": 0.5950779036827195, | |
| "eval_e3_equal": 0.8269298866855525, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9907932011331445, | |
| "eval_e3_heavy_atoms_difference": 0.020095609065155808, | |
| "eval_e3_heavy_atoms_difference_norm": -0.005098161544823606, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9907932011331445, | |
| "eval_has_all_attachment_points": 0.9908817280453258, | |
| "eval_has_three_substructures": 0.9994688385269122, | |
| "eval_heavy_atoms_difference": 4.166873229461756, | |
| "eval_heavy_atoms_difference_norm": 0.056463252320257784, | |
| "eval_linker_equal": 0.8476451841359773, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9973441926345609, | |
| "eval_linker_heavy_atoms_difference": 0.18369334277620397, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0003880178606031887, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9973441926345609, | |
| "eval_loss": 0.39738067984580994, | |
| "eval_num_fragments": 3.0001770538243626, | |
| "eval_poi_equal": 0.7907223796033994, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9626416430594901, | |
| "eval_poi_heavy_atoms_difference": 1.1489907932011332, | |
| "eval_poi_heavy_atoms_difference_norm": 0.034174711553182684, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9626416430594901, | |
| "eval_reassembly": 0.6020715297450425, | |
| "eval_reassembly_nostereo": 0.6333215297450425, | |
| "eval_runtime": 2310.7745, | |
| "eval_samples_per_second": 4.888, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9522839943342776, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 8.187019135924245, | |
| "grad_norm": 0.010825219564139843, | |
| "learning_rate": 3.4970733201485363e-06, | |
| "loss": 0.0002, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 8.23633852830933, | |
| "grad_norm": 0.02623472549021244, | |
| "learning_rate": 3.2992246426767598e-06, | |
| "loss": 0.0002, | |
| "step": 83500 | |
| }, | |
| { | |
| "epoch": 8.285657920694417, | |
| "grad_norm": 0.00604247534647584, | |
| "learning_rate": 3.1067410313554478e-06, | |
| "loss": 0.0002, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 8.334977313079502, | |
| "grad_norm": 0.010682527907192707, | |
| "learning_rate": 2.9196700737611133e-06, | |
| "loss": 0.0002, | |
| "step": 84500 | |
| }, | |
| { | |
| "epoch": 8.384296705464589, | |
| "grad_norm": 0.006912824232131243, | |
| "learning_rate": 2.7384157653582875e-06, | |
| "loss": 0.0002, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 8.384296705464589, | |
| "eval_all_ligands_equal": 0.5959631728045326, | |
| "eval_e3_equal": 0.8273725212464589, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9899964589235127, | |
| "eval_e3_heavy_atoms_difference": 0.11968838526912182, | |
| "eval_e3_heavy_atoms_difference_norm": -0.0009492417791004414, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9899964589235127, | |
| "eval_has_all_attachment_points": 0.9893767705382436, | |
| "eval_has_three_substructures": 0.9993803116147308, | |
| "eval_heavy_atoms_difference": 4.283817280453258, | |
| "eval_heavy_atoms_difference_norm": 0.05811594320539324, | |
| "eval_linker_equal": 0.8491501416430595, | |
| "eval_linker_graph_edit_distance": 2.9213881019830026e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9970786118980169, | |
| "eval_linker_heavy_atoms_difference": 0.1703257790368272, | |
| "eval_linker_heavy_atoms_difference_norm": -0.000978229223694918, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9970786118980169, | |
| "eval_loss": 0.39642128348350525, | |
| "eval_num_fragments": 3.000619688385269, | |
| "eval_poi_equal": 0.7903682719546742, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9614907932011332, | |
| "eval_poi_heavy_atoms_difference": 1.1409348441926346, | |
| "eval_poi_heavy_atoms_difference_norm": 0.033616156238003056, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9614907932011332, | |
| "eval_reassembly": 0.6024256373937678, | |
| "eval_reassembly_nostereo": 0.6345609065155807, | |
| "eval_runtime": 2317.1994, | |
| "eval_samples_per_second": 4.875, | |
| "eval_steps_per_second": 0.076, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9502478753541076, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 8.433616097849674, | |
| "grad_norm": 0.011240696534514427, | |
| "learning_rate": 2.5622964623348843e-06, | |
| "loss": 0.0002, | |
| "step": 85500 | |
| }, | |
| { | |
| "epoch": 8.48293549023476, | |
| "grad_norm": 0.01842450723052025, | |
| "learning_rate": 2.3917244156542534e-06, | |
| "loss": 0.0002, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 8.532254882619846, | |
| "grad_norm": 0.005196568090468645, | |
| "learning_rate": 2.2267417957134883e-06, | |
| "loss": 0.0002, | |
| "step": 86500 | |
| }, | |
| { | |
| "epoch": 8.581574275004932, | |
| "grad_norm": 0.009838773868978024, | |
| "learning_rate": 2.0673893910399504e-06, | |
| "loss": 0.0002, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 8.630893667390017, | |
| "grad_norm": 0.04019368812441826, | |
| "learning_rate": 1.914008280031923e-06, | |
| "loss": 0.0002, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 8.630893667390017, | |
| "eval_all_ligands_equal": 0.5953434844192634, | |
| "eval_e3_equal": 0.8263101983002833, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9887570821529745, | |
| "eval_e3_heavy_atoms_difference": 0.1730701133144476, | |
| "eval_e3_heavy_atoms_difference_norm": 0.0008003837390301783, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9887570821529745, | |
| "eval_has_all_attachment_points": 0.9884029745042493, | |
| "eval_has_three_substructures": 0.9995573654390935, | |
| "eval_heavy_atoms_difference": 4.526203966005665, | |
| "eval_heavy_atoms_difference_norm": 0.061568743782646304, | |
| "eval_linker_equal": 0.8501239376770539, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9972556657223796, | |
| "eval_linker_heavy_atoms_difference": 0.20635623229461755, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0013172185920579528, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9972556657223796, | |
| "eval_loss": 0.3990042209625244, | |
| "eval_num_fragments": 3.0000885269121813, | |
| "eval_poi_equal": 0.7899256373937678, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9604284702549575, | |
| "eval_poi_heavy_atoms_difference": 1.170325779036827, | |
| "eval_poi_heavy_atoms_difference_norm": 0.03487419309053563, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9604284702549575, | |
| "eval_reassembly": 0.6017174220963173, | |
| "eval_reassembly_nostereo": 0.632878895184136, | |
| "eval_runtime": 2317.1407, | |
| "eval_samples_per_second": 4.875, | |
| "eval_steps_per_second": 0.076, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9478576487252125, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 8.680213059775104, | |
| "grad_norm": 0.009999396279454231, | |
| "learning_rate": 1.7660216417237113e-06, | |
| "loss": 0.0002, | |
| "step": 88000 | |
| }, | |
| { | |
| "epoch": 8.72953245216019, | |
| "grad_norm": 0.0032589335460215807, | |
| "learning_rate": 1.6237791221752203e-06, | |
| "loss": 0.0002, | |
| "step": 88500 | |
| }, | |
| { | |
| "epoch": 8.778851844545276, | |
| "grad_norm": 0.007267610169947147, | |
| "learning_rate": 1.4873158878961374e-06, | |
| "loss": 0.0002, | |
| "step": 89000 | |
| }, | |
| { | |
| "epoch": 8.82817123693036, | |
| "grad_norm": 0.02816702425479889, | |
| "learning_rate": 1.3566656765878621e-06, | |
| "loss": 0.0002, | |
| "step": 89500 | |
| }, | |
| { | |
| "epoch": 8.877490629315448, | |
| "grad_norm": 0.009792421944439411, | |
| "learning_rate": 1.2318607888025413e-06, | |
| "loss": 0.0002, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 8.877490629315448, | |
| "eval_all_ligands_equal": 0.5956090651558074, | |
| "eval_e3_equal": 0.8270184135977338, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9861898016997167, | |
| "eval_e3_heavy_atoms_difference": 0.2705382436260623, | |
| "eval_e3_heavy_atoms_difference_norm": 0.0048639786608157655, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9861898016997167, | |
| "eval_has_all_attachment_points": 0.9884029745042493, | |
| "eval_has_three_substructures": 0.9995573654390935, | |
| "eval_heavy_atoms_difference": 4.702195467422096, | |
| "eval_heavy_atoms_difference_norm": 0.06408037436572298, | |
| "eval_linker_equal": 0.8497698300283286, | |
| "eval_linker_graph_edit_distance": 3.009915014164306e+61, | |
| "eval_linker_graph_edit_distance_norm": 0.02778354797312324, | |
| "eval_linker_has_attachment_point(s)": 0.9969900849858357, | |
| "eval_linker_heavy_atoms_difference": 0.1989199716713881, | |
| "eval_linker_heavy_atoms_difference_norm": 0.00136169255914937, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9969900849858357, | |
| "eval_loss": 0.39773058891296387, | |
| "eval_num_fragments": 3.000265580736544, | |
| "eval_poi_equal": 0.7905453257790368, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9607825779036827, | |
| "eval_poi_heavy_atoms_difference": 1.1535056657223797, | |
| "eval_poi_heavy_atoms_difference_norm": 0.03463791665241003, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9607825779036827, | |
| "eval_reassembly": 0.6021600566572238, | |
| "eval_reassembly_nostereo": 0.6329674220963173, | |
| "eval_runtime": 2308.9117, | |
| "eval_samples_per_second": 4.892, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9456444759206799, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 8.926810021700533, | |
| "grad_norm": 0.007008455228060484, | |
| "learning_rate": 1.1129320799574572e-06, | |
| "loss": 0.0002, | |
| "step": 90500 | |
| }, | |
| { | |
| "epoch": 8.97612941408562, | |
| "grad_norm": 0.02331043779850006, | |
| "learning_rate": 9.999089527066409e-07, | |
| "loss": 0.0002, | |
| "step": 91000 | |
| }, | |
| { | |
| "epoch": 9.025448806470704, | |
| "grad_norm": 0.006514385808259249, | |
| "learning_rate": 8.930275893781537e-07, | |
| "loss": 0.0002, | |
| "step": 91500 | |
| }, | |
| { | |
| "epoch": 9.07476819885579, | |
| "grad_norm": 0.0031390993390232325, | |
| "learning_rate": 7.91886040795814e-07, | |
| "loss": 0.0002, | |
| "step": 92000 | |
| }, | |
| { | |
| "epoch": 9.124087591240876, | |
| "grad_norm": 0.0029660104773938656, | |
| "learning_rate": 6.967294457752959e-07, | |
| "loss": 0.0002, | |
| "step": 92500 | |
| }, | |
| { | |
| "epoch": 9.124087591240876, | |
| "eval_all_ligands_equal": 0.5965828611898017, | |
| "eval_e3_equal": 0.8274610481586402, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.988314447592068, | |
| "eval_e3_heavy_atoms_difference": 0.2075070821529745, | |
| "eval_e3_heavy_atoms_difference_norm": 0.0022690262383540734, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.988314447592068, | |
| "eval_has_all_attachment_points": 0.9898194050991501, | |
| "eval_has_three_substructures": 0.9995573654390935, | |
| "eval_heavy_atoms_difference": 4.5609065155807365, | |
| "eval_heavy_atoms_difference_norm": 0.06199652946669472, | |
| "eval_linker_equal": 0.8513633144475921, | |
| "eval_linker_graph_edit_distance": 3.186968838526912e+61, | |
| "eval_linker_graph_edit_distance_norm": 0.027842974986650174, | |
| "eval_linker_has_attachment_point(s)": 0.9968130311614731, | |
| "eval_linker_heavy_atoms_difference": 0.1783817280453258, | |
| "eval_linker_heavy_atoms_difference_norm": -0.00041578205572987336, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9968130311614731, | |
| "eval_loss": 0.39806732535362244, | |
| "eval_num_fragments": 3.000265580736544, | |
| "eval_poi_equal": 0.7913420679886686, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9606055240793201, | |
| "eval_poi_heavy_atoms_difference": 1.1666076487252124, | |
| "eval_poi_heavy_atoms_difference_norm": 0.035157453659247065, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9606055240793201, | |
| "eval_reassembly": 0.6030453257790368, | |
| "eval_reassembly_nostereo": 0.6340297450424929, | |
| "eval_runtime": 2310.5417, | |
| "eval_samples_per_second": 4.889, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9470609065155807, | |
| "step": 92500 | |
| }, | |
| { | |
| "epoch": 9.173406983625961, | |
| "grad_norm": 0.013469894416630268, | |
| "learning_rate": 6.07581329809051e-07, | |
| "loss": 0.0002, | |
| "step": 93000 | |
| }, | |
| { | |
| "epoch": 9.222726376011048, | |
| "grad_norm": 0.00611234363168478, | |
| "learning_rate": 5.246239357464988e-07, | |
| "loss": 0.0002, | |
| "step": 93500 | |
| }, | |
| { | |
| "epoch": 9.272045768396133, | |
| "grad_norm": 0.0028775690589100122, | |
| "learning_rate": 4.4754528537962746e-07, | |
| "loss": 0.0002, | |
| "step": 94000 | |
| }, | |
| { | |
| "epoch": 9.32136516078122, | |
| "grad_norm": 0.01138852909207344, | |
| "learning_rate": 3.765367196770675e-07, | |
| "loss": 0.0002, | |
| "step": 94500 | |
| }, | |
| { | |
| "epoch": 9.370684553166305, | |
| "grad_norm": 0.0019448177190497518, | |
| "learning_rate": 3.1161579403239385e-07, | |
| "loss": 0.0002, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 9.370684553166305, | |
| "eval_all_ligands_equal": 0.5964058073654391, | |
| "eval_e3_equal": 0.8270184135977338, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9884029745042493, | |
| "eval_e3_heavy_atoms_difference": 0.199185552407932, | |
| "eval_e3_heavy_atoms_difference_norm": 0.0013211183148759823, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9884029745042493, | |
| "eval_has_all_attachment_points": 0.9898194050991501, | |
| "eval_has_three_substructures": 0.9994688385269122, | |
| "eval_heavy_atoms_difference": 4.647751416430595, | |
| "eval_heavy_atoms_difference_norm": 0.06302557595947754, | |
| "eval_linker_equal": 0.8510092067988668, | |
| "eval_linker_graph_edit_distance": 3.452549575070822e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9965474504249292, | |
| "eval_linker_heavy_atoms_difference": 0.2017528328611898, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0011362630141366337, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9965474504249292, | |
| "eval_loss": 0.39840996265411377, | |
| "eval_num_fragments": 3.0003541076487252, | |
| "eval_poi_equal": 0.7906338526912181, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9596317280453258, | |
| "eval_poi_heavy_atoms_difference": 1.2269830028328612, | |
| "eval_poi_heavy_atoms_difference_norm": 0.037509484595162255, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9596317280453258, | |
| "eval_reassembly": 0.6029567988668555, | |
| "eval_reassembly_nostereo": 0.6342953257790368, | |
| "eval_runtime": 2306.5495, | |
| "eval_samples_per_second": 4.897, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9463526912181303, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 9.420003945551391, | |
| "grad_norm": 0.0025114284362643957, | |
| "learning_rate": 2.529100918672256e-07, | |
| "loss": 0.0002, | |
| "step": 95500 | |
| }, | |
| { | |
| "epoch": 9.469323337936476, | |
| "grad_norm": 0.001364166266284883, | |
| "learning_rate": 2.0019883840954324e-07, | |
| "loss": 0.0002, | |
| "step": 96000 | |
| }, | |
| { | |
| "epoch": 9.518642730321563, | |
| "grad_norm": 0.010334326885640621, | |
| "learning_rate": 1.5361882088972267e-07, | |
| "loss": 0.0002, | |
| "step": 96500 | |
| }, | |
| { | |
| "epoch": 9.567962122706648, | |
| "grad_norm": 0.008032439276576042, | |
| "learning_rate": 1.1318155524998597e-07, | |
| "loss": 0.0002, | |
| "step": 97000 | |
| }, | |
| { | |
| "epoch": 9.617281515091735, | |
| "grad_norm": 0.00466285552829504, | |
| "learning_rate": 7.889703876469978e-08, | |
| "loss": 0.0002, | |
| "step": 97500 | |
| }, | |
| { | |
| "epoch": 9.617281515091735, | |
| "eval_all_ligands_equal": 0.5960516997167139, | |
| "eval_e3_equal": 0.8272839943342776, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9888456090651558, | |
| "eval_e3_heavy_atoms_difference": 0.1958215297450425, | |
| "eval_e3_heavy_atoms_difference_norm": 0.0009669152686349697, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9888456090651558, | |
| "eval_has_all_attachment_points": 0.9904390934844193, | |
| "eval_has_three_substructures": 0.9994688385269122, | |
| "eval_heavy_atoms_difference": 4.558870396600566, | |
| "eval_heavy_atoms_difference_norm": 0.06187204058293781, | |
| "eval_linker_equal": 0.8502124645892352, | |
| "eval_linker_graph_edit_distance": 3.0984419263456094e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9969015580736544, | |
| "eval_linker_heavy_atoms_difference": 0.19821175637393768, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0008445360181919918, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9969015580736544, | |
| "eval_loss": 0.39866751432418823, | |
| "eval_num_fragments": 3.0003541076487252, | |
| "eval_poi_equal": 0.790899433427762, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9602514164305949, | |
| "eval_poi_heavy_atoms_difference": 1.1934313031161472, | |
| "eval_poi_heavy_atoms_difference_norm": 0.03607653809133688, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9602514164305949, | |
| "eval_reassembly": 0.6026026912181303, | |
| "eval_reassembly_nostereo": 0.6339412181303116, | |
| "eval_runtime": 2314.9163, | |
| "eval_samples_per_second": 4.88, | |
| "eval_steps_per_second": 0.076, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9475920679886686, | |
| "step": 97500 | |
| }, | |
| { | |
| "epoch": 9.66660090747682, | |
| "grad_norm": 0.0011036242358386517, | |
| "learning_rate": 5.077374756873865e-08, | |
| "loss": 0.0002, | |
| "step": 98000 | |
| }, | |
| { | |
| "epoch": 9.715920299861907, | |
| "grad_norm": 0.0038821862544864416, | |
| "learning_rate": 2.8818634561947334e-08, | |
| "loss": 0.0002, | |
| "step": 98500 | |
| }, | |
| { | |
| "epoch": 9.765239692246992, | |
| "grad_norm": 0.008210109546780586, | |
| "learning_rate": 1.3037127690179729e-08, | |
| "loss": 0.0002, | |
| "step": 99000 | |
| }, | |
| { | |
| "epoch": 9.814559084632076, | |
| "grad_norm": 0.004091315437108278, | |
| "learning_rate": 3.44616961621691e-09, | |
| "loss": 0.0002, | |
| "step": 99500 | |
| }, | |
| { | |
| "epoch": 9.863878477017163, | |
| "grad_norm": 0.0069528305903077126, | |
| "learning_rate": 9.691585167725947e-12, | |
| "loss": 0.0002, | |
| "step": 100000 | |
| }, | |
| { | |
| "epoch": 9.863878477017163, | |
| "eval_all_ligands_equal": 0.5960516997167139, | |
| "eval_e3_equal": 0.8272839943342776, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9886685552407932, | |
| "eval_e3_heavy_atoms_difference": 0.19573300283286119, | |
| "eval_e3_heavy_atoms_difference_norm": 0.001114461012300652, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9886685552407932, | |
| "eval_has_all_attachment_points": 0.9904390934844193, | |
| "eval_has_three_substructures": 0.9994688385269122, | |
| "eval_heavy_atoms_difference": 4.570024787535411, | |
| "eval_heavy_atoms_difference_norm": 0.0620230810464109, | |
| "eval_linker_equal": 0.8502124645892352, | |
| "eval_linker_graph_edit_distance": 3.009915014164306e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9969900849858357, | |
| "eval_linker_heavy_atoms_difference": 0.19785764872521247, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0009133688030441173, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9969900849858357, | |
| "eval_loss": 0.3986365497112274, | |
| "eval_num_fragments": 3.0003541076487252, | |
| "eval_poi_equal": 0.7909879603399433, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9601628895184136, | |
| "eval_poi_heavy_atoms_difference": 1.1950247875354107, | |
| "eval_poi_heavy_atoms_difference_norm": 0.03611503910692388, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9601628895184136, | |
| "eval_reassembly": 0.6025141643059491, | |
| "eval_reassembly_nostereo": 0.6337641643059491, | |
| "eval_runtime": 2312.0798, | |
| "eval_samples_per_second": 4.886, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.947415014164306, | |
| "step": 100000 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 100000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 10, | |
| "save_steps": 5000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 8.728946664336123e+17, | |
| "train_batch_size": 128, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |