| { | |
| "best_metric": 0.5964058073654391, | |
| "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine_restarts/checkpoint-100000", | |
| "epoch": 9.863878477017163, | |
| "eval_steps": 2500, | |
| "global_step": 100000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.049319392385085814, | |
| "grad_norm": 0.7575409412384033, | |
| "learning_rate": 4.465132736856968e-05, | |
| "loss": 1.4495, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.09863878477017163, | |
| "grad_norm": 0.6125224232673645, | |
| "learning_rate": 8.628481651367859e-06, | |
| "loss": 0.3014, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.14795817715525744, | |
| "grad_norm": 0.5499436259269714, | |
| "learning_rate": 4.3093350952626764e-05, | |
| "loss": 0.1515, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.19727756954034326, | |
| "grad_norm": 0.28163060545921326, | |
| "learning_rate": 6.906649047373284e-06, | |
| "loss": 0.0885, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.24659696192542907, | |
| "grad_norm": 0.26792627573013306, | |
| "learning_rate": 4.137151834863209e-05, | |
| "loss": 0.0576, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.2959163543105149, | |
| "grad_norm": 0.23422876000404358, | |
| "learning_rate": 5.348672631430346e-06, | |
| "loss": 0.0408, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.3452357466956007, | |
| "grad_norm": 0.25026798248291016, | |
| "learning_rate": 3.9501422739279976e-05, | |
| "loss": 0.0302, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.3945551390806865, | |
| "grad_norm": 0.18633298575878143, | |
| "learning_rate": 3.968661679220468e-06, | |
| "loss": 0.0237, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.4438745314657723, | |
| "grad_norm": 0.21757744252681732, | |
| "learning_rate": 3.750000000000008e-05, | |
| "loss": 0.0193, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.49319392385085814, | |
| "grad_norm": 0.14472977817058563, | |
| "learning_rate": 2.77911378362688e-06, | |
| "loss": 0.0156, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.49319392385085814, | |
| "eval_all_ligands_equal": 0.45963172804532576, | |
| "eval_e3_equal": 0.7815155807365439, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9915899433427762, | |
| "eval_e3_heavy_atoms_difference": 0.39128895184135976, | |
| "eval_e3_heavy_atoms_difference_norm": 0.006079141519917903, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9915899433427762, | |
| "eval_has_all_attachment_points": 0.986278328611898, | |
| "eval_has_three_substructures": 0.9988491501416431, | |
| "eval_heavy_atoms_difference": 7.0030099150141645, | |
| "eval_heavy_atoms_difference_norm": 0.09184213866701363, | |
| "eval_linker_equal": 0.6622698300283286, | |
| "eval_linker_graph_edit_distance": 3.5410764872521247e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9964589235127479, | |
| "eval_linker_heavy_atoms_difference": 0.38048866855524077, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0004655148396369131, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9964589235127479, | |
| "eval_loss": 0.28911474347114563, | |
| "eval_num_fragments": 3.000619688385269, | |
| "eval_poi_equal": 0.7265403682719547, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9251947592067988, | |
| "eval_poi_heavy_atoms_difference": 2.3129426345609065, | |
| "eval_poi_heavy_atoms_difference_norm": 0.07392839153429423, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9251947592067988, | |
| "eval_reassembly": 0.4665368271954674, | |
| "eval_reassembly_nostereo": 0.4910587818696884, | |
| "eval_runtime": 2426.7651, | |
| "eval_samples_per_second": 4.655, | |
| "eval_steps_per_second": 0.073, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9168732294617564, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.542513316235944, | |
| "grad_norm": 0.21385140717029572, | |
| "learning_rate": 3.5385375325047186e-05, | |
| "loss": 0.0135, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.5918327086210298, | |
| "grad_norm": 0.1257910281419754, | |
| "learning_rate": 1.7908016745981804e-06, | |
| "loss": 0.011, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.6411521010061156, | |
| "grad_norm": 0.11779133975505829, | |
| "learning_rate": 3.3176699082935504e-05, | |
| "loss": 0.0099, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.6904714933912014, | |
| "grad_norm": 0.13609211146831512, | |
| "learning_rate": 1.0126756596375464e-06, | |
| "loss": 0.0089, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.7397908857762873, | |
| "grad_norm": 0.13676239550113678, | |
| "learning_rate": 3.089397338773558e-05, | |
| "loss": 0.0077, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.7397908857762873, | |
| "eval_all_ligands_equal": 0.4898194050991501, | |
| "eval_e3_equal": 0.7908109065155807, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9856586402266289, | |
| "eval_e3_heavy_atoms_difference": 0.48450779036827196, | |
| "eval_e3_heavy_atoms_difference_norm": 0.008794338048794908, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9856586402266289, | |
| "eval_has_all_attachment_points": 0.9812322946175638, | |
| "eval_has_three_substructures": 0.9994688385269122, | |
| "eval_heavy_atoms_difference": 8.28390580736544, | |
| "eval_heavy_atoms_difference_norm": 0.10934640012787276, | |
| "eval_linker_equal": 0.7095432011331445, | |
| "eval_linker_graph_edit_distance": 5.6657223796034e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9943342776203966, | |
| "eval_linker_heavy_atoms_difference": 0.448300283286119, | |
| "eval_linker_heavy_atoms_difference_norm": 0.008959724750780267, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9943342776203966, | |
| "eval_loss": 0.30725741386413574, | |
| "eval_num_fragments": 3.0001770538243626, | |
| "eval_poi_equal": 0.736278328611898, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9160764872521246, | |
| "eval_poi_heavy_atoms_difference": 2.5749822946175636, | |
| "eval_poi_heavy_atoms_difference_norm": 0.0832007262765964, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9160764872521246, | |
| "eval_reassembly": 0.49752124645892354, | |
| "eval_reassembly_nostereo": 0.5220432011331445, | |
| "eval_runtime": 2398.3138, | |
| "eval_samples_per_second": 4.71, | |
| "eval_steps_per_second": 0.074, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.8987252124645893, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.789110278161373, | |
| "grad_norm": 0.12578321993350983, | |
| "learning_rate": 4.517825684323074e-07, | |
| "loss": 0.0066, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.8384296705464589, | |
| "grad_norm": 0.12456193566322327, | |
| "learning_rate": 2.8557870956831956e-05, | |
| "loss": 0.0065, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.8877490629315447, | |
| "grad_norm": 0.1042216420173645, | |
| "learning_rate": 1.1320193567289083e-07, | |
| "loss": 0.0054, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.9370684553166305, | |
| "grad_norm": 0.142706498503685, | |
| "learning_rate": 2.6189547895593596e-05, | |
| "loss": 0.0055, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.9863878477017163, | |
| "grad_norm": 0.10423340648412704, | |
| "learning_rate": 0.0, | |
| "loss": 0.0046, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.9863878477017163, | |
| "eval_all_ligands_equal": 0.5462110481586402, | |
| "eval_e3_equal": 0.8045325779036827, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9896423512747875, | |
| "eval_e3_heavy_atoms_difference": 0.5553293201133145, | |
| "eval_e3_heavy_atoms_difference_norm": 0.013095491925927988, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9896423512747875, | |
| "eval_has_all_attachment_points": 0.9836225212464589, | |
| "eval_has_three_substructures": 0.9991147308781869, | |
| "eval_heavy_atoms_difference": 7.0101805949008495, | |
| "eval_heavy_atoms_difference_norm": 0.09392218371981209, | |
| "eval_linker_equal": 0.7666430594900849, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9961048158640227, | |
| "eval_linker_heavy_atoms_difference": 0.31444759206798867, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0032526462468232837, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9961048158640227, | |
| "eval_loss": 0.3085635304450989, | |
| "eval_num_fragments": 2.9998229461756374, | |
| "eval_poi_equal": 0.7679709631728046, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9272308781869688, | |
| "eval_poi_heavy_atoms_difference": 2.120839235127479, | |
| "eval_poi_heavy_atoms_difference_norm": 0.07192788253687527, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9272308781869688, | |
| "eval_reassembly": 0.5543555240793201, | |
| "eval_reassembly_nostereo": 0.5795856940509915, | |
| "eval_runtime": 2370.5232, | |
| "eval_samples_per_second": 4.765, | |
| "eval_steps_per_second": 0.075, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9157223796033994, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 1.0357072400868022, | |
| "grad_norm": 0.046021342277526855, | |
| "learning_rate": 5.441173572077518e-07, | |
| "loss": 0.0031, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 1.085026632471888, | |
| "grad_norm": 0.13969920575618744, | |
| "learning_rate": 4.9120608413395366e-05, | |
| "loss": 0.0069, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 1.1343460248569737, | |
| "grad_norm": 0.09333090484142303, | |
| "learning_rate": 4.3443815219214587e-05, | |
| "loss": 0.0065, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 1.1836654172420595, | |
| "grad_norm": 0.0680941715836525, | |
| "learning_rate": 3.3682631321120504e-05, | |
| "loss": 0.0036, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 1.2329848096271454, | |
| "grad_norm": 0.06363417953252792, | |
| "learning_rate": 2.1998675033908916e-05, | |
| "loss": 0.0028, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 1.2823042020122313, | |
| "grad_norm": 0.04241478443145752, | |
| "learning_rate": 1.0979363433559891e-05, | |
| "loss": 0.0022, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 1.3316235943973171, | |
| "grad_norm": 0.04060855880379677, | |
| "learning_rate": 3.0649277482143607e-06, | |
| "loss": 0.0018, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.3809429867824028, | |
| "grad_norm": 0.05137120559811592, | |
| "learning_rate": 8.023117589237017e-09, | |
| "loss": 0.0016, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 1.4302623791674887, | |
| "grad_norm": 0.06896722316741943, | |
| "learning_rate": 4.7514397631165766e-05, | |
| "loss": 0.0061, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 1.4795817715525745, | |
| "grad_norm": 0.0655803456902504, | |
| "learning_rate": 4.005099547419455e-05, | |
| "loss": 0.0034, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 1.5289011639376602, | |
| "grad_norm": 0.060984883457422256, | |
| "learning_rate": 2.92545439850333e-05, | |
| "loss": 0.0025, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 1.578220556322746, | |
| "grad_norm": 0.027485696598887444, | |
| "learning_rate": 1.751592192566606e-05, | |
| "loss": 0.0019, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 1.627539948707832, | |
| "grad_norm": 0.027819139882922173, | |
| "learning_rate": 7.434652130505285e-06, | |
| "loss": 0.0015, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 1.6768593410929178, | |
| "grad_norm": 0.02593580074608326, | |
| "learning_rate": 1.2432360720283142e-06, | |
| "loss": 0.0013, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 1.7261787334780037, | |
| "grad_norm": 0.1803756207227707, | |
| "learning_rate": 4.9687234537299765e-05, | |
| "loss": 0.0028, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 1.7754981258630895, | |
| "grad_norm": 0.0707838386297226, | |
| "learning_rate": 4.5150706586350105e-05, | |
| "loss": 0.0046, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 1.8248175182481752, | |
| "grad_norm": 0.04294075071811676, | |
| "learning_rate": 3.6151796114194655e-05, | |
| "loss": 0.0022, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 1.874136910633261, | |
| "grad_norm": 0.033499088138341904, | |
| "learning_rate": 2.4683315662109008e-05, | |
| "loss": 0.0017, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 1.9234563030183467, | |
| "grad_norm": 0.028059741482138634, | |
| "learning_rate": 1.3284965090982499e-05, | |
| "loss": 0.0013, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 1.9727756954034326, | |
| "grad_norm": 0.016390886157751083, | |
| "learning_rate": 4.480913969818085e-06, | |
| "loss": 0.0011, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 2.0220950877885184, | |
| "grad_norm": 0.02433215267956257, | |
| "learning_rate": 2.2082315395764165e-07, | |
| "loss": 0.0009, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 2.0714144801736043, | |
| "grad_norm": 0.04352913051843643, | |
| "learning_rate": 4.855190843835337e-05, | |
| "loss": 0.0041, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 2.12073387255869, | |
| "grad_norm": 0.03815969452261925, | |
| "learning_rate": 4.210905991658837e-05, | |
| "loss": 0.0028, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 2.170053264943776, | |
| "grad_norm": 0.020274870097637177, | |
| "learning_rate": 3.1877402818861966e-05, | |
| "loss": 0.0016, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 2.219372657328862, | |
| "grad_norm": 0.021578455343842506, | |
| "learning_rate": 2.0122741949596797e-05, | |
| "loss": 0.0012, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 2.2686920497139473, | |
| "grad_norm": 0.018344903364777565, | |
| "learning_rate": 9.448151944460648e-06, | |
| "loss": 0.001, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 2.318011442099033, | |
| "grad_norm": 0.023422742262482643, | |
| "learning_rate": 2.2175252881083937e-06, | |
| "loss": 0.0009, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 2.367330834484119, | |
| "grad_norm": 0.04089091718196869, | |
| "learning_rate": 4.996791267927632e-05, | |
| "loss": 0.0008, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 2.416650226869205, | |
| "grad_norm": 0.040251534432172775, | |
| "learning_rate": 4.663610541397389e-05, | |
| "loss": 0.0039, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 2.465969619254291, | |
| "grad_norm": 0.04687808081507683, | |
| "learning_rate": 3.8511781678751205e-05, | |
| "loss": 0.0018, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 2.5152890116393767, | |
| "grad_norm": 0.053063150495290756, | |
| "learning_rate": 2.739526815925299e-05, | |
| "loss": 0.0013, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 2.5646084040244626, | |
| "grad_norm": 0.026560774073004723, | |
| "learning_rate": 1.574832148984774e-05, | |
| "loss": 0.001, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 2.6139277964095484, | |
| "grad_norm": 0.011897539719939232, | |
| "learning_rate": 6.150162950953281e-06, | |
| "loss": 0.0008, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 2.6632471887946343, | |
| "grad_norm": 0.026670893654227257, | |
| "learning_rate": 7.263088224173791e-07, | |
| "loss": 0.0007, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 2.7125665811797197, | |
| "grad_norm": 0.08923093229532242, | |
| "learning_rate": 4.932760869209433e-05, | |
| "loss": 0.0021, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 2.7618859735648056, | |
| "grad_norm": 0.03337786719202995, | |
| "learning_rate": 4.401529007475872e-05, | |
| "loss": 0.003, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 2.8112053659498915, | |
| "grad_norm": 0.01894857920706272, | |
| "learning_rate": 3.448656676980336e-05, | |
| "loss": 0.0014, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 2.8605247583349773, | |
| "grad_norm": 0.02331797406077385, | |
| "learning_rate": 2.2857039229600964e-05, | |
| "loss": 0.001, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 2.909844150720063, | |
| "grad_norm": 0.01910516433417797, | |
| "learning_rate": 1.1702071261018568e-05, | |
| "loss": 0.0008, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 2.959163543105149, | |
| "grad_norm": 0.021712150424718857, | |
| "learning_rate": 3.4919352524861338e-06, | |
| "loss": 0.0007, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 3.008482935490235, | |
| "grad_norm": 0.02103266492486, | |
| "learning_rate": 4.476930572223925e-08, | |
| "loss": 0.0006, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 3.0578023278753204, | |
| "grad_norm": 0.05578412860631943, | |
| "learning_rate": 4.787605021812086e-05, | |
| "loss": 0.0031, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 3.1071217202604062, | |
| "grad_norm": 0.028496012091636658, | |
| "learning_rate": 4.073095871744838e-05, | |
| "loss": 0.0016, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 3.156441112645492, | |
| "grad_norm": 0.01569642685353756, | |
| "learning_rate": 3.01022397359359e-05, | |
| "loss": 0.001, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 3.205760505030578, | |
| "grad_norm": 0.043978385627269745, | |
| "learning_rate": 1.8343627600599105e-05, | |
| "loss": 0.0008, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 3.255079897415664, | |
| "grad_norm": 0.045741546899080276, | |
| "learning_rate": 8.05907195665312e-06, | |
| "loss": 0.0006, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 3.3043992898007497, | |
| "grad_norm": 0.01765783689916134, | |
| "learning_rate": 1.526092014020153e-06, | |
| "loss": 0.0006, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 3.3537186821858356, | |
| "grad_norm": 0.06937197595834732, | |
| "learning_rate": 4.980858104780184e-05, | |
| "loss": 0.0009, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 3.4030380745709214, | |
| "grad_norm": 0.03800417482852936, | |
| "learning_rate": 4.567612427657311e-05, | |
| "loss": 0.0025, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 3.4523574669560073, | |
| "grad_norm": 0.015555808320641518, | |
| "learning_rate": 3.6959108562141376e-05, | |
| "loss": 0.0012, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 3.501676859341093, | |
| "grad_norm": 0.021087462082505226, | |
| "learning_rate": 2.559374318373061e-05, | |
| "loss": 0.0009, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 3.5509962517261786, | |
| "grad_norm": 0.022866034880280495, | |
| "learning_rate": 1.409689312542472e-05, | |
| "loss": 0.0007, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 3.6003156441112645, | |
| "grad_norm": 0.042014122009277344, | |
| "learning_rate": 5.014540709444851e-06, | |
| "loss": 0.0006, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 3.6496350364963503, | |
| "grad_norm": 0.012297836132347584, | |
| "learning_rate": 3.5797673092126393e-07, | |
| "loss": 0.0005, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 3.698954428881436, | |
| "grad_norm": 0.06086747720837593, | |
| "learning_rate": 4.884160074185018e-05, | |
| "loss": 0.002, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 3.748273821266522, | |
| "grad_norm": 0.03979231417179108, | |
| "learning_rate": 4.279484393132507e-05, | |
| "loss": 0.0017, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 3.797593213651608, | |
| "grad_norm": 0.020537426695227623, | |
| "learning_rate": 3.2793161758248515e-05, | |
| "loss": 0.0011, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 3.8469126060366934, | |
| "grad_norm": 0.012818257324397564, | |
| "learning_rate": 2.1065680608053368e-05, | |
| "loss": 0.0007, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 3.8962319984217793, | |
| "grad_norm": 0.009783191606402397, | |
| "learning_rate": 1.0209456156090146e-05, | |
| "loss": 0.0006, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 3.945551390806865, | |
| "grad_norm": 0.01907140202820301, | |
| "learning_rate": 2.628603912310604e-06, | |
| "loss": 0.0005, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 3.994870783191951, | |
| "grad_norm": 0.02355390600860119, | |
| "learning_rate": 4.999809317820512e-05, | |
| "loss": 0.0005, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 4.044190175577037, | |
| "grad_norm": 0.030535969883203506, | |
| "learning_rate": 4.708895193055257e-05, | |
| "loss": 0.0023, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 4.093509567962123, | |
| "grad_norm": 0.030475422739982605, | |
| "learning_rate": 3.9288202908042935e-05, | |
| "loss": 0.0011, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 4.142828960347209, | |
| "grad_norm": 0.03414959833025932, | |
| "learning_rate": 2.8323325290683895e-05, | |
| "loss": 0.0008, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 4.1921483527322945, | |
| "grad_norm": 0.03184778243303299, | |
| "learning_rate": 1.6622495877992905e-05, | |
| "loss": 0.0006, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 4.24146774511738, | |
| "grad_norm": 0.036285772919654846, | |
| "learning_rate": 6.776868306745027e-06, | |
| "loss": 0.0005, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 4.290787137502466, | |
| "grad_norm": 0.020429372787475586, | |
| "learning_rate": 9.667609789756588e-07, | |
| "loss": 0.0005, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 4.340106529887552, | |
| "grad_norm": 0.10811227560043335, | |
| "learning_rate": 4.952579070517538e-05, | |
| "loss": 0.0014, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 4.389425922272638, | |
| "grad_norm": 0.05198413506150246, | |
| "learning_rate": 4.460839084373921e-05, | |
| "loss": 0.0019, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 4.438745314657724, | |
| "grad_norm": 0.03819139301776886, | |
| "learning_rate": 3.534411131835637e-05, | |
| "loss": 0.001, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 4.48806470704281, | |
| "grad_norm": 0.019216470420360565, | |
| "learning_rate": 2.3789123953690796e-05, | |
| "loss": 0.0007, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 4.537384099427895, | |
| "grad_norm": 0.02267817035317421, | |
| "learning_rate": 1.2502285603117597e-05, | |
| "loss": 0.0005, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 4.5867034918129805, | |
| "grad_norm": 0.01543174497783184, | |
| "learning_rate": 3.983071407096459e-06, | |
| "loss": 0.0005, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 4.636022884198066, | |
| "grad_norm": 0.020069127902388573, | |
| "learning_rate": 1.180649630719205e-07, | |
| "loss": 0.0004, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 4.685342276583152, | |
| "grad_norm": 0.06315533071756363, | |
| "learning_rate": 4.8236826412442815e-05, | |
| "loss": 0.0016, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 4.734661668968238, | |
| "grad_norm": 0.05815032124519348, | |
| "learning_rate": 4.1445912722173056e-05, | |
| "loss": 0.0011, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 4.783981061353324, | |
| "grad_norm": 0.03868298977613449, | |
| "learning_rate": 3.103609665799435e-05, | |
| "loss": 0.0009, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 4.83330045373841, | |
| "grad_norm": 0.026825400069355965, | |
| "learning_rate": 1.9271702808963772e-05, | |
| "loss": 0.0006, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 4.882619846123496, | |
| "grad_norm": 0.022513261064887047, | |
| "learning_rate": 8.775842802754727e-06, | |
| "loss": 0.0005, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 4.931939238508582, | |
| "grad_norm": 0.04292171820998192, | |
| "learning_rate": 1.8728292899447775e-06, | |
| "loss": 0.0004, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 4.9812586308936675, | |
| "grad_norm": 0.09393975138664246, | |
| "learning_rate": 4.9908662458888896e-05, | |
| "loss": 0.0005, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 5.030578023278753, | |
| "grad_norm": 0.019034981727600098, | |
| "learning_rate": 4.61993375092424e-05, | |
| "loss": 0.0019, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 5.079897415663839, | |
| "grad_norm": 0.008647086098790169, | |
| "learning_rate": 3.7791398273089494e-05, | |
| "loss": 0.0009, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 5.129216808048925, | |
| "grad_norm": 0.013575778342783451, | |
| "learning_rate": 2.655079846342193e-05, | |
| "loss": 0.0006, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 5.178536200434011, | |
| "grad_norm": 0.009713593870401382, | |
| "learning_rate": 1.496677367620188e-05, | |
| "loss": 0.0005, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 5.227855592819097, | |
| "grad_norm": 0.024831971153616905, | |
| "learning_rate": 5.604611114346173e-06, | |
| "loss": 0.0004, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 5.277174985204182, | |
| "grad_norm": 0.005304230377078056, | |
| "learning_rate": 5.37565648481686e-07, | |
| "loss": 0.0004, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 5.326494377589268, | |
| "grad_norm": 0.055664077401161194, | |
| "learning_rate": 4.911226335360363e-05, | |
| "loss": 0.0012, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 5.375813769974354, | |
| "grad_norm": 0.050345055758953094, | |
| "learning_rate": 4.342242140793571e-05, | |
| "loss": 0.0011, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 5.425133162359439, | |
| "grad_norm": 0.015438569709658623, | |
| "learning_rate": 3.3652926426937384e-05, | |
| "loss": 0.0008, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 5.474452554744525, | |
| "grad_norm": 0.007345585618168116, | |
| "learning_rate": 2.1967237218238608e-05, | |
| "loss": 0.0006, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 5.523771947129611, | |
| "grad_norm": 0.016498006880283356, | |
| "learning_rate": 1.0953154614018291e-05, | |
| "loss": 0.0005, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 5.573091339514697, | |
| "grad_norm": 0.023072505369782448, | |
| "learning_rate": 3.049751879006049e-06, | |
| "loss": 0.0004, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 5.622410731899783, | |
| "grad_norm": 0.018147263675928116, | |
| "learning_rate": 7.823803658785211e-09, | |
| "loss": 0.0004, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 5.671730124284869, | |
| "grad_norm": 0.03701692819595337, | |
| "learning_rate": 4.752127641360646e-05, | |
| "loss": 0.0017, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 5.721049516669955, | |
| "grad_norm": 0.03156555816531181, | |
| "learning_rate": 4.006363587728823e-05, | |
| "loss": 0.0009, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 5.7703689090550405, | |
| "grad_norm": 0.003923078067600727, | |
| "learning_rate": 2.9270146786152846e-05, | |
| "loss": 0.0006, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 5.819688301440126, | |
| "grad_norm": 0.012611697427928448, | |
| "learning_rate": 1.7531031877885103e-05, | |
| "loss": 0.0005, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 5.869007693825212, | |
| "grad_norm": 0.008357984013855457, | |
| "learning_rate": 7.445923128495683e-06, | |
| "loss": 0.0004, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 5.918327086210298, | |
| "grad_norm": 0.01403570082038641, | |
| "learning_rate": 1.2481721485038734e-06, | |
| "loss": 0.0004, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 5.967646478595384, | |
| "grad_norm": 0.09385232627391815, | |
| "learning_rate": 4.968972647856831e-05, | |
| "loss": 0.0008, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 6.01696587098047, | |
| "grad_norm": 0.03150836378335953, | |
| "learning_rate": 4.5188130514991015e-05, | |
| "loss": 0.0014, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 6.066285263365556, | |
| "grad_norm": 0.01714126579463482, | |
| "learning_rate": 3.620844809525939e-05, | |
| "loss": 0.0007, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 6.115604655750641, | |
| "grad_norm": 0.007772900629788637, | |
| "learning_rate": 2.4746650090359677e-05, | |
| "loss": 0.0005, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 6.164924048135727, | |
| "grad_norm": 0.006022776942700148, | |
| "learning_rate": 1.3340956530417454e-05, | |
| "loss": 0.0004, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 6.2142434405208125, | |
| "grad_norm": 0.009159950539469719, | |
| "learning_rate": 4.517163092423821e-06, | |
| "loss": 0.0004, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 6.263562832905898, | |
| "grad_norm": 0.010325823910534382, | |
| "learning_rate": 2.2930257615656205e-07, | |
| "loss": 0.0003, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 6.312882225290984, | |
| "grad_norm": 0.04627057537436485, | |
| "learning_rate": 4.857307648820405e-05, | |
| "loss": 0.0013, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 6.36220161767607, | |
| "grad_norm": 0.028987212106585503, | |
| "learning_rate": 4.2172457334497206e-05, | |
| "loss": 0.0011, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 6.411521010061156, | |
| "grad_norm": 0.009136197157204151, | |
| "learning_rate": 3.198390069020853e-05, | |
| "loss": 0.0007, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 6.460840402446242, | |
| "grad_norm": 0.013344330713152885, | |
| "learning_rate": 2.0231502206670133e-05, | |
| "loss": 0.0005, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 6.510159794831328, | |
| "grad_norm": 0.015842510387301445, | |
| "learning_rate": 9.535089582196754e-06, | |
| "loss": 0.0004, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 6.5594791872164135, | |
| "grad_norm": 0.003152635879814625, | |
| "learning_rate": 2.263387929753233e-06, | |
| "loss": 0.0004, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 6.608798579601499, | |
| "grad_norm": 0.05013451725244522, | |
| "learning_rate": 4.997328133516944e-05, | |
| "loss": 0.0004, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 6.658117971986585, | |
| "grad_norm": 0.0076034897938370705, | |
| "learning_rate": 4.669142524548852e-05, | |
| "loss": 0.002, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 6.707437364371671, | |
| "grad_norm": 0.019150741398334503, | |
| "learning_rate": 3.860490727350733e-05, | |
| "loss": 0.0009, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 6.756756756756757, | |
| "grad_norm": 0.011379432864487171, | |
| "learning_rate": 2.750557681478918e-05, | |
| "loss": 0.0005, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 6.806076149141843, | |
| "grad_norm": 0.042575038969516754, | |
| "learning_rate": 1.585138530760425e-05, | |
| "loss": 0.0004, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 6.855395541526929, | |
| "grad_norm": 0.007366618607193232, | |
| "learning_rate": 6.223158404769164e-06, | |
| "loss": 0.0004, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 6.904714933912015, | |
| "grad_norm": 0.006865040399134159, | |
| "learning_rate": 7.530710380818595e-07, | |
| "loss": 0.0003, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 6.904714933912015, | |
| "eval_all_ligands_equal": 0.5913597733711048, | |
| "eval_e3_equal": 0.8265757790368272, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9850389518413598, | |
| "eval_e3_heavy_atoms_difference": 0.5427584985835694, | |
| "eval_e3_heavy_atoms_difference_norm": 0.013521375343157706, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9850389518413598, | |
| "eval_has_all_attachment_points": 0.9859242209631728, | |
| "eval_has_three_substructures": 1.0, | |
| "eval_heavy_atoms_difference": 7.366501416430595, | |
| "eval_heavy_atoms_difference_norm": 0.09873397870398803, | |
| "eval_linker_equal": 0.8408286118980169, | |
| "eval_linker_graph_edit_distance": 2.8328611898017003e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9971671388101983, | |
| "eval_linker_heavy_atoms_difference": 0.20856940509915015, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0012705826125932818, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9971671388101983, | |
| "eval_loss": 0.34158065915107727, | |
| "eval_num_fragments": 3.0, | |
| "eval_poi_equal": 0.7900141643059491, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9246635977337111, | |
| "eval_poi_heavy_atoms_difference": 2.174929178470255, | |
| "eval_poi_heavy_atoms_difference_norm": 0.07429925518971324, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9246635977337111, | |
| "eval_reassembly": 0.5992386685552408, | |
| "eval_reassembly_nostereo": 0.6233179886685553, | |
| "eval_runtime": 2295.0337, | |
| "eval_samples_per_second": 4.922, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9081975920679887, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 6.9540343262971, | |
| "grad_norm": 0.05451405048370361, | |
| "learning_rate": 4.9347523624767874e-05, | |
| "loss": 0.0008, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 7.0033537186821855, | |
| "grad_norm": 0.02272048220038414, | |
| "learning_rate": 4.40563488530207e-05, | |
| "loss": 0.0011, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 7.052673111067271, | |
| "grad_norm": 0.013967321254312992, | |
| "learning_rate": 3.4545137544325615e-05, | |
| "loss": 0.0006, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 7.101992503452357, | |
| "grad_norm": 0.01299150288105011, | |
| "learning_rate": 2.2920151477520506e-05, | |
| "loss": 0.0005, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 7.151311895837443, | |
| "grad_norm": 0.013562225736677647, | |
| "learning_rate": 1.1755748748262491e-05, | |
| "loss": 0.0004, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 7.151311895837443, | |
| "eval_all_ligands_equal": 0.5941926345609065, | |
| "eval_e3_equal": 0.8275495750708215, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9848618980169972, | |
| "eval_e3_heavy_atoms_difference": 0.5763101983002833, | |
| "eval_e3_heavy_atoms_difference_norm": 0.012249421889539655, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9848618980169972, | |
| "eval_has_all_attachment_points": 0.9845963172804533, | |
| "eval_has_three_substructures": 1.0, | |
| "eval_heavy_atoms_difference": 7.1870573654390935, | |
| "eval_heavy_atoms_difference_norm": 0.0960046616450944, | |
| "eval_linker_equal": 0.8450779036827195, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9970786118980169, | |
| "eval_linker_heavy_atoms_difference": 0.22237960339943344, | |
| "eval_linker_heavy_atoms_difference_norm": 0.002666489083035178, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9970786118980169, | |
| "eval_loss": 0.336382657289505, | |
| "eval_num_fragments": 3.0, | |
| "eval_poi_equal": 0.7933781869688386, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9270538243626062, | |
| "eval_poi_heavy_atoms_difference": 2.0790545325779037, | |
| "eval_poi_heavy_atoms_difference_norm": 0.07067489172114745, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9270538243626062, | |
| "eval_reassembly": 0.6018944759206799, | |
| "eval_reassembly_nostereo": 0.6242032577903682, | |
| "eval_runtime": 2299.7907, | |
| "eval_samples_per_second": 4.912, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9109419263456091, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 7.200631288222529, | |
| "grad_norm": 0.007629915606230497, | |
| "learning_rate": 3.524291076734654e-06, | |
| "loss": 0.0003, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 7.249950680607615, | |
| "grad_norm": 0.01034531369805336, | |
| "learning_rate": 4.863825597072047e-08, | |
| "loss": 0.0003, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 7.299270072992701, | |
| "grad_norm": 0.01756499893963337, | |
| "learning_rate": 4.7920534490309e-05, | |
| "loss": 0.0014, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 7.348589465377787, | |
| "grad_norm": 0.02920692414045334, | |
| "learning_rate": 4.081695196932368e-05, | |
| "loss": 0.0007, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 7.397908857762872, | |
| "grad_norm": 0.013725588098168373, | |
| "learning_rate": 3.02106987253934e-05, | |
| "loss": 0.0005, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 7.397908857762872, | |
| "eval_all_ligands_equal": 0.5900318696883853, | |
| "eval_e3_equal": 0.8248937677053825, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9838881019830028, | |
| "eval_e3_heavy_atoms_difference": 0.6882082152974505, | |
| "eval_e3_heavy_atoms_difference_norm": 0.01871438676379277, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9838881019830028, | |
| "eval_has_all_attachment_points": 0.9873406515580736, | |
| "eval_has_three_substructures": 0.9997344192634561, | |
| "eval_heavy_atoms_difference": 7.455913597733711, | |
| "eval_heavy_atoms_difference_norm": 0.0985007505593061, | |
| "eval_linker_equal": 0.8362252124645893, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9970786118980169, | |
| "eval_linker_heavy_atoms_difference": 0.2505311614730878, | |
| "eval_linker_heavy_atoms_difference_norm": 0.002953837656202626, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9970786118980169, | |
| "eval_loss": 0.3489196300506592, | |
| "eval_num_fragments": 2.999734419263456, | |
| "eval_poi_equal": 0.7870042492917847, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9274079320113314, | |
| "eval_poi_heavy_atoms_difference": 2.143059490084986, | |
| "eval_poi_heavy_atoms_difference_norm": 0.07038631697218085, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9274079320113314, | |
| "eval_reassembly": 0.5971140226628895, | |
| "eval_reassembly_nostereo": 0.6221671388101983, | |
| "eval_runtime": 2290.38, | |
| "eval_samples_per_second": 4.932, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9107648725212465, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 7.447228250147958, | |
| "grad_norm": 0.009434948675334454, | |
| "learning_rate": 1.8450534038416463e-05, | |
| "loss": 0.0004, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 7.496547642533044, | |
| "grad_norm": 0.00249391607940197, | |
| "learning_rate": 8.140751366827746e-06, | |
| "loss": 0.0004, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 7.54586703491813, | |
| "grad_norm": 0.0030246416572481394, | |
| "learning_rate": 1.5644564565212016e-06, | |
| "loss": 0.0003, | |
| "step": 76500 | |
| }, | |
| { | |
| "epoch": 7.595186427303216, | |
| "grad_norm": 0.06065337359905243, | |
| "learning_rate": 4.982202739503969e-05, | |
| "loss": 0.0005, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 7.644505819688302, | |
| "grad_norm": 0.004869398195296526, | |
| "learning_rate": 4.5711662750799136e-05, | |
| "loss": 0.0012, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 7.644505819688302, | |
| "eval_all_ligands_equal": 0.5848087818696884, | |
| "eval_e3_equal": 0.8249822946175638, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9820290368271954, | |
| "eval_e3_heavy_atoms_difference": 0.6423512747875354, | |
| "eval_e3_heavy_atoms_difference_norm": 0.014855219684020874, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9820290368271954, | |
| "eval_has_all_attachment_points": 0.9863668555240793, | |
| "eval_has_three_substructures": 0.9997344192634561, | |
| "eval_heavy_atoms_difference": 7.4127124645892355, | |
| "eval_heavy_atoms_difference_norm": 0.09827559741989533, | |
| "eval_linker_equal": 0.8247167138810199, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9964589235127479, | |
| "eval_linker_heavy_atoms_difference": 0.3464943342776204, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0071817416289111135, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9964589235127479, | |
| "eval_loss": 0.3375075161457062, | |
| "eval_num_fragments": 2.999734419263456, | |
| "eval_poi_equal": 0.7898371104815864, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9271423512747875, | |
| "eval_poi_heavy_atoms_difference": 2.2372521246458925, | |
| "eval_poi_heavy_atoms_difference_norm": 0.07599291909561608, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9271423512747875, | |
| "eval_reassembly": 0.5920679886685553, | |
| "eval_reassembly_nostereo": 0.6126947592067988, | |
| "eval_runtime": 2298.8398, | |
| "eval_samples_per_second": 4.914, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9095254957507082, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 7.693825212073387, | |
| "grad_norm": 0.008935153484344482, | |
| "learning_rate": 3.701469160478539e-05, | |
| "loss": 0.0007, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 7.7431446044584735, | |
| "grad_norm": 0.00356725649908185, | |
| "learning_rate": 2.5680805361916853e-05, | |
| "loss": 0.0005, | |
| "step": 78500 | |
| }, | |
| { | |
| "epoch": 7.7924639968435585, | |
| "grad_norm": 0.009323718026280403, | |
| "learning_rate": 1.417533070530511e-05, | |
| "loss": 0.0004, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 7.841783389228644, | |
| "grad_norm": 0.010333404876291752, | |
| "learning_rate": 5.066983655682358e-06, | |
| "loss": 0.0003, | |
| "step": 79500 | |
| }, | |
| { | |
| "epoch": 7.89110278161373, | |
| "grad_norm": 0.011676914989948273, | |
| "learning_rate": 3.7281153082609763e-07, | |
| "loss": 0.0003, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 7.89110278161373, | |
| "eval_all_ligands_equal": 0.595520538243626, | |
| "eval_e3_equal": 0.8263101983002833, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9783109065155807, | |
| "eval_e3_heavy_atoms_difference": 0.716979461756374, | |
| "eval_e3_heavy_atoms_difference_norm": 0.017889919151187116, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9783109065155807, | |
| "eval_has_all_attachment_points": 0.9868094900849859, | |
| "eval_has_three_substructures": 0.9994688385269122, | |
| "eval_heavy_atoms_difference": 7.486898016997167, | |
| "eval_heavy_atoms_difference_norm": 0.09974410367244255, | |
| "eval_linker_equal": 0.8450779036827195, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9970786118980169, | |
| "eval_linker_heavy_atoms_difference": 0.21441218130311615, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0019166933809747133, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9970786118980169, | |
| "eval_loss": 0.34107285737991333, | |
| "eval_num_fragments": 2.9998229461756374, | |
| "eval_poi_equal": 0.7927584985835694, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9295325779036827, | |
| "eval_poi_heavy_atoms_difference": 2.1622698300283285, | |
| "eval_poi_heavy_atoms_difference_norm": 0.07501641382843047, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9295325779036827, | |
| "eval_reassembly": 0.6029567988668555, | |
| "eval_reassembly_nostereo": 0.6278328611898017, | |
| "eval_runtime": 2289.8594, | |
| "eval_samples_per_second": 4.933, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9079320113314447, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 7.940422173998816, | |
| "grad_norm": 0.03382929041981697, | |
| "learning_rate": 4.887471587353222e-05, | |
| "loss": 0.0011, | |
| "step": 80500 | |
| }, | |
| { | |
| "epoch": 7.989741566383902, | |
| "grad_norm": 0.032275088131427765, | |
| "learning_rate": 4.2872523262296e-05, | |
| "loss": 0.0011, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 8.039060958768989, | |
| "grad_norm": 0.02223849669098854, | |
| "learning_rate": 3.289840422610638e-05, | |
| "loss": 0.0005, | |
| "step": 81500 | |
| }, | |
| { | |
| "epoch": 8.088380351154074, | |
| "grad_norm": 0.014037691988050938, | |
| "learning_rate": 2.1175180223648972e-05, | |
| "loss": 0.0004, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 8.137699743539159, | |
| "grad_norm": 0.012884391471743584, | |
| "learning_rate": 1.0298964183101326e-05, | |
| "loss": 0.0004, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 8.137699743539159, | |
| "eval_all_ligands_equal": 0.5924220963172805, | |
| "eval_e3_equal": 0.8275495750708215, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9795502832861189, | |
| "eval_e3_heavy_atoms_difference": 0.6832507082152974, | |
| "eval_e3_heavy_atoms_difference_norm": 0.017574528722493352, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9795502832861189, | |
| "eval_has_all_attachment_points": 0.986278328611898, | |
| "eval_has_three_substructures": 0.9999114730878187, | |
| "eval_heavy_atoms_difference": 7.748849150141643, | |
| "eval_heavy_atoms_difference_norm": 0.10325215265036036, | |
| "eval_linker_equal": 0.8377301699716714, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9966359773371105, | |
| "eval_linker_heavy_atoms_difference": 0.21414660056657223, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0009361178008556331, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9966359773371105, | |
| "eval_loss": 0.35050350427627563, | |
| "eval_num_fragments": 3.0000885269121813, | |
| "eval_poi_equal": 0.7926699716713881, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9267882436260623, | |
| "eval_poi_heavy_atoms_difference": 2.1771423512747874, | |
| "eval_poi_heavy_atoms_difference_norm": 0.0742126942729648, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9267882436260623, | |
| "eval_reassembly": 0.5996813031161473, | |
| "eval_reassembly_nostereo": 0.6242917847025495, | |
| "eval_runtime": 2288.2269, | |
| "eval_samples_per_second": 4.937, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9047450424929179, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 8.187019135924245, | |
| "grad_norm": 0.0096856988966465, | |
| "learning_rate": 2.678298760187492e-06, | |
| "loss": 0.0003, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 8.23633852830933, | |
| "grad_norm": 0.006260647438466549, | |
| "learning_rate": 4.999921645460451e-05, | |
| "loss": 0.0003, | |
| "step": 83500 | |
| }, | |
| { | |
| "epoch": 8.285657920694417, | |
| "grad_norm": 0.058494918048381805, | |
| "learning_rate": 4.7140644581193054e-05, | |
| "loss": 0.0012, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 8.334977313079502, | |
| "grad_norm": 0.008097327314317226, | |
| "learning_rate": 3.9379017573513076e-05, | |
| "loss": 0.0007, | |
| "step": 84500 | |
| }, | |
| { | |
| "epoch": 8.384296705464589, | |
| "grad_norm": 0.02641126699745655, | |
| "learning_rate": 2.8433151024792702e-05, | |
| "loss": 0.0005, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 8.384296705464589, | |
| "eval_all_ligands_equal": 0.5915368271954674, | |
| "eval_e3_equal": 0.8260446175637394, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9817634560906515, | |
| "eval_e3_heavy_atoms_difference": 0.6590828611898017, | |
| "eval_e3_heavy_atoms_difference_norm": 0.01584419899552303, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9817634560906515, | |
| "eval_has_all_attachment_points": 0.9820290368271954, | |
| "eval_has_three_substructures": 0.9996458923512748, | |
| "eval_heavy_atoms_difference": 8.116678470254957, | |
| "eval_heavy_atoms_difference_norm": 0.10816132188057066, | |
| "eval_linker_equal": 0.8387039660056658, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9963703966005666, | |
| "eval_linker_heavy_atoms_difference": 0.23919971671388102, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0021445197832858022, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9963703966005666, | |
| "eval_loss": 0.34624338150024414, | |
| "eval_num_fragments": 3.0, | |
| "eval_poi_equal": 0.7883321529745042, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9173158640226629, | |
| "eval_poi_heavy_atoms_difference": 2.409968130311615, | |
| "eval_poi_heavy_atoms_difference_norm": 0.08244812364841567, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9173158640226629, | |
| "eval_reassembly": 0.5984419263456091, | |
| "eval_reassembly_nostereo": 0.6215474504249292, | |
| "eval_runtime": 2286.0335, | |
| "eval_samples_per_second": 4.941, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.8986366855524079, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 8.433616097849674, | |
| "grad_norm": 0.017572874203324318, | |
| "learning_rate": 1.6727011725335507e-05, | |
| "loss": 0.0004, | |
| "step": 85500 | |
| }, | |
| { | |
| "epoch": 8.48293549023476, | |
| "grad_norm": 0.005754662211984396, | |
| "learning_rate": 6.852929188593271e-06, | |
| "loss": 0.0004, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 8.532254882619846, | |
| "grad_norm": 0.013183694332838058, | |
| "learning_rate": 9.975231807800883e-07, | |
| "loss": 0.0003, | |
| "step": 86500 | |
| }, | |
| { | |
| "epoch": 8.581574275004932, | |
| "grad_norm": 0.08640342950820923, | |
| "learning_rate": 4.954703613923131e-05, | |
| "loss": 0.0008, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 8.630893667390017, | |
| "grad_norm": 0.056753478944301605, | |
| "learning_rate": 4.4662297201220896e-05, | |
| "loss": 0.0009, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 8.630893667390017, | |
| "eval_all_ligands_equal": 0.5887039660056658, | |
| "eval_e3_equal": 0.8239199716713881, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9795502832861189, | |
| "eval_e3_heavy_atoms_difference": 0.6268590651558074, | |
| "eval_e3_heavy_atoms_difference_norm": 0.01645704704948741, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9795502832861189, | |
| "eval_has_all_attachment_points": 0.9834454674220963, | |
| "eval_has_three_substructures": 0.9996458923512748, | |
| "eval_heavy_atoms_difference": 7.879426345609065, | |
| "eval_heavy_atoms_difference_norm": 0.1044045921838352, | |
| "eval_linker_equal": 0.8345432011331445, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9971671388101983, | |
| "eval_linker_heavy_atoms_difference": 0.2329143059490085, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0019770889270123897, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9971671388101983, | |
| "eval_loss": 0.3325388729572296, | |
| "eval_num_fragments": 2.9996458923512748, | |
| "eval_poi_equal": 0.7900141643059491, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9225389518413598, | |
| "eval_poi_heavy_atoms_difference": 2.3541076487252126, | |
| "eval_poi_heavy_atoms_difference_norm": 0.07942848698757946, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9225389518413598, | |
| "eval_reassembly": 0.5965828611898017, | |
| "eval_reassembly_nostereo": 0.6195998583569405, | |
| "eval_runtime": 2290.0019, | |
| "eval_samples_per_second": 4.933, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9019121813031161, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 8.680213059775104, | |
| "grad_norm": 0.017264124006032944, | |
| "learning_rate": 3.542333420751848e-05, | |
| "loss": 0.0006, | |
| "step": 88000 | |
| }, | |
| { | |
| "epoch": 8.72953245216019, | |
| "grad_norm": 0.03280142694711685, | |
| "learning_rate": 2.38761194321228e-05, | |
| "loss": 0.0004, | |
| "step": 88500 | |
| }, | |
| { | |
| "epoch": 8.778851844545276, | |
| "grad_norm": 0.01054272148758173, | |
| "learning_rate": 1.257778848521878e-05, | |
| "loss": 0.0004, | |
| "step": 89000 | |
| }, | |
| { | |
| "epoch": 8.82817123693036, | |
| "grad_norm": 0.026974298059940338, | |
| "learning_rate": 4.030361547597777e-06, | |
| "loss": 0.0003, | |
| "step": 89500 | |
| }, | |
| { | |
| "epoch": 8.877490629315448, | |
| "grad_norm": 0.019895225763320923, | |
| "learning_rate": 1.2666994029975898e-07, | |
| "loss": 0.0003, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 8.877490629315448, | |
| "eval_all_ligands_equal": 0.5953434844192634, | |
| "eval_e3_equal": 0.8287004249291785, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9800814447592068, | |
| "eval_e3_heavy_atoms_difference": 0.6809490084985835, | |
| "eval_e3_heavy_atoms_difference_norm": 0.017487914834098253, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9800814447592068, | |
| "eval_has_all_attachment_points": 0.9846848441926346, | |
| "eval_has_three_substructures": 0.9999114730878187, | |
| "eval_heavy_atoms_difference": 7.618714589235127, | |
| "eval_heavy_atoms_difference_norm": 0.10135026066621577, | |
| "eval_linker_equal": 0.8446352691218131, | |
| "eval_linker_graph_edit_distance": 2.5672804532577907e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9974327195467422, | |
| "eval_linker_heavy_atoms_difference": 0.221671388101983, | |
| "eval_linker_heavy_atoms_difference_norm": 0.002011927125712562, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9974327195467422, | |
| "eval_loss": 0.33729150891304016, | |
| "eval_num_fragments": 2.9999114730878187, | |
| "eval_poi_equal": 0.7946175637393768, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.925814447592068, | |
| "eval_poi_heavy_atoms_difference": 2.1887393767705383, | |
| "eval_poi_heavy_atoms_difference_norm": 0.07438328075051615, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.925814447592068, | |
| "eval_reassembly": 0.6028682719546742, | |
| "eval_reassembly_nostereo": 0.6284525495750708, | |
| "eval_runtime": 2283.6349, | |
| "eval_samples_per_second": 4.946, | |
| "eval_steps_per_second": 0.078, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9058073654390935, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 8.926810021700533, | |
| "grad_norm": 0.05070818215608597, | |
| "learning_rate": 4.827748657192423e-05, | |
| "loss": 0.0011, | |
| "step": 90500 | |
| }, | |
| { | |
| "epoch": 8.97612941408562, | |
| "grad_norm": 0.030104659497737885, | |
| "learning_rate": 4.154704514310835e-05, | |
| "loss": 0.0008, | |
| "step": 91000 | |
| }, | |
| { | |
| "epoch": 9.025448806470704, | |
| "grad_norm": 0.005397017113864422, | |
| "learning_rate": 3.11436001621374e-05, | |
| "loss": 0.0005, | |
| "step": 91500 | |
| }, | |
| { | |
| "epoch": 9.07476819885579, | |
| "grad_norm": 0.0016132403397932649, | |
| "learning_rate": 1.9379652319228854e-05, | |
| "loss": 0.0004, | |
| "step": 92000 | |
| }, | |
| { | |
| "epoch": 9.124087591240876, | |
| "grad_norm": 0.026641881093382835, | |
| "learning_rate": 8.86033285430268e-06, | |
| "loss": 0.0003, | |
| "step": 92500 | |
| }, | |
| { | |
| "epoch": 9.124087591240876, | |
| "eval_all_ligands_equal": 0.5941041076487252, | |
| "eval_e3_equal": 0.8294971671388102, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9816749291784702, | |
| "eval_e3_heavy_atoms_difference": 0.6019830028328612, | |
| "eval_e3_heavy_atoms_difference_norm": 0.015736074666804267, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9816749291784702, | |
| "eval_has_all_attachment_points": 0.9828257790368272, | |
| "eval_has_three_substructures": 0.9997344192634561, | |
| "eval_heavy_atoms_difference": 7.330736543909349, | |
| "eval_heavy_atoms_difference_norm": 0.09710332347209924, | |
| "eval_linker_equal": 0.8425991501416431, | |
| "eval_linker_graph_edit_distance": 3.186968838526912e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9968130311614731, | |
| "eval_linker_heavy_atoms_difference": 0.2689447592067989, | |
| "eval_linker_heavy_atoms_difference_norm": 0.003959269090028944, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9968130311614731, | |
| "eval_loss": 0.341238796710968, | |
| "eval_num_fragments": 2.999734419263456, | |
| "eval_poi_equal": 0.7867386685552408, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9301522662889519, | |
| "eval_poi_heavy_atoms_difference": 2.0254957507082154, | |
| "eval_poi_heavy_atoms_difference_norm": 0.06838961592083195, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9301522662889519, | |
| "eval_reassembly": 0.6010977337110481, | |
| "eval_reassembly_nostereo": 0.6212818696883853, | |
| "eval_runtime": 2307.3699, | |
| "eval_samples_per_second": 4.896, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9105878186968839, | |
| "step": 92500 | |
| }, | |
| { | |
| "epoch": 9.173406983625961, | |
| "grad_norm": 0.011141962371766567, | |
| "learning_rate": 1.9151495250208456e-06, | |
| "loss": 0.0003, | |
| "step": 93000 | |
| }, | |
| { | |
| "epoch": 9.222726376011048, | |
| "grad_norm": 0.034191377460956573, | |
| "learning_rate": 4.9917883874293875e-05, | |
| "loss": 0.0003, | |
| "step": 93500 | |
| }, | |
| { | |
| "epoch": 9.272045768396133, | |
| "grad_norm": 0.07056564092636108, | |
| "learning_rate": 4.625787974910562e-05, | |
| "loss": 0.0018, | |
| "step": 94000 | |
| }, | |
| { | |
| "epoch": 9.32136516078122, | |
| "grad_norm": 0.003852763446047902, | |
| "learning_rate": 3.788650703787805e-05, | |
| "loss": 0.0006, | |
| "step": 94500 | |
| }, | |
| { | |
| "epoch": 9.370684553166305, | |
| "grad_norm": 0.02334916777908802, | |
| "learning_rate": 2.6661411876829478e-05, | |
| "loss": 0.0004, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 9.370684553166305, | |
| "eval_all_ligands_equal": 0.5910941926345609, | |
| "eval_e3_equal": 0.8256905099150141, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9795502832861189, | |
| "eval_e3_heavy_atoms_difference": 0.6888279036827195, | |
| "eval_e3_heavy_atoms_difference_norm": 0.018894557591006497, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9795502832861189, | |
| "eval_has_all_attachment_points": 0.9868094900849859, | |
| "eval_has_three_substructures": 0.9996458923512748, | |
| "eval_heavy_atoms_difference": 7.51292492917847, | |
| "eval_heavy_atoms_difference_norm": 0.1005060568497407, | |
| "eval_linker_equal": 0.8351628895184136, | |
| "eval_linker_graph_edit_distance": 2.3902266288951842e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9976097733711048, | |
| "eval_linker_heavy_atoms_difference": 0.3182542492917847, | |
| "eval_linker_heavy_atoms_difference_norm": 0.00709696659420846, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9976097733711048, | |
| "eval_loss": 0.34200209379196167, | |
| "eval_num_fragments": 2.9998229461756374, | |
| "eval_poi_equal": 0.7879780453257791, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9274964589235127, | |
| "eval_poi_heavy_atoms_difference": 2.1637747875354107, | |
| "eval_poi_heavy_atoms_difference_norm": 0.07346859452024099, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9274964589235127, | |
| "eval_reassembly": 0.5992386685552408, | |
| "eval_reassembly_nostereo": 0.6217245042492918, | |
| "eval_runtime": 2300.4132, | |
| "eval_samples_per_second": 4.91, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9073123229461756, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 9.420003945551391, | |
| "grad_norm": 0.0024575276765972376, | |
| "learning_rate": 1.5068396350908711e-05, | |
| "loss": 0.0003, | |
| "step": 95500 | |
| }, | |
| { | |
| "epoch": 9.469323337936476, | |
| "grad_norm": 0.006234079599380493, | |
| "learning_rate": 5.674738665931575e-06, | |
| "loss": 0.0003, | |
| "step": 96000 | |
| }, | |
| { | |
| "epoch": 9.518642730321563, | |
| "grad_norm": 0.0032219027634710073, | |
| "learning_rate": 5.606683008998137e-07, | |
| "loss": 0.0003, | |
| "step": 96500 | |
| }, | |
| { | |
| "epoch": 9.567962122706648, | |
| "grad_norm": 0.06151190027594566, | |
| "learning_rate": 4.914130169253062e-05, | |
| "loss": 0.0008, | |
| "step": 97000 | |
| }, | |
| { | |
| "epoch": 9.617281515091735, | |
| "grad_norm": 0.012705793604254723, | |
| "learning_rate": 4.349717018573489e-05, | |
| "loss": 0.001, | |
| "step": 97500 | |
| }, | |
| { | |
| "epoch": 9.617281515091735, | |
| "eval_all_ligands_equal": 0.5900318696883853, | |
| "eval_e3_equal": 0.8251593484419264, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9770715297450425, | |
| "eval_e3_heavy_atoms_difference": 0.7837287535410765, | |
| "eval_e3_heavy_atoms_difference_norm": 0.021666618023752426, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9770715297450425, | |
| "eval_has_all_attachment_points": 0.9895538243626062, | |
| "eval_has_three_substructures": 0.9992917847025495, | |
| "eval_heavy_atoms_difference": 8.148016997167138, | |
| "eval_heavy_atoms_difference_norm": 0.10804380735986156, | |
| "eval_linker_equal": 0.8340120396600567, | |
| "eval_linker_graph_edit_distance": 3.5410764872521247e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9964589235127479, | |
| "eval_linker_heavy_atoms_difference": 0.23999645892351273, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0019764159599654626, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9964589235127479, | |
| "eval_loss": 0.3328213393688202, | |
| "eval_num_fragments": 3.0003541076487252, | |
| "eval_poi_equal": 0.7890403682719547, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9206798866855525, | |
| "eval_poi_heavy_atoms_difference": 2.272220254957507, | |
| "eval_poi_heavy_atoms_difference_norm": 0.07798413320246032, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9206798866855525, | |
| "eval_reassembly": 0.5977337110481586, | |
| "eval_reassembly_nostereo": 0.6202195467422096, | |
| "eval_runtime": 2296.0078, | |
| "eval_samples_per_second": 4.92, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.8981940509915014, | |
| "step": 97500 | |
| }, | |
| { | |
| "epoch": 9.66660090747682, | |
| "grad_norm": 0.004747785162180662, | |
| "learning_rate": 3.375683249505763e-05, | |
| "loss": 0.0005, | |
| "step": 98000 | |
| }, | |
| { | |
| "epoch": 9.715920299861907, | |
| "grad_norm": 0.007481692358851433, | |
| "learning_rate": 2.2077290533818858e-05, | |
| "loss": 0.0004, | |
| "step": 98500 | |
| }, | |
| { | |
| "epoch": 9.765239692246992, | |
| "grad_norm": 0.009572615846991539, | |
| "learning_rate": 1.1044983823640434e-05, | |
| "loss": 0.0003, | |
| "step": 99000 | |
| }, | |
| { | |
| "epoch": 9.814559084632076, | |
| "grad_norm": 0.005006254650652409, | |
| "learning_rate": 3.103021365246661e-06, | |
| "loss": 0.0003, | |
| "step": 99500 | |
| }, | |
| { | |
| "epoch": 9.863878477017163, | |
| "grad_norm": 0.005209687165915966, | |
| "learning_rate": 1.015411392275567e-08, | |
| "loss": 0.0003, | |
| "step": 100000 | |
| }, | |
| { | |
| "epoch": 9.863878477017163, | |
| "eval_all_ligands_equal": 0.5964058073654391, | |
| "eval_e3_equal": 0.8259560906515581, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9784879603399433, | |
| "eval_e3_heavy_atoms_difference": 0.6771423512747875, | |
| "eval_e3_heavy_atoms_difference_norm": 0.01736375434990478, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9784879603399433, | |
| "eval_has_all_attachment_points": 0.9868094900849859, | |
| "eval_has_three_substructures": 0.9998229461756374, | |
| "eval_heavy_atoms_difference": 7.545945467422096, | |
| "eval_heavy_atoms_difference_norm": 0.10099267663653283, | |
| "eval_linker_equal": 0.845520538243626, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9975212464589235, | |
| "eval_linker_heavy_atoms_difference": 0.22822237960339944, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0015094216065725523, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9975212464589235, | |
| "eval_loss": 0.3421395421028137, | |
| "eval_num_fragments": 3.0001770538243626, | |
| "eval_poi_equal": 0.7916961756373938, | |
| "eval_poi_graph_edit_distance": 7.37429178470255e+62, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9262570821529745, | |
| "eval_poi_heavy_atoms_difference": 2.1195998583569406, | |
| "eval_poi_heavy_atoms_difference_norm": 0.07347695826275151, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9262570821529745, | |
| "eval_reassembly": 0.6034879603399433, | |
| "eval_reassembly_nostereo": 0.6257967422096318, | |
| "eval_runtime": 2304.8548, | |
| "eval_samples_per_second": 4.901, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9041253541076487, | |
| "step": 100000 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 100000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 10, | |
| "save_steps": 5000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 8.663173632603694e+17, | |
| "train_batch_size": 128, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |