| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.9987096774193547, | |
| "eval_steps": 500, | |
| "global_step": 3486, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.008602150537634409, | |
| "grad_norm": 1.435989601453093, | |
| "learning_rate": 5.730659025787966e-06, | |
| "loss": 1.2534, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.017204301075268817, | |
| "grad_norm": 0.956311142058486, | |
| "learning_rate": 1.1461318051575932e-05, | |
| "loss": 1.1941, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.025806451612903226, | |
| "grad_norm": 0.5955272986463238, | |
| "learning_rate": 1.7191977077363898e-05, | |
| "loss": 1.0416, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.034408602150537634, | |
| "grad_norm": 0.5371220864077184, | |
| "learning_rate": 2.2922636103151864e-05, | |
| "loss": 0.9252, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.043010752688172046, | |
| "grad_norm": 0.3298242735866968, | |
| "learning_rate": 2.8653295128939826e-05, | |
| "loss": 0.8582, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.05161290322580645, | |
| "grad_norm": 0.3017426075394883, | |
| "learning_rate": 3.4383954154727795e-05, | |
| "loss": 0.7998, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.060215053763440864, | |
| "grad_norm": 0.3375526814194791, | |
| "learning_rate": 4.011461318051576e-05, | |
| "loss": 0.7638, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.06881720430107527, | |
| "grad_norm": 0.2705319023822148, | |
| "learning_rate": 4.584527220630373e-05, | |
| "loss": 0.731, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.07741935483870968, | |
| "grad_norm": 0.2857694097360681, | |
| "learning_rate": 5.157593123209169e-05, | |
| "loss": 0.7083, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.08602150537634409, | |
| "grad_norm": 0.3098635398347818, | |
| "learning_rate": 5.730659025787965e-05, | |
| "loss": 0.6841, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.09462365591397849, | |
| "grad_norm": 0.3198686339975185, | |
| "learning_rate": 6.303724928366761e-05, | |
| "loss": 0.6606, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.1032258064516129, | |
| "grad_norm": 0.3190855908911574, | |
| "learning_rate": 6.876790830945559e-05, | |
| "loss": 0.6557, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.11182795698924732, | |
| "grad_norm": 0.3381628025718819, | |
| "learning_rate": 7.449856733524355e-05, | |
| "loss": 0.649, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.12043010752688173, | |
| "grad_norm": 0.31598591910637613, | |
| "learning_rate": 8.022922636103152e-05, | |
| "loss": 0.6497, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.12903225806451613, | |
| "grad_norm": 0.3128921107543011, | |
| "learning_rate": 8.595988538681948e-05, | |
| "loss": 0.6392, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.13763440860215054, | |
| "grad_norm": 0.32601075629670023, | |
| "learning_rate": 9.169054441260745e-05, | |
| "loss": 0.6329, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.14623655913978495, | |
| "grad_norm": 0.3810140632755748, | |
| "learning_rate": 9.742120343839543e-05, | |
| "loss": 0.6246, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.15483870967741936, | |
| "grad_norm": 0.32108686980228696, | |
| "learning_rate": 0.00010315186246418338, | |
| "loss": 0.6341, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.16344086021505377, | |
| "grad_norm": 0.35140482524130334, | |
| "learning_rate": 0.00010888252148997136, | |
| "loss": 0.6341, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.17204301075268819, | |
| "grad_norm": 0.29928696988494263, | |
| "learning_rate": 0.0001146131805157593, | |
| "loss": 0.6256, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.18064516129032257, | |
| "grad_norm": 0.29203351329015653, | |
| "learning_rate": 0.0001203438395415473, | |
| "loss": 0.6206, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.18924731182795698, | |
| "grad_norm": 0.269908780529402, | |
| "learning_rate": 0.00012607449856733523, | |
| "loss": 0.6201, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.1978494623655914, | |
| "grad_norm": 0.2918951896599236, | |
| "learning_rate": 0.0001318051575931232, | |
| "loss": 0.6181, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.2064516129032258, | |
| "grad_norm": 0.2708453980927917, | |
| "learning_rate": 0.00013753581661891118, | |
| "loss": 0.6053, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.21505376344086022, | |
| "grad_norm": 0.2662594416135698, | |
| "learning_rate": 0.00014326647564469916, | |
| "loss": 0.6033, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.22365591397849463, | |
| "grad_norm": 0.2776129880103733, | |
| "learning_rate": 0.0001489971346704871, | |
| "loss": 0.618, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.23225806451612904, | |
| "grad_norm": 0.279068656400945, | |
| "learning_rate": 0.00015472779369627508, | |
| "loss": 0.6111, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.24086021505376345, | |
| "grad_norm": 0.2524990953581738, | |
| "learning_rate": 0.00016045845272206303, | |
| "loss": 0.5888, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.24946236559139784, | |
| "grad_norm": 0.27628464942088166, | |
| "learning_rate": 0.000166189111747851, | |
| "loss": 0.6051, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.25806451612903225, | |
| "grad_norm": 0.2738432329342113, | |
| "learning_rate": 0.00017191977077363896, | |
| "loss": 0.6025, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.26666666666666666, | |
| "grad_norm": 0.25607005690614404, | |
| "learning_rate": 0.00017765042979942693, | |
| "loss": 0.6035, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.2752688172043011, | |
| "grad_norm": 0.247801611216806, | |
| "learning_rate": 0.0001833810888252149, | |
| "loss": 0.6118, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.2838709677419355, | |
| "grad_norm": 0.24971595808191485, | |
| "learning_rate": 0.00018911174785100288, | |
| "loss": 0.6003, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.2924731182795699, | |
| "grad_norm": 0.2411785853482113, | |
| "learning_rate": 0.00019484240687679086, | |
| "loss": 0.6019, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.3010752688172043, | |
| "grad_norm": 0.24693597842223552, | |
| "learning_rate": 0.0001999999498534943, | |
| "loss": 0.5912, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.3096774193548387, | |
| "grad_norm": 0.23649676383007667, | |
| "learning_rate": 0.0001999939323336644, | |
| "loss": 0.6032, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.31827956989247314, | |
| "grad_norm": 0.23340962816367694, | |
| "learning_rate": 0.0001999778862042167, | |
| "loss": 0.5867, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.32688172043010755, | |
| "grad_norm": 0.228292747503904, | |
| "learning_rate": 0.0001999518130744525, | |
| "loss": 0.5869, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.33548387096774196, | |
| "grad_norm": 0.23546242335669995, | |
| "learning_rate": 0.0001999157155593029, | |
| "loss": 0.585, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.34408602150537637, | |
| "grad_norm": 0.23135215802858206, | |
| "learning_rate": 0.0001998695972790664, | |
| "loss": 0.5929, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.35268817204301073, | |
| "grad_norm": 0.22524299635343834, | |
| "learning_rate": 0.00019981346285904595, | |
| "loss": 0.5976, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.36129032258064514, | |
| "grad_norm": 0.21756192058074578, | |
| "learning_rate": 0.00019974731792908506, | |
| "loss": 0.587, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.36989247311827955, | |
| "grad_norm": 0.23763473038875538, | |
| "learning_rate": 0.000199671169123003, | |
| "loss": 0.5934, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.37849462365591396, | |
| "grad_norm": 0.23701431523985933, | |
| "learning_rate": 0.00019958502407792963, | |
| "loss": 0.6009, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.3870967741935484, | |
| "grad_norm": 0.22134697304917678, | |
| "learning_rate": 0.00019948889143353948, | |
| "loss": 0.5994, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.3956989247311828, | |
| "grad_norm": 0.2268467235014266, | |
| "learning_rate": 0.00019938278083118517, | |
| "loss": 0.5982, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.4043010752688172, | |
| "grad_norm": 0.2247581060586749, | |
| "learning_rate": 0.00019926670291293055, | |
| "loss": 0.5782, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.4129032258064516, | |
| "grad_norm": 0.21243655986167898, | |
| "learning_rate": 0.00019914066932048317, | |
| "loss": 0.5874, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.421505376344086, | |
| "grad_norm": 0.21385338375703783, | |
| "learning_rate": 0.00019900469269402708, | |
| "loss": 0.5813, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.43010752688172044, | |
| "grad_norm": 0.21415772228830113, | |
| "learning_rate": 0.00019885878667095472, | |
| "loss": 0.5934, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.43870967741935485, | |
| "grad_norm": 0.23153816801674154, | |
| "learning_rate": 0.00019870296588449948, | |
| "loss": 0.5766, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.44731182795698926, | |
| "grad_norm": 0.21768913976009108, | |
| "learning_rate": 0.00019853724596226795, | |
| "loss": 0.5964, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.4559139784946237, | |
| "grad_norm": 0.21481929780656356, | |
| "learning_rate": 0.00019836164352467263, | |
| "loss": 0.5748, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.4645161290322581, | |
| "grad_norm": 0.29328457757364856, | |
| "learning_rate": 0.00019817617618326516, | |
| "loss": 0.5863, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.4731182795698925, | |
| "grad_norm": 0.2140715456768234, | |
| "learning_rate": 0.00019798086253896974, | |
| "loss": 0.5787, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.4817204301075269, | |
| "grad_norm": 0.22386132897333755, | |
| "learning_rate": 0.00019777572218021789, | |
| "loss": 0.5863, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.49032258064516127, | |
| "grad_norm": 0.23289220288123677, | |
| "learning_rate": 0.00019756077568098372, | |
| "loss": 0.5638, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.4989247311827957, | |
| "grad_norm": 0.21115266625450407, | |
| "learning_rate": 0.00019733604459872055, | |
| "loss": 0.5744, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.5075268817204301, | |
| "grad_norm": 0.2207246383657776, | |
| "learning_rate": 0.00019710155147219886, | |
| "loss": 0.5789, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.5161290322580645, | |
| "grad_norm": 0.2127412426650313, | |
| "learning_rate": 0.00019685731981924592, | |
| "loss": 0.5764, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.524731182795699, | |
| "grad_norm": 0.2256297779595538, | |
| "learning_rate": 0.00019660337413438697, | |
| "loss": 0.5798, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.5333333333333333, | |
| "grad_norm": 0.22243638154383857, | |
| "learning_rate": 0.00019633973988638877, | |
| "loss": 0.5767, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.5419354838709678, | |
| "grad_norm": 0.21425852582870433, | |
| "learning_rate": 0.00019606644351570518, | |
| "loss": 0.5736, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.5505376344086022, | |
| "grad_norm": 0.20506928092936336, | |
| "learning_rate": 0.00019578351243182545, | |
| "loss": 0.5766, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.5591397849462365, | |
| "grad_norm": 0.2175808639953357, | |
| "learning_rate": 0.0001954909750105252, | |
| "loss": 0.5883, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.567741935483871, | |
| "grad_norm": 0.21972174494018643, | |
| "learning_rate": 0.00019518886059102062, | |
| "loss": 0.5774, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.5763440860215053, | |
| "grad_norm": 0.2222552681428519, | |
| "learning_rate": 0.0001948771994730259, | |
| "loss": 0.5694, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.5849462365591398, | |
| "grad_norm": 0.21552094411185982, | |
| "learning_rate": 0.0001945560229137145, | |
| "loss": 0.5657, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.5935483870967742, | |
| "grad_norm": 0.21243053577901072, | |
| "learning_rate": 0.0001942253631245842, | |
| "loss": 0.5776, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.6021505376344086, | |
| "grad_norm": 0.21257424919590467, | |
| "learning_rate": 0.00019388525326822665, | |
| "loss": 0.5764, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.610752688172043, | |
| "grad_norm": 0.21485742957719725, | |
| "learning_rate": 0.0001935357274550013, | |
| "loss": 0.5821, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.6193548387096774, | |
| "grad_norm": 0.2184569847068338, | |
| "learning_rate": 0.00019317682073961444, | |
| "loss": 0.5674, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.6279569892473118, | |
| "grad_norm": 0.20912150886711225, | |
| "learning_rate": 0.00019280856911760354, | |
| "loss": 0.559, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.6365591397849463, | |
| "grad_norm": 0.21433879312995557, | |
| "learning_rate": 0.00019243100952172723, | |
| "loss": 0.5615, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.6451612903225806, | |
| "grad_norm": 0.22214855400737676, | |
| "learning_rate": 0.00019204417981826091, | |
| "loss": 0.5761, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.6537634408602151, | |
| "grad_norm": 0.2196194509755129, | |
| "learning_rate": 0.00019164811880319958, | |
| "loss": 0.5718, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.6623655913978495, | |
| "grad_norm": 0.2089872234950888, | |
| "learning_rate": 0.00019124286619836637, | |
| "loss": 0.5819, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.6709677419354839, | |
| "grad_norm": 0.20958335214398927, | |
| "learning_rate": 0.00019082846264742917, | |
| "loss": 0.5663, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.6795698924731183, | |
| "grad_norm": 0.22250020784958888, | |
| "learning_rate": 0.00019040494971182413, | |
| "loss": 0.572, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.6881720430107527, | |
| "grad_norm": 0.2162793985278103, | |
| "learning_rate": 0.00018997236986658753, | |
| "loss": 0.5578, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.6967741935483871, | |
| "grad_norm": 0.21650281403806723, | |
| "learning_rate": 0.00018953076649609564, | |
| "loss": 0.5698, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.7053763440860215, | |
| "grad_norm": 0.21882852535131322, | |
| "learning_rate": 0.000189080183889714, | |
| "loss": 0.5712, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.7139784946236559, | |
| "grad_norm": 0.21577728486568806, | |
| "learning_rate": 0.00018862066723735512, | |
| "loss": 0.5676, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.7225806451612903, | |
| "grad_norm": 0.20731869739137426, | |
| "learning_rate": 0.00018815226262494655, | |
| "loss": 0.567, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.7311827956989247, | |
| "grad_norm": 0.21966746890763253, | |
| "learning_rate": 0.00018767501702980874, | |
| "loss": 0.5738, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.7397849462365591, | |
| "grad_norm": 0.2199440815107056, | |
| "learning_rate": 0.00018718897831594355, | |
| "loss": 0.5691, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.7483870967741936, | |
| "grad_norm": 0.21722437739899045, | |
| "learning_rate": 0.00018669419522923393, | |
| "loss": 0.5753, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.7569892473118279, | |
| "grad_norm": 0.2218916153172574, | |
| "learning_rate": 0.00018619071739255506, | |
| "loss": 0.5572, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.7655913978494624, | |
| "grad_norm": 0.20612913226427296, | |
| "learning_rate": 0.00018567859530079753, | |
| "loss": 0.567, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.7741935483870968, | |
| "grad_norm": 0.2296891108499081, | |
| "learning_rate": 0.00018515788031580317, | |
| "loss": 0.5735, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.7827956989247312, | |
| "grad_norm": 0.20791733581224098, | |
| "learning_rate": 0.0001846286246612138, | |
| "loss": 0.56, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.7913978494623656, | |
| "grad_norm": 0.2130914822501822, | |
| "learning_rate": 0.00018409088141723364, | |
| "loss": 0.5588, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.21931892060132505, | |
| "learning_rate": 0.00018354470451530574, | |
| "loss": 0.5718, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.8086021505376344, | |
| "grad_norm": 0.20510223609898656, | |
| "learning_rate": 0.00018299014873270314, | |
| "loss": 0.5693, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.8172043010752689, | |
| "grad_norm": 0.20754250893604087, | |
| "learning_rate": 0.00018242726968703505, | |
| "loss": 0.5657, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.8258064516129032, | |
| "grad_norm": 0.2119047697743998, | |
| "learning_rate": 0.00018185612383066893, | |
| "loss": 0.5523, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.8344086021505376, | |
| "grad_norm": 0.21665086623022414, | |
| "learning_rate": 0.00018127676844506874, | |
| "loss": 0.5564, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.843010752688172, | |
| "grad_norm": 0.22240006490076028, | |
| "learning_rate": 0.00018068926163505, | |
| "loss": 0.5661, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.8516129032258064, | |
| "grad_norm": 0.20224509449303205, | |
| "learning_rate": 0.00018009366232295235, | |
| "loss": 0.5709, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.8602150537634409, | |
| "grad_norm": 0.20943438346218987, | |
| "learning_rate": 0.00017949003024273015, | |
| "loss": 0.5567, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.8688172043010752, | |
| "grad_norm": 0.2119577392060311, | |
| "learning_rate": 0.0001788784259339616, | |
| "loss": 0.5726, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.8774193548387097, | |
| "grad_norm": 0.21671277914581005, | |
| "learning_rate": 0.0001782589107357771, | |
| "loss": 0.5721, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.886021505376344, | |
| "grad_norm": 0.21042958546072746, | |
| "learning_rate": 0.00017763154678070733, | |
| "loss": 0.5682, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.8946236559139785, | |
| "grad_norm": 0.21134103487495276, | |
| "learning_rate": 0.0001769963969884521, | |
| "loss": 0.5578, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.9032258064516129, | |
| "grad_norm": 0.2138330925142033, | |
| "learning_rate": 0.0001763535250595696, | |
| "loss": 0.5622, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.9118279569892473, | |
| "grad_norm": 0.20191022832768318, | |
| "learning_rate": 0.00017570299546908812, | |
| "loss": 0.5543, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.9204301075268817, | |
| "grad_norm": 0.20993539174391937, | |
| "learning_rate": 0.0001750448734600394, | |
| "loss": 0.5687, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.9290322580645162, | |
| "grad_norm": 0.21448241595075324, | |
| "learning_rate": 0.0001743792250369155, | |
| "loss": 0.5628, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.9376344086021505, | |
| "grad_norm": 0.21562115673441815, | |
| "learning_rate": 0.00017370611695904895, | |
| "loss": 0.5707, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.946236559139785, | |
| "grad_norm": 0.20800257947078069, | |
| "learning_rate": 0.00017302561673391732, | |
| "loss": 0.5682, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.9548387096774194, | |
| "grad_norm": 0.2142950321860787, | |
| "learning_rate": 0.00017233779261037268, | |
| "loss": 0.5644, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.9634408602150538, | |
| "grad_norm": 0.22988301024268806, | |
| "learning_rate": 0.00017164271357179698, | |
| "loss": 0.5747, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.9720430107526882, | |
| "grad_norm": 0.2044546345524181, | |
| "learning_rate": 0.00017094044932918336, | |
| "loss": 0.5627, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.9806451612903225, | |
| "grad_norm": 0.20801733651601362, | |
| "learning_rate": 0.00017023107031414477, | |
| "loss": 0.5655, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.989247311827957, | |
| "grad_norm": 0.21348629669321226, | |
| "learning_rate": 0.00016951464767185013, | |
| "loss": 0.5569, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.9978494623655914, | |
| "grad_norm": 0.21098078403488563, | |
| "learning_rate": 0.0001687912532538892, | |
| "loss": 0.5686, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.0064516129032257, | |
| "grad_norm": 0.23232328451409845, | |
| "learning_rate": 0.00016806095961106632, | |
| "loss": 0.5352, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.0150537634408603, | |
| "grad_norm": 0.21849412389861325, | |
| "learning_rate": 0.00016732383998612407, | |
| "loss": 0.5166, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.0236559139784946, | |
| "grad_norm": 0.22255257432573036, | |
| "learning_rate": 0.00016657996830639774, | |
| "loss": 0.5286, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.032258064516129, | |
| "grad_norm": 0.23162311614487038, | |
| "learning_rate": 0.0001658294191764009, | |
| "loss": 0.5209, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.0408602150537634, | |
| "grad_norm": 0.23679830183057837, | |
| "learning_rate": 0.0001650722678703432, | |
| "loss": 0.5155, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.049462365591398, | |
| "grad_norm": 0.24431579126710756, | |
| "learning_rate": 0.00016430859032458086, | |
| "loss": 0.5239, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.0580645161290323, | |
| "grad_norm": 0.21925666538005387, | |
| "learning_rate": 0.00016353846313000098, | |
| "loss": 0.5142, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.0666666666666667, | |
| "grad_norm": 0.23676074055821078, | |
| "learning_rate": 0.00016276196352434, | |
| "loss": 0.5267, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.075268817204301, | |
| "grad_norm": 0.23532815378000083, | |
| "learning_rate": 0.00016197916938443733, | |
| "loss": 0.5211, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.0838709677419356, | |
| "grad_norm": 0.23216395268938977, | |
| "learning_rate": 0.00016119015921842503, | |
| "loss": 0.5161, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 1.09247311827957, | |
| "grad_norm": 0.23822646088799643, | |
| "learning_rate": 0.0001603950121578539, | |
| "loss": 0.5077, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 1.1010752688172043, | |
| "grad_norm": 0.2290007557059369, | |
| "learning_rate": 0.00015959380794975734, | |
| "loss": 0.5156, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 1.1096774193548387, | |
| "grad_norm": 0.22484506558131512, | |
| "learning_rate": 0.00015878662694865327, | |
| "loss": 0.5205, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 1.118279569892473, | |
| "grad_norm": 0.22971689352325197, | |
| "learning_rate": 0.00015797355010848519, | |
| "loss": 0.5218, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.1268817204301076, | |
| "grad_norm": 0.23491219834804952, | |
| "learning_rate": 0.00015715465897450317, | |
| "loss": 0.5248, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 1.135483870967742, | |
| "grad_norm": 0.23555948179310204, | |
| "learning_rate": 0.00015633003567508557, | |
| "loss": 0.524, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 1.1440860215053763, | |
| "grad_norm": 0.2275140567256555, | |
| "learning_rate": 0.00015549976291350204, | |
| "loss": 0.517, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 1.1526881720430107, | |
| "grad_norm": 0.22712302146976637, | |
| "learning_rate": 0.00015466392395961908, | |
| "loss": 0.5243, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 1.1612903225806452, | |
| "grad_norm": 0.22097671231695554, | |
| "learning_rate": 0.0001538226026415489, | |
| "loss": 0.5118, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 1.1698924731182796, | |
| "grad_norm": 0.22706370204760404, | |
| "learning_rate": 0.00015297588333724183, | |
| "loss": 0.5235, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 1.178494623655914, | |
| "grad_norm": 0.2309770159886724, | |
| "learning_rate": 0.00015212385096602415, | |
| "loss": 0.5202, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 1.1870967741935483, | |
| "grad_norm": 0.23885726185135214, | |
| "learning_rate": 0.0001512665909800811, | |
| "loss": 0.5201, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 1.1956989247311829, | |
| "grad_norm": 0.23447118491147034, | |
| "learning_rate": 0.00015040418935588682, | |
| "loss": 0.5103, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 1.2043010752688172, | |
| "grad_norm": 0.24148993682260503, | |
| "learning_rate": 0.0001495367325855816, | |
| "loss": 0.5229, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 1.2129032258064516, | |
| "grad_norm": 0.22483070355440393, | |
| "learning_rate": 0.00014866430766829743, | |
| "loss": 0.5132, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 1.221505376344086, | |
| "grad_norm": 0.2254845362019036, | |
| "learning_rate": 0.00014778700210143242, | |
| "loss": 0.5231, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 1.2301075268817203, | |
| "grad_norm": 0.23952864821588327, | |
| "learning_rate": 0.00014690490387187584, | |
| "loss": 0.5157, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 1.238709677419355, | |
| "grad_norm": 0.2257464446574845, | |
| "learning_rate": 0.00014601810144718345, | |
| "loss": 0.537, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 1.2473118279569892, | |
| "grad_norm": 0.23617654028249638, | |
| "learning_rate": 0.0001451266837667051, | |
| "loss": 0.5187, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 1.2559139784946236, | |
| "grad_norm": 0.23682708876074962, | |
| "learning_rate": 0.00014423074023266463, | |
| "loss": 0.52, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 1.2645161290322582, | |
| "grad_norm": 0.22896552654447108, | |
| "learning_rate": 0.00014333036070119363, | |
| "loss": 0.5163, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 1.2731182795698925, | |
| "grad_norm": 0.24027250250612717, | |
| "learning_rate": 0.00014242563547331956, | |
| "loss": 0.517, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 1.281720430107527, | |
| "grad_norm": 0.23395780934976781, | |
| "learning_rate": 0.00014151665528590925, | |
| "loss": 0.5215, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 1.2903225806451613, | |
| "grad_norm": 0.24728067879450083, | |
| "learning_rate": 0.0001406035113025687, | |
| "loss": 0.5208, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.2989247311827956, | |
| "grad_norm": 0.22394688366493132, | |
| "learning_rate": 0.0001396862951045001, | |
| "loss": 0.5274, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 1.3075268817204302, | |
| "grad_norm": 0.23212309124084482, | |
| "learning_rate": 0.00013876509868131692, | |
| "loss": 0.5172, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 1.3161290322580645, | |
| "grad_norm": 0.23245525514987117, | |
| "learning_rate": 0.0001378400144218181, | |
| "loss": 0.521, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 1.324731182795699, | |
| "grad_norm": 0.24340497246468923, | |
| "learning_rate": 0.00013691113510472212, | |
| "loss": 0.533, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 1.3333333333333333, | |
| "grad_norm": 0.2321928544760521, | |
| "learning_rate": 0.00013597855388936217, | |
| "loss": 0.5283, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 1.3419354838709676, | |
| "grad_norm": 0.23326523784187517, | |
| "learning_rate": 0.00013504236430634286, | |
| "loss": 0.5416, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 1.3505376344086022, | |
| "grad_norm": 0.2262845978211386, | |
| "learning_rate": 0.0001341026602481597, | |
| "loss": 0.5231, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 1.3591397849462366, | |
| "grad_norm": 0.2371091538942524, | |
| "learning_rate": 0.00013315953595978287, | |
| "loss": 0.5328, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 1.367741935483871, | |
| "grad_norm": 0.2404491874213034, | |
| "learning_rate": 0.00013221308602920468, | |
| "loss": 0.5273, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 1.3763440860215055, | |
| "grad_norm": 0.2382492165329928, | |
| "learning_rate": 0.00013126340537795343, | |
| "loss": 0.529, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.3849462365591398, | |
| "grad_norm": 0.22921228109841923, | |
| "learning_rate": 0.0001303105892515734, | |
| "loss": 0.5198, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 1.3935483870967742, | |
| "grad_norm": 0.22823020260902263, | |
| "learning_rate": 0.00012935473321007255, | |
| "loss": 0.5215, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 1.4021505376344086, | |
| "grad_norm": 0.23341543795313835, | |
| "learning_rate": 0.0001283959331183386, | |
| "loss": 0.5179, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 1.410752688172043, | |
| "grad_norm": 0.23483174318348113, | |
| "learning_rate": 0.00012743428513652442, | |
| "loss": 0.5242, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 1.4193548387096775, | |
| "grad_norm": 0.23015296236307278, | |
| "learning_rate": 0.00012646988571040398, | |
| "loss": 0.522, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 1.4279569892473118, | |
| "grad_norm": 0.24213814226406005, | |
| "learning_rate": 0.00012550283156169965, | |
| "loss": 0.5238, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 1.4365591397849462, | |
| "grad_norm": 0.23239320809561437, | |
| "learning_rate": 0.0001245332196783817, | |
| "loss": 0.5156, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 1.4451612903225808, | |
| "grad_norm": 0.23180542434985962, | |
| "learning_rate": 0.00012356114730494123, | |
| "loss": 0.5184, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 1.453763440860215, | |
| "grad_norm": 0.23588148083202282, | |
| "learning_rate": 0.00012258671193263716, | |
| "loss": 0.5155, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 1.4623655913978495, | |
| "grad_norm": 0.23435001337205247, | |
| "learning_rate": 0.00012161001128971879, | |
| "loss": 0.5293, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.4709677419354839, | |
| "grad_norm": 0.22453229677185074, | |
| "learning_rate": 0.00012063114333162438, | |
| "loss": 0.5151, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 1.4795698924731182, | |
| "grad_norm": 0.2513082837299489, | |
| "learning_rate": 0.00011965020623115688, | |
| "loss": 0.5416, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 1.4881720430107528, | |
| "grad_norm": 0.23485488435380322, | |
| "learning_rate": 0.0001186672983686381, | |
| "loss": 0.5114, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 1.4967741935483871, | |
| "grad_norm": 0.2339236059203922, | |
| "learning_rate": 0.00011768251832204187, | |
| "loss": 0.5189, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 1.5053763440860215, | |
| "grad_norm": 0.2363738308894046, | |
| "learning_rate": 0.00011669596485710741, | |
| "loss": 0.5236, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.513978494623656, | |
| "grad_norm": 0.23458902400188686, | |
| "learning_rate": 0.00011570773691743397, | |
| "loss": 0.5284, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 1.5225806451612902, | |
| "grad_norm": 0.23836545061013703, | |
| "learning_rate": 0.0001147179336145575, | |
| "loss": 0.5121, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 1.5311827956989248, | |
| "grad_norm": 0.23790373487214236, | |
| "learning_rate": 0.00011372665421801056, | |
| "loss": 0.5165, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 1.5397849462365591, | |
| "grad_norm": 0.23929874324218836, | |
| "learning_rate": 0.0001127339981453664, | |
| "loss": 0.5257, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 1.5483870967741935, | |
| "grad_norm": 0.23994815774490452, | |
| "learning_rate": 0.00011174006495226812, | |
| "loss": 0.5125, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 1.556989247311828, | |
| "grad_norm": 0.22986043036890028, | |
| "learning_rate": 0.00011074495432244397, | |
| "loss": 0.5142, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 1.5655913978494622, | |
| "grad_norm": 0.2276160657387438, | |
| "learning_rate": 0.0001097487660577099, | |
| "loss": 0.5198, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 1.5741935483870968, | |
| "grad_norm": 0.23590750768735258, | |
| "learning_rate": 0.00010875160006796024, | |
| "loss": 0.5203, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 1.5827956989247312, | |
| "grad_norm": 0.24238494113010198, | |
| "learning_rate": 0.00010775355636114755, | |
| "loss": 0.5288, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 1.5913978494623655, | |
| "grad_norm": 0.23079601762605023, | |
| "learning_rate": 0.00010675473503325245, | |
| "loss": 0.516, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "grad_norm": 0.23473655518093983, | |
| "learning_rate": 0.00010575523625824488, | |
| "loss": 0.5245, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 1.6086021505376344, | |
| "grad_norm": 0.2297499982621768, | |
| "learning_rate": 0.00010475516027803751, | |
| "loss": 0.5162, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 1.6172043010752688, | |
| "grad_norm": 0.23292165198339548, | |
| "learning_rate": 0.00010375460739243215, | |
| "loss": 0.5246, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 1.6258064516129034, | |
| "grad_norm": 0.24340787369255004, | |
| "learning_rate": 0.00010275367794906044, | |
| "loss": 0.5099, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 1.6344086021505375, | |
| "grad_norm": 0.24374482359627925, | |
| "learning_rate": 0.00010175247233331989, | |
| "loss": 0.5097, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.643010752688172, | |
| "grad_norm": 0.23616305839556445, | |
| "learning_rate": 0.00010075109095830584, | |
| "loss": 0.5231, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 1.6516129032258065, | |
| "grad_norm": 0.23231972508384582, | |
| "learning_rate": 9.974963425474106e-05, | |
| "loss": 0.5213, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 1.6602150537634408, | |
| "grad_norm": 0.24023523508198666, | |
| "learning_rate": 9.874820266090303e-05, | |
| "loss": 0.5145, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 1.6688172043010754, | |
| "grad_norm": 0.24186555744758984, | |
| "learning_rate": 9.774689661255106e-05, | |
| "loss": 0.5256, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 1.6774193548387095, | |
| "grad_norm": 0.2372332084270041, | |
| "learning_rate": 9.67458165328531e-05, | |
| "loss": 0.5094, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 1.686021505376344, | |
| "grad_norm": 0.2350639603238352, | |
| "learning_rate": 9.574506282231433e-05, | |
| "loss": 0.5119, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 1.6946236559139785, | |
| "grad_norm": 0.23559422057113213, | |
| "learning_rate": 9.474473584870757e-05, | |
| "loss": 0.5155, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 1.7032258064516128, | |
| "grad_norm": 0.2336022634685686, | |
| "learning_rate": 9.374493593700723e-05, | |
| "loss": 0.5113, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 1.7118279569892474, | |
| "grad_norm": 0.23539681381001779, | |
| "learning_rate": 9.274576335932767e-05, | |
| "loss": 0.5189, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 1.7204301075268817, | |
| "grad_norm": 0.23650024951926094, | |
| "learning_rate": 9.174731832486648e-05, | |
| "loss": 0.5082, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.729032258064516, | |
| "grad_norm": 0.23916533165546086, | |
| "learning_rate": 9.074970096985427e-05, | |
| "loss": 0.5284, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 1.7376344086021507, | |
| "grad_norm": 0.23608187656307292, | |
| "learning_rate": 8.975301134751202e-05, | |
| "loss": 0.5172, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 1.7462365591397848, | |
| "grad_norm": 0.23344322072994478, | |
| "learning_rate": 8.87573494180163e-05, | |
| "loss": 0.511, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 1.7548387096774194, | |
| "grad_norm": 0.24030220361622895, | |
| "learning_rate": 8.77628150384741e-05, | |
| "loss": 0.5254, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 1.7634408602150538, | |
| "grad_norm": 0.23404925065530963, | |
| "learning_rate": 8.676950795290802e-05, | |
| "loss": 0.5189, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 1.772043010752688, | |
| "grad_norm": 0.23053178874827976, | |
| "learning_rate": 8.57775277822526e-05, | |
| "loss": 0.5252, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 1.7806451612903227, | |
| "grad_norm": 0.23781516582822665, | |
| "learning_rate": 8.478697401436323e-05, | |
| "loss": 0.5146, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 1.789247311827957, | |
| "grad_norm": 0.2276955393985772, | |
| "learning_rate": 8.379794599403836e-05, | |
| "loss": 0.5169, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 1.7978494623655914, | |
| "grad_norm": 0.23946966468353473, | |
| "learning_rate": 8.281054291305566e-05, | |
| "loss": 0.5181, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 1.8064516129032258, | |
| "grad_norm": 0.22844184267941248, | |
| "learning_rate": 8.182486380022426e-05, | |
| "loss": 0.5009, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 1.8150537634408601, | |
| "grad_norm": 0.2386146449957716, | |
| "learning_rate": 8.084100751145277e-05, | |
| "loss": 0.5167, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 1.8236559139784947, | |
| "grad_norm": 0.22996861067382443, | |
| "learning_rate": 7.985907271983467e-05, | |
| "loss": 0.5172, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 1.832258064516129, | |
| "grad_norm": 0.23122069068423765, | |
| "learning_rate": 7.887915790575241e-05, | |
| "loss": 0.512, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 1.8408602150537634, | |
| "grad_norm": 0.226577342689351, | |
| "learning_rate": 7.790136134700042e-05, | |
| "loss": 0.5181, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 1.849462365591398, | |
| "grad_norm": 0.24205636496242264, | |
| "learning_rate": 7.692578110892876e-05, | |
| "loss": 0.5206, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 1.8580645161290321, | |
| "grad_norm": 0.24349873078348327, | |
| "learning_rate": 7.595251503460778e-05, | |
| "loss": 0.5226, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 1.8666666666666667, | |
| "grad_norm": 0.23758597762780134, | |
| "learning_rate": 7.498166073501529e-05, | |
| "loss": 0.5244, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 1.875268817204301, | |
| "grad_norm": 0.2284958254909026, | |
| "learning_rate": 7.401331557924707e-05, | |
| "loss": 0.505, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 1.8838709677419354, | |
| "grad_norm": 0.24598875919420438, | |
| "learning_rate": 7.304757668475122e-05, | |
| "loss": 0.5317, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 1.89247311827957, | |
| "grad_norm": 0.24113280579774263, | |
| "learning_rate": 7.208454090758832e-05, | |
| "loss": 0.5178, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 1.9010752688172043, | |
| "grad_norm": 0.2362286661628526, | |
| "learning_rate": 7.112430483271746e-05, | |
| "loss": 0.5202, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 1.9096774193548387, | |
| "grad_norm": 0.23413327678165968, | |
| "learning_rate": 7.016696476430931e-05, | |
| "loss": 0.4944, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 1.9182795698924733, | |
| "grad_norm": 0.23501629015109748, | |
| "learning_rate": 6.921261671608791e-05, | |
| "loss": 0.5153, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 1.9268817204301074, | |
| "grad_norm": 0.27066815871872996, | |
| "learning_rate": 6.826135640170101e-05, | |
| "loss": 0.5252, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 1.935483870967742, | |
| "grad_norm": 0.2439503871040344, | |
| "learning_rate": 6.731327922512074e-05, | |
| "loss": 0.5112, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 1.9440860215053763, | |
| "grad_norm": 0.23161029568643865, | |
| "learning_rate": 6.636848027107544e-05, | |
| "loss": 0.5307, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 1.9526881720430107, | |
| "grad_norm": 0.24210935717094484, | |
| "learning_rate": 6.54270542955134e-05, | |
| "loss": 0.5011, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 1.9612903225806453, | |
| "grad_norm": 0.24063211639329316, | |
| "learning_rate": 6.44890957160994e-05, | |
| "loss": 0.5234, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 1.9698924731182794, | |
| "grad_norm": 0.23615678793407338, | |
| "learning_rate": 6.355469860274574e-05, | |
| "loss": 0.5046, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 1.978494623655914, | |
| "grad_norm": 0.22842985809994715, | |
| "learning_rate": 6.262395666817724e-05, | |
| "loss": 0.5012, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 1.9870967741935484, | |
| "grad_norm": 0.23784853454359128, | |
| "learning_rate": 6.169696325853312e-05, | |
| "loss": 0.519, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 1.9956989247311827, | |
| "grad_norm": 0.2342469521451372, | |
| "learning_rate": 6.077381134400462e-05, | |
| "loss": 0.5002, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 2.0043010752688173, | |
| "grad_norm": 0.22833969780513486, | |
| "learning_rate": 5.985459350951121e-05, | |
| "loss": 0.4845, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 2.0129032258064514, | |
| "grad_norm": 0.2544744796985262, | |
| "learning_rate": 5.893940194541492e-05, | |
| "loss": 0.464, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 2.021505376344086, | |
| "grad_norm": 0.24382320744489763, | |
| "learning_rate": 5.802832843827419e-05, | |
| "loss": 0.455, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 2.0301075268817206, | |
| "grad_norm": 0.24697875997895144, | |
| "learning_rate": 5.712146436163863e-05, | |
| "loss": 0.4598, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 2.0387096774193547, | |
| "grad_norm": 0.2513303656347981, | |
| "learning_rate": 5.6218900666884975e-05, | |
| "loss": 0.4515, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 2.0473118279569893, | |
| "grad_norm": 0.26236895102262536, | |
| "learning_rate": 5.5320727874095014e-05, | |
| "loss": 0.4598, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 2.055913978494624, | |
| "grad_norm": 0.25978280232463796, | |
| "learning_rate": 5.4427036062977744e-05, | |
| "loss": 0.451, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 2.064516129032258, | |
| "grad_norm": 0.26162376545024385, | |
| "learning_rate": 5.3537914863834374e-05, | |
| "loss": 0.4563, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 2.0731182795698926, | |
| "grad_norm": 0.27200603174186766, | |
| "learning_rate": 5.265345344856979e-05, | |
| "loss": 0.4595, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 2.0817204301075267, | |
| "grad_norm": 0.2516133186216482, | |
| "learning_rate": 5.1773740521748793e-05, | |
| "loss": 0.4543, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 2.0903225806451613, | |
| "grad_norm": 0.2559266512535566, | |
| "learning_rate": 5.089886431169999e-05, | |
| "loss": 0.4576, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 2.098924731182796, | |
| "grad_norm": 0.26643890557668787, | |
| "learning_rate": 5.0028912561667104e-05, | |
| "loss": 0.4651, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 2.10752688172043, | |
| "grad_norm": 0.2594035563202495, | |
| "learning_rate": 4.916397252100892e-05, | |
| "loss": 0.4581, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 2.1161290322580646, | |
| "grad_norm": 0.2589087763537241, | |
| "learning_rate": 4.830413093644913e-05, | |
| "loss": 0.4508, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 2.1247311827956987, | |
| "grad_norm": 0.25564217019549307, | |
| "learning_rate": 4.744947404337605e-05, | |
| "loss": 0.4498, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 2.1333333333333333, | |
| "grad_norm": 0.25846858564022296, | |
| "learning_rate": 4.660008755719397e-05, | |
| "loss": 0.4439, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 2.141935483870968, | |
| "grad_norm": 0.2615309859057683, | |
| "learning_rate": 4.5756056664726554e-05, | |
| "loss": 0.4621, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 2.150537634408602, | |
| "grad_norm": 0.2751137583558904, | |
| "learning_rate": 4.491746601567343e-05, | |
| "loss": 0.4542, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 2.1591397849462366, | |
| "grad_norm": 0.2640093266145403, | |
| "learning_rate": 4.408439971412013e-05, | |
| "loss": 0.4709, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 2.167741935483871, | |
| "grad_norm": 0.26804533689921306, | |
| "learning_rate": 4.325694131010346e-05, | |
| "loss": 0.4545, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 2.1763440860215053, | |
| "grad_norm": 0.26953837584010415, | |
| "learning_rate": 4.243517379123193e-05, | |
| "loss": 0.447, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 2.18494623655914, | |
| "grad_norm": 0.26331494847521747, | |
| "learning_rate": 4.161917957436271e-05, | |
| "loss": 0.4454, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 2.193548387096774, | |
| "grad_norm": 0.26650178799391877, | |
| "learning_rate": 4.080904049733607e-05, | |
| "loss": 0.4452, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 2.2021505376344086, | |
| "grad_norm": 0.2596397199813554, | |
| "learning_rate": 4.0004837810767294e-05, | |
| "loss": 0.455, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 2.210752688172043, | |
| "grad_norm": 0.2855359315189352, | |
| "learning_rate": 3.9206652169898364e-05, | |
| "loss": 0.4619, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 2.2193548387096773, | |
| "grad_norm": 0.2715169650377671, | |
| "learning_rate": 3.841456362650837e-05, | |
| "loss": 0.4615, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 2.227956989247312, | |
| "grad_norm": 0.2628755048951637, | |
| "learning_rate": 3.7628651620885444e-05, | |
| "loss": 0.4645, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 2.236559139784946, | |
| "grad_norm": 0.28592361434576347, | |
| "learning_rate": 3.6848994973859105e-05, | |
| "loss": 0.4692, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 2.2451612903225806, | |
| "grad_norm": 0.26764492546621405, | |
| "learning_rate": 3.607567187889538e-05, | |
| "loss": 0.4563, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 2.253763440860215, | |
| "grad_norm": 0.271421407542708, | |
| "learning_rate": 3.5308759894254496e-05, | |
| "loss": 0.4671, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 2.2623655913978493, | |
| "grad_norm": 0.2653346076557134, | |
| "learning_rate": 3.45483359352125e-05, | |
| "loss": 0.4535, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 2.270967741935484, | |
| "grad_norm": 0.2672124362703016, | |
| "learning_rate": 3.379447626634712e-05, | |
| "loss": 0.4523, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 2.279569892473118, | |
| "grad_norm": 0.2815570906050588, | |
| "learning_rate": 3.304725649388919e-05, | |
| "loss": 0.468, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 2.2881720430107526, | |
| "grad_norm": 0.27533793807613155, | |
| "learning_rate": 3.230675155813979e-05, | |
| "loss": 0.4587, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 2.296774193548387, | |
| "grad_norm": 0.2718010567194757, | |
| "learning_rate": 3.1573035725954344e-05, | |
| "loss": 0.4518, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 2.3053763440860213, | |
| "grad_norm": 0.26881948505806064, | |
| "learning_rate": 3.084618258329443e-05, | |
| "loss": 0.4511, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 2.313978494623656, | |
| "grad_norm": 0.2735905144552066, | |
| "learning_rate": 3.012626502784729e-05, | |
| "loss": 0.4437, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 2.3225806451612905, | |
| "grad_norm": 0.2753569674239171, | |
| "learning_rate": 2.9413355261715192e-05, | |
| "loss": 0.457, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 2.3311827956989246, | |
| "grad_norm": 0.27264140236415646, | |
| "learning_rate": 2.87075247841738e-05, | |
| "loss": 0.4562, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 2.339784946236559, | |
| "grad_norm": 0.2750362048193341, | |
| "learning_rate": 2.8008844384501566e-05, | |
| "loss": 0.4578, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 2.3483870967741938, | |
| "grad_norm": 0.2743768587791329, | |
| "learning_rate": 2.7317384134879965e-05, | |
| "loss": 0.4569, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 2.356989247311828, | |
| "grad_norm": 0.27994254494622967, | |
| "learning_rate": 2.6633213383365906e-05, | |
| "loss": 0.4554, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 2.3655913978494625, | |
| "grad_norm": 0.273412382006339, | |
| "learning_rate": 2.595640074693664e-05, | |
| "loss": 0.4521, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 2.3741935483870966, | |
| "grad_norm": 0.26693176903025, | |
| "learning_rate": 2.5287014104607975e-05, | |
| "loss": 0.4472, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 2.382795698924731, | |
| "grad_norm": 0.26563176753886897, | |
| "learning_rate": 2.4625120590626595e-05, | |
| "loss": 0.4487, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 2.3913978494623658, | |
| "grad_norm": 0.27008804180981577, | |
| "learning_rate": 2.397078658773699e-05, | |
| "loss": 0.4611, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "grad_norm": 0.2624172113992881, | |
| "learning_rate": 2.3324077720523785e-05, | |
| "loss": 0.4461, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 2.4086021505376345, | |
| "grad_norm": 0.27677007427727796, | |
| "learning_rate": 2.2685058848830076e-05, | |
| "loss": 0.4536, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 2.4172043010752686, | |
| "grad_norm": 0.27569949386731046, | |
| "learning_rate": 2.2053794061252675e-05, | |
| "loss": 0.4474, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 2.425806451612903, | |
| "grad_norm": 0.267987384900051, | |
| "learning_rate": 2.1430346668714175e-05, | |
| "loss": 0.4487, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 2.434408602150538, | |
| "grad_norm": 0.2669869480120869, | |
| "learning_rate": 2.0814779198113687e-05, | |
| "loss": 0.455, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 2.443010752688172, | |
| "grad_norm": 0.26354591329030985, | |
| "learning_rate": 2.020715338605581e-05, | |
| "loss": 0.4616, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 2.4516129032258065, | |
| "grad_norm": 0.2785720489510887, | |
| "learning_rate": 1.9607530172658715e-05, | |
| "loss": 0.4419, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 2.4602150537634406, | |
| "grad_norm": 0.27851326696395934, | |
| "learning_rate": 1.9015969695442704e-05, | |
| "loss": 0.4572, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 2.468817204301075, | |
| "grad_norm": 0.27236418612561275, | |
| "learning_rate": 1.8432531283298458e-05, | |
| "loss": 0.453, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 2.47741935483871, | |
| "grad_norm": 0.2816067593156165, | |
| "learning_rate": 1.7857273450537227e-05, | |
| "loss": 0.4607, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 2.486021505376344, | |
| "grad_norm": 0.27424768533392857, | |
| "learning_rate": 1.7290253891022e-05, | |
| "loss": 0.4385, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 2.4946236559139785, | |
| "grad_norm": 0.2766562755277204, | |
| "learning_rate": 1.673152947238139e-05, | |
| "loss": 0.4567, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 2.5032258064516126, | |
| "grad_norm": 0.2614813998070769, | |
| "learning_rate": 1.618115623030625e-05, | |
| "loss": 0.4428, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 2.511827956989247, | |
| "grad_norm": 0.272212168117098, | |
| "learning_rate": 1.5639189362929695e-05, | |
| "loss": 0.453, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 2.520430107526882, | |
| "grad_norm": 0.2747575768340182, | |
| "learning_rate": 1.5105683225291211e-05, | |
| "loss": 0.4586, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 2.5290322580645164, | |
| "grad_norm": 0.2711344401379175, | |
| "learning_rate": 1.4580691323885209e-05, | |
| "loss": 0.4506, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 2.5376344086021505, | |
| "grad_norm": 0.27821711929840565, | |
| "learning_rate": 1.4064266311294793e-05, | |
| "loss": 0.4669, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 2.546236559139785, | |
| "grad_norm": 0.2807569627250837, | |
| "learning_rate": 1.3556459980911085e-05, | |
| "loss": 0.4567, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 2.554838709677419, | |
| "grad_norm": 0.27478294823892596, | |
| "learning_rate": 1.305732326173882e-05, | |
| "loss": 0.4629, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 2.563440860215054, | |
| "grad_norm": 0.2694458161205595, | |
| "learning_rate": 1.2566906213288388e-05, | |
| "loss": 0.4518, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 2.5720430107526884, | |
| "grad_norm": 0.26379184313364523, | |
| "learning_rate": 1.2085258020555556e-05, | |
| "loss": 0.4489, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 2.5806451612903225, | |
| "grad_norm": 0.27978762666210216, | |
| "learning_rate": 1.1612426989088232e-05, | |
| "loss": 0.4493, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 2.589247311827957, | |
| "grad_norm": 0.28005079300575625, | |
| "learning_rate": 1.1148460540142125e-05, | |
| "loss": 0.4501, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 2.5978494623655912, | |
| "grad_norm": 0.2765800997942907, | |
| "learning_rate": 1.0693405205924579e-05, | |
| "loss": 0.4599, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 2.606451612903226, | |
| "grad_norm": 0.26856411845690764, | |
| "learning_rate": 1.0247306624927789e-05, | |
| "loss": 0.4489, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 2.6150537634408604, | |
| "grad_norm": 0.2758438510594412, | |
| "learning_rate": 9.810209537351645e-06, | |
| "loss": 0.4427, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 2.6236559139784945, | |
| "grad_norm": 0.2808673768423207, | |
| "learning_rate": 9.382157780616606e-06, | |
| "loss": 0.4369, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 2.632258064516129, | |
| "grad_norm": 0.26355013415952777, | |
| "learning_rate": 8.963194284967202e-06, | |
| "loss": 0.4543, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 2.6408602150537632, | |
| "grad_norm": 0.26682463219677827, | |
| "learning_rate": 8.553361069166388e-06, | |
| "loss": 0.4448, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 2.649462365591398, | |
| "grad_norm": 0.2824919005760602, | |
| "learning_rate": 8.15269923628147e-06, | |
| "loss": 0.463, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 2.6580645161290324, | |
| "grad_norm": 0.2606472077942361, | |
| "learning_rate": 7.761248969561729e-06, | |
| "loss": 0.4435, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 2.6666666666666665, | |
| "grad_norm": 0.2716672911173004, | |
| "learning_rate": 7.379049528408433e-06, | |
| "loss": 0.4436, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 2.675268817204301, | |
| "grad_norm": 0.2692189798747163, | |
| "learning_rate": 7.00613924443726e-06, | |
| "loss": 0.456, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 2.6838709677419352, | |
| "grad_norm": 0.2751219149255623, | |
| "learning_rate": 6.642555517634197e-06, | |
| "loss": 0.4486, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 2.69247311827957, | |
| "grad_norm": 0.26558006981794996, | |
| "learning_rate": 6.288334812604324e-06, | |
| "loss": 0.448, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 2.7010752688172044, | |
| "grad_norm": 0.2859552474881587, | |
| "learning_rate": 5.943512654914951e-06, | |
| "loss": 0.4357, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 2.709677419354839, | |
| "grad_norm": 0.27548323875619696, | |
| "learning_rate": 5.6081236275325355e-06, | |
| "loss": 0.4624, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 2.718279569892473, | |
| "grad_norm": 0.2806542534666262, | |
| "learning_rate": 5.282201367354245e-06, | |
| "loss": 0.4505, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 2.7268817204301077, | |
| "grad_norm": 0.26475293341781525, | |
| "learning_rate": 4.965778561834644e-06, | |
| "loss": 0.4386, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 2.735483870967742, | |
| "grad_norm": 0.28440947626379526, | |
| "learning_rate": 4.658886945707164e-06, | |
| "loss": 0.4542, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 2.7440860215053764, | |
| "grad_norm": 0.2696410246213196, | |
| "learning_rate": 4.361557297801499e-06, | |
| "loss": 0.4416, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 2.752688172043011, | |
| "grad_norm": 0.27708040093077324, | |
| "learning_rate": 4.073819437956694e-06, | |
| "loss": 0.4471, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 2.761290322580645, | |
| "grad_norm": 0.2615573424318419, | |
| "learning_rate": 3.7957022240304173e-06, | |
| "loss": 0.4505, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 2.7698924731182797, | |
| "grad_norm": 0.2776744589099636, | |
| "learning_rate": 3.5272335490047937e-06, | |
| "loss": 0.4416, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 2.778494623655914, | |
| "grad_norm": 0.28362883428775265, | |
| "learning_rate": 3.2684403381889272e-06, | |
| "loss": 0.4562, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 2.7870967741935484, | |
| "grad_norm": 0.2659089658104915, | |
| "learning_rate": 3.019348546518508e-06, | |
| "loss": 0.4542, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 2.795698924731183, | |
| "grad_norm": 0.27817617897785124, | |
| "learning_rate": 2.7799831559527258e-06, | |
| "loss": 0.4535, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 2.804301075268817, | |
| "grad_norm": 0.2767257855534767, | |
| "learning_rate": 2.550368172968809e-06, | |
| "loss": 0.4438, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 2.8129032258064517, | |
| "grad_norm": 0.2659520383981123, | |
| "learning_rate": 2.3305266261542945e-06, | |
| "loss": 0.4648, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 2.821505376344086, | |
| "grad_norm": 0.26860959961537617, | |
| "learning_rate": 2.1204805638975646e-06, | |
| "loss": 0.4478, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 2.8301075268817204, | |
| "grad_norm": 0.2772290842479072, | |
| "learning_rate": 1.9202510521763696e-06, | |
| "loss": 0.4556, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 2.838709677419355, | |
| "grad_norm": 0.2773729267266774, | |
| "learning_rate": 1.7298581724452978e-06, | |
| "loss": 0.4546, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 2.847311827956989, | |
| "grad_norm": 0.288243164539903, | |
| "learning_rate": 1.5493210196216079e-06, | |
| "loss": 0.457, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 2.8559139784946237, | |
| "grad_norm": 0.27792696429990915, | |
| "learning_rate": 1.378657700170205e-06, | |
| "loss": 0.4585, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 2.864516129032258, | |
| "grad_norm": 0.27882422779333554, | |
| "learning_rate": 1.2178853302877159e-06, | |
| "loss": 0.4486, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 2.8731182795698924, | |
| "grad_norm": 0.28133695329057534, | |
| "learning_rate": 1.0670200341858394e-06, | |
| "loss": 0.4502, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 2.881720430107527, | |
| "grad_norm": 0.2859131700306455, | |
| "learning_rate": 9.260769424742633e-07, | |
| "loss": 0.4452, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 2.8903225806451616, | |
| "grad_norm": 0.26707708683830167, | |
| "learning_rate": 7.950701906431324e-07, | |
| "loss": 0.4498, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 2.8989247311827957, | |
| "grad_norm": 0.2716338526967156, | |
| "learning_rate": 6.740129176453725e-07, | |
| "loss": 0.4568, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 2.90752688172043, | |
| "grad_norm": 0.29044410591630965, | |
| "learning_rate": 5.629172645789882e-07, | |
| "loss": 0.4517, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 2.9161290322580644, | |
| "grad_norm": 0.27054199049752875, | |
| "learning_rate": 4.617943734694152e-07, | |
| "loss": 0.4559, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 2.924731182795699, | |
| "grad_norm": 0.27515263341966606, | |
| "learning_rate": 3.7065438615198066e-07, | |
| "loss": 0.4576, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 2.9333333333333336, | |
| "grad_norm": 0.2653013119464242, | |
| "learning_rate": 2.8950644325485e-07, | |
| "loss": 0.449, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 2.9419354838709677, | |
| "grad_norm": 0.2791971697586357, | |
| "learning_rate": 2.183586832822493e-07, | |
| "loss": 0.4562, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 2.9505376344086023, | |
| "grad_norm": 0.2668878384341789, | |
| "learning_rate": 1.572182417982515e-07, | |
| "loss": 0.447, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 2.9591397849462364, | |
| "grad_norm": 0.2759015749441583, | |
| "learning_rate": 1.0609125071109338e-07, | |
| "loss": 0.4567, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 2.967741935483871, | |
| "grad_norm": 0.283562302360861, | |
| "learning_rate": 6.49828376582673e-08, | |
| "loss": 0.4503, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 2.9763440860215056, | |
| "grad_norm": 0.28059419780497596, | |
| "learning_rate": 3.3897125492188266e-08, | |
| "loss": 0.4525, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 2.9849462365591397, | |
| "grad_norm": 0.2757143204657877, | |
| "learning_rate": 1.2837231866746902e-08, | |
| "loss": 0.4455, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 2.9935483870967743, | |
| "grad_norm": 0.26729797929321086, | |
| "learning_rate": 1.8052689246150779e-09, | |
| "loss": 0.4443, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 2.9987096774193547, | |
| "step": 3486, | |
| "total_flos": 6514990399881216.0, | |
| "train_loss": 0.5279938033170147, | |
| "train_runtime": 9440.7969, | |
| "train_samples_per_second": 5.91, | |
| "train_steps_per_second": 0.369 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 3486, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 6514990399881216.0, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |