| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.0, | |
| "eval_steps": 500, | |
| "global_step": 1946, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.010277492291880781, | |
| "grad_norm": 1.4090419657971816, | |
| "learning_rate": 3.0612244897959185e-06, | |
| "loss": 0.7119, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.020554984583761562, | |
| "grad_norm": 1.5528273772375807, | |
| "learning_rate": 6.122448979591837e-06, | |
| "loss": 0.6333, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.030832476875642344, | |
| "grad_norm": 0.6875726812709237, | |
| "learning_rate": 9.183673469387756e-06, | |
| "loss": 0.609, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.041109969167523124, | |
| "grad_norm": 0.5630238576069669, | |
| "learning_rate": 1.2244897959183674e-05, | |
| "loss": 0.5998, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.051387461459403906, | |
| "grad_norm": 0.49796436120814924, | |
| "learning_rate": 1.530612244897959e-05, | |
| "loss": 0.5692, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.06166495375128469, | |
| "grad_norm": 0.48386904448400564, | |
| "learning_rate": 1.836734693877551e-05, | |
| "loss": 0.5833, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.07194244604316546, | |
| "grad_norm": 0.5317730488962762, | |
| "learning_rate": 2.1428571428571428e-05, | |
| "loss": 0.5632, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.08221993833504625, | |
| "grad_norm": 0.4884815339298427, | |
| "learning_rate": 2.448979591836735e-05, | |
| "loss": 0.5804, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.09249743062692703, | |
| "grad_norm": 0.4738318035314195, | |
| "learning_rate": 2.7551020408163265e-05, | |
| "loss": 0.5805, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.10277492291880781, | |
| "grad_norm": 0.5201543078674129, | |
| "learning_rate": 2.9999219710414462e-05, | |
| "loss": 0.5904, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.1130524152106886, | |
| "grad_norm": 0.5194572571460542, | |
| "learning_rate": 2.9971918097900504e-05, | |
| "loss": 0.5891, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.12332990750256938, | |
| "grad_norm": 0.5451888545802674, | |
| "learning_rate": 2.9905683148398642e-05, | |
| "loss": 0.5912, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.13360739979445016, | |
| "grad_norm": 0.4551734320613588, | |
| "learning_rate": 2.9800687100869334e-05, | |
| "loss": 0.5834, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.14388489208633093, | |
| "grad_norm": 0.4201748639994036, | |
| "learning_rate": 2.9657202989567393e-05, | |
| "loss": 0.564, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.15416238437821173, | |
| "grad_norm": 0.39502869903115356, | |
| "learning_rate": 2.94756039340371e-05, | |
| "loss": 0.5842, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.1644398766700925, | |
| "grad_norm": 0.45207720954882885, | |
| "learning_rate": 2.9256362168843153e-05, | |
| "loss": 0.5748, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.1747173689619733, | |
| "grad_norm": 0.5380473565961726, | |
| "learning_rate": 2.90000478155605e-05, | |
| "loss": 0.5773, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.18499486125385406, | |
| "grad_norm": 0.43815842467576244, | |
| "learning_rate": 2.870732740021648e-05, | |
| "loss": 0.5639, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.19527235354573483, | |
| "grad_norm": 0.4809575724437436, | |
| "learning_rate": 2.837896212004041e-05, | |
| "loss": 0.5681, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.20554984583761562, | |
| "grad_norm": 0.43270245528917933, | |
| "learning_rate": 2.801580586402798e-05, | |
| "loss": 0.5714, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.2158273381294964, | |
| "grad_norm": 0.430262537558555, | |
| "learning_rate": 2.7618802992467718e-05, | |
| "loss": 0.5576, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.2261048304213772, | |
| "grad_norm": 0.44110443094179325, | |
| "learning_rate": 2.7188985881203685e-05, | |
| "loss": 0.5691, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.23638232271325796, | |
| "grad_norm": 0.4032486283659276, | |
| "learning_rate": 2.672747223702045e-05, | |
| "loss": 0.5762, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.24665981500513876, | |
| "grad_norm": 0.4242787937283615, | |
| "learning_rate": 2.623546219113139e-05, | |
| "loss": 0.5643, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.2569373072970195, | |
| "grad_norm": 0.4836565922888775, | |
| "learning_rate": 2.5714235178328554e-05, | |
| "loss": 0.5774, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.2672147995889003, | |
| "grad_norm": 0.42526900764591785, | |
| "learning_rate": 2.516514660990954e-05, | |
| "loss": 0.5627, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.2774922918807811, | |
| "grad_norm": 0.3683992684534281, | |
| "learning_rate": 2.4589624349033273e-05, | |
| "loss": 0.5594, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.28776978417266186, | |
| "grad_norm": 0.41496288894802913, | |
| "learning_rate": 2.3989164997670207e-05, | |
| "loss": 0.5562, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.29804727646454265, | |
| "grad_norm": 0.3635108549251573, | |
| "learning_rate": 2.3365330004802443e-05, | |
| "loss": 0.5301, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.30832476875642345, | |
| "grad_norm": 0.3913157769758501, | |
| "learning_rate": 2.271974160599415e-05, | |
| "loss": 0.5506, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.3186022610483042, | |
| "grad_norm": 0.4061061729653975, | |
| "learning_rate": 2.2054078604891052e-05, | |
| "loss": 0.5517, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.328879753340185, | |
| "grad_norm": 0.38228520965937396, | |
| "learning_rate": 2.1370072007618947e-05, | |
| "loss": 0.5622, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.3391572456320658, | |
| "grad_norm": 0.4320279036786554, | |
| "learning_rate": 2.06695005214336e-05, | |
| "loss": 0.5447, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.3494347379239466, | |
| "grad_norm": 0.37703622035095746, | |
| "learning_rate": 1.995418592932751e-05, | |
| "loss": 0.541, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.3597122302158273, | |
| "grad_norm": 0.34784635261686314, | |
| "learning_rate": 1.9225988352621445e-05, | |
| "loss": 0.5433, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.3699897225077081, | |
| "grad_norm": 0.40884262314646, | |
| "learning_rate": 1.8486801413860122e-05, | |
| "loss": 0.5416, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.3802672147995889, | |
| "grad_norm": 0.3872918353959618, | |
| "learning_rate": 1.7738547312590426e-05, | |
| "loss": 0.5455, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.39054470709146966, | |
| "grad_norm": 0.3909142802295015, | |
| "learning_rate": 1.6983171826827357e-05, | |
| "loss": 0.5331, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.40082219938335045, | |
| "grad_norm": 0.4247312946608561, | |
| "learning_rate": 1.6222639253205944e-05, | |
| "loss": 0.5468, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.41109969167523125, | |
| "grad_norm": 0.3779654218810271, | |
| "learning_rate": 1.5458927298976893e-05, | |
| "loss": 0.5274, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.42137718396711205, | |
| "grad_norm": 0.3835548049313215, | |
| "learning_rate": 1.4694021939128921e-05, | |
| "loss": 0.5433, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.4316546762589928, | |
| "grad_norm": 0.3430867289416931, | |
| "learning_rate": 1.3929912252011512e-05, | |
| "loss": 0.5292, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.4419321685508736, | |
| "grad_norm": 0.33530951933279407, | |
| "learning_rate": 1.3168585246887604e-05, | |
| "loss": 0.5288, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.4522096608427544, | |
| "grad_norm": 0.38479043994508066, | |
| "learning_rate": 1.2412020696866724e-05, | |
| "loss": 0.5356, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.4624871531346352, | |
| "grad_norm": 0.38638629022508425, | |
| "learning_rate": 1.1662185990655285e-05, | |
| "loss": 0.5491, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.4727646454265159, | |
| "grad_norm": 0.35722768346449285, | |
| "learning_rate": 1.0921031016511509e-05, | |
| "loss": 0.5253, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.4830421377183967, | |
| "grad_norm": 0.3735863030025159, | |
| "learning_rate": 1.0190483091708966e-05, | |
| "loss": 0.5246, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.4933196300102775, | |
| "grad_norm": 0.38885388838001816, | |
| "learning_rate": 9.472441950694257e-06, | |
| "loss": 0.5311, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.5035971223021583, | |
| "grad_norm": 0.3260936892474308, | |
| "learning_rate": 8.768774804971705e-06, | |
| "loss": 0.5341, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.513874614594039, | |
| "grad_norm": 0.3552425546911982, | |
| "learning_rate": 8.081311487561545e-06, | |
| "loss": 0.5141, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.5241521068859198, | |
| "grad_norm": 0.34385819298403936, | |
| "learning_rate": 7.411839694657976e-06, | |
| "loss": 0.5165, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.5344295991778006, | |
| "grad_norm": 0.3585285471621075, | |
| "learning_rate": 6.762100336860873e-06, | |
| "loss": 0.5085, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.5447070914696814, | |
| "grad_norm": 0.3607978964114949, | |
| "learning_rate": 6.133783012069853e-06, | |
| "loss": 0.5158, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.5549845837615622, | |
| "grad_norm": 0.3297200319919343, | |
| "learning_rate": 5.528521611813133e-06, | |
| "loss": 0.5313, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.5652620760534429, | |
| "grad_norm": 0.3379144676775949, | |
| "learning_rate": 4.947890072436609e-06, | |
| "loss": 0.5193, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.5755395683453237, | |
| "grad_norm": 0.333040604472348, | |
| "learning_rate": 4.393398282201788e-06, | |
| "loss": 0.5108, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.5858170606372045, | |
| "grad_norm": 0.3389958077738962, | |
| "learning_rate": 3.866488154935951e-06, | |
| "loss": 0.5356, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.5960945529290853, | |
| "grad_norm": 0.3518113489771542, | |
| "learning_rate": 3.3685298804446406e-06, | |
| "loss": 0.4915, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.6063720452209661, | |
| "grad_norm": 0.39595048545111644, | |
| "learning_rate": 2.900818361437053e-06, | |
| "loss": 0.5135, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.6166495375128469, | |
| "grad_norm": 0.3359427514687364, | |
| "learning_rate": 2.4645698462297583e-06, | |
| "loss": 0.5052, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.6269270298047277, | |
| "grad_norm": 0.3227737246533257, | |
| "learning_rate": 2.0609187659852914e-06, | |
| "loss": 0.5023, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.6372045220966084, | |
| "grad_norm": 0.32500420569729616, | |
| "learning_rate": 1.6909147847099526e-06, | |
| "loss": 0.5359, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.6474820143884892, | |
| "grad_norm": 0.3322817521202919, | |
| "learning_rate": 1.3555200696822235e-06, | |
| "loss": 0.5073, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.65775950668037, | |
| "grad_norm": 0.3496806144821667, | |
| "learning_rate": 1.0556067894097571e-06, | |
| "loss": 0.5242, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.6680369989722508, | |
| "grad_norm": 0.30896880735642235, | |
| "learning_rate": 7.919548456213516e-07, | |
| "loss": 0.5155, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.6783144912641316, | |
| "grad_norm": 0.3257972583698041, | |
| "learning_rate": 5.652498451916799e-07, | |
| "loss": 0.5158, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.6885919835560124, | |
| "grad_norm": 0.3201104743635745, | |
| "learning_rate": 3.760813172726457e-07, | |
| "loss": 0.5142, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.6988694758478932, | |
| "grad_norm": 0.34686855669764965, | |
| "learning_rate": 2.2494118026754551e-07, | |
| "loss": 0.5232, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.7091469681397738, | |
| "grad_norm": 0.30350285406793465, | |
| "learning_rate": 1.1222246263458469e-07, | |
| "loss": 0.5148, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.7194244604316546, | |
| "grad_norm": 0.31292172166688387, | |
| "learning_rate": 3.821828084619727e-08, | |
| "loss": 0.5118, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.7297019527235354, | |
| "grad_norm": 0.32011342982016167, | |
| "learning_rate": 3.1210771619027966e-09, | |
| "loss": 0.5009, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.7399794450154162, | |
| "grad_norm": 0.35647392118546595, | |
| "learning_rate": 7.022119196808396e-09, | |
| "loss": 0.5148, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.750256937307297, | |
| "grad_norm": 0.3191068876583651, | |
| "learning_rate": 4.991126258710177e-08, | |
| "loss": 0.5122, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.7605344295991778, | |
| "grad_norm": 0.32588243816148815, | |
| "learning_rate": 1.3167697736969798e-07, | |
| "loss": 0.4926, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.7708119218910586, | |
| "grad_norm": 0.3201190610520271, | |
| "learning_rate": 2.5210663800745493e-07, | |
| "loss": 0.5149, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.7810894141829393, | |
| "grad_norm": 0.35476449861767284, | |
| "learning_rate": 4.108870763057343e-07, | |
| "loss": 0.5118, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.7913669064748201, | |
| "grad_norm": 0.3125164507289312, | |
| "learning_rate": 6.076053957825395e-07, | |
| "loss": 0.5157, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.8016443987667009, | |
| "grad_norm": 0.3366477883256346, | |
| "learning_rate": 8.417500453744864e-07, | |
| "loss": 0.5292, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.8119218910585817, | |
| "grad_norm": 0.3269719144972129, | |
| "learning_rate": 1.1127121496865256e-06, | |
| "loss": 0.5143, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.8221993833504625, | |
| "grad_norm": 0.3126014006138838, | |
| "learning_rate": 1.419787092326219e-06, | |
| "loss": 0.5155, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.8324768756423433, | |
| "grad_norm": 0.36354641808871235, | |
| "learning_rate": 1.7621763482051812e-06, | |
| "loss": 0.5044, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.8427543679342241, | |
| "grad_norm": 0.36699145399738387, | |
| "learning_rate": 2.138989560043002e-06, | |
| "loss": 0.4822, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.8530318602261048, | |
| "grad_norm": 0.33037450883968417, | |
| "learning_rate": 2.549246853673793e-06, | |
| "loss": 0.5087, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.8633093525179856, | |
| "grad_norm": 0.3257167825813466, | |
| "learning_rate": 2.991881386134589e-06, | |
| "loss": 0.5192, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.8735868448098664, | |
| "grad_norm": 0.3100164686825301, | |
| "learning_rate": 3.465742119909566e-06, | |
| "loss": 0.498, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.8838643371017472, | |
| "grad_norm": 0.34697143051842105, | |
| "learning_rate": 3.969596816115712e-06, | |
| "loss": 0.506, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.894141829393628, | |
| "grad_norm": 0.33866121968347435, | |
| "learning_rate": 4.502135238846574e-06, | |
| "loss": 0.5293, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.9044193216855088, | |
| "grad_norm": 0.36845378999111805, | |
| "learning_rate": 5.061972562341309e-06, | |
| "loss": 0.4909, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.9146968139773896, | |
| "grad_norm": 0.387252650542103, | |
| "learning_rate": 5.647652972118995e-06, | |
| "loss": 0.5312, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.9249743062692704, | |
| "grad_norm": 0.3647647773617935, | |
| "learning_rate": 6.257653450713748e-06, | |
| "loss": 0.5101, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.935251798561151, | |
| "grad_norm": 0.33897356731666334, | |
| "learning_rate": 6.890387738166038e-06, | |
| "loss": 0.5166, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.9455292908530318, | |
| "grad_norm": 0.3868822782515743, | |
| "learning_rate": 7.5442104569713904e-06, | |
| "loss": 0.4984, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.9558067831449126, | |
| "grad_norm": 0.35781401043924416, | |
| "learning_rate": 8.217421390759708e-06, | |
| "loss": 0.4994, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.9660842754367934, | |
| "grad_norm": 0.384761931416951, | |
| "learning_rate": 8.908269905578991e-06, | |
| "loss": 0.5099, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.9763617677286742, | |
| "grad_norm": 0.42104665274918485, | |
| "learning_rate": 9.614959502286018e-06, | |
| "loss": 0.5258, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.986639260020555, | |
| "grad_norm": 0.4029794401605718, | |
| "learning_rate": 1.033565248820611e-05, | |
| "loss": 0.5263, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.9969167523124358, | |
| "grad_norm": 0.39772123505901325, | |
| "learning_rate": 1.106847475591347e-05, | |
| "loss": 0.5244, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 1.0071942446043165, | |
| "grad_norm": 0.4048130684054564, | |
| "learning_rate": 1.1811520656705356e-05, | |
| "loss": 0.4667, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 1.0174717368961974, | |
| "grad_norm": 0.3967838935343299, | |
| "learning_rate": 1.256285795609691e-05, | |
| "loss": 0.4647, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 1.027749229188078, | |
| "grad_norm": 0.43071529156369454, | |
| "learning_rate": 1.3320532858450379e-05, | |
| "loss": 0.4419, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.0380267214799588, | |
| "grad_norm": 0.39440960560899796, | |
| "learning_rate": 1.408257508767236e-05, | |
| "loss": 0.4559, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.0483042137718397, | |
| "grad_norm": 0.3955707807238948, | |
| "learning_rate": 1.4847003010767309e-05, | |
| "loss": 0.468, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.0585817060637204, | |
| "grad_norm": 0.46986854711677106, | |
| "learning_rate": 1.5611828790923776e-05, | |
| "loss": 0.4536, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.0688591983556013, | |
| "grad_norm": 0.4115140300945968, | |
| "learning_rate": 1.6375063556733252e-05, | |
| "loss": 0.4564, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.079136690647482, | |
| "grad_norm": 0.44106204752327893, | |
| "learning_rate": 1.7134722574099274e-05, | |
| "loss": 0.458, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.0894141829393629, | |
| "grad_norm": 0.47583761779999495, | |
| "learning_rate": 1.7888830407387902e-05, | |
| "loss": 0.4429, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.0996916752312436, | |
| "grad_norm": 0.4144253978666447, | |
| "learning_rate": 1.8635426056398183e-05, | |
| "loss": 0.4724, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.1099691675231242, | |
| "grad_norm": 0.4487588852746433, | |
| "learning_rate": 1.9372568055794383e-05, | |
| "loss": 0.4671, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.1202466598150052, | |
| "grad_norm": 0.4360867503615642, | |
| "learning_rate": 2.0098339523739255e-05, | |
| "loss": 0.4515, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.1305241521068858, | |
| "grad_norm": 0.4534710394135783, | |
| "learning_rate": 2.081085314659985e-05, | |
| "loss": 0.4687, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.1408016443987667, | |
| "grad_norm": 0.407838255680345, | |
| "learning_rate": 2.1508256086763382e-05, | |
| "loss": 0.4498, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.1510791366906474, | |
| "grad_norm": 0.4377752649694931, | |
| "learning_rate": 2.218873480080084e-05, | |
| "loss": 0.4605, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.1613566289825283, | |
| "grad_norm": 0.45327786293937533, | |
| "learning_rate": 2.285051975544917e-05, | |
| "loss": 0.4748, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.171634121274409, | |
| "grad_norm": 0.47562670380204264, | |
| "learning_rate": 2.3491890029148105e-05, | |
| "loss": 0.4829, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.1819116135662897, | |
| "grad_norm": 0.43638700959699767, | |
| "learning_rate": 2.4111177787166208e-05, | |
| "loss": 0.4795, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.1921891058581706, | |
| "grad_norm": 0.5211794580103576, | |
| "learning_rate": 2.4706772618678505e-05, | |
| "loss": 0.4762, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.2024665981500513, | |
| "grad_norm": 0.414897294023872, | |
| "learning_rate": 2.5277125724517662e-05, | |
| "loss": 0.4851, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.2127440904419322, | |
| "grad_norm": 0.4010723068335088, | |
| "learning_rate": 2.582075394470868e-05, | |
| "loss": 0.4896, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.223021582733813, | |
| "grad_norm": 0.42016718369332373, | |
| "learning_rate": 2.6336243615313876e-05, | |
| "loss": 0.473, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.2332990750256938, | |
| "grad_norm": 0.43068136492845654, | |
| "learning_rate": 2.682225424455871e-05, | |
| "loss": 0.4673, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.2435765673175745, | |
| "grad_norm": 0.4177444143376056, | |
| "learning_rate": 2.7277521998678908e-05, | |
| "loss": 0.4753, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.2538540596094552, | |
| "grad_norm": 0.40201721497527027, | |
| "learning_rate": 2.7700862988424254e-05, | |
| "loss": 0.4707, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.264131551901336, | |
| "grad_norm": 0.4491628984476387, | |
| "learning_rate": 2.8091176347672836e-05, | |
| "loss": 0.4896, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.274409044193217, | |
| "grad_norm": 0.48516694721767883, | |
| "learning_rate": 2.8447447096149756e-05, | |
| "loss": 0.4675, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.2846865364850977, | |
| "grad_norm": 0.4709039922274958, | |
| "learning_rate": 2.8768748778806387e-05, | |
| "loss": 0.4941, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.2949640287769784, | |
| "grad_norm": 0.4204135412519858, | |
| "learning_rate": 2.9054245874996426e-05, | |
| "loss": 0.4701, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 1.3052415210688593, | |
| "grad_norm": 0.40339521660066563, | |
| "learning_rate": 2.9303195971183912e-05, | |
| "loss": 0.4869, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 1.31551901336074, | |
| "grad_norm": 0.4613749457606698, | |
| "learning_rate": 2.951495169153333e-05, | |
| "loss": 0.476, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 1.3257965056526206, | |
| "grad_norm": 0.403722598944795, | |
| "learning_rate": 2.9688962381361317e-05, | |
| "loss": 0.4754, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 1.3360739979445015, | |
| "grad_norm": 0.4143629787165823, | |
| "learning_rate": 2.9824775539072402e-05, | |
| "loss": 0.4752, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.3463514902363825, | |
| "grad_norm": 0.3799384001189195, | |
| "learning_rate": 2.992203799285506e-05, | |
| "loss": 0.4872, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 1.3566289825282631, | |
| "grad_norm": 0.40123402762720445, | |
| "learning_rate": 2.9980496819078232e-05, | |
| "loss": 0.471, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 1.3669064748201438, | |
| "grad_norm": 0.4269435577016413, | |
| "learning_rate": 3e-05, | |
| "loss": 0.4809, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 1.3771839671120247, | |
| "grad_norm": 0.4859736842291524, | |
| "learning_rate": 2.9980496819078232e-05, | |
| "loss": 0.4948, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 1.3874614594039054, | |
| "grad_norm": 0.4696636005599923, | |
| "learning_rate": 2.9922037992855063e-05, | |
| "loss": 0.4767, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 1.397738951695786, | |
| "grad_norm": 0.4387419339812974, | |
| "learning_rate": 2.9824775539072402e-05, | |
| "loss": 0.489, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 1.408016443987667, | |
| "grad_norm": 0.3997269476655739, | |
| "learning_rate": 2.9688962381361317e-05, | |
| "loss": 0.476, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 1.418293936279548, | |
| "grad_norm": 0.456762318598325, | |
| "learning_rate": 2.9514951691533335e-05, | |
| "loss": 0.4859, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 1.4285714285714286, | |
| "grad_norm": 0.3988144971120737, | |
| "learning_rate": 2.930319597118392e-05, | |
| "loss": 0.4811, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 1.4388489208633093, | |
| "grad_norm": 0.469809339558885, | |
| "learning_rate": 2.905424587499643e-05, | |
| "loss": 0.4576, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 1.4491264131551902, | |
| "grad_norm": 0.40204250371661987, | |
| "learning_rate": 2.8768748778806387e-05, | |
| "loss": 0.4792, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 1.4594039054470709, | |
| "grad_norm": 0.43278844680785417, | |
| "learning_rate": 2.8447447096149766e-05, | |
| "loss": 0.471, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 1.4696813977389516, | |
| "grad_norm": 0.43570757632960133, | |
| "learning_rate": 2.8091176347672846e-05, | |
| "loss": 0.4503, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 1.4799588900308325, | |
| "grad_norm": 0.3801533344075315, | |
| "learning_rate": 2.7700862988424265e-05, | |
| "loss": 0.4865, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 1.4902363823227134, | |
| "grad_norm": 0.3636715368816546, | |
| "learning_rate": 2.7277521998678904e-05, | |
| "loss": 0.4729, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 1.500513874614594, | |
| "grad_norm": 0.4110765263131838, | |
| "learning_rate": 2.682225424455871e-05, | |
| "loss": 0.4738, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 1.5107913669064748, | |
| "grad_norm": 0.43381733872982453, | |
| "learning_rate": 2.6336243615313872e-05, | |
| "loss": 0.4829, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 1.5210688591983557, | |
| "grad_norm": 0.40182611171324867, | |
| "learning_rate": 2.5820753944708684e-05, | |
| "loss": 0.4559, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 1.5313463514902363, | |
| "grad_norm": 0.40535624187153857, | |
| "learning_rate": 2.527712572451767e-05, | |
| "loss": 0.4643, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 1.541623843782117, | |
| "grad_norm": 0.4115337818793653, | |
| "learning_rate": 2.470677261867851e-05, | |
| "loss": 0.4597, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.551901336073998, | |
| "grad_norm": 0.407199487411736, | |
| "learning_rate": 2.411117778716621e-05, | |
| "loss": 0.4501, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 1.5621788283658788, | |
| "grad_norm": 0.3743288001977607, | |
| "learning_rate": 2.3491890029148122e-05, | |
| "loss": 0.4744, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 1.5724563206577595, | |
| "grad_norm": 0.3699307654421427, | |
| "learning_rate": 2.285051975544919e-05, | |
| "loss": 0.4565, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 1.5827338129496402, | |
| "grad_norm": 0.4028551311302927, | |
| "learning_rate": 2.2188734800800862e-05, | |
| "loss": 0.4556, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 1.5930113052415211, | |
| "grad_norm": 0.4747323206923733, | |
| "learning_rate": 2.150825608676338e-05, | |
| "loss": 0.4694, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 1.6032887975334018, | |
| "grad_norm": 0.38235625826277586, | |
| "learning_rate": 2.0810853146599847e-05, | |
| "loss": 0.4722, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 1.6135662898252825, | |
| "grad_norm": 0.37462427392587805, | |
| "learning_rate": 2.0098339523739252e-05, | |
| "loss": 0.4494, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 1.6238437821171634, | |
| "grad_norm": 0.4138622513625198, | |
| "learning_rate": 1.9372568055794376e-05, | |
| "loss": 0.4344, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 1.6341212744090443, | |
| "grad_norm": 0.3402046797002805, | |
| "learning_rate": 1.863542605639818e-05, | |
| "loss": 0.4407, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 1.644398766700925, | |
| "grad_norm": 0.3414266415166782, | |
| "learning_rate": 1.78888304073879e-05, | |
| "loss": 0.4544, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.6546762589928057, | |
| "grad_norm": 0.390886428434792, | |
| "learning_rate": 1.713472257409928e-05, | |
| "loss": 0.4341, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 1.6649537512846866, | |
| "grad_norm": 0.40001689741621593, | |
| "learning_rate": 1.6375063556733273e-05, | |
| "loss": 0.4521, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 1.6752312435765673, | |
| "grad_norm": 0.35804788765594897, | |
| "learning_rate": 1.5611828790923796e-05, | |
| "loss": 0.4387, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 1.685508735868448, | |
| "grad_norm": 0.36748214820791725, | |
| "learning_rate": 1.4847003010767315e-05, | |
| "loss": 0.4646, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 1.6957862281603289, | |
| "grad_norm": 0.36341313104520045, | |
| "learning_rate": 1.4082575087672367e-05, | |
| "loss": 0.4415, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 1.7060637204522098, | |
| "grad_norm": 0.3576653140862902, | |
| "learning_rate": 1.3320532858450387e-05, | |
| "loss": 0.4399, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 1.7163412127440905, | |
| "grad_norm": 0.36581212655600825, | |
| "learning_rate": 1.2562857956096917e-05, | |
| "loss": 0.4369, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 1.7266187050359711, | |
| "grad_norm": 0.45119158137057136, | |
| "learning_rate": 1.1811520656705362e-05, | |
| "loss": 0.4248, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 1.736896197327852, | |
| "grad_norm": 0.3724511924577398, | |
| "learning_rate": 1.1068474755913466e-05, | |
| "loss": 0.4395, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 1.7471736896197327, | |
| "grad_norm": 0.3751212959461515, | |
| "learning_rate": 1.0335652488206117e-05, | |
| "loss": 0.4201, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.7574511819116134, | |
| "grad_norm": 0.3464962966358675, | |
| "learning_rate": 9.614959502286013e-06, | |
| "loss": 0.4043, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 1.7677286742034943, | |
| "grad_norm": 0.3663266639181022, | |
| "learning_rate": 8.908269905578998e-06, | |
| "loss": 0.4129, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 1.7780061664953752, | |
| "grad_norm": 0.35132549079437864, | |
| "learning_rate": 8.217421390759727e-06, | |
| "loss": 0.4164, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 1.788283658787256, | |
| "grad_norm": 0.3404967583222996, | |
| "learning_rate": 7.544210456971385e-06, | |
| "loss": 0.4265, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 1.7985611510791366, | |
| "grad_norm": 0.469045408430661, | |
| "learning_rate": 6.890387738166045e-06, | |
| "loss": 0.4104, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.8088386433710175, | |
| "grad_norm": 0.3604653210084856, | |
| "learning_rate": 6.257653450713743e-06, | |
| "loss": 0.4256, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 1.8191161356628982, | |
| "grad_norm": 0.34250220546967325, | |
| "learning_rate": 5.647652972119002e-06, | |
| "loss": 0.4136, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 1.829393627954779, | |
| "grad_norm": 0.36630040080056403, | |
| "learning_rate": 5.061972562341319e-06, | |
| "loss": 0.4158, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 1.8396711202466598, | |
| "grad_norm": 0.41269762392169684, | |
| "learning_rate": 4.502135238846574e-06, | |
| "loss": 0.4169, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 1.8499486125385407, | |
| "grad_norm": 0.36458609289471444, | |
| "learning_rate": 3.96959681611572e-06, | |
| "loss": 0.3983, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 1.8602261048304214, | |
| "grad_norm": 0.3384192979617524, | |
| "learning_rate": 3.465742119909566e-06, | |
| "loss": 0.4187, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 1.870503597122302, | |
| "grad_norm": 0.3558048886988858, | |
| "learning_rate": 2.9918813861345975e-06, | |
| "loss": 0.4191, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 1.880781089414183, | |
| "grad_norm": 0.35735445996159837, | |
| "learning_rate": 2.5492468536738013e-06, | |
| "loss": 0.3931, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 1.8910585817060637, | |
| "grad_norm": 0.41549465965742105, | |
| "learning_rate": 2.138989560043004e-06, | |
| "loss": 0.406, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 1.9013360739979444, | |
| "grad_norm": 0.35888882880957634, | |
| "learning_rate": 1.7621763482051878e-06, | |
| "loss": 0.4213, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.9116135662898253, | |
| "grad_norm": 0.3876466527861625, | |
| "learning_rate": 1.419787092326219e-06, | |
| "loss": 0.4205, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 1.9218910585817062, | |
| "grad_norm": 0.3916893545346833, | |
| "learning_rate": 1.1127121496865321e-06, | |
| "loss": 0.3946, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 1.9321685508735869, | |
| "grad_norm": 0.3473782488747357, | |
| "learning_rate": 8.417500453744848e-07, | |
| "loss": 0.4043, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 1.9424460431654675, | |
| "grad_norm": 0.3347296327648612, | |
| "learning_rate": 6.076053957825395e-07, | |
| "loss": 0.4002, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 1.9527235354573484, | |
| "grad_norm": 0.3341213934258483, | |
| "learning_rate": 4.1088707630573096e-07, | |
| "loss": 0.4121, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.9630010277492291, | |
| "grad_norm": 0.3441476454719243, | |
| "learning_rate": 2.5210663800745493e-07, | |
| "loss": 0.4047, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 1.9732785200411098, | |
| "grad_norm": 0.41400797241467135, | |
| "learning_rate": 1.316769773697013e-07, | |
| "loss": 0.3995, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 1.9835560123329907, | |
| "grad_norm": 0.37953330032711113, | |
| "learning_rate": 4.991126258710344e-08, | |
| "loss": 0.4065, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 1.9938335046248716, | |
| "grad_norm": 0.3420466537478671, | |
| "learning_rate": 7.022119196808396e-09, | |
| "loss": 0.4018, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "step": 1946, | |
| "total_flos": 1232698903363584.0, | |
| "train_loss": 0.49459491236726894, | |
| "train_runtime": 32220.0255, | |
| "train_samples_per_second": 1.932, | |
| "train_steps_per_second": 0.06 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 1946, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 2000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1232698903363584.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |