| { | |
| "best_metric": 0.48741263151168823, | |
| "best_model_checkpoint": "/home/ray/default/save/checkpoint-1000", | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 1154, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.004332755632582322, | |
| "grad_norm": 0.10313185304403305, | |
| "learning_rate": 8.333333333333333e-07, | |
| "loss": 1.0706, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.008665511265164644, | |
| "grad_norm": 0.10810094326734543, | |
| "learning_rate": 1.6666666666666667e-06, | |
| "loss": 1.0694, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.012998266897746967, | |
| "grad_norm": 0.09731286019086838, | |
| "learning_rate": 2.5e-06, | |
| "loss": 1.0706, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.01733102253032929, | |
| "grad_norm": 0.11459868401288986, | |
| "learning_rate": 3.3333333333333333e-06, | |
| "loss": 1.0772, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.021663778162911613, | |
| "grad_norm": 0.10845116525888443, | |
| "learning_rate": 4.166666666666667e-06, | |
| "loss": 1.0809, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.025996533795493933, | |
| "grad_norm": 0.12091381102800369, | |
| "learning_rate": 5e-06, | |
| "loss": 1.0737, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.030329289428076257, | |
| "grad_norm": 0.12315661460161209, | |
| "learning_rate": 5.833333333333334e-06, | |
| "loss": 1.0675, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.03466204506065858, | |
| "grad_norm": 0.13054250180721283, | |
| "learning_rate": 6.666666666666667e-06, | |
| "loss": 1.0638, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.0389948006932409, | |
| "grad_norm": 0.13414187729358673, | |
| "learning_rate": 7.5e-06, | |
| "loss": 1.0588, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.043327556325823226, | |
| "grad_norm": 0.12992985546588898, | |
| "learning_rate": 8.333333333333334e-06, | |
| "loss": 1.0512, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.047660311958405546, | |
| "grad_norm": 0.11500700563192368, | |
| "learning_rate": 9.166666666666666e-06, | |
| "loss": 1.0545, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.05199306759098787, | |
| "grad_norm": 0.10958714783191681, | |
| "learning_rate": 1e-05, | |
| "loss": 1.0357, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.05632582322357019, | |
| "grad_norm": 0.10853663086891174, | |
| "learning_rate": 1.0833333333333334e-05, | |
| "loss": 1.0268, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.060658578856152515, | |
| "grad_norm": 0.12498235702514648, | |
| "learning_rate": 1.1666666666666668e-05, | |
| "loss": 1.0135, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.06499133448873484, | |
| "grad_norm": 0.11716682463884354, | |
| "learning_rate": 1.25e-05, | |
| "loss": 1.0014, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.06932409012131716, | |
| "grad_norm": 0.09530466794967651, | |
| "learning_rate": 1.3333333333333333e-05, | |
| "loss": 0.9824, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.07365684575389948, | |
| "grad_norm": 0.09298353642225266, | |
| "learning_rate": 1.4166666666666668e-05, | |
| "loss": 0.9676, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.0779896013864818, | |
| "grad_norm": 0.09433547407388687, | |
| "learning_rate": 1.5e-05, | |
| "loss": 0.9463, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.08232235701906412, | |
| "grad_norm": 0.09255563467741013, | |
| "learning_rate": 1.5833333333333333e-05, | |
| "loss": 0.9341, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.08665511265164645, | |
| "grad_norm": 0.09951213002204895, | |
| "learning_rate": 1.6666666666666667e-05, | |
| "loss": 0.9209, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.09098786828422877, | |
| "grad_norm": 0.10951012372970581, | |
| "learning_rate": 1.75e-05, | |
| "loss": 0.8927, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.09532062391681109, | |
| "grad_norm": 0.110866479575634, | |
| "learning_rate": 1.8333333333333333e-05, | |
| "loss": 0.8755, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.09965337954939342, | |
| "grad_norm": 0.11648208647966385, | |
| "learning_rate": 1.9166666666666667e-05, | |
| "loss": 0.8453, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.10398613518197573, | |
| "grad_norm": 0.12016862630844116, | |
| "learning_rate": 2e-05, | |
| "loss": 0.8081, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.10831889081455806, | |
| "grad_norm": 0.12387488037347794, | |
| "learning_rate": 2.0833333333333336e-05, | |
| "loss": 0.7784, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.11265164644714037, | |
| "grad_norm": 0.12779255211353302, | |
| "learning_rate": 2.1666666666666667e-05, | |
| "loss": 0.7353, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.1169844020797227, | |
| "grad_norm": 0.12649372220039368, | |
| "learning_rate": 2.25e-05, | |
| "loss": 0.7085, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.12131715771230503, | |
| "grad_norm": 0.1445430964231491, | |
| "learning_rate": 2.3333333333333336e-05, | |
| "loss": 0.6753, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.12564991334488734, | |
| "grad_norm": 0.1329505294561386, | |
| "learning_rate": 2.4166666666666667e-05, | |
| "loss": 0.6448, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.12998266897746968, | |
| "grad_norm": 0.13544394075870514, | |
| "learning_rate": 2.5e-05, | |
| "loss": 0.621, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.134315424610052, | |
| "grad_norm": 0.1351090520620346, | |
| "learning_rate": 2.5833333333333336e-05, | |
| "loss": 0.5997, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.1386481802426343, | |
| "grad_norm": 0.127303346991539, | |
| "learning_rate": 2.6666666666666667e-05, | |
| "loss": 0.5791, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.14298093587521662, | |
| "grad_norm": 0.1476767510175705, | |
| "learning_rate": 2.7500000000000004e-05, | |
| "loss": 0.57, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.14731369150779897, | |
| "grad_norm": 0.13856437802314758, | |
| "learning_rate": 2.8333333333333335e-05, | |
| "loss": 0.5645, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.15164644714038128, | |
| "grad_norm": 0.1533508449792862, | |
| "learning_rate": 2.916666666666667e-05, | |
| "loss": 0.5583, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.1559792027729636, | |
| "grad_norm": 0.13325001299381256, | |
| "learning_rate": 3e-05, | |
| "loss": 0.555, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.16031195840554593, | |
| "grad_norm": 0.13416974246501923, | |
| "learning_rate": 3.0833333333333335e-05, | |
| "loss": 0.5439, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.16464471403812825, | |
| "grad_norm": 0.1278882771730423, | |
| "learning_rate": 3.1666666666666666e-05, | |
| "loss": 0.537, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.16897746967071056, | |
| "grad_norm": 0.14047101140022278, | |
| "learning_rate": 3.2500000000000004e-05, | |
| "loss": 0.5381, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.1733102253032929, | |
| "grad_norm": 0.13340455293655396, | |
| "learning_rate": 3.3333333333333335e-05, | |
| "loss": 0.5344, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.17764298093587522, | |
| "grad_norm": 0.13049094378948212, | |
| "learning_rate": 3.4166666666666666e-05, | |
| "loss": 0.5291, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.18197573656845753, | |
| "grad_norm": 0.16296444833278656, | |
| "learning_rate": 3.5e-05, | |
| "loss": 0.5342, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.18630849220103987, | |
| "grad_norm": 0.1682613343000412, | |
| "learning_rate": 3.5833333333333335e-05, | |
| "loss": 0.5285, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.19064124783362218, | |
| "grad_norm": 0.1439386010169983, | |
| "learning_rate": 3.6666666666666666e-05, | |
| "loss": 0.5268, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.1949740034662045, | |
| "grad_norm": 0.15248768031597137, | |
| "learning_rate": 3.7500000000000003e-05, | |
| "loss": 0.5252, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.19930675909878684, | |
| "grad_norm": 0.1604214906692505, | |
| "learning_rate": 3.8333333333333334e-05, | |
| "loss": 0.5201, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.20363951473136915, | |
| "grad_norm": 0.16192543506622314, | |
| "learning_rate": 3.9166666666666665e-05, | |
| "loss": 0.5222, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.20797227036395147, | |
| "grad_norm": 0.14945088326931, | |
| "learning_rate": 4e-05, | |
| "loss": 0.5158, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.2123050259965338, | |
| "grad_norm": 0.16204427182674408, | |
| "learning_rate": 4.0833333333333334e-05, | |
| "loss": 0.5127, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.21663778162911612, | |
| "grad_norm": 0.1618306040763855, | |
| "learning_rate": 4.166666666666667e-05, | |
| "loss": 0.5181, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.22097053726169844, | |
| "grad_norm": 0.14193885028362274, | |
| "learning_rate": 4.25e-05, | |
| "loss": 0.5164, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.22530329289428075, | |
| "grad_norm": 0.12552691996097565, | |
| "learning_rate": 4.3333333333333334e-05, | |
| "loss": 0.5149, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.2296360485268631, | |
| "grad_norm": 0.14471225440502167, | |
| "learning_rate": 4.4166666666666665e-05, | |
| "loss": 0.5137, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.2339688041594454, | |
| "grad_norm": 0.13988590240478516, | |
| "learning_rate": 4.5e-05, | |
| "loss": 0.5066, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.23830155979202772, | |
| "grad_norm": 0.13964875042438507, | |
| "learning_rate": 4.5833333333333334e-05, | |
| "loss": 0.5116, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.24263431542461006, | |
| "grad_norm": 0.12847208976745605, | |
| "learning_rate": 4.666666666666667e-05, | |
| "loss": 0.5095, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.24696707105719237, | |
| "grad_norm": 0.13142219185829163, | |
| "learning_rate": 4.75e-05, | |
| "loss": 0.5077, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.2512998266897747, | |
| "grad_norm": 0.19098567962646484, | |
| "learning_rate": 4.8333333333333334e-05, | |
| "loss": 0.5099, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.255632582322357, | |
| "grad_norm": 0.1430283784866333, | |
| "learning_rate": 4.9166666666666665e-05, | |
| "loss": 0.5036, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.25996533795493937, | |
| "grad_norm": 0.12594453990459442, | |
| "learning_rate": 5e-05, | |
| "loss": 0.5085, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.26429809358752165, | |
| "grad_norm": 0.14377984404563904, | |
| "learning_rate": 4.999577115486055e-05, | |
| "loss": 0.4978, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.268630849220104, | |
| "grad_norm": 0.12468158453702927, | |
| "learning_rate": 4.998308605009268e-05, | |
| "loss": 0.5095, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.2729636048526863, | |
| "grad_norm": 0.12854167819023132, | |
| "learning_rate": 4.996194897716389e-05, | |
| "loss": 0.5037, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.2772963604852686, | |
| "grad_norm": 0.13529527187347412, | |
| "learning_rate": 4.993236708690683e-05, | |
| "loss": 0.5058, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.28162911611785096, | |
| "grad_norm": 0.1377994418144226, | |
| "learning_rate": 4.9894350387100126e-05, | |
| "loss": 0.4998, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.28596187175043325, | |
| "grad_norm": 0.14942322671413422, | |
| "learning_rate": 4.984791173908267e-05, | |
| "loss": 0.5007, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.2902946273830156, | |
| "grad_norm": 0.1366725116968155, | |
| "learning_rate": 4.9793066853402536e-05, | |
| "loss": 0.5038, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.29462738301559793, | |
| "grad_norm": 0.13133087754249573, | |
| "learning_rate": 4.9729834284501995e-05, | |
| "loss": 0.5062, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.2989601386481802, | |
| "grad_norm": 0.11725670844316483, | |
| "learning_rate": 4.965823542444037e-05, | |
| "loss": 0.5025, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.30329289428076256, | |
| "grad_norm": 0.10628046840429306, | |
| "learning_rate": 4.9578294495656965e-05, | |
| "loss": 0.4999, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.3076256499133449, | |
| "grad_norm": 0.13826170563697815, | |
| "learning_rate": 4.949003854277644e-05, | |
| "loss": 0.4978, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.3119584055459272, | |
| "grad_norm": 0.1305851936340332, | |
| "learning_rate": 4.9393497423459376e-05, | |
| "loss": 0.4997, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.31629116117850953, | |
| "grad_norm": 0.11465763300657272, | |
| "learning_rate": 4.928870379830124e-05, | |
| "loss": 0.5037, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.32062391681109187, | |
| "grad_norm": 0.15975706279277802, | |
| "learning_rate": 4.9175693119783013e-05, | |
| "loss": 0.4982, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.32495667244367415, | |
| "grad_norm": 0.15360799431800842, | |
| "learning_rate": 4.905450362027738e-05, | |
| "loss": 0.5013, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.3292894280762565, | |
| "grad_norm": 0.14006198942661285, | |
| "learning_rate": 4.8925176299114416e-05, | |
| "loss": 0.5008, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.33362218370883884, | |
| "grad_norm": 0.14255651831626892, | |
| "learning_rate": 4.878775490871121e-05, | |
| "loss": 0.4975, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.3379549393414211, | |
| "grad_norm": 0.12012791633605957, | |
| "learning_rate": 4.864228593977006e-05, | |
| "loss": 0.5047, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.34228769497400346, | |
| "grad_norm": 0.11433300375938416, | |
| "learning_rate": 4.848881860555035e-05, | |
| "loss": 0.4986, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.3466204506065858, | |
| "grad_norm": 0.11102011054754257, | |
| "learning_rate": 4.832740482521931e-05, | |
| "loss": 0.4981, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.3509532062391681, | |
| "grad_norm": 0.12340573221445084, | |
| "learning_rate": 4.815809920628738e-05, | |
| "loss": 0.4984, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.35528596187175043, | |
| "grad_norm": 0.1148650050163269, | |
| "learning_rate": 4.7980959026134044e-05, | |
| "loss": 0.4942, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.3596187175043328, | |
| "grad_norm": 0.10840712487697601, | |
| "learning_rate": 4.7796044212630486e-05, | |
| "loss": 0.4903, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.36395147313691506, | |
| "grad_norm": 0.11093516647815704, | |
| "learning_rate": 4.7603417323865547e-05, | |
| "loss": 0.4957, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.3682842287694974, | |
| "grad_norm": 0.1237047016620636, | |
| "learning_rate": 4.74031435269818e-05, | |
| "loss": 0.499, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.37261698440207974, | |
| "grad_norm": 0.11613244563341141, | |
| "learning_rate": 4.7195290576129034e-05, | |
| "loss": 0.4959, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.37694974003466203, | |
| "grad_norm": 0.10890854150056839, | |
| "learning_rate": 4.697992878954255e-05, | |
| "loss": 0.4944, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.38128249566724437, | |
| "grad_norm": 0.11364572495222092, | |
| "learning_rate": 4.6757131025753886e-05, | |
| "loss": 0.4909, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.3856152512998267, | |
| "grad_norm": 0.12619757652282715, | |
| "learning_rate": 4.652697265894228e-05, | |
| "loss": 0.4966, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.389948006932409, | |
| "grad_norm": 0.11669816076755524, | |
| "learning_rate": 4.628953155343499e-05, | |
| "loss": 0.4956, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.39428076256499134, | |
| "grad_norm": 0.13808482885360718, | |
| "learning_rate": 4.604488803736523e-05, | |
| "loss": 0.4973, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.3986135181975737, | |
| "grad_norm": 0.11171045899391174, | |
| "learning_rate": 4.579312487549649e-05, | |
| "loss": 0.4903, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.40294627383015597, | |
| "grad_norm": 0.11149395257234573, | |
| "learning_rate": 4.553432724122265e-05, | |
| "loss": 0.4999, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.4072790294627383, | |
| "grad_norm": 0.11308333277702332, | |
| "learning_rate": 4.526858268775313e-05, | |
| "loss": 0.4967, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.41161178509532065, | |
| "grad_norm": 0.10610105097293854, | |
| "learning_rate": 4.499598111849299e-05, | |
| "loss": 0.4936, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.41594454072790293, | |
| "grad_norm": 0.11356962472200394, | |
| "learning_rate": 4.471661475662792e-05, | |
| "loss": 0.493, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.4202772963604853, | |
| "grad_norm": 0.10310888290405273, | |
| "learning_rate": 4.443057811392445e-05, | |
| "loss": 0.5002, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.4246100519930676, | |
| "grad_norm": 0.11358631402254105, | |
| "learning_rate": 4.413796795875586e-05, | |
| "loss": 0.4983, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.4289428076256499, | |
| "grad_norm": 0.12575574219226837, | |
| "learning_rate": 4.383888328336476e-05, | |
| "loss": 0.4949, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.43327556325823224, | |
| "grad_norm": 0.09878399968147278, | |
| "learning_rate": 4.3533425270373216e-05, | |
| "loss": 0.4953, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.43327556325823224, | |
| "eval_loss": 0.4938061535358429, | |
| "eval_runtime": 140.4236, | |
| "eval_samples_per_second": 2.685, | |
| "eval_steps_per_second": 0.677, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.4376083188908146, | |
| "grad_norm": 0.11693672835826874, | |
| "learning_rate": 4.3221697258551906e-05, | |
| "loss": 0.4934, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.44194107452339687, | |
| "grad_norm": 0.11737816035747528, | |
| "learning_rate": 4.2903804707859835e-05, | |
| "loss": 0.4934, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.4462738301559792, | |
| "grad_norm": 0.11542918533086777, | |
| "learning_rate": 4.257985516376644e-05, | |
| "loss": 0.4927, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.4506065857885615, | |
| "grad_norm": 0.10959002375602722, | |
| "learning_rate": 4.224995822086812e-05, | |
| "loss": 0.4935, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.45493934142114384, | |
| "grad_norm": 0.10852424800395966, | |
| "learning_rate": 4.191422548581154e-05, | |
| "loss": 0.4947, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.4592720970537262, | |
| "grad_norm": 0.11614954471588135, | |
| "learning_rate": 4.157277053953631e-05, | |
| "loss": 0.4935, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.46360485268630847, | |
| "grad_norm": 0.11209428310394287, | |
| "learning_rate": 4.1225708898849594e-05, | |
| "loss": 0.4975, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.4679376083188908, | |
| "grad_norm": 0.11714442819356918, | |
| "learning_rate": 4.0873157977346e-05, | |
| "loss": 0.4923, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.47227036395147315, | |
| "grad_norm": 0.1179489716887474, | |
| "learning_rate": 4.051523704568557e-05, | |
| "loss": 0.4939, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.47660311958405543, | |
| "grad_norm": 0.11872310936450958, | |
| "learning_rate": 4.0152067191243696e-05, | |
| "loss": 0.4927, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.4809358752166378, | |
| "grad_norm": 0.1182030588388443, | |
| "learning_rate": 3.978377127714628e-05, | |
| "loss": 0.4908, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.4852686308492201, | |
| "grad_norm": 0.11203285306692123, | |
| "learning_rate": 3.941047390070419e-05, | |
| "loss": 0.4898, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.4896013864818024, | |
| "grad_norm": 0.10341402143239975, | |
| "learning_rate": 3.903230135126113e-05, | |
| "loss": 0.4899, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.49393414211438474, | |
| "grad_norm": 0.11410869657993317, | |
| "learning_rate": 3.864938156746891e-05, | |
| "loss": 0.4914, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.4982668977469671, | |
| "grad_norm": 0.11390075832605362, | |
| "learning_rate": 3.8261844094004815e-05, | |
| "loss": 0.4922, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.5025996533795494, | |
| "grad_norm": 0.10876569896936417, | |
| "learning_rate": 3.7869820037745776e-05, | |
| "loss": 0.4964, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.5069324090121318, | |
| "grad_norm": 0.11602895706892014, | |
| "learning_rate": 3.747344202341373e-05, | |
| "loss": 0.4945, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.511265164644714, | |
| "grad_norm": 0.12035666406154633, | |
| "learning_rate": 3.707284414870786e-05, | |
| "loss": 0.499, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.5155979202772963, | |
| "grad_norm": 0.1848023533821106, | |
| "learning_rate": 3.666816193893817e-05, | |
| "loss": 0.4947, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.5199306759098787, | |
| "grad_norm": 0.12017575651407242, | |
| "learning_rate": 3.6259532301176335e-05, | |
| "loss": 0.4905, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.524263431542461, | |
| "grad_norm": 0.10070477426052094, | |
| "learning_rate": 3.5847093477938956e-05, | |
| "loss": 0.4974, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.5285961871750433, | |
| "grad_norm": 0.11581992357969284, | |
| "learning_rate": 3.543098500041906e-05, | |
| "loss": 0.4927, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.5329289428076257, | |
| "grad_norm": 0.09777580201625824, | |
| "learning_rate": 3.501134764128167e-05, | |
| "loss": 0.4907, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.537261698440208, | |
| "grad_norm": 0.10121456533670425, | |
| "learning_rate": 3.458832336703929e-05, | |
| "loss": 0.491, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.5415944540727903, | |
| "grad_norm": 0.10312589257955551, | |
| "learning_rate": 3.416205529002363e-05, | |
| "loss": 0.4942, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.5459272097053726, | |
| "grad_norm": 0.10299069434404373, | |
| "learning_rate": 3.37326876199695e-05, | |
| "loss": 0.493, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.550259965337955, | |
| "grad_norm": 0.11771389842033386, | |
| "learning_rate": 3.3300365615227685e-05, | |
| "loss": 0.4942, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.5545927209705372, | |
| "grad_norm": 0.1121087372303009, | |
| "learning_rate": 3.286523553362287e-05, | |
| "loss": 0.4946, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.5589254766031195, | |
| "grad_norm": 0.10604743659496307, | |
| "learning_rate": 3.242744458297348e-05, | |
| "loss": 0.4907, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.5632582322357019, | |
| "grad_norm": 0.10975582152605057, | |
| "learning_rate": 3.1987140871290236e-05, | |
| "loss": 0.4889, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.5675909878682842, | |
| "grad_norm": 0.10636895149946213, | |
| "learning_rate": 3.154447335667001e-05, | |
| "loss": 0.4935, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.5719237435008665, | |
| "grad_norm": 0.10076680779457092, | |
| "learning_rate": 3.1099591796902215e-05, | |
| "loss": 0.4925, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.5762564991334489, | |
| "grad_norm": 0.10885344445705414, | |
| "learning_rate": 3.0652646698804585e-05, | |
| "loss": 0.4892, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.5805892547660312, | |
| "grad_norm": 0.11372784525156021, | |
| "learning_rate": 3.0203789267305567e-05, | |
| "loss": 0.4918, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.5849220103986135, | |
| "grad_norm": 0.11742879450321198, | |
| "learning_rate": 2.975317135429056e-05, | |
| "loss": 0.4885, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.5892547660311959, | |
| "grad_norm": 0.09908663481473923, | |
| "learning_rate": 2.930094540722927e-05, | |
| "loss": 0.4889, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.5935875216637782, | |
| "grad_norm": 0.11032566428184509, | |
| "learning_rate": 2.884726441760155e-05, | |
| "loss": 0.4936, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.5979202772963604, | |
| "grad_norm": 0.11033419519662857, | |
| "learning_rate": 2.8392281869139213e-05, | |
| "loss": 0.4925, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.6022530329289428, | |
| "grad_norm": 0.12115441262722015, | |
| "learning_rate": 2.7936151685901373e-05, | |
| "loss": 0.49, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.6065857885615251, | |
| "grad_norm": 0.13125091791152954, | |
| "learning_rate": 2.747902818020067e-05, | |
| "loss": 0.4919, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.6109185441941074, | |
| "grad_norm": 0.11565785109996796, | |
| "learning_rate": 2.7021066000398344e-05, | |
| "loss": 0.4928, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.6152512998266898, | |
| "grad_norm": 0.12640348076820374, | |
| "learning_rate": 2.6562420078585433e-05, | |
| "loss": 0.489, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.6195840554592721, | |
| "grad_norm": 0.12297876179218292, | |
| "learning_rate": 2.6103245578168106e-05, | |
| "loss": 0.4919, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.6239168110918544, | |
| "grad_norm": 0.10511300712823868, | |
| "learning_rate": 2.564369784137472e-05, | |
| "loss": 0.4883, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.6282495667244368, | |
| "grad_norm": 0.10803347080945969, | |
| "learning_rate": 2.5183932336702297e-05, | |
| "loss": 0.4909, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.6325823223570191, | |
| "grad_norm": 0.11415662616491318, | |
| "learning_rate": 2.4724104606320445e-05, | |
| "loss": 0.4878, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.6369150779896013, | |
| "grad_norm": 0.10404873639345169, | |
| "learning_rate": 2.426437021345015e-05, | |
| "loss": 0.4901, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.6412478336221837, | |
| "grad_norm": 0.11284226924180984, | |
| "learning_rate": 2.3804884689735642e-05, | |
| "loss": 0.4906, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.645580589254766, | |
| "grad_norm": 0.1138242855668068, | |
| "learning_rate": 2.3345803482626797e-05, | |
| "loss": 0.494, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.6499133448873483, | |
| "grad_norm": 0.18223117291927338, | |
| "learning_rate": 2.288728190279014e-05, | |
| "loss": 0.4886, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.6542461005199307, | |
| "grad_norm": 0.1119702160358429, | |
| "learning_rate": 2.2429475071565987e-05, | |
| "loss": 0.4908, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.658578856152513, | |
| "grad_norm": 0.11668991297483444, | |
| "learning_rate": 2.1972537868489797e-05, | |
| "loss": 0.4896, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.6629116117850953, | |
| "grad_norm": 0.10587523132562637, | |
| "learning_rate": 2.151662487889518e-05, | |
| "loss": 0.4885, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.6672443674176777, | |
| "grad_norm": 0.10824766755104065, | |
| "learning_rate": 2.1061890341616558e-05, | |
| "loss": 0.4873, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.67157712305026, | |
| "grad_norm": 0.11111487448215485, | |
| "learning_rate": 2.060848809680893e-05, | |
| "loss": 0.4853, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.6759098786828422, | |
| "grad_norm": 0.10754924267530441, | |
| "learning_rate": 2.0156571533902627e-05, | |
| "loss": 0.4849, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.6802426343154246, | |
| "grad_norm": 0.10407605022192001, | |
| "learning_rate": 1.97062935397105e-05, | |
| "loss": 0.488, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.6845753899480069, | |
| "grad_norm": 0.1077013909816742, | |
| "learning_rate": 1.9257806446705116e-05, | |
| "loss": 0.4867, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.6889081455805892, | |
| "grad_norm": 0.11459195613861084, | |
| "learning_rate": 1.8811261981483548e-05, | |
| "loss": 0.4911, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.6932409012131716, | |
| "grad_norm": 0.12345458567142487, | |
| "learning_rate": 1.8366811213437092e-05, | |
| "loss": 0.4888, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.6975736568457539, | |
| "grad_norm": 0.10806908458471298, | |
| "learning_rate": 1.7924604503643367e-05, | |
| "loss": 0.4903, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.7019064124783362, | |
| "grad_norm": 0.12911449372768402, | |
| "learning_rate": 1.7484791453998006e-05, | |
| "loss": 0.4874, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.7062391681109186, | |
| "grad_norm": 0.11117111146450043, | |
| "learning_rate": 1.7047520856603183e-05, | |
| "loss": 0.4895, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.7105719237435009, | |
| "grad_norm": 0.10993503779172897, | |
| "learning_rate": 1.6612940643430138e-05, | |
| "loss": 0.4909, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.7149046793760832, | |
| "grad_norm": 0.10971593856811523, | |
| "learning_rate": 1.618119783627263e-05, | |
| "loss": 0.4898, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.7192374350086655, | |
| "grad_norm": 0.12124701589345932, | |
| "learning_rate": 1.5752438497008405e-05, | |
| "loss": 0.4886, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.7235701906412478, | |
| "grad_norm": 0.10207706689834595, | |
| "learning_rate": 1.5326807678185267e-05, | |
| "loss": 0.4893, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.7279029462738301, | |
| "grad_norm": 0.10885774344205856, | |
| "learning_rate": 1.490444937394879e-05, | |
| "loss": 0.4891, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.7322357019064125, | |
| "grad_norm": 0.10901923477649689, | |
| "learning_rate": 1.4485506471327914e-05, | |
| "loss": 0.4968, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.7365684575389948, | |
| "grad_norm": 0.11790922284126282, | |
| "learning_rate": 1.407012070189524e-05, | |
| "loss": 0.4861, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.7409012131715771, | |
| "grad_norm": 0.11575620621442795, | |
| "learning_rate": 1.3658432593818149e-05, | |
| "loss": 0.4888, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.7452339688041595, | |
| "grad_norm": 0.11284902691841125, | |
| "learning_rate": 1.325058142431701e-05, | |
| "loss": 0.4879, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.7495667244367418, | |
| "grad_norm": 0.11264630407094955, | |
| "learning_rate": 1.2846705172546675e-05, | |
| "loss": 0.4909, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.7538994800693241, | |
| "grad_norm": 0.11645620316267014, | |
| "learning_rate": 1.2446940472917099e-05, | |
| "loss": 0.485, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.7582322357019065, | |
| "grad_norm": 0.10355979204177856, | |
| "learning_rate": 1.2051422568868833e-05, | |
| "loss": 0.4891, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.7625649913344887, | |
| "grad_norm": 0.10600101947784424, | |
| "learning_rate": 1.1660285267119167e-05, | |
| "loss": 0.4903, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.766897746967071, | |
| "grad_norm": 0.11851372569799423, | |
| "learning_rate": 1.1273660892394297e-05, | |
| "loss": 0.4895, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.7712305025996534, | |
| "grad_norm": 0.11533911526203156, | |
| "learning_rate": 1.0891680242662835e-05, | |
| "loss": 0.4843, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.7755632582322357, | |
| "grad_norm": 0.09746929258108139, | |
| "learning_rate": 1.051447254488591e-05, | |
| "loss": 0.4894, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.779896013864818, | |
| "grad_norm": 0.1193188726902008, | |
| "learning_rate": 1.0142165411298662e-05, | |
| "loss": 0.4884, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.7842287694974004, | |
| "grad_norm": 0.11370085179805756, | |
| "learning_rate": 9.774884796238085e-06, | |
| "loss": 0.49, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.7885615251299827, | |
| "grad_norm": 0.1011599749326706, | |
| "learning_rate": 9.412754953531663e-06, | |
| "loss": 0.4884, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.792894280762565, | |
| "grad_norm": 0.11001604050397873, | |
| "learning_rate": 9.055898394461423e-06, | |
| "loss": 0.4883, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.7972270363951474, | |
| "grad_norm": 0.11630392074584961, | |
| "learning_rate": 8.704435846317386e-06, | |
| "loss": 0.4869, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.8015597920277296, | |
| "grad_norm": 0.10578346997499466, | |
| "learning_rate": 8.358486211554637e-06, | |
| "loss": 0.4887, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.8058925476603119, | |
| "grad_norm": 0.11845128238201141, | |
| "learning_rate": 8.018166527567672e-06, | |
| "loss": 0.4852, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.8102253032928943, | |
| "grad_norm": 0.10032304376363754, | |
| "learning_rate": 7.683591927095824e-06, | |
| "loss": 0.4865, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.8145580589254766, | |
| "grad_norm": 0.10029744356870651, | |
| "learning_rate": 7.354875599272928e-06, | |
| "loss": 0.4902, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.8188908145580589, | |
| "grad_norm": 0.10729606449604034, | |
| "learning_rate": 7.0321287513346074e-06, | |
| "loss": 0.4879, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.8232235701906413, | |
| "grad_norm": 0.11077585816383362, | |
| "learning_rate": 6.715460570995988e-06, | |
| "loss": 0.4905, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.8275563258232236, | |
| "grad_norm": 0.112645223736763, | |
| "learning_rate": 6.404978189512681e-06, | |
| "loss": 0.492, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.8318890814558059, | |
| "grad_norm": 0.11094575375318527, | |
| "learning_rate": 6.100786645437481e-06, | |
| "loss": 0.4855, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.8362218370883883, | |
| "grad_norm": 0.10790159553289413, | |
| "learning_rate": 5.8029888490850005e-06, | |
| "loss": 0.4942, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.8405545927209706, | |
| "grad_norm": 0.11253953725099564, | |
| "learning_rate": 5.511685547716328e-06, | |
| "loss": 0.4829, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.8448873483535528, | |
| "grad_norm": 0.11395300179719925, | |
| "learning_rate": 5.226975291455477e-06, | |
| "loss": 0.4856, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.8492201039861352, | |
| "grad_norm": 0.1128508597612381, | |
| "learning_rate": 4.9489543999491045e-06, | |
| "loss": 0.4852, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.8535528596187175, | |
| "grad_norm": 0.11124628782272339, | |
| "learning_rate": 4.67771692978087e-06, | |
| "loss": 0.4885, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.8578856152512998, | |
| "grad_norm": 0.10993985086679459, | |
| "learning_rate": 4.413354642651369e-06, | |
| "loss": 0.4867, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.8622183708838822, | |
| "grad_norm": 0.1036507710814476, | |
| "learning_rate": 4.1559569743344405e-06, | |
| "loss": 0.4839, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.8665511265164645, | |
| "grad_norm": 0.10577788203954697, | |
| "learning_rate": 3.90561100442036e-06, | |
| "loss": 0.4909, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.8665511265164645, | |
| "eval_loss": 0.48741263151168823, | |
| "eval_runtime": 139.1562, | |
| "eval_samples_per_second": 2.709, | |
| "eval_steps_per_second": 0.683, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.8708838821490468, | |
| "grad_norm": 0.10687436163425446, | |
| "learning_rate": 3.662401426856177e-06, | |
| "loss": 0.4912, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.8752166377816292, | |
| "grad_norm": 0.10468524694442749, | |
| "learning_rate": 3.4264105212930915e-06, | |
| "loss": 0.4882, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.8795493934142115, | |
| "grad_norm": 0.11409106850624084, | |
| "learning_rate": 3.197718125250618e-06, | |
| "loss": 0.4834, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.8838821490467937, | |
| "grad_norm": 0.1069907397031784, | |
| "learning_rate": 2.9764016071069434e-06, | |
| "loss": 0.4881, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.8882149046793761, | |
| "grad_norm": 0.10976656526327133, | |
| "learning_rate": 2.7625358399246376e-06, | |
| "loss": 0.4852, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.8925476603119584, | |
| "grad_norm": 0.11071466654539108, | |
| "learning_rate": 2.5561931761205082e-06, | |
| "loss": 0.4879, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.8968804159445407, | |
| "grad_norm": 0.10794007033109665, | |
| "learning_rate": 2.3574434229882145e-06, | |
| "loss": 0.4865, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.901213171577123, | |
| "grad_norm": 0.11333485692739487, | |
| "learning_rate": 2.166353819081968e-06, | |
| "loss": 0.4855, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.9055459272097054, | |
| "grad_norm": 0.10130605846643448, | |
| "learning_rate": 1.982989011469172e-06, | |
| "loss": 0.4868, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.9098786828422877, | |
| "grad_norm": 0.10454142093658447, | |
| "learning_rate": 1.8074110338598682e-06, | |
| "loss": 0.487, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.91421143847487, | |
| "grad_norm": 0.10157745331525803, | |
| "learning_rate": 1.6396792856202298e-06, | |
| "loss": 0.4881, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.9185441941074524, | |
| "grad_norm": 0.11562693864107132, | |
| "learning_rate": 1.479850511677322e-06, | |
| "loss": 0.4877, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.9228769497400346, | |
| "grad_norm": 0.12777337431907654, | |
| "learning_rate": 1.3279787833218537e-06, | |
| "loss": 0.4886, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.9272097053726169, | |
| "grad_norm": 0.10074026137590408, | |
| "learning_rate": 1.1841154799154374e-06, | |
| "loss": 0.4856, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.9315424610051993, | |
| "grad_norm": 0.11187005788087845, | |
| "learning_rate": 1.0483092715085879e-06, | |
| "loss": 0.4872, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.9358752166377816, | |
| "grad_norm": 0.10271576046943665, | |
| "learning_rate": 9.206061023752516e-07, | |
| "loss": 0.4867, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.9402079722703639, | |
| "grad_norm": 0.11686773598194122, | |
| "learning_rate": 8.010491754695177e-07, | |
| "loss": 0.4868, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 0.9445407279029463, | |
| "grad_norm": 0.12068697810173035, | |
| "learning_rate": 6.896789378097179e-07, | |
| "loss": 0.4911, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.9488734835355286, | |
| "grad_norm": 0.10469616949558258, | |
| "learning_rate": 5.865330667949115e-07, | |
| "loss": 0.485, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.9532062391681109, | |
| "grad_norm": 0.11302938312292099, | |
| "learning_rate": 4.916464574583251e-07, | |
| "loss": 0.4864, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.9575389948006933, | |
| "grad_norm": 0.11449886858463287, | |
| "learning_rate": 4.050512106620913e-07, | |
| "loss": 0.4873, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 0.9618717504332756, | |
| "grad_norm": 0.1062023788690567, | |
| "learning_rate": 3.267766222372931e-07, | |
| "loss": 0.4885, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.9662045060658578, | |
| "grad_norm": 0.11473377794027328, | |
| "learning_rate": 2.568491730729539e-07, | |
| "loss": 0.4869, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 0.9705372616984402, | |
| "grad_norm": 0.10777969658374786, | |
| "learning_rate": 1.95292520157353e-07, | |
| "loss": 0.489, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.9748700173310225, | |
| "grad_norm": 0.10438723862171173, | |
| "learning_rate": 1.4212748857468926e-07, | |
| "loss": 0.4871, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.9792027729636048, | |
| "grad_norm": 0.10945618152618408, | |
| "learning_rate": 9.737206445979696e-08, | |
| "loss": 0.4875, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.9835355285961872, | |
| "grad_norm": 0.10462047904729843, | |
| "learning_rate": 6.104138891329659e-08, | |
| "loss": 0.4895, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 0.9878682842287695, | |
| "grad_norm": 0.1181010752916336, | |
| "learning_rate": 3.314775287923677e-08, | |
| "loss": 0.4857, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.9922010398613518, | |
| "grad_norm": 0.10257267206907272, | |
| "learning_rate": 1.3700592986998217e-08, | |
| "loss": 0.4866, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 0.9965337954939342, | |
| "grad_norm": 0.10488082468509674, | |
| "learning_rate": 2.7064883587807345e-09, | |
| "loss": 0.4893, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 1154, | |
| "total_flos": 8.066906520991826e+17, | |
| "train_loss": 0.5547944433354423, | |
| "train_runtime": 22221.3478, | |
| "train_samples_per_second": 0.831, | |
| "train_steps_per_second": 0.052 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 1154, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 8.066906520991826e+17, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |