| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 8.0, | |
| "eval_steps": 500, | |
| "global_step": 3504, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0228310502283105, | |
| "grad_norm": 2.228510856628418, | |
| "learning_rate": 4.987157534246575e-05, | |
| "loss": 6.6129, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.045662100456621, | |
| "grad_norm": 1.832844614982605, | |
| "learning_rate": 4.9728881278538815e-05, | |
| "loss": 3.1197, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.0684931506849315, | |
| "grad_norm": 1.3700411319732666, | |
| "learning_rate": 4.958618721461187e-05, | |
| "loss": 2.1387, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.091324200913242, | |
| "grad_norm": 1.3453809022903442, | |
| "learning_rate": 4.9443493150684935e-05, | |
| "loss": 1.4287, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.1141552511415525, | |
| "grad_norm": 1.195481777191162, | |
| "learning_rate": 4.930079908675799e-05, | |
| "loss": 0.9585, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.136986301369863, | |
| "grad_norm": 1.0389988422393799, | |
| "learning_rate": 4.9158105022831055e-05, | |
| "loss": 0.5942, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.1598173515981735, | |
| "grad_norm": 0.9327354431152344, | |
| "learning_rate": 4.901541095890411e-05, | |
| "loss": 0.4014, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.182648401826484, | |
| "grad_norm": 0.791429340839386, | |
| "learning_rate": 4.8872716894977175e-05, | |
| "loss": 0.2836, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.2054794520547945, | |
| "grad_norm": 0.7389950156211853, | |
| "learning_rate": 4.873002283105023e-05, | |
| "loss": 0.1995, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.228310502283105, | |
| "grad_norm": 0.6029524207115173, | |
| "learning_rate": 4.8587328767123295e-05, | |
| "loss": 0.1431, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.2511415525114155, | |
| "grad_norm": 0.5200534462928772, | |
| "learning_rate": 4.844463470319635e-05, | |
| "loss": 0.1019, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.273972602739726, | |
| "grad_norm": 0.6404463648796082, | |
| "learning_rate": 4.830194063926941e-05, | |
| "loss": 0.0846, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.2968036529680365, | |
| "grad_norm": 0.4442310631275177, | |
| "learning_rate": 4.815924657534247e-05, | |
| "loss": 0.0654, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.319634703196347, | |
| "grad_norm": 0.4031509757041931, | |
| "learning_rate": 4.801655251141553e-05, | |
| "loss": 0.0542, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.3424657534246575, | |
| "grad_norm": 0.4141863286495209, | |
| "learning_rate": 4.7873858447488584e-05, | |
| "loss": 0.0508, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.365296803652968, | |
| "grad_norm": 0.3440322279930115, | |
| "learning_rate": 4.773116438356164e-05, | |
| "loss": 0.0391, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.3881278538812785, | |
| "grad_norm": 0.6025522947311401, | |
| "learning_rate": 4.7588470319634704e-05, | |
| "loss": 0.0381, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.410958904109589, | |
| "grad_norm": 0.34279781579971313, | |
| "learning_rate": 4.744577625570776e-05, | |
| "loss": 0.0339, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.4337899543378995, | |
| "grad_norm": 0.3098609149456024, | |
| "learning_rate": 4.7303082191780824e-05, | |
| "loss": 0.0301, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.45662100456621, | |
| "grad_norm": 0.29201066493988037, | |
| "learning_rate": 4.716038812785388e-05, | |
| "loss": 0.0287, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.4794520547945205, | |
| "grad_norm": 0.2325911968946457, | |
| "learning_rate": 4.7017694063926944e-05, | |
| "loss": 0.0258, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.502283105022831, | |
| "grad_norm": 0.2296101152896881, | |
| "learning_rate": 4.6875e-05, | |
| "loss": 0.0227, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.5251141552511416, | |
| "grad_norm": 0.24569889903068542, | |
| "learning_rate": 4.6732305936073064e-05, | |
| "loss": 0.0232, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.547945205479452, | |
| "grad_norm": 0.22321540117263794, | |
| "learning_rate": 4.658961187214612e-05, | |
| "loss": 0.018, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.5707762557077626, | |
| "grad_norm": 0.26079800724983215, | |
| "learning_rate": 4.6446917808219184e-05, | |
| "loss": 0.0192, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.593607305936073, | |
| "grad_norm": 0.29277801513671875, | |
| "learning_rate": 4.630422374429224e-05, | |
| "loss": 0.02, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.6164383561643836, | |
| "grad_norm": 0.2620185315608978, | |
| "learning_rate": 4.61615296803653e-05, | |
| "loss": 0.0172, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.639269406392694, | |
| "grad_norm": 0.32734254002571106, | |
| "learning_rate": 4.601883561643836e-05, | |
| "loss": 0.0191, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.6621004566210046, | |
| "grad_norm": 0.34528958797454834, | |
| "learning_rate": 4.587614155251142e-05, | |
| "loss": 0.0151, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.684931506849315, | |
| "grad_norm": 0.16524860262870789, | |
| "learning_rate": 4.5733447488584474e-05, | |
| "loss": 0.0154, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.7077625570776256, | |
| "grad_norm": 0.17619894444942474, | |
| "learning_rate": 4.559075342465753e-05, | |
| "loss": 0.0141, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.730593607305936, | |
| "grad_norm": 0.18591953814029694, | |
| "learning_rate": 4.5448059360730594e-05, | |
| "loss": 0.0128, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.7534246575342466, | |
| "grad_norm": 0.17972640693187714, | |
| "learning_rate": 4.530536529680365e-05, | |
| "loss": 0.0119, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.776255707762557, | |
| "grad_norm": 0.2308385968208313, | |
| "learning_rate": 4.5162671232876714e-05, | |
| "loss": 0.0114, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.7990867579908676, | |
| "grad_norm": 0.13175222277641296, | |
| "learning_rate": 4.501997716894977e-05, | |
| "loss": 0.0107, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.821917808219178, | |
| "grad_norm": 0.1399480700492859, | |
| "learning_rate": 4.4877283105022834e-05, | |
| "loss": 0.0097, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.8447488584474886, | |
| "grad_norm": 0.12605524063110352, | |
| "learning_rate": 4.473458904109589e-05, | |
| "loss": 0.0112, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.867579908675799, | |
| "grad_norm": 0.16242380440235138, | |
| "learning_rate": 4.4591894977168954e-05, | |
| "loss": 0.0099, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.8904109589041096, | |
| "grad_norm": 0.12138999998569489, | |
| "learning_rate": 4.444920091324201e-05, | |
| "loss": 0.0089, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.91324200913242, | |
| "grad_norm": 0.22154438495635986, | |
| "learning_rate": 4.4306506849315074e-05, | |
| "loss": 0.009, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.9360730593607306, | |
| "grad_norm": 0.09284752607345581, | |
| "learning_rate": 4.416381278538813e-05, | |
| "loss": 0.0079, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.958904109589041, | |
| "grad_norm": 0.2589144706726074, | |
| "learning_rate": 4.4021118721461194e-05, | |
| "loss": 0.0077, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.9817351598173516, | |
| "grad_norm": 0.18040631711483002, | |
| "learning_rate": 4.387842465753425e-05, | |
| "loss": 0.0073, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 0.0013460684567689896, | |
| "eval_runtime": 71.4572, | |
| "eval_samples_per_second": 41.983, | |
| "eval_steps_per_second": 2.631, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 1.004566210045662, | |
| "grad_norm": 0.16953226923942566, | |
| "learning_rate": 4.373573059360731e-05, | |
| "loss": 0.0071, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.0273972602739727, | |
| "grad_norm": 0.13055773079395294, | |
| "learning_rate": 4.359303652968037e-05, | |
| "loss": 0.0076, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.0502283105022832, | |
| "grad_norm": 0.11131970584392548, | |
| "learning_rate": 4.345034246575343e-05, | |
| "loss": 0.0072, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.0730593607305936, | |
| "grad_norm": 0.1048307716846466, | |
| "learning_rate": 4.3307648401826484e-05, | |
| "loss": 0.0053, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.095890410958904, | |
| "grad_norm": 0.191425159573555, | |
| "learning_rate": 4.316495433789954e-05, | |
| "loss": 0.0061, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.1187214611872145, | |
| "grad_norm": 0.1289547234773636, | |
| "learning_rate": 4.3022260273972604e-05, | |
| "loss": 0.0057, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.1415525114155252, | |
| "grad_norm": 0.09019248932600021, | |
| "learning_rate": 4.287956621004566e-05, | |
| "loss": 0.0049, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.1643835616438356, | |
| "grad_norm": 0.22927595674991608, | |
| "learning_rate": 4.2736872146118724e-05, | |
| "loss": 0.0051, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.187214611872146, | |
| "grad_norm": 0.1198509931564331, | |
| "learning_rate": 4.259417808219178e-05, | |
| "loss": 0.0047, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.2100456621004567, | |
| "grad_norm": 0.1271675080060959, | |
| "learning_rate": 4.2451484018264844e-05, | |
| "loss": 0.0051, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.2328767123287672, | |
| "grad_norm": 0.13385379314422607, | |
| "learning_rate": 4.23087899543379e-05, | |
| "loss": 0.0054, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.2557077625570776, | |
| "grad_norm": 0.06279306858778, | |
| "learning_rate": 4.2166095890410964e-05, | |
| "loss": 0.0046, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.278538812785388, | |
| "grad_norm": 0.12206210196018219, | |
| "learning_rate": 4.202340182648402e-05, | |
| "loss": 0.0046, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.3013698630136985, | |
| "grad_norm": 0.08528764545917511, | |
| "learning_rate": 4.1880707762557084e-05, | |
| "loss": 0.0049, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.3242009132420092, | |
| "grad_norm": 0.11532297730445862, | |
| "learning_rate": 4.173801369863014e-05, | |
| "loss": 0.0044, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.3470319634703196, | |
| "grad_norm": 0.1577356606721878, | |
| "learning_rate": 4.1595319634703204e-05, | |
| "loss": 0.0047, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.36986301369863, | |
| "grad_norm": 0.0654304102063179, | |
| "learning_rate": 4.145262557077626e-05, | |
| "loss": 0.0043, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.3926940639269407, | |
| "grad_norm": 0.13026106357574463, | |
| "learning_rate": 4.130993150684932e-05, | |
| "loss": 0.0046, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.4155251141552512, | |
| "grad_norm": 0.1981119066476822, | |
| "learning_rate": 4.1167237442922374e-05, | |
| "loss": 0.0036, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.4383561643835616, | |
| "grad_norm": 0.0626244992017746, | |
| "learning_rate": 4.102454337899544e-05, | |
| "loss": 0.0029, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.461187214611872, | |
| "grad_norm": 0.09634381532669067, | |
| "learning_rate": 4.0881849315068494e-05, | |
| "loss": 0.0035, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.4840182648401825, | |
| "grad_norm": 0.10267172008752823, | |
| "learning_rate": 4.073915525114155e-05, | |
| "loss": 0.0029, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.5068493150684932, | |
| "grad_norm": 0.1381852775812149, | |
| "learning_rate": 4.0596461187214614e-05, | |
| "loss": 0.0034, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.5296803652968036, | |
| "grad_norm": 0.05667397379875183, | |
| "learning_rate": 4.045376712328767e-05, | |
| "loss": 0.0034, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.5525114155251143, | |
| "grad_norm": 0.09371186792850494, | |
| "learning_rate": 4.0311073059360734e-05, | |
| "loss": 0.0035, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.5753424657534247, | |
| "grad_norm": 0.09487811475992203, | |
| "learning_rate": 4.016837899543379e-05, | |
| "loss": 0.0032, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.5981735159817352, | |
| "grad_norm": 0.09329680353403091, | |
| "learning_rate": 4.0025684931506853e-05, | |
| "loss": 0.0034, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.6210045662100456, | |
| "grad_norm": 0.10534738004207611, | |
| "learning_rate": 3.988299086757991e-05, | |
| "loss": 0.003, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.643835616438356, | |
| "grad_norm": 0.1082799881696701, | |
| "learning_rate": 3.9740296803652973e-05, | |
| "loss": 0.0032, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.6666666666666665, | |
| "grad_norm": 0.06728220731019974, | |
| "learning_rate": 3.959760273972603e-05, | |
| "loss": 0.0031, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.6894977168949772, | |
| "grad_norm": 0.22814679145812988, | |
| "learning_rate": 3.9454908675799093e-05, | |
| "loss": 0.003, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.7123287671232876, | |
| "grad_norm": 0.10962113738059998, | |
| "learning_rate": 3.931221461187215e-05, | |
| "loss": 0.003, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.7351598173515983, | |
| "grad_norm": 0.054916515946388245, | |
| "learning_rate": 3.916952054794521e-05, | |
| "loss": 0.0024, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.7579908675799087, | |
| "grad_norm": 0.08513092249631882, | |
| "learning_rate": 3.902682648401826e-05, | |
| "loss": 0.0027, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.7808219178082192, | |
| "grad_norm": 0.03006896749138832, | |
| "learning_rate": 3.888413242009133e-05, | |
| "loss": 0.0026, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.8036529680365296, | |
| "grad_norm": 0.0818442776799202, | |
| "learning_rate": 3.874143835616438e-05, | |
| "loss": 0.0025, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.82648401826484, | |
| "grad_norm": 0.06740756332874298, | |
| "learning_rate": 3.859874429223744e-05, | |
| "loss": 0.0019, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.8493150684931505, | |
| "grad_norm": 0.022570671513676643, | |
| "learning_rate": 3.84560502283105e-05, | |
| "loss": 0.0026, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.8721461187214612, | |
| "grad_norm": 0.06723079085350037, | |
| "learning_rate": 3.831335616438356e-05, | |
| "loss": 0.0021, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.8949771689497716, | |
| "grad_norm": 0.12359272688627243, | |
| "learning_rate": 3.817066210045662e-05, | |
| "loss": 0.0023, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.9178082191780823, | |
| "grad_norm": 0.07307197153568268, | |
| "learning_rate": 3.802796803652968e-05, | |
| "loss": 0.0024, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.9406392694063928, | |
| "grad_norm": 0.11846048384904861, | |
| "learning_rate": 3.788527397260274e-05, | |
| "loss": 0.0026, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.9634703196347032, | |
| "grad_norm": 0.043823979794979095, | |
| "learning_rate": 3.77425799086758e-05, | |
| "loss": 0.0026, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.9863013698630136, | |
| "grad_norm": 0.06740803271532059, | |
| "learning_rate": 3.759988584474886e-05, | |
| "loss": 0.0026, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 0.0005470711039379239, | |
| "eval_runtime": 70.9617, | |
| "eval_samples_per_second": 42.276, | |
| "eval_steps_per_second": 2.649, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 2.009132420091324, | |
| "grad_norm": 0.07825632393360138, | |
| "learning_rate": 3.745719178082192e-05, | |
| "loss": 0.0023, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 2.0319634703196345, | |
| "grad_norm": 0.07475470006465912, | |
| "learning_rate": 3.731449771689498e-05, | |
| "loss": 0.0026, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 2.0547945205479454, | |
| "grad_norm": 0.027003251016139984, | |
| "learning_rate": 3.717180365296804e-05, | |
| "loss": 0.002, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 2.077625570776256, | |
| "grad_norm": 0.02263038419187069, | |
| "learning_rate": 3.70291095890411e-05, | |
| "loss": 0.0024, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 2.1004566210045663, | |
| "grad_norm": 0.10978245735168457, | |
| "learning_rate": 3.688641552511416e-05, | |
| "loss": 0.0016, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 2.1232876712328768, | |
| "grad_norm": 0.10029231756925583, | |
| "learning_rate": 3.6743721461187216e-05, | |
| "loss": 0.0019, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 2.146118721461187, | |
| "grad_norm": 0.13055719435214996, | |
| "learning_rate": 3.660102739726027e-05, | |
| "loss": 0.0024, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 2.1689497716894977, | |
| "grad_norm": 0.02834697626531124, | |
| "learning_rate": 3.6458333333333336e-05, | |
| "loss": 0.0021, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 2.191780821917808, | |
| "grad_norm": 0.03155607730150223, | |
| "learning_rate": 3.631563926940639e-05, | |
| "loss": 0.0017, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 2.2146118721461185, | |
| "grad_norm": 0.012841666117310524, | |
| "learning_rate": 3.617294520547945e-05, | |
| "loss": 0.0014, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 2.237442922374429, | |
| "grad_norm": 0.06208495423197746, | |
| "learning_rate": 3.603025114155251e-05, | |
| "loss": 0.0018, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 2.26027397260274, | |
| "grad_norm": 0.03939255326986313, | |
| "learning_rate": 3.588755707762557e-05, | |
| "loss": 0.0015, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 2.2831050228310503, | |
| "grad_norm": 0.0387921966612339, | |
| "learning_rate": 3.574486301369863e-05, | |
| "loss": 0.0016, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 2.3059360730593608, | |
| "grad_norm": 0.037367332726716995, | |
| "learning_rate": 3.560216894977169e-05, | |
| "loss": 0.0017, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 2.328767123287671, | |
| "grad_norm": 0.09546244889497757, | |
| "learning_rate": 3.545947488584475e-05, | |
| "loss": 0.0016, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 2.3515981735159817, | |
| "grad_norm": 0.05310087651014328, | |
| "learning_rate": 3.531678082191781e-05, | |
| "loss": 0.0017, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 2.374429223744292, | |
| "grad_norm": 0.13905133306980133, | |
| "learning_rate": 3.517408675799087e-05, | |
| "loss": 0.0016, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 2.3972602739726026, | |
| "grad_norm": 0.13682198524475098, | |
| "learning_rate": 3.503139269406393e-05, | |
| "loss": 0.0015, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 2.4200913242009134, | |
| "grad_norm": 0.07560446858406067, | |
| "learning_rate": 3.488869863013699e-05, | |
| "loss": 0.0019, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 2.442922374429224, | |
| "grad_norm": 0.024818843230605125, | |
| "learning_rate": 3.474600456621005e-05, | |
| "loss": 0.0019, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 2.4657534246575343, | |
| "grad_norm": 0.2794188857078552, | |
| "learning_rate": 3.4603310502283106e-05, | |
| "loss": 0.0014, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 2.4885844748858448, | |
| "grad_norm": 0.03945288807153702, | |
| "learning_rate": 3.446061643835616e-05, | |
| "loss": 0.0013, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 2.5114155251141552, | |
| "grad_norm": 0.028486991301178932, | |
| "learning_rate": 3.4317922374429226e-05, | |
| "loss": 0.0015, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 2.5342465753424657, | |
| "grad_norm": 0.19093026220798492, | |
| "learning_rate": 3.417522831050228e-05, | |
| "loss": 0.0016, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 2.557077625570776, | |
| "grad_norm": 0.01798168569803238, | |
| "learning_rate": 3.4032534246575346e-05, | |
| "loss": 0.0012, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 2.5799086757990866, | |
| "grad_norm": 0.03545854985713959, | |
| "learning_rate": 3.38898401826484e-05, | |
| "loss": 0.0012, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 2.602739726027397, | |
| "grad_norm": 0.0346076525747776, | |
| "learning_rate": 3.374714611872146e-05, | |
| "loss": 0.0013, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 2.625570776255708, | |
| "grad_norm": 0.031317390501499176, | |
| "learning_rate": 3.360445205479452e-05, | |
| "loss": 0.0011, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 2.6484018264840183, | |
| "grad_norm": 0.018010897561907768, | |
| "learning_rate": 3.346175799086758e-05, | |
| "loss": 0.0013, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 2.671232876712329, | |
| "grad_norm": 0.04990324005484581, | |
| "learning_rate": 3.331906392694064e-05, | |
| "loss": 0.0017, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 2.6940639269406392, | |
| "grad_norm": 0.10162694752216339, | |
| "learning_rate": 3.31763698630137e-05, | |
| "loss": 0.0015, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 2.7168949771689497, | |
| "grad_norm": 0.05713967978954315, | |
| "learning_rate": 3.303367579908676e-05, | |
| "loss": 0.0014, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 2.73972602739726, | |
| "grad_norm": 0.12799331545829773, | |
| "learning_rate": 3.289098173515982e-05, | |
| "loss": 0.0013, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 2.762557077625571, | |
| "grad_norm": 0.022338515147566795, | |
| "learning_rate": 3.274828767123288e-05, | |
| "loss": 0.0013, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 2.7853881278538815, | |
| "grad_norm": 0.01897992566227913, | |
| "learning_rate": 3.260559360730594e-05, | |
| "loss": 0.0011, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 2.808219178082192, | |
| "grad_norm": 0.04136025533080101, | |
| "learning_rate": 3.2462899543379e-05, | |
| "loss": 0.0017, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 2.8310502283105023, | |
| "grad_norm": 0.09205514937639236, | |
| "learning_rate": 3.232020547945205e-05, | |
| "loss": 0.0012, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 2.853881278538813, | |
| "grad_norm": 0.05605219677090645, | |
| "learning_rate": 3.2177511415525116e-05, | |
| "loss": 0.0013, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 2.8767123287671232, | |
| "grad_norm": 0.20732928812503815, | |
| "learning_rate": 3.203481735159817e-05, | |
| "loss": 0.0016, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 2.8995433789954337, | |
| "grad_norm": 0.01767803728580475, | |
| "learning_rate": 3.1892123287671236e-05, | |
| "loss": 0.0011, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 2.922374429223744, | |
| "grad_norm": 0.05888301134109497, | |
| "learning_rate": 3.174942922374429e-05, | |
| "loss": 0.0013, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 2.9452054794520546, | |
| "grad_norm": 0.060406286269426346, | |
| "learning_rate": 3.160673515981735e-05, | |
| "loss": 0.0012, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 2.968036529680365, | |
| "grad_norm": 0.016563862562179565, | |
| "learning_rate": 3.146404109589041e-05, | |
| "loss": 0.0011, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 2.990867579908676, | |
| "grad_norm": 0.032936833798885345, | |
| "learning_rate": 3.132134703196347e-05, | |
| "loss": 0.0012, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 2.3043350665830076e-05, | |
| "eval_runtime": 71.3592, | |
| "eval_samples_per_second": 42.041, | |
| "eval_steps_per_second": 2.635, | |
| "step": 1314 | |
| }, | |
| { | |
| "epoch": 3.0136986301369864, | |
| "grad_norm": 0.043910350650548935, | |
| "learning_rate": 3.117865296803653e-05, | |
| "loss": 0.0012, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 3.036529680365297, | |
| "grad_norm": 0.051233988255262375, | |
| "learning_rate": 3.103595890410959e-05, | |
| "loss": 0.0012, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 3.0593607305936072, | |
| "grad_norm": 0.0460955835878849, | |
| "learning_rate": 3.089326484018265e-05, | |
| "loss": 0.0012, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 3.0821917808219177, | |
| "grad_norm": 0.11139486730098724, | |
| "learning_rate": 3.075057077625571e-05, | |
| "loss": 0.0015, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 3.105022831050228, | |
| "grad_norm": 0.020884834229946136, | |
| "learning_rate": 3.060787671232877e-05, | |
| "loss": 0.0009, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 3.127853881278539, | |
| "grad_norm": 0.04865699261426926, | |
| "learning_rate": 3.046518264840183e-05, | |
| "loss": 0.0009, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 3.1506849315068495, | |
| "grad_norm": 0.04672854766249657, | |
| "learning_rate": 3.032248858447489e-05, | |
| "loss": 0.0009, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 3.17351598173516, | |
| "grad_norm": 0.04292335361242294, | |
| "learning_rate": 3.017979452054795e-05, | |
| "loss": 0.0009, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 3.1963470319634704, | |
| "grad_norm": 0.06898372620344162, | |
| "learning_rate": 3.0037100456621002e-05, | |
| "loss": 0.0009, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 3.219178082191781, | |
| "grad_norm": 0.08381140232086182, | |
| "learning_rate": 2.9894406392694062e-05, | |
| "loss": 0.0012, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 3.2420091324200913, | |
| "grad_norm": 0.009932724758982658, | |
| "learning_rate": 2.9751712328767122e-05, | |
| "loss": 0.0009, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 3.2648401826484017, | |
| "grad_norm": 0.055948104709386826, | |
| "learning_rate": 2.9609018264840182e-05, | |
| "loss": 0.0012, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 3.287671232876712, | |
| "grad_norm": 0.028575940057635307, | |
| "learning_rate": 2.9466324200913242e-05, | |
| "loss": 0.0011, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 3.3105022831050226, | |
| "grad_norm": 0.023319421336054802, | |
| "learning_rate": 2.9323630136986302e-05, | |
| "loss": 0.0009, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 3.3333333333333335, | |
| "grad_norm": 0.05601884797215462, | |
| "learning_rate": 2.9180936073059362e-05, | |
| "loss": 0.0009, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 3.356164383561644, | |
| "grad_norm": 0.017812812700867653, | |
| "learning_rate": 2.9038242009132422e-05, | |
| "loss": 0.0008, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 3.3789954337899544, | |
| "grad_norm": 0.015839802101254463, | |
| "learning_rate": 2.8895547945205482e-05, | |
| "loss": 0.0009, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 3.401826484018265, | |
| "grad_norm": 0.013905179686844349, | |
| "learning_rate": 2.8752853881278542e-05, | |
| "loss": 0.0009, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 3.4246575342465753, | |
| "grad_norm": 0.034228548407554626, | |
| "learning_rate": 2.8610159817351602e-05, | |
| "loss": 0.0009, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 3.4474885844748857, | |
| "grad_norm": 0.1382512003183365, | |
| "learning_rate": 2.846746575342466e-05, | |
| "loss": 0.001, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 3.470319634703196, | |
| "grad_norm": 0.026061363518238068, | |
| "learning_rate": 2.832477168949772e-05, | |
| "loss": 0.0007, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 3.493150684931507, | |
| "grad_norm": 0.033779580146074295, | |
| "learning_rate": 2.818207762557078e-05, | |
| "loss": 0.0007, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 3.5159817351598175, | |
| "grad_norm": 0.01672547496855259, | |
| "learning_rate": 2.803938356164384e-05, | |
| "loss": 0.0009, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 3.538812785388128, | |
| "grad_norm": 0.03841786831617355, | |
| "learning_rate": 2.7896689497716895e-05, | |
| "loss": 0.0008, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 3.5616438356164384, | |
| "grad_norm": 0.042960572987794876, | |
| "learning_rate": 2.7753995433789952e-05, | |
| "loss": 0.0008, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 3.584474885844749, | |
| "grad_norm": 0.016713028773665428, | |
| "learning_rate": 2.7611301369863012e-05, | |
| "loss": 0.0007, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 3.6073059360730593, | |
| "grad_norm": 0.04062510281801224, | |
| "learning_rate": 2.7468607305936072e-05, | |
| "loss": 0.0007, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 3.6301369863013697, | |
| "grad_norm": 0.013800432905554771, | |
| "learning_rate": 2.7325913242009132e-05, | |
| "loss": 0.0009, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 3.65296803652968, | |
| "grad_norm": 0.03654768317937851, | |
| "learning_rate": 2.7183219178082192e-05, | |
| "loss": 0.0008, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 3.6757990867579906, | |
| "grad_norm": 0.04716182500123978, | |
| "learning_rate": 2.7040525114155252e-05, | |
| "loss": 0.0007, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 3.6986301369863015, | |
| "grad_norm": 0.006938883103430271, | |
| "learning_rate": 2.6897831050228312e-05, | |
| "loss": 0.0005, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 3.721461187214612, | |
| "grad_norm": 0.07364363968372345, | |
| "learning_rate": 2.6755136986301372e-05, | |
| "loss": 0.0007, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 3.7442922374429224, | |
| "grad_norm": 0.028122154995799065, | |
| "learning_rate": 2.6612442922374432e-05, | |
| "loss": 0.0006, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 3.767123287671233, | |
| "grad_norm": 0.008902553468942642, | |
| "learning_rate": 2.6469748858447492e-05, | |
| "loss": 0.0006, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 3.7899543378995433, | |
| "grad_norm": 0.013538227416574955, | |
| "learning_rate": 2.6327054794520552e-05, | |
| "loss": 0.0007, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 3.8127853881278537, | |
| "grad_norm": 0.01429641991853714, | |
| "learning_rate": 2.6184360730593612e-05, | |
| "loss": 0.0007, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 3.8356164383561646, | |
| "grad_norm": 0.03232923895120621, | |
| "learning_rate": 2.604166666666667e-05, | |
| "loss": 0.0006, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 3.858447488584475, | |
| "grad_norm": 0.008046794682741165, | |
| "learning_rate": 2.589897260273973e-05, | |
| "loss": 0.0008, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 3.8812785388127855, | |
| "grad_norm": 0.04098529741168022, | |
| "learning_rate": 2.575627853881279e-05, | |
| "loss": 0.0006, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 3.904109589041096, | |
| "grad_norm": 0.06163308396935463, | |
| "learning_rate": 2.5613584474885845e-05, | |
| "loss": 0.0007, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 3.9269406392694064, | |
| "grad_norm": 0.022219218313694, | |
| "learning_rate": 2.54708904109589e-05, | |
| "loss": 0.0006, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 3.949771689497717, | |
| "grad_norm": 0.09443829208612442, | |
| "learning_rate": 2.532819634703196e-05, | |
| "loss": 0.0007, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 3.9726027397260273, | |
| "grad_norm": 0.03068207949399948, | |
| "learning_rate": 2.518550228310502e-05, | |
| "loss": 0.0007, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 3.9954337899543377, | |
| "grad_norm": 0.09690997749567032, | |
| "learning_rate": 2.504280821917808e-05, | |
| "loss": 0.0007, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_loss": 9.936958122125361e-06, | |
| "eval_runtime": 71.0483, | |
| "eval_samples_per_second": 42.225, | |
| "eval_steps_per_second": 2.646, | |
| "step": 1752 | |
| }, | |
| { | |
| "epoch": 4.018264840182648, | |
| "grad_norm": 0.011336731724441051, | |
| "learning_rate": 2.490011415525114e-05, | |
| "loss": 0.0006, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 4.041095890410959, | |
| "grad_norm": 0.015287565998733044, | |
| "learning_rate": 2.47574200913242e-05, | |
| "loss": 0.0008, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 4.063926940639269, | |
| "grad_norm": 0.008016168139874935, | |
| "learning_rate": 2.461472602739726e-05, | |
| "loss": 0.0007, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 4.0867579908675795, | |
| "grad_norm": 0.06609506160020828, | |
| "learning_rate": 2.447203196347032e-05, | |
| "loss": 0.0006, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 4.109589041095891, | |
| "grad_norm": 0.009175929240882397, | |
| "learning_rate": 2.432933789954338e-05, | |
| "loss": 0.0007, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 4.132420091324201, | |
| "grad_norm": 0.0126581359654665, | |
| "learning_rate": 2.418664383561644e-05, | |
| "loss": 0.0005, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 4.155251141552512, | |
| "grad_norm": 0.08836681395769119, | |
| "learning_rate": 2.40439497716895e-05, | |
| "loss": 0.001, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 4.178082191780822, | |
| "grad_norm": 0.025666864588856697, | |
| "learning_rate": 2.3901255707762558e-05, | |
| "loss": 0.0006, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 4.200913242009133, | |
| "grad_norm": 0.03877547010779381, | |
| "learning_rate": 2.3758561643835618e-05, | |
| "loss": 0.0005, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 4.223744292237443, | |
| "grad_norm": 0.016599468886852264, | |
| "learning_rate": 2.3615867579908678e-05, | |
| "loss": 0.0006, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 4.2465753424657535, | |
| "grad_norm": 0.06378292292356491, | |
| "learning_rate": 2.3473173515981735e-05, | |
| "loss": 0.0012, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 4.269406392694064, | |
| "grad_norm": 0.014281521551311016, | |
| "learning_rate": 2.3330479452054795e-05, | |
| "loss": 0.0008, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 4.292237442922374, | |
| "grad_norm": 0.09428286552429199, | |
| "learning_rate": 2.3187785388127855e-05, | |
| "loss": 0.0009, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 4.315068493150685, | |
| "grad_norm": 0.006857575848698616, | |
| "learning_rate": 2.3045091324200915e-05, | |
| "loss": 0.0006, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 4.337899543378995, | |
| "grad_norm": 0.08501231670379639, | |
| "learning_rate": 2.2902397260273975e-05, | |
| "loss": 0.0006, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 4.360730593607306, | |
| "grad_norm": 0.01724555715918541, | |
| "learning_rate": 2.275970319634703e-05, | |
| "loss": 0.0006, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 4.383561643835616, | |
| "grad_norm": 0.008119191974401474, | |
| "learning_rate": 2.261700913242009e-05, | |
| "loss": 0.0006, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 4.406392694063927, | |
| "grad_norm": 0.026137014850974083, | |
| "learning_rate": 2.247431506849315e-05, | |
| "loss": 0.0005, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 4.429223744292237, | |
| "grad_norm": 0.005971621256321669, | |
| "learning_rate": 2.233162100456621e-05, | |
| "loss": 0.0006, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 4.4520547945205475, | |
| "grad_norm": 0.04149805009365082, | |
| "learning_rate": 2.218892694063927e-05, | |
| "loss": 0.0006, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 4.474885844748858, | |
| "grad_norm": 0.00725915003567934, | |
| "learning_rate": 2.204623287671233e-05, | |
| "loss": 0.0005, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 4.497716894977169, | |
| "grad_norm": 0.04520484060049057, | |
| "learning_rate": 2.190353881278539e-05, | |
| "loss": 0.0006, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 4.52054794520548, | |
| "grad_norm": 0.029336770996451378, | |
| "learning_rate": 2.1760844748858448e-05, | |
| "loss": 0.0005, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 4.54337899543379, | |
| "grad_norm": 0.09512148797512054, | |
| "learning_rate": 2.1618150684931508e-05, | |
| "loss": 0.0008, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 4.566210045662101, | |
| "grad_norm": 0.042485009878873825, | |
| "learning_rate": 2.1475456621004568e-05, | |
| "loss": 0.0005, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 4.589041095890411, | |
| "grad_norm": 0.013867770321667194, | |
| "learning_rate": 2.1332762557077628e-05, | |
| "loss": 0.0006, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 4.6118721461187215, | |
| "grad_norm": 0.05775037035346031, | |
| "learning_rate": 2.1190068493150684e-05, | |
| "loss": 0.0006, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 4.634703196347032, | |
| "grad_norm": 0.010395473800599575, | |
| "learning_rate": 2.1047374429223744e-05, | |
| "loss": 0.0006, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 4.657534246575342, | |
| "grad_norm": 0.008579747751355171, | |
| "learning_rate": 2.0904680365296804e-05, | |
| "loss": 0.0005, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 4.680365296803653, | |
| "grad_norm": 0.011320821940898895, | |
| "learning_rate": 2.0761986301369864e-05, | |
| "loss": 0.0004, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 4.703196347031963, | |
| "grad_norm": 0.038680724799633026, | |
| "learning_rate": 2.061929223744292e-05, | |
| "loss": 0.0005, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 4.726027397260274, | |
| "grad_norm": 0.014130471274256706, | |
| "learning_rate": 2.047659817351598e-05, | |
| "loss": 0.0005, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 4.748858447488584, | |
| "grad_norm": 0.01286408118903637, | |
| "learning_rate": 2.033390410958904e-05, | |
| "loss": 0.0006, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 4.771689497716895, | |
| "grad_norm": 0.06255431473255157, | |
| "learning_rate": 2.01912100456621e-05, | |
| "loss": 0.0006, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 4.794520547945205, | |
| "grad_norm": 0.033248186111450195, | |
| "learning_rate": 2.004851598173516e-05, | |
| "loss": 0.0007, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 4.817351598173516, | |
| "grad_norm": 0.008136480115354061, | |
| "learning_rate": 1.990582191780822e-05, | |
| "loss": 0.0007, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 4.840182648401827, | |
| "grad_norm": 0.0758700743317604, | |
| "learning_rate": 1.976312785388128e-05, | |
| "loss": 0.0006, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 4.863013698630137, | |
| "grad_norm": 0.008333769626915455, | |
| "learning_rate": 1.962043378995434e-05, | |
| "loss": 0.0005, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 4.885844748858448, | |
| "grad_norm": 0.010320034809410572, | |
| "learning_rate": 1.9477739726027397e-05, | |
| "loss": 0.0004, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 4.908675799086758, | |
| "grad_norm": 0.00360821932554245, | |
| "learning_rate": 1.9335045662100457e-05, | |
| "loss": 0.0006, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 4.931506849315069, | |
| "grad_norm": 0.02473396062850952, | |
| "learning_rate": 1.9192351598173517e-05, | |
| "loss": 0.0005, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 4.954337899543379, | |
| "grad_norm": 0.022373970597982407, | |
| "learning_rate": 1.9049657534246577e-05, | |
| "loss": 0.0006, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 4.9771689497716896, | |
| "grad_norm": 0.04363209009170532, | |
| "learning_rate": 1.8906963470319637e-05, | |
| "loss": 0.0005, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 0.0052858320996165276, | |
| "learning_rate": 1.8764269406392694e-05, | |
| "loss": 0.0005, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_loss": 5.627029167953879e-06, | |
| "eval_runtime": 71.2259, | |
| "eval_samples_per_second": 42.12, | |
| "eval_steps_per_second": 2.639, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 5.0228310502283104, | |
| "grad_norm": 0.00950402021408081, | |
| "learning_rate": 1.8621575342465754e-05, | |
| "loss": 0.0005, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 5.045662100456621, | |
| "grad_norm": 0.010009740479290485, | |
| "learning_rate": 1.8478881278538814e-05, | |
| "loss": 0.0005, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 5.068493150684931, | |
| "grad_norm": 0.032611336559057236, | |
| "learning_rate": 1.833618721461187e-05, | |
| "loss": 0.0004, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 5.091324200913242, | |
| "grad_norm": 0.00459344731643796, | |
| "learning_rate": 1.819349315068493e-05, | |
| "loss": 0.0005, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 5.114155251141552, | |
| "grad_norm": 0.03434896096587181, | |
| "learning_rate": 1.805079908675799e-05, | |
| "loss": 0.0008, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 5.136986301369863, | |
| "grad_norm": 0.030723148956894875, | |
| "learning_rate": 1.790810502283105e-05, | |
| "loss": 0.0004, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 5.159817351598173, | |
| "grad_norm": 0.0043680015951395035, | |
| "learning_rate": 1.776541095890411e-05, | |
| "loss": 0.0003, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 5.182648401826484, | |
| "grad_norm": 0.011316012591123581, | |
| "learning_rate": 1.762271689497717e-05, | |
| "loss": 0.0006, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 5.205479452054795, | |
| "grad_norm": 0.006297079846262932, | |
| "learning_rate": 1.748002283105023e-05, | |
| "loss": 0.0005, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 5.228310502283105, | |
| "grad_norm": 0.018946554511785507, | |
| "learning_rate": 1.733732876712329e-05, | |
| "loss": 0.0004, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 5.251141552511416, | |
| "grad_norm": 0.017262902110815048, | |
| "learning_rate": 1.7194634703196347e-05, | |
| "loss": 0.0004, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 5.273972602739726, | |
| "grad_norm": 0.010157715529203415, | |
| "learning_rate": 1.7051940639269407e-05, | |
| "loss": 0.0004, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 5.296803652968037, | |
| "grad_norm": 0.012058730237185955, | |
| "learning_rate": 1.6909246575342467e-05, | |
| "loss": 0.0008, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 5.319634703196347, | |
| "grad_norm": 0.006543818395584822, | |
| "learning_rate": 1.6766552511415527e-05, | |
| "loss": 0.0003, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 5.342465753424658, | |
| "grad_norm": 0.010772732086479664, | |
| "learning_rate": 1.6623858447488587e-05, | |
| "loss": 0.0004, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 5.365296803652968, | |
| "grad_norm": 0.0169979240745306, | |
| "learning_rate": 1.6481164383561644e-05, | |
| "loss": 0.0003, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 5.3881278538812785, | |
| "grad_norm": 0.009800782427191734, | |
| "learning_rate": 1.6338470319634704e-05, | |
| "loss": 0.0003, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 5.410958904109589, | |
| "grad_norm": 0.04885419085621834, | |
| "learning_rate": 1.6195776255707764e-05, | |
| "loss": 0.0004, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 5.433789954337899, | |
| "grad_norm": 0.024770863354206085, | |
| "learning_rate": 1.605308219178082e-05, | |
| "loss": 0.0005, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 5.45662100456621, | |
| "grad_norm": 0.0030876509845256805, | |
| "learning_rate": 1.591038812785388e-05, | |
| "loss": 0.0004, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 5.47945205479452, | |
| "grad_norm": 0.018670551478862762, | |
| "learning_rate": 1.576769406392694e-05, | |
| "loss": 0.0004, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 5.502283105022831, | |
| "grad_norm": 0.011356856673955917, | |
| "learning_rate": 1.5625e-05, | |
| "loss": 0.0006, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 5.525114155251142, | |
| "grad_norm": 0.0847870260477066, | |
| "learning_rate": 1.548230593607306e-05, | |
| "loss": 0.0004, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 5.5479452054794525, | |
| "grad_norm": 0.017403950914740562, | |
| "learning_rate": 1.533961187214612e-05, | |
| "loss": 0.0004, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 5.570776255707763, | |
| "grad_norm": 0.17497682571411133, | |
| "learning_rate": 1.519691780821918e-05, | |
| "loss": 0.0006, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 5.593607305936073, | |
| "grad_norm": 0.032679907977581024, | |
| "learning_rate": 1.5054223744292237e-05, | |
| "loss": 0.0003, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 5.616438356164384, | |
| "grad_norm": 0.011194907128810883, | |
| "learning_rate": 1.4911529680365297e-05, | |
| "loss": 0.0004, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 5.639269406392694, | |
| "grad_norm": 0.017079461365938187, | |
| "learning_rate": 1.4768835616438357e-05, | |
| "loss": 0.0004, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 5.662100456621005, | |
| "grad_norm": 0.09856831282377243, | |
| "learning_rate": 1.4626141552511415e-05, | |
| "loss": 0.0004, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 5.684931506849315, | |
| "grad_norm": 0.1313595324754715, | |
| "learning_rate": 1.4483447488584475e-05, | |
| "loss": 0.0007, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 5.707762557077626, | |
| "grad_norm": 0.006652920041233301, | |
| "learning_rate": 1.4340753424657535e-05, | |
| "loss": 0.0004, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 5.730593607305936, | |
| "grad_norm": 0.009232467040419579, | |
| "learning_rate": 1.4198059360730595e-05, | |
| "loss": 0.0004, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 5.7534246575342465, | |
| "grad_norm": 0.013255695812404156, | |
| "learning_rate": 1.4055365296803655e-05, | |
| "loss": 0.0005, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 5.776255707762557, | |
| "grad_norm": 0.017853327095508575, | |
| "learning_rate": 1.3912671232876712e-05, | |
| "loss": 0.0005, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 5.799086757990867, | |
| "grad_norm": 0.011152304708957672, | |
| "learning_rate": 1.3769977168949772e-05, | |
| "loss": 0.0004, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 5.821917808219178, | |
| "grad_norm": 0.00643182685598731, | |
| "learning_rate": 1.3627283105022832e-05, | |
| "loss": 0.0004, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 5.844748858447488, | |
| "grad_norm": 0.016502410173416138, | |
| "learning_rate": 1.348458904109589e-05, | |
| "loss": 0.0005, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 5.867579908675799, | |
| "grad_norm": 0.01771596260368824, | |
| "learning_rate": 1.334189497716895e-05, | |
| "loss": 0.0005, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 5.890410958904109, | |
| "grad_norm": 0.024237055331468582, | |
| "learning_rate": 1.319920091324201e-05, | |
| "loss": 0.0004, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 5.91324200913242, | |
| "grad_norm": 0.03518439084291458, | |
| "learning_rate": 1.305650684931507e-05, | |
| "loss": 0.0003, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 5.936073059360731, | |
| "grad_norm": 0.014818640425801277, | |
| "learning_rate": 1.291381278538813e-05, | |
| "loss": 0.0004, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 5.958904109589041, | |
| "grad_norm": 0.006442646961659193, | |
| "learning_rate": 1.2771118721461187e-05, | |
| "loss": 0.0005, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 5.981735159817352, | |
| "grad_norm": 0.0037180872168391943, | |
| "learning_rate": 1.2628424657534247e-05, | |
| "loss": 0.0004, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_loss": 3.286777655375772e-06, | |
| "eval_runtime": 71.001, | |
| "eval_samples_per_second": 42.253, | |
| "eval_steps_per_second": 2.648, | |
| "step": 2628 | |
| }, | |
| { | |
| "epoch": 6.004566210045662, | |
| "grad_norm": 0.00996373500674963, | |
| "learning_rate": 1.2485730593607307e-05, | |
| "loss": 0.0003, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 6.027397260273973, | |
| "grad_norm": 0.0067086792550981045, | |
| "learning_rate": 1.2343036529680365e-05, | |
| "loss": 0.0005, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 6.050228310502283, | |
| "grad_norm": 0.009166865609586239, | |
| "learning_rate": 1.2200342465753425e-05, | |
| "loss": 0.0005, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 6.073059360730594, | |
| "grad_norm": 0.005400144029408693, | |
| "learning_rate": 1.2057648401826485e-05, | |
| "loss": 0.0003, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 6.095890410958904, | |
| "grad_norm": 0.0033110773656517267, | |
| "learning_rate": 1.1914954337899543e-05, | |
| "loss": 0.0003, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 6.1187214611872145, | |
| "grad_norm": 0.017671756446361542, | |
| "learning_rate": 1.1772260273972603e-05, | |
| "loss": 0.0003, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 6.141552511415525, | |
| "grad_norm": 0.016568470746278763, | |
| "learning_rate": 1.1629566210045663e-05, | |
| "loss": 0.0005, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 6.164383561643835, | |
| "grad_norm": 0.11980602890253067, | |
| "learning_rate": 1.1486872146118723e-05, | |
| "loss": 0.0004, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 6.187214611872146, | |
| "grad_norm": 0.03170336037874222, | |
| "learning_rate": 1.1344178082191781e-05, | |
| "loss": 0.0004, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 6.210045662100456, | |
| "grad_norm": 0.01595723256468773, | |
| "learning_rate": 1.1201484018264841e-05, | |
| "loss": 0.0006, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 6.232876712328767, | |
| "grad_norm": 0.05525317043066025, | |
| "learning_rate": 1.10587899543379e-05, | |
| "loss": 0.0004, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 6.255707762557078, | |
| "grad_norm": 0.00609338004142046, | |
| "learning_rate": 1.091609589041096e-05, | |
| "loss": 0.0004, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 6.2785388127853885, | |
| "grad_norm": 0.061812903732061386, | |
| "learning_rate": 1.0773401826484018e-05, | |
| "loss": 0.0004, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 6.301369863013699, | |
| "grad_norm": 0.004055253695696592, | |
| "learning_rate": 1.0630707762557078e-05, | |
| "loss": 0.0003, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 6.324200913242009, | |
| "grad_norm": 0.010857106186449528, | |
| "learning_rate": 1.0488013698630138e-05, | |
| "loss": 0.0004, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 6.34703196347032, | |
| "grad_norm": 0.008729905821383, | |
| "learning_rate": 1.0345319634703198e-05, | |
| "loss": 0.0003, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 6.36986301369863, | |
| "grad_norm": 0.003697366453707218, | |
| "learning_rate": 1.0202625570776256e-05, | |
| "loss": 0.0003, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 6.392694063926941, | |
| "grad_norm": 0.0052353376522660255, | |
| "learning_rate": 1.0059931506849316e-05, | |
| "loss": 0.0003, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 6.415525114155251, | |
| "grad_norm": 0.06486661732196808, | |
| "learning_rate": 9.917237442922375e-06, | |
| "loss": 0.0003, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 6.438356164383562, | |
| "grad_norm": 0.036493659019470215, | |
| "learning_rate": 9.774543378995435e-06, | |
| "loss": 0.0004, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 6.461187214611872, | |
| "grad_norm": 0.010287893004715443, | |
| "learning_rate": 9.631849315068493e-06, | |
| "loss": 0.0005, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 6.4840182648401825, | |
| "grad_norm": 0.04603143036365509, | |
| "learning_rate": 9.489155251141553e-06, | |
| "loss": 0.0005, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 6.506849315068493, | |
| "grad_norm": 0.016138222068548203, | |
| "learning_rate": 9.346461187214613e-06, | |
| "loss": 0.0004, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 6.529680365296803, | |
| "grad_norm": 0.018696075305342674, | |
| "learning_rate": 9.203767123287671e-06, | |
| "loss": 0.0003, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 6.552511415525114, | |
| "grad_norm": 0.009262233041226864, | |
| "learning_rate": 9.061073059360731e-06, | |
| "loss": 0.0003, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 6.575342465753424, | |
| "grad_norm": 0.005299023352563381, | |
| "learning_rate": 8.918378995433791e-06, | |
| "loss": 0.0003, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 6.598173515981735, | |
| "grad_norm": 0.00924585945904255, | |
| "learning_rate": 8.77568493150685e-06, | |
| "loss": 0.0005, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 6.621004566210045, | |
| "grad_norm": 0.023650696501135826, | |
| "learning_rate": 8.632990867579908e-06, | |
| "loss": 0.0003, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 6.6438356164383565, | |
| "grad_norm": 0.005081293638795614, | |
| "learning_rate": 8.490296803652968e-06, | |
| "loss": 0.0003, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 6.666666666666667, | |
| "grad_norm": 0.012082752771675587, | |
| "learning_rate": 8.347602739726028e-06, | |
| "loss": 0.0003, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 6.689497716894977, | |
| "grad_norm": 0.004923298954963684, | |
| "learning_rate": 8.204908675799088e-06, | |
| "loss": 0.0004, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 6.712328767123288, | |
| "grad_norm": 0.012247243896126747, | |
| "learning_rate": 8.062214611872146e-06, | |
| "loss": 0.0003, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 6.735159817351598, | |
| "grad_norm": 0.003992004320025444, | |
| "learning_rate": 7.919520547945206e-06, | |
| "loss": 0.0004, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 6.757990867579909, | |
| "grad_norm": 0.11996075510978699, | |
| "learning_rate": 7.776826484018266e-06, | |
| "loss": 0.0004, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 6.780821917808219, | |
| "grad_norm": 0.005172852426767349, | |
| "learning_rate": 7.634132420091324e-06, | |
| "loss": 0.0003, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 6.80365296803653, | |
| "grad_norm": 0.006295430473983288, | |
| "learning_rate": 7.4914383561643835e-06, | |
| "loss": 0.0004, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 6.82648401826484, | |
| "grad_norm": 0.004225455224514008, | |
| "learning_rate": 7.3487442922374434e-06, | |
| "loss": 0.0003, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 6.8493150684931505, | |
| "grad_norm": 0.0032389916013926268, | |
| "learning_rate": 7.206050228310503e-06, | |
| "loss": 0.0003, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 6.872146118721461, | |
| "grad_norm": 0.038855355232954025, | |
| "learning_rate": 7.063356164383563e-06, | |
| "loss": 0.0003, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 6.894977168949771, | |
| "grad_norm": 0.003689020639285445, | |
| "learning_rate": 6.920662100456621e-06, | |
| "loss": 0.0004, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 6.917808219178082, | |
| "grad_norm": 0.01840381883084774, | |
| "learning_rate": 6.777968036529681e-06, | |
| "loss": 0.0003, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 6.940639269406392, | |
| "grad_norm": 0.00360796507447958, | |
| "learning_rate": 6.63527397260274e-06, | |
| "loss": 0.0003, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 6.963470319634704, | |
| "grad_norm": 0.01734967716038227, | |
| "learning_rate": 6.4925799086758e-06, | |
| "loss": 0.0003, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 6.986301369863014, | |
| "grad_norm": 0.03626614063978195, | |
| "learning_rate": 6.349885844748858e-06, | |
| "loss": 0.0003, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_loss": 2.8402284897310892e-06, | |
| "eval_runtime": 70.9798, | |
| "eval_samples_per_second": 42.266, | |
| "eval_steps_per_second": 2.649, | |
| "step": 3066 | |
| }, | |
| { | |
| "epoch": 7.0091324200913245, | |
| "grad_norm": 0.029538586735725403, | |
| "learning_rate": 6.207191780821918e-06, | |
| "loss": 0.0003, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 7.031963470319635, | |
| "grad_norm": 0.011442271992564201, | |
| "learning_rate": 6.0644977168949774e-06, | |
| "loss": 0.0003, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 7.054794520547945, | |
| "grad_norm": 0.009785550646483898, | |
| "learning_rate": 5.921803652968037e-06, | |
| "loss": 0.0003, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 7.077625570776256, | |
| "grad_norm": 0.048744283616542816, | |
| "learning_rate": 5.779109589041097e-06, | |
| "loss": 0.0006, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 7.100456621004566, | |
| "grad_norm": 0.007315410766750574, | |
| "learning_rate": 5.636415525114156e-06, | |
| "loss": 0.0002, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 7.123287671232877, | |
| "grad_norm": 0.007827061228454113, | |
| "learning_rate": 5.493721461187215e-06, | |
| "loss": 0.0004, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 7.146118721461187, | |
| "grad_norm": 0.0036791411694139242, | |
| "learning_rate": 5.351027397260274e-06, | |
| "loss": 0.0003, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 7.168949771689498, | |
| "grad_norm": 0.010082785040140152, | |
| "learning_rate": 5.208333333333334e-06, | |
| "loss": 0.0003, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 7.191780821917808, | |
| "grad_norm": 0.008839433081448078, | |
| "learning_rate": 5.065639269406393e-06, | |
| "loss": 0.0003, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 7.2146118721461185, | |
| "grad_norm": 0.009164104238152504, | |
| "learning_rate": 4.922945205479452e-06, | |
| "loss": 0.0003, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 7.237442922374429, | |
| "grad_norm": 0.06758883595466614, | |
| "learning_rate": 4.7802511415525114e-06, | |
| "loss": 0.0003, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 7.260273972602739, | |
| "grad_norm": 0.0034141126088798046, | |
| "learning_rate": 4.6375570776255714e-06, | |
| "loss": 0.0004, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 7.28310502283105, | |
| "grad_norm": 0.009116005152463913, | |
| "learning_rate": 4.494863013698631e-06, | |
| "loss": 0.0003, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 7.30593607305936, | |
| "grad_norm": 0.014750838279724121, | |
| "learning_rate": 4.35216894977169e-06, | |
| "loss": 0.0003, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 7.328767123287671, | |
| "grad_norm": 0.013239488005638123, | |
| "learning_rate": 4.209474885844749e-06, | |
| "loss": 0.0003, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 7.351598173515982, | |
| "grad_norm": 0.009504084475338459, | |
| "learning_rate": 4.066780821917809e-06, | |
| "loss": 0.0004, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 7.3744292237442925, | |
| "grad_norm": 0.018879897892475128, | |
| "learning_rate": 3.924086757990868e-06, | |
| "loss": 0.0004, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 7.397260273972603, | |
| "grad_norm": 0.01805570162832737, | |
| "learning_rate": 3.7813926940639276e-06, | |
| "loss": 0.0003, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 7.420091324200913, | |
| "grad_norm": 0.08645830303430557, | |
| "learning_rate": 3.6386986301369863e-06, | |
| "loss": 0.0005, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 7.442922374429224, | |
| "grad_norm": 0.01047169417142868, | |
| "learning_rate": 3.4960045662100463e-06, | |
| "loss": 0.0003, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 7.465753424657534, | |
| "grad_norm": 0.00378281413577497, | |
| "learning_rate": 3.353310502283105e-06, | |
| "loss": 0.0003, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 7.488584474885845, | |
| "grad_norm": 0.008809005841612816, | |
| "learning_rate": 3.210616438356165e-06, | |
| "loss": 0.0003, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 7.511415525114155, | |
| "grad_norm": 0.006725851446390152, | |
| "learning_rate": 3.0679223744292237e-06, | |
| "loss": 0.0003, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 7.534246575342466, | |
| "grad_norm": 0.011108343489468098, | |
| "learning_rate": 2.9252283105022833e-06, | |
| "loss": 0.0003, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 7.557077625570776, | |
| "grad_norm": 0.006323591805994511, | |
| "learning_rate": 2.7825342465753424e-06, | |
| "loss": 0.0003, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 7.579908675799087, | |
| "grad_norm": 0.01762283407151699, | |
| "learning_rate": 2.639840182648402e-06, | |
| "loss": 0.0003, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 7.602739726027397, | |
| "grad_norm": 0.003676204476505518, | |
| "learning_rate": 2.497146118721461e-06, | |
| "loss": 0.0002, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 7.6255707762557075, | |
| "grad_norm": 0.04348291829228401, | |
| "learning_rate": 2.3544520547945207e-06, | |
| "loss": 0.0003, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 7.648401826484018, | |
| "grad_norm": 0.01973540708422661, | |
| "learning_rate": 2.21175799086758e-06, | |
| "loss": 0.0004, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 7.671232876712329, | |
| "grad_norm": 0.0020602825097739697, | |
| "learning_rate": 2.0690639269406394e-06, | |
| "loss": 0.0003, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 7.69406392694064, | |
| "grad_norm": 0.031328316777944565, | |
| "learning_rate": 1.9263698630136986e-06, | |
| "loss": 0.0003, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 7.71689497716895, | |
| "grad_norm": 0.004346560686826706, | |
| "learning_rate": 1.7836757990867581e-06, | |
| "loss": 0.0003, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 7.739726027397261, | |
| "grad_norm": 0.020926913246512413, | |
| "learning_rate": 1.6409817351598175e-06, | |
| "loss": 0.0003, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 7.762557077625571, | |
| "grad_norm": 0.012918849475681782, | |
| "learning_rate": 1.4982876712328766e-06, | |
| "loss": 0.0005, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 7.7853881278538815, | |
| "grad_norm": 0.03088468685746193, | |
| "learning_rate": 1.355593607305936e-06, | |
| "loss": 0.0004, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 7.808219178082192, | |
| "grad_norm": 0.02280372381210327, | |
| "learning_rate": 1.2128995433789954e-06, | |
| "loss": 0.0006, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 7.831050228310502, | |
| "grad_norm": 0.0038483564276248217, | |
| "learning_rate": 1.0702054794520547e-06, | |
| "loss": 0.0003, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 7.853881278538813, | |
| "grad_norm": 0.026072148233652115, | |
| "learning_rate": 9.275114155251142e-07, | |
| "loss": 0.0004, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 7.876712328767123, | |
| "grad_norm": 0.05634905397891998, | |
| "learning_rate": 7.848173515981735e-07, | |
| "loss": 0.0003, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 7.899543378995434, | |
| "grad_norm": 0.003738977015018463, | |
| "learning_rate": 6.421232876712329e-07, | |
| "loss": 0.0003, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 7.922374429223744, | |
| "grad_norm": 0.00907827913761139, | |
| "learning_rate": 4.994292237442923e-07, | |
| "loss": 0.0003, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 7.945205479452055, | |
| "grad_norm": 0.02188229374587536, | |
| "learning_rate": 3.567351598173516e-07, | |
| "loss": 0.0003, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 7.968036529680365, | |
| "grad_norm": 0.020626889541745186, | |
| "learning_rate": 2.1404109589041096e-07, | |
| "loss": 0.0003, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 7.9908675799086755, | |
| "grad_norm": 0.010182246565818787, | |
| "learning_rate": 7.134703196347032e-08, | |
| "loss": 0.0004, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_loss": 2.446558710289537e-06, | |
| "eval_runtime": 70.9518, | |
| "eval_samples_per_second": 42.282, | |
| "eval_steps_per_second": 2.65, | |
| "step": 3504 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 3504, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 8, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 8742142387077120.0, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |