| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.999537251272559, |
| "eval_steps": 100, |
| "global_step": 900, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.005552984729291994, |
| "grad_norm": 5.378993948282789, |
| "learning_rate": 1.111111111111111e-06, |
| "loss": 1.3618, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.011105969458583989, |
| "grad_norm": 4.007932344194533, |
| "learning_rate": 2.222222222222222e-06, |
| "loss": 1.3497, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.016658954187875982, |
| "grad_norm": 2.870467768621527, |
| "learning_rate": 3.3333333333333333e-06, |
| "loss": 1.3292, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.022211938917167977, |
| "grad_norm": 2.9129656624638147, |
| "learning_rate": 4.444444444444444e-06, |
| "loss": 1.3086, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.027764923646459973, |
| "grad_norm": 2.0216771653661767, |
| "learning_rate": 5.555555555555557e-06, |
| "loss": 1.1997, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.033317908375751965, |
| "grad_norm": 1.7768511495820114, |
| "learning_rate": 6.666666666666667e-06, |
| "loss": 1.1742, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.03887089310504396, |
| "grad_norm": 1.4804724166332308, |
| "learning_rate": 7.77777777777778e-06, |
| "loss": 1.1244, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.044423877834335955, |
| "grad_norm": 1.5180593968939422, |
| "learning_rate": 8.888888888888888e-06, |
| "loss": 1.0946, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.04997686256362795, |
| "grad_norm": 1.4879318494570208, |
| "learning_rate": 1e-05, |
| "loss": 1.1032, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.055529847292919945, |
| "grad_norm": 1.449589161585785, |
| "learning_rate": 1.1111111111111113e-05, |
| "loss": 1.0751, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.06108283202221194, |
| "grad_norm": 1.3300428861282843, |
| "learning_rate": 1.2222222222222224e-05, |
| "loss": 1.0704, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.06663581675150393, |
| "grad_norm": 1.5071782473179958, |
| "learning_rate": 1.3333333333333333e-05, |
| "loss": 1.0561, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.07218880148079593, |
| "grad_norm": 1.3957616523065333, |
| "learning_rate": 1.4444444444444446e-05, |
| "loss": 1.0527, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.07774178621008793, |
| "grad_norm": 1.3254556004203786, |
| "learning_rate": 1.555555555555556e-05, |
| "loss": 1.0107, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.08329477093937991, |
| "grad_norm": 1.5821507688192253, |
| "learning_rate": 1.6666666666666667e-05, |
| "loss": 1.035, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.08884775566867191, |
| "grad_norm": 1.5770964291038414, |
| "learning_rate": 1.7777777777777777e-05, |
| "loss": 1.0285, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.09440074039796391, |
| "grad_norm": 1.5156588992406825, |
| "learning_rate": 1.888888888888889e-05, |
| "loss": 1.0462, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.0999537251272559, |
| "grad_norm": 1.8368814946604881, |
| "learning_rate": 2e-05, |
| "loss": 1.0321, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.10550670985654789, |
| "grad_norm": 1.6154848357095026, |
| "learning_rate": 1.9998119704485016e-05, |
| "loss": 0.9962, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.11105969458583989, |
| "grad_norm": 1.6218644276479086, |
| "learning_rate": 1.9992479525042305e-05, |
| "loss": 1.0216, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.11105969458583989, |
| "eval_loss": 1.0477724075317383, |
| "eval_runtime": 14.0681, |
| "eval_samples_per_second": 18.197, |
| "eval_steps_per_second": 4.549, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.11661267931513189, |
| "grad_norm": 1.4399255902202484, |
| "learning_rate": 1.9983081582712684e-05, |
| "loss": 1.0344, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.12216566404442387, |
| "grad_norm": 1.8069324642784876, |
| "learning_rate": 1.996992941167792e-05, |
| "loss": 1.0204, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.12771864877371586, |
| "grad_norm": 2.1003495945173865, |
| "learning_rate": 1.9953027957931658e-05, |
| "loss": 1.0169, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.13327163350300786, |
| "grad_norm": 1.9735058045205416, |
| "learning_rate": 1.9932383577419432e-05, |
| "loss": 1.0039, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.13882461823229986, |
| "grad_norm": 1.4822914643295109, |
| "learning_rate": 1.9908004033648452e-05, |
| "loss": 1.0025, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.14437760296159186, |
| "grad_norm": 1.5730749273183444, |
| "learning_rate": 1.9879898494768093e-05, |
| "loss": 1.0014, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.14993058769088385, |
| "grad_norm": 1.4317675236519043, |
| "learning_rate": 1.9848077530122083e-05, |
| "loss": 1.0055, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.15548357242017585, |
| "grad_norm": 1.4305310143482486, |
| "learning_rate": 1.9812553106273848e-05, |
| "loss": 0.9772, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.16103655714946785, |
| "grad_norm": 1.835206300273569, |
| "learning_rate": 1.9773338582506357e-05, |
| "loss": 1.0041, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.16658954187875982, |
| "grad_norm": 1.346307644251197, |
| "learning_rate": 1.973044870579824e-05, |
| "loss": 0.9836, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.17214252660805182, |
| "grad_norm": 1.3087431013115765, |
| "learning_rate": 1.9683899605278062e-05, |
| "loss": 0.9906, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.17769551133734382, |
| "grad_norm": 1.6304479084212757, |
| "learning_rate": 1.9633708786158803e-05, |
| "loss": 0.9846, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.18324849606663582, |
| "grad_norm": 1.4079238364588627, |
| "learning_rate": 1.957989512315489e-05, |
| "loss": 0.9953, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.18880148079592782, |
| "grad_norm": 1.4160494597971853, |
| "learning_rate": 1.9522478853384154e-05, |
| "loss": 0.9728, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.19435446552521982, |
| "grad_norm": 1.3831279332888735, |
| "learning_rate": 1.946148156875751e-05, |
| "loss": 0.9786, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.1999074502545118, |
| "grad_norm": 1.3223005814385944, |
| "learning_rate": 1.9396926207859085e-05, |
| "loss": 0.9915, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.20546043498380379, |
| "grad_norm": 7.733682148553893, |
| "learning_rate": 1.932883704732001e-05, |
| "loss": 1.0145, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.21101341971309578, |
| "grad_norm": 1.6669211194032283, |
| "learning_rate": 1.9257239692688907e-05, |
| "loss": 0.9862, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.21656640444238778, |
| "grad_norm": 1.4461392109939817, |
| "learning_rate": 1.9182161068802742e-05, |
| "loss": 0.9944, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.22211938917167978, |
| "grad_norm": 1.3816905654957743, |
| "learning_rate": 1.9103629409661468e-05, |
| "loss": 0.99, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.22211938917167978, |
| "eval_loss": 1.0106741189956665, |
| "eval_runtime": 13.0423, |
| "eval_samples_per_second": 19.628, |
| "eval_steps_per_second": 4.907, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.22767237390097178, |
| "grad_norm": 1.3296649623180647, |
| "learning_rate": 1.902167424781038e-05, |
| "loss": 0.9874, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.23322535863026378, |
| "grad_norm": 1.4147575392387848, |
| "learning_rate": 1.8936326403234125e-05, |
| "loss": 0.9906, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.23877834335955575, |
| "grad_norm": 1.5048777388561694, |
| "learning_rate": 1.8847617971766577e-05, |
| "loss": 0.9721, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.24433132808884775, |
| "grad_norm": 1.3281954721580007, |
| "learning_rate": 1.8755582313020912e-05, |
| "loss": 0.9848, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.24988431281813975, |
| "grad_norm": 1.3865529123681517, |
| "learning_rate": 1.866025403784439e-05, |
| "loss": 0.9885, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.2554372975474317, |
| "grad_norm": 1.468591497661801, |
| "learning_rate": 1.8561668995302668e-05, |
| "loss": 0.9713, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.2609902822767237, |
| "grad_norm": 1.3686298836564843, |
| "learning_rate": 1.845986425919841e-05, |
| "loss": 0.9579, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.2665432670060157, |
| "grad_norm": 1.4148255169077197, |
| "learning_rate": 1.8354878114129368e-05, |
| "loss": 0.9506, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.2720962517353077, |
| "grad_norm": 1.5172086489276786, |
| "learning_rate": 1.824675004109107e-05, |
| "loss": 0.99, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.2776492364645997, |
| "grad_norm": 1.3436284741737878, |
| "learning_rate": 1.8135520702629677e-05, |
| "loss": 0.9654, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.2832022211938917, |
| "grad_norm": 1.4414606449378646, |
| "learning_rate": 1.802123192755044e-05, |
| "loss": 0.9668, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.2887552059231837, |
| "grad_norm": 1.4031899462530004, |
| "learning_rate": 1.7903926695187595e-05, |
| "loss": 0.9626, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.2943081906524757, |
| "grad_norm": 1.4639958980701138, |
| "learning_rate": 1.7783649119241603e-05, |
| "loss": 0.9459, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.2998611753817677, |
| "grad_norm": 1.426651386589118, |
| "learning_rate": 1.766044443118978e-05, |
| "loss": 0.9863, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.3054141601110597, |
| "grad_norm": 1.4219614478552796, |
| "learning_rate": 1.7534358963276606e-05, |
| "loss": 0.9719, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.3109671448403517, |
| "grad_norm": 1.3151103765065284, |
| "learning_rate": 1.740544013109005e-05, |
| "loss": 0.9903, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.3165201295696437, |
| "grad_norm": 1.3354926740055781, |
| "learning_rate": 1.7273736415730488e-05, |
| "loss": 0.9681, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.3220731142989357, |
| "grad_norm": 1.3717221395296357, |
| "learning_rate": 1.7139297345578992e-05, |
| "loss": 0.9456, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.32762609902822765, |
| "grad_norm": 1.6411114205659896, |
| "learning_rate": 1.7002173477671685e-05, |
| "loss": 0.9591, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.33317908375751965, |
| "grad_norm": 1.2826425067775913, |
| "learning_rate": 1.686241637868734e-05, |
| "loss": 0.9328, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.33317908375751965, |
| "eval_loss": 0.99033522605896, |
| "eval_runtime": 13.2544, |
| "eval_samples_per_second": 19.314, |
| "eval_steps_per_second": 4.829, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.33873206848681164, |
| "grad_norm": 1.4005676511844571, |
| "learning_rate": 1.6720078605555227e-05, |
| "loss": 0.9803, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.34428505321610364, |
| "grad_norm": 1.3022121862564202, |
| "learning_rate": 1.657521368569064e-05, |
| "loss": 0.9622, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.34983803794539564, |
| "grad_norm": 1.4147938551822972, |
| "learning_rate": 1.6427876096865394e-05, |
| "loss": 0.9733, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.35539102267468764, |
| "grad_norm": 1.4563599805627543, |
| "learning_rate": 1.627812124672099e-05, |
| "loss": 0.9695, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.36094400740397964, |
| "grad_norm": 1.2666931064627047, |
| "learning_rate": 1.6126005451932028e-05, |
| "loss": 0.9512, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.36649699213327164, |
| "grad_norm": 1.3459612867606927, |
| "learning_rate": 1.5971585917027864e-05, |
| "loss": 0.9962, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.37204997686256364, |
| "grad_norm": 1.4053767425847852, |
| "learning_rate": 1.5814920712880267e-05, |
| "loss": 0.9456, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.37760296159185563, |
| "grad_norm": 1.2542630494816203, |
| "learning_rate": 1.5656068754865388e-05, |
| "loss": 0.9625, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.38315594632114763, |
| "grad_norm": 1.2467268401516984, |
| "learning_rate": 1.5495089780708062e-05, |
| "loss": 0.9416, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.38870893105043963, |
| "grad_norm": 1.3642420280912566, |
| "learning_rate": 1.5332044328016916e-05, |
| "loss": 0.9745, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.39426191577973163, |
| "grad_norm": 1.370419991211419, |
| "learning_rate": 1.5166993711518631e-05, |
| "loss": 0.9235, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.3998149005090236, |
| "grad_norm": 1.4175545037228292, |
| "learning_rate": 1.5000000000000002e-05, |
| "loss": 0.9419, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.4053678852383156, |
| "grad_norm": 1.5230125115585145, |
| "learning_rate": 1.4831125992966386e-05, |
| "loss": 0.9482, |
| "step": 365 |
| }, |
| { |
| "epoch": 0.41092086996760757, |
| "grad_norm": 1.3243119971736406, |
| "learning_rate": 1.4660435197025391e-05, |
| "loss": 0.9516, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.41647385469689957, |
| "grad_norm": 1.3168758353005081, |
| "learning_rate": 1.4487991802004625e-05, |
| "loss": 0.9274, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.42202683942619157, |
| "grad_norm": 1.3316158417964403, |
| "learning_rate": 1.4313860656812537e-05, |
| "loss": 0.9279, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.42757982415548357, |
| "grad_norm": 1.2112610899992784, |
| "learning_rate": 1.4138107245051394e-05, |
| "loss": 0.9604, |
| "step": 385 |
| }, |
| { |
| "epoch": 0.43313280888477557, |
| "grad_norm": 1.4256068345744652, |
| "learning_rate": 1.396079766039157e-05, |
| "loss": 0.9504, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.43868579361406757, |
| "grad_norm": 1.48676297842706, |
| "learning_rate": 1.3781998581716427e-05, |
| "loss": 0.9783, |
| "step": 395 |
| }, |
| { |
| "epoch": 0.44423877834335956, |
| "grad_norm": 1.3917645840392392, |
| "learning_rate": 1.3601777248047105e-05, |
| "loss": 0.9428, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.44423877834335956, |
| "eval_loss": 0.975586473941803, |
| "eval_runtime": 13.0584, |
| "eval_samples_per_second": 19.604, |
| "eval_steps_per_second": 4.901, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.44979176307265156, |
| "grad_norm": 1.3230682794396744, |
| "learning_rate": 1.342020143325669e-05, |
| "loss": 0.9477, |
| "step": 405 |
| }, |
| { |
| "epoch": 0.45534474780194356, |
| "grad_norm": 1.2045267312184351, |
| "learning_rate": 1.3237339420583213e-05, |
| "loss": 0.9568, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.46089773253123556, |
| "grad_norm": 1.2839147189555775, |
| "learning_rate": 1.3053259976951134e-05, |
| "loss": 0.9256, |
| "step": 415 |
| }, |
| { |
| "epoch": 0.46645071726052756, |
| "grad_norm": 1.284551507215065, |
| "learning_rate": 1.2868032327110904e-05, |
| "loss": 0.9246, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.4720037019898195, |
| "grad_norm": 1.3021372211969566, |
| "learning_rate": 1.2681726127606374e-05, |
| "loss": 0.9527, |
| "step": 425 |
| }, |
| { |
| "epoch": 0.4775566867191115, |
| "grad_norm": 1.3264954697360052, |
| "learning_rate": 1.2494411440579814e-05, |
| "loss": 0.9659, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.4831096714484035, |
| "grad_norm": 1.2931203310254533, |
| "learning_rate": 1.2306158707424402e-05, |
| "loss": 0.9044, |
| "step": 435 |
| }, |
| { |
| "epoch": 0.4886626561776955, |
| "grad_norm": 1.5033726081614123, |
| "learning_rate": 1.211703872229411e-05, |
| "loss": 0.9135, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.4942156409069875, |
| "grad_norm": 1.2465062347279259, |
| "learning_rate": 1.1927122605480899e-05, |
| "loss": 0.9382, |
| "step": 445 |
| }, |
| { |
| "epoch": 0.4997686256362795, |
| "grad_norm": 1.2206738144949079, |
| "learning_rate": 1.1736481776669307e-05, |
| "loss": 0.9429, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.5053216103655715, |
| "grad_norm": 1.268246804784844, |
| "learning_rate": 1.1545187928078407e-05, |
| "loss": 0.9163, |
| "step": 455 |
| }, |
| { |
| "epoch": 0.5108745950948634, |
| "grad_norm": 1.1842171344447763, |
| "learning_rate": 1.1353312997501313e-05, |
| "loss": 0.9296, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.5164275798241554, |
| "grad_norm": 1.3329158977326516, |
| "learning_rate": 1.1160929141252303e-05, |
| "loss": 0.9536, |
| "step": 465 |
| }, |
| { |
| "epoch": 0.5219805645534474, |
| "grad_norm": 1.3140264320853574, |
| "learning_rate": 1.0968108707031792e-05, |
| "loss": 0.9314, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.5275335492827394, |
| "grad_norm": 1.2112452453436886, |
| "learning_rate": 1.077492420671931e-05, |
| "loss": 0.8858, |
| "step": 475 |
| }, |
| { |
| "epoch": 0.5330865340120314, |
| "grad_norm": 1.244066114064373, |
| "learning_rate": 1.0581448289104759e-05, |
| "loss": 0.938, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.5386395187413234, |
| "grad_norm": 1.2958379438943313, |
| "learning_rate": 1.038775371256817e-05, |
| "loss": 0.9366, |
| "step": 485 |
| }, |
| { |
| "epoch": 0.5441925034706154, |
| "grad_norm": 1.2706792533588414, |
| "learning_rate": 1.0193913317718245e-05, |
| "loss": 0.9809, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.5497454881999074, |
| "grad_norm": 1.2373793128051882, |
| "learning_rate": 1e-05, |
| "loss": 0.9107, |
| "step": 495 |
| }, |
| { |
| "epoch": 0.5552984729291994, |
| "grad_norm": 1.3259462807230906, |
| "learning_rate": 9.806086682281759e-06, |
| "loss": 0.9478, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.5552984729291994, |
| "eval_loss": 0.9645185470581055, |
| "eval_runtime": 14.0973, |
| "eval_samples_per_second": 18.16, |
| "eval_steps_per_second": 4.54, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.5608514576584914, |
| "grad_norm": 1.2818749562353056, |
| "learning_rate": 9.612246287431832e-06, |
| "loss": 0.9157, |
| "step": 505 |
| }, |
| { |
| "epoch": 0.5664044423877834, |
| "grad_norm": 1.2789066996692675, |
| "learning_rate": 9.418551710895243e-06, |
| "loss": 0.9715, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.5719574271170754, |
| "grad_norm": 1.3159690301384173, |
| "learning_rate": 9.225075793280693e-06, |
| "loss": 0.9564, |
| "step": 515 |
| }, |
| { |
| "epoch": 0.5775104118463674, |
| "grad_norm": 1.1820999233185365, |
| "learning_rate": 9.03189129296821e-06, |
| "loss": 0.9333, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.5830633965756594, |
| "grad_norm": 2.9705367947063595, |
| "learning_rate": 8.839070858747697e-06, |
| "loss": 0.9393, |
| "step": 525 |
| }, |
| { |
| "epoch": 0.5886163813049514, |
| "grad_norm": 1.1578484132538005, |
| "learning_rate": 8.646687002498692e-06, |
| "loss": 0.9386, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.5941693660342434, |
| "grad_norm": 1.2495396181990488, |
| "learning_rate": 8.454812071921597e-06, |
| "loss": 0.9202, |
| "step": 535 |
| }, |
| { |
| "epoch": 0.5997223507635354, |
| "grad_norm": 1.18243382811664, |
| "learning_rate": 8.263518223330698e-06, |
| "loss": 0.9222, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.6052753354928274, |
| "grad_norm": 1.2247856999180733, |
| "learning_rate": 8.072877394519103e-06, |
| "loss": 0.9426, |
| "step": 545 |
| }, |
| { |
| "epoch": 0.6108283202221194, |
| "grad_norm": 1.220145252189745, |
| "learning_rate": 7.882961277705897e-06, |
| "loss": 0.9161, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.6163813049514114, |
| "grad_norm": 1.2450054960684753, |
| "learning_rate": 7.6938412925756e-06, |
| "loss": 0.9419, |
| "step": 555 |
| }, |
| { |
| "epoch": 0.6219342896807034, |
| "grad_norm": 1.2450637061923184, |
| "learning_rate": 7.505588559420188e-06, |
| "loss": 0.9471, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.6274872744099954, |
| "grad_norm": 1.2160705969336116, |
| "learning_rate": 7.3182738723936255e-06, |
| "loss": 0.9446, |
| "step": 565 |
| }, |
| { |
| "epoch": 0.6330402591392874, |
| "grad_norm": 1.2857156644799776, |
| "learning_rate": 7.131967672889101e-06, |
| "loss": 0.9473, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.6385932438685794, |
| "grad_norm": 1.1688923012156514, |
| "learning_rate": 6.94674002304887e-06, |
| "loss": 0.9193, |
| "step": 575 |
| }, |
| { |
| "epoch": 0.6441462285978714, |
| "grad_norm": 1.3240345620915759, |
| "learning_rate": 6.762660579416791e-06, |
| "loss": 0.955, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.6496992133271634, |
| "grad_norm": 1.171487612102084, |
| "learning_rate": 6.579798566743314e-06, |
| "loss": 0.9345, |
| "step": 585 |
| }, |
| { |
| "epoch": 0.6552521980564553, |
| "grad_norm": 1.1809918233177483, |
| "learning_rate": 6.3982227519528986e-06, |
| "loss": 0.9317, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.6608051827857473, |
| "grad_norm": 1.1849135550680583, |
| "learning_rate": 6.218001418283577e-06, |
| "loss": 0.9282, |
| "step": 595 |
| }, |
| { |
| "epoch": 0.6663581675150393, |
| "grad_norm": 1.2353428612054926, |
| "learning_rate": 6.039202339608432e-06, |
| "loss": 0.9186, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.6663581675150393, |
| "eval_loss": 0.9549762010574341, |
| "eval_runtime": 13.1058, |
| "eval_samples_per_second": 19.533, |
| "eval_steps_per_second": 4.883, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.6719111522443313, |
| "grad_norm": 1.2260206730700363, |
| "learning_rate": 5.8618927549486095e-06, |
| "loss": 0.91, |
| "step": 605 |
| }, |
| { |
| "epoch": 0.6774641369736233, |
| "grad_norm": 1.2511346515975978, |
| "learning_rate": 5.686139343187468e-06, |
| "loss": 0.9445, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.6830171217029153, |
| "grad_norm": 1.1415920400151276, |
| "learning_rate": 5.512008197995379e-06, |
| "loss": 0.9267, |
| "step": 615 |
| }, |
| { |
| "epoch": 0.6885701064322073, |
| "grad_norm": 1.1718028560057823, |
| "learning_rate": 5.339564802974615e-06, |
| "loss": 0.9173, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.6941230911614993, |
| "grad_norm": 1.1427826499364968, |
| "learning_rate": 5.168874007033615e-06, |
| "loss": 0.9113, |
| "step": 625 |
| }, |
| { |
| "epoch": 0.6996760758907913, |
| "grad_norm": 1.1646711162888848, |
| "learning_rate": 5.000000000000003e-06, |
| "loss": 0.9314, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.7052290606200833, |
| "grad_norm": 1.1666389482954498, |
| "learning_rate": 4.8330062884813714e-06, |
| "loss": 0.949, |
| "step": 635 |
| }, |
| { |
| "epoch": 0.7107820453493753, |
| "grad_norm": 1.2669383225264854, |
| "learning_rate": 4.66795567198309e-06, |
| "loss": 0.9298, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.7163350300786673, |
| "grad_norm": 1.1727436218210467, |
| "learning_rate": 4.504910219291941e-06, |
| "loss": 0.9384, |
| "step": 645 |
| }, |
| { |
| "epoch": 0.7218880148079593, |
| "grad_norm": 1.160141429869342, |
| "learning_rate": 4.343931245134616e-06, |
| "loss": 0.9231, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.7274409995372513, |
| "grad_norm": 1.1571744280380951, |
| "learning_rate": 4.185079287119733e-06, |
| "loss": 0.9379, |
| "step": 655 |
| }, |
| { |
| "epoch": 0.7329939842665433, |
| "grad_norm": 1.2080957370089618, |
| "learning_rate": 4.028414082972141e-06, |
| "loss": 0.9087, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.7385469689958353, |
| "grad_norm": 1.2262375492125892, |
| "learning_rate": 3.873994548067972e-06, |
| "loss": 0.9175, |
| "step": 665 |
| }, |
| { |
| "epoch": 0.7440999537251273, |
| "grad_norm": 1.1868816861234752, |
| "learning_rate": 3.7218787532790167e-06, |
| "loss": 0.915, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.7496529384544193, |
| "grad_norm": 1.2149836180874647, |
| "learning_rate": 3.5721239031346067e-06, |
| "loss": 0.9359, |
| "step": 675 |
| }, |
| { |
| "epoch": 0.7552059231837113, |
| "grad_norm": 1.1401883150941166, |
| "learning_rate": 3.424786314309365e-06, |
| "loss": 0.8996, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.7607589079130033, |
| "grad_norm": 1.1918737445166034, |
| "learning_rate": 3.279921394444776e-06, |
| "loss": 0.9478, |
| "step": 685 |
| }, |
| { |
| "epoch": 0.7663118926422953, |
| "grad_norm": 1.1601543056853199, |
| "learning_rate": 3.1375836213126653e-06, |
| "loss": 0.9144, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.7718648773715873, |
| "grad_norm": 1.1849230333131153, |
| "learning_rate": 2.9978265223283152e-06, |
| "loss": 0.9134, |
| "step": 695 |
| }, |
| { |
| "epoch": 0.7774178621008793, |
| "grad_norm": 1.1922764155084293, |
| "learning_rate": 2.8607026544210115e-06, |
| "loss": 0.9184, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.7774178621008793, |
| "eval_loss": 0.9487817287445068, |
| "eval_runtime": 12.9966, |
| "eval_samples_per_second": 19.697, |
| "eval_steps_per_second": 4.924, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.7829708468301713, |
| "grad_norm": 1.0668112491118575, |
| "learning_rate": 2.726263584269513e-06, |
| "loss": 0.9038, |
| "step": 705 |
| }, |
| { |
| "epoch": 0.7885238315594633, |
| "grad_norm": 1.191610088893989, |
| "learning_rate": 2.594559868909956e-06, |
| "loss": 0.9003, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.7940768162887553, |
| "grad_norm": 1.2465046516243472, |
| "learning_rate": 2.4656410367233928e-06, |
| "loss": 0.9215, |
| "step": 715 |
| }, |
| { |
| "epoch": 0.7996298010180471, |
| "grad_norm": 1.1463175579772371, |
| "learning_rate": 2.339555568810221e-06, |
| "loss": 0.9303, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.8051827857473391, |
| "grad_norm": 1.1279074050372055, |
| "learning_rate": 2.2163508807584e-06, |
| "loss": 0.9294, |
| "step": 725 |
| }, |
| { |
| "epoch": 0.8107357704766311, |
| "grad_norm": 1.1668600336379225, |
| "learning_rate": 2.0960733048124082e-06, |
| "loss": 0.9082, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.8162887552059231, |
| "grad_norm": 1.1280124295582716, |
| "learning_rate": 1.9787680724495617e-06, |
| "loss": 0.898, |
| "step": 735 |
| }, |
| { |
| "epoch": 0.8218417399352151, |
| "grad_norm": 1.2693061690074205, |
| "learning_rate": 1.8644792973703252e-06, |
| "loss": 0.9473, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.8273947246645071, |
| "grad_norm": 1.0935688113313613, |
| "learning_rate": 1.7532499589089324e-06, |
| "loss": 0.9312, |
| "step": 745 |
| }, |
| { |
| "epoch": 0.8329477093937991, |
| "grad_norm": 1.1491375579091732, |
| "learning_rate": 1.6451218858706374e-06, |
| "loss": 0.9295, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.8385006941230911, |
| "grad_norm": 1.1971381545639446, |
| "learning_rate": 1.5401357408015893e-06, |
| "loss": 0.8977, |
| "step": 755 |
| }, |
| { |
| "epoch": 0.8440536788523831, |
| "grad_norm": 1.2118976466189748, |
| "learning_rate": 1.4383310046973365e-06, |
| "loss": 0.9201, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.8496066635816751, |
| "grad_norm": 1.1129578498310642, |
| "learning_rate": 1.339745962155613e-06, |
| "loss": 0.9235, |
| "step": 765 |
| }, |
| { |
| "epoch": 0.8551596483109671, |
| "grad_norm": 1.155034707390975, |
| "learning_rate": 1.2444176869790925e-06, |
| "loss": 0.906, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.8607126330402591, |
| "grad_norm": 1.1211050312494577, |
| "learning_rate": 1.152382028233422e-06, |
| "loss": 0.9027, |
| "step": 775 |
| }, |
| { |
| "epoch": 0.8662656177695511, |
| "grad_norm": 1.161462499770876, |
| "learning_rate": 1.0636735967658785e-06, |
| "loss": 0.902, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.8718186024988431, |
| "grad_norm": 1.2466679714276248, |
| "learning_rate": 9.783257521896228e-07, |
| "loss": 0.9312, |
| "step": 785 |
| }, |
| { |
| "epoch": 0.8773715872281351, |
| "grad_norm": 1.1688203046154668, |
| "learning_rate": 8.963705903385344e-07, |
| "loss": 0.9231, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.8829245719574271, |
| "grad_norm": 1.1446515266652288, |
| "learning_rate": 8.178389311972612e-07, |
| "loss": 0.924, |
| "step": 795 |
| }, |
| { |
| "epoch": 0.8884775566867191, |
| "grad_norm": 1.1971045442791841, |
| "learning_rate": 7.427603073110967e-07, |
| "loss": 0.8564, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.8884775566867191, |
| "eval_loss": 0.9455364346504211, |
| "eval_runtime": 13.3345, |
| "eval_samples_per_second": 19.198, |
| "eval_steps_per_second": 4.8, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.8940305414160111, |
| "grad_norm": 1.158461858290895, |
| "learning_rate": 6.711629526799946e-07, |
| "loss": 0.9399, |
| "step": 805 |
| }, |
| { |
| "epoch": 0.8995835261453031, |
| "grad_norm": 1.1264872944659339, |
| "learning_rate": 6.030737921409169e-07, |
| "loss": 0.8946, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.9051365108745951, |
| "grad_norm": 1.122295252126309, |
| "learning_rate": 5.385184312424973e-07, |
| "loss": 0.9228, |
| "step": 815 |
| }, |
| { |
| "epoch": 0.9106894956038871, |
| "grad_norm": 1.130327163633901, |
| "learning_rate": 4.775211466158469e-07, |
| "loss": 0.9278, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.9162424803331791, |
| "grad_norm": 1.152980775746581, |
| "learning_rate": 4.2010487684511105e-07, |
| "loss": 0.9105, |
| "step": 825 |
| }, |
| { |
| "epoch": 0.9217954650624711, |
| "grad_norm": 1.2630255671771575, |
| "learning_rate": 3.662912138411967e-07, |
| "loss": 0.9103, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.9273484497917631, |
| "grad_norm": 1.1212305446295874, |
| "learning_rate": 3.161003947219421e-07, |
| "loss": 0.8847, |
| "step": 835 |
| }, |
| { |
| "epoch": 0.9329014345210551, |
| "grad_norm": 1.1635748807768116, |
| "learning_rate": 2.6955129420176193e-07, |
| "loss": 0.9139, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.9384544192503471, |
| "grad_norm": 1.101552910147697, |
| "learning_rate": 2.2666141749364434e-07, |
| "loss": 0.9233, |
| "step": 845 |
| }, |
| { |
| "epoch": 0.944007403979639, |
| "grad_norm": 1.1104286718600382, |
| "learning_rate": 1.874468937261531e-07, |
| "loss": 0.8872, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.949560388708931, |
| "grad_norm": 1.1513796874736284, |
| "learning_rate": 1.519224698779198e-07, |
| "loss": 0.938, |
| "step": 855 |
| }, |
| { |
| "epoch": 0.955113373438223, |
| "grad_norm": 1.172996631179348, |
| "learning_rate": 1.201015052319099e-07, |
| "loss": 0.9366, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.960666358167515, |
| "grad_norm": 1.178813337349509, |
| "learning_rate": 9.199596635154684e-08, |
| "loss": 0.9382, |
| "step": 865 |
| }, |
| { |
| "epoch": 0.966219342896807, |
| "grad_norm": 1.123260974533791, |
| "learning_rate": 6.761642258056977e-08, |
| "loss": 0.8969, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.971772327626099, |
| "grad_norm": 1.2440553964560856, |
| "learning_rate": 4.6972042068341714e-08, |
| "loss": 0.9125, |
| "step": 875 |
| }, |
| { |
| "epoch": 0.977325312355391, |
| "grad_norm": 1.100622329056596, |
| "learning_rate": 3.0070588322079765e-08, |
| "loss": 0.9106, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.982878297084683, |
| "grad_norm": 1.224884941200071, |
| "learning_rate": 1.6918417287318245e-08, |
| "loss": 0.9045, |
| "step": 885 |
| }, |
| { |
| "epoch": 0.988431281813975, |
| "grad_norm": 1.1506168883594277, |
| "learning_rate": 7.520474957699586e-09, |
| "loss": 0.9341, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.993984266543267, |
| "grad_norm": 1.3030322318546892, |
| "learning_rate": 1.8802955149865854e-09, |
| "loss": 0.9093, |
| "step": 895 |
| }, |
| { |
| "epoch": 0.999537251272559, |
| "grad_norm": 1.164399511709177, |
| "learning_rate": 0.0, |
| "loss": 0.943, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.999537251272559, |
| "eval_loss": 0.9448966979980469, |
| "eval_runtime": 12.8793, |
| "eval_samples_per_second": 19.877, |
| "eval_steps_per_second": 4.969, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.999537251272559, |
| "step": 900, |
| "total_flos": 37979261239296.0, |
| "train_loss": 0.9636553647783067, |
| "train_runtime": 8704.1469, |
| "train_samples_per_second": 4.965, |
| "train_steps_per_second": 0.103 |
| } |
| ], |
| "logging_steps": 5, |
| "max_steps": 900, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": false, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 37979261239296.0, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|