{ "best_global_step": 875, "best_metric": 0.41610679030418396, "best_model_checkpoint": "/home/ricoiban/GEMMA/mnlp_chatsplaining/results_model/try_ft/checkpoint-875", "epoch": 0.5460448642266824, "eval_steps": 25, "global_step": 925, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0029515938606847697, "grad_norm": 274.0, "learning_rate": 4.800000000000001e-06, "loss": 10.7873, "step": 5 }, { "epoch": 0.0059031877213695395, "grad_norm": 160.0, "learning_rate": 1.08e-05, "loss": 8.2818, "step": 10 }, { "epoch": 0.00885478158205431, "grad_norm": 96.0, "learning_rate": 1.6800000000000002e-05, "loss": 5.4171, "step": 15 }, { "epoch": 0.011806375442739079, "grad_norm": 141.0, "learning_rate": 2.2800000000000002e-05, "loss": 4.8734, "step": 20 }, { "epoch": 0.01475796930342385, "grad_norm": 111.0, "learning_rate": 2.88e-05, "loss": 4.5143, "step": 25 }, { "epoch": 0.01475796930342385, "eval_loss": 4.148748874664307, "eval_runtime": 22.5903, "eval_samples_per_second": 88.534, "eval_steps_per_second": 88.534, "step": 25 }, { "epoch": 0.01770956316410862, "grad_norm": 234.0, "learning_rate": 3.48e-05, "loss": 4.0827, "step": 30 }, { "epoch": 0.02066115702479339, "grad_norm": 173.0, "learning_rate": 4.08e-05, "loss": 3.624, "step": 35 }, { "epoch": 0.023612750885478158, "grad_norm": 91.0, "learning_rate": 4.6800000000000006e-05, "loss": 2.6036, "step": 40 }, { "epoch": 0.026564344746162927, "grad_norm": 163.0, "learning_rate": 5.28e-05, "loss": 1.9856, "step": 45 }, { "epoch": 0.0295159386068477, "grad_norm": 65.0, "learning_rate": 5.88e-05, "loss": 1.1888, "step": 50 }, { "epoch": 0.0295159386068477, "eval_loss": 0.8685765862464905, "eval_runtime": 22.0353, "eval_samples_per_second": 90.764, "eval_steps_per_second": 90.764, "step": 50 }, { "epoch": 0.032467532467532464, "grad_norm": 31.125, "learning_rate": 5.9999123594193744e-05, "loss": 0.7088, "step": 55 }, { "epoch": 0.03541912632821724, "grad_norm": 25.625, "learning_rate": 5.9995563283365586e-05, "loss": 0.4428, "step": 60 }, { "epoch": 0.03837072018890201, "grad_norm": 38.25, "learning_rate": 5.998926461693058e-05, "loss": 0.7334, "step": 65 }, { "epoch": 0.04132231404958678, "grad_norm": 13.375, "learning_rate": 5.9980228169906714e-05, "loss": 0.6483, "step": 70 }, { "epoch": 0.04427390791027155, "grad_norm": 41.5, "learning_rate": 5.9968454767249506e-05, "loss": 0.5933, "step": 75 }, { "epoch": 0.04427390791027155, "eval_loss": 0.6194470524787903, "eval_runtime": 22.0585, "eval_samples_per_second": 90.668, "eval_steps_per_second": 90.668, "step": 75 }, { "epoch": 0.047225501770956316, "grad_norm": 26.0, "learning_rate": 5.995394548377669e-05, "loss": 0.7882, "step": 80 }, { "epoch": 0.050177095631641085, "grad_norm": 17.875, "learning_rate": 5.993670164407008e-05, "loss": 0.6423, "step": 85 }, { "epoch": 0.053128689492325853, "grad_norm": 19.75, "learning_rate": 5.991672482235466e-05, "loss": 0.6952, "step": 90 }, { "epoch": 0.05608028335301062, "grad_norm": 19.875, "learning_rate": 5.9894016842354855e-05, "loss": 0.6287, "step": 95 }, { "epoch": 0.0590318772136954, "grad_norm": 19.75, "learning_rate": 5.986857977712809e-05, "loss": 0.6606, "step": 100 }, { "epoch": 0.0590318772136954, "eval_loss": 0.5243311524391174, "eval_runtime": 22.0497, "eval_samples_per_second": 90.704, "eval_steps_per_second": 90.704, "step": 100 }, { "epoch": 0.06198347107438017, "grad_norm": 26.25, "learning_rate": 5.9840415948875444e-05, "loss": 0.6193, "step": 105 }, { "epoch": 0.06493506493506493, "grad_norm": 9.9375, "learning_rate": 5.980952792872975e-05, "loss": 0.5167, "step": 110 }, { "epoch": 0.0678866587957497, "grad_norm": 35.0, "learning_rate": 5.9775918536520786e-05, "loss": 0.5922, "step": 115 }, { "epoch": 0.07083825265643448, "grad_norm": 35.25, "learning_rate": 5.973959084051791e-05, "loss": 0.7114, "step": 120 }, { "epoch": 0.07378984651711924, "grad_norm": 8.5625, "learning_rate": 5.970054815714995e-05, "loss": 0.6309, "step": 125 }, { "epoch": 0.07378984651711924, "eval_loss": 0.644968569278717, "eval_runtime": 22.0752, "eval_samples_per_second": 90.6, "eval_steps_per_second": 90.6, "step": 125 }, { "epoch": 0.07674144037780402, "grad_norm": 19.375, "learning_rate": 5.965879405070235e-05, "loss": 0.622, "step": 130 }, { "epoch": 0.07969303423848878, "grad_norm": 15.4375, "learning_rate": 5.961433233299193e-05, "loss": 0.5902, "step": 135 }, { "epoch": 0.08264462809917356, "grad_norm": 19.375, "learning_rate": 5.956716706301877e-05, "loss": 0.6647, "step": 140 }, { "epoch": 0.08559622195985832, "grad_norm": 11.0, "learning_rate": 5.951730254659569e-05, "loss": 0.7817, "step": 145 }, { "epoch": 0.0885478158205431, "grad_norm": 16.875, "learning_rate": 5.946474333595521e-05, "loss": 0.6154, "step": 150 }, { "epoch": 0.0885478158205431, "eval_loss": 0.5631881356239319, "eval_runtime": 22.058, "eval_samples_per_second": 90.67, "eval_steps_per_second": 90.67, "step": 150 }, { "epoch": 0.09149940968122787, "grad_norm": 17.625, "learning_rate": 5.9409494229333904e-05, "loss": 0.6532, "step": 155 }, { "epoch": 0.09445100354191263, "grad_norm": 22.25, "learning_rate": 5.935156027053442e-05, "loss": 0.6099, "step": 160 }, { "epoch": 0.09740259740259741, "grad_norm": 25.5, "learning_rate": 5.929094674846495e-05, "loss": 0.7848, "step": 165 }, { "epoch": 0.10035419126328217, "grad_norm": 22.0, "learning_rate": 5.922765919665644e-05, "loss": 0.5597, "step": 170 }, { "epoch": 0.10330578512396695, "grad_norm": 11.9375, "learning_rate": 5.916170339275745e-05, "loss": 0.6166, "step": 175 }, { "epoch": 0.10330578512396695, "eval_loss": 0.5547081828117371, "eval_runtime": 22.0339, "eval_samples_per_second": 90.769, "eval_steps_per_second": 90.769, "step": 175 }, { "epoch": 0.10625737898465171, "grad_norm": 12.625, "learning_rate": 5.909308535800664e-05, "loss": 0.5548, "step": 180 }, { "epoch": 0.10920897284533648, "grad_norm": 32.5, "learning_rate": 5.90218113566831e-05, "loss": 0.5793, "step": 185 }, { "epoch": 0.11216056670602124, "grad_norm": 19.0, "learning_rate": 5.8947887895534504e-05, "loss": 0.8419, "step": 190 }, { "epoch": 0.11511216056670602, "grad_norm": 12.5625, "learning_rate": 5.8871321723183046e-05, "loss": 0.6571, "step": 195 }, { "epoch": 0.1180637544273908, "grad_norm": 18.375, "learning_rate": 5.879211982950937e-05, "loss": 0.6579, "step": 200 }, { "epoch": 0.1180637544273908, "eval_loss": 0.5850950479507446, "eval_runtime": 22.079, "eval_samples_per_second": 90.584, "eval_steps_per_second": 90.584, "step": 200 }, { "epoch": 0.12101534828807556, "grad_norm": 20.0, "learning_rate": 5.871028944501446e-05, "loss": 0.5835, "step": 205 }, { "epoch": 0.12396694214876033, "grad_norm": 20.25, "learning_rate": 5.862583804015953e-05, "loss": 0.5418, "step": 210 }, { "epoch": 0.1269185360094451, "grad_norm": 27.375, "learning_rate": 5.853877332468404e-05, "loss": 0.6755, "step": 215 }, { "epoch": 0.12987012987012986, "grad_norm": 29.5, "learning_rate": 5.844910324690187e-05, "loss": 0.6172, "step": 220 }, { "epoch": 0.13282172373081463, "grad_norm": 25.375, "learning_rate": 5.835683599297568e-05, "loss": 0.5844, "step": 225 }, { "epoch": 0.13282172373081463, "eval_loss": 0.5579066276550293, "eval_runtime": 22.0218, "eval_samples_per_second": 90.819, "eval_steps_per_second": 90.819, "step": 225 }, { "epoch": 0.1357733175914994, "grad_norm": 16.5, "learning_rate": 5.8261979986169596e-05, "loss": 0.6147, "step": 230 }, { "epoch": 0.13872491145218419, "grad_norm": 10.6875, "learning_rate": 5.816454388608023e-05, "loss": 0.6352, "step": 235 }, { "epoch": 0.14167650531286896, "grad_norm": 16.125, "learning_rate": 5.8064536587846115e-05, "loss": 0.7107, "step": 240 }, { "epoch": 0.1446280991735537, "grad_norm": 23.5, "learning_rate": 5.7961967221335674e-05, "loss": 0.6752, "step": 245 }, { "epoch": 0.14757969303423848, "grad_norm": 27.625, "learning_rate": 5.7856845150313716e-05, "loss": 0.6039, "step": 250 }, { "epoch": 0.14757969303423848, "eval_loss": 0.5708025693893433, "eval_runtime": 22.0736, "eval_samples_per_second": 90.606, "eval_steps_per_second": 90.606, "step": 250 }, { "epoch": 0.15053128689492326, "grad_norm": 19.375, "learning_rate": 5.7749179971586596e-05, "loss": 0.6029, "step": 255 }, { "epoch": 0.15348288075560804, "grad_norm": 24.75, "learning_rate": 5.763898151412613e-05, "loss": 0.4986, "step": 260 }, { "epoch": 0.15643447461629278, "grad_norm": 36.0, "learning_rate": 5.752625983817225e-05, "loss": 0.6049, "step": 265 }, { "epoch": 0.15938606847697756, "grad_norm": 22.25, "learning_rate": 5.7411025234314634e-05, "loss": 0.5895, "step": 270 }, { "epoch": 0.16233766233766234, "grad_norm": 13.0, "learning_rate": 5.729328822255319e-05, "loss": 0.6957, "step": 275 }, { "epoch": 0.16233766233766234, "eval_loss": 0.5568270087242126, "eval_runtime": 22.052, "eval_samples_per_second": 90.695, "eval_steps_per_second": 90.695, "step": 275 }, { "epoch": 0.1652892561983471, "grad_norm": 27.75, "learning_rate": 5.717305955133773e-05, "loss": 0.7125, "step": 280 }, { "epoch": 0.1682408500590319, "grad_norm": 12.8125, "learning_rate": 5.7050350196586686e-05, "loss": 0.5977, "step": 285 }, { "epoch": 0.17119244391971664, "grad_norm": 18.75, "learning_rate": 5.692517136068511e-05, "loss": 0.5908, "step": 290 }, { "epoch": 0.1741440377804014, "grad_norm": 24.25, "learning_rate": 5.679753447146195e-05, "loss": 0.5334, "step": 295 }, { "epoch": 0.1770956316410862, "grad_norm": 11.625, "learning_rate": 5.666745118114688e-05, "loss": 0.4347, "step": 300 }, { "epoch": 0.1770956316410862, "eval_loss": 0.5026609301567078, "eval_runtime": 22.0534, "eval_samples_per_second": 90.689, "eval_steps_per_second": 90.689, "step": 300 }, { "epoch": 0.18004722550177096, "grad_norm": 24.75, "learning_rate": 5.6534933365306394e-05, "loss": 0.5473, "step": 305 }, { "epoch": 0.18299881936245574, "grad_norm": 27.75, "learning_rate": 5.6399993121759797e-05, "loss": 0.5315, "step": 310 }, { "epoch": 0.1859504132231405, "grad_norm": 11.3125, "learning_rate": 5.626264276947469e-05, "loss": 0.4026, "step": 315 }, { "epoch": 0.18890200708382526, "grad_norm": 27.125, "learning_rate": 5.612289484744238e-05, "loss": 0.55, "step": 320 }, { "epoch": 0.19185360094451004, "grad_norm": 29.375, "learning_rate": 5.5980762113533166e-05, "loss": 0.5988, "step": 325 }, { "epoch": 0.19185360094451004, "eval_loss": 0.5229803919792175, "eval_runtime": 22.0483, "eval_samples_per_second": 90.71, "eval_steps_per_second": 90.71, "step": 325 }, { "epoch": 0.19480519480519481, "grad_norm": 18.0, "learning_rate": 5.5836257543331644e-05, "loss": 0.5174, "step": 330 }, { "epoch": 0.19775678866587956, "grad_norm": 17.75, "learning_rate": 5.568939432895213e-05, "loss": 0.6662, "step": 335 }, { "epoch": 0.20070838252656434, "grad_norm": 13.4375, "learning_rate": 5.554018587783435e-05, "loss": 0.6594, "step": 340 }, { "epoch": 0.20365997638724911, "grad_norm": 20.125, "learning_rate": 5.538864581151943e-05, "loss": 0.5776, "step": 345 }, { "epoch": 0.2066115702479339, "grad_norm": 29.0, "learning_rate": 5.523478796440633e-05, "loss": 0.6647, "step": 350 }, { "epoch": 0.2066115702479339, "eval_loss": 0.5277854204177856, "eval_runtime": 22.0408, "eval_samples_per_second": 90.741, "eval_steps_per_second": 90.741, "step": 350 }, { "epoch": 0.20956316410861867, "grad_norm": 17.625, "learning_rate": 5.507862638248896e-05, "loss": 0.5446, "step": 355 }, { "epoch": 0.21251475796930341, "grad_norm": 15.9375, "learning_rate": 5.49201753220738e-05, "loss": 0.5346, "step": 360 }, { "epoch": 0.2154663518299882, "grad_norm": 13.8125, "learning_rate": 5.475944924847845e-05, "loss": 0.4782, "step": 365 }, { "epoch": 0.21841794569067297, "grad_norm": 46.0, "learning_rate": 5.459646283471106e-05, "loss": 0.6363, "step": 370 }, { "epoch": 0.22136953955135774, "grad_norm": 12.625, "learning_rate": 5.443123096013083e-05, "loss": 0.5603, "step": 375 }, { "epoch": 0.22136953955135774, "eval_loss": 0.49818602204322815, "eval_runtime": 22.0724, "eval_samples_per_second": 90.611, "eval_steps_per_second": 90.611, "step": 375 }, { "epoch": 0.2243211334120425, "grad_norm": 25.125, "learning_rate": 5.426376870908959e-05, "loss": 0.6536, "step": 380 }, { "epoch": 0.22727272727272727, "grad_norm": 17.25, "learning_rate": 5.409409136955476e-05, "loss": 0.5464, "step": 385 }, { "epoch": 0.23022432113341204, "grad_norm": 23.375, "learning_rate": 5.3922214431713654e-05, "loss": 0.5587, "step": 390 }, { "epoch": 0.23317591499409682, "grad_norm": 10.75, "learning_rate": 5.3748153586559385e-05, "loss": 0.5231, "step": 395 }, { "epoch": 0.2361275088547816, "grad_norm": 19.875, "learning_rate": 5.357192472445835e-05, "loss": 0.503, "step": 400 }, { "epoch": 0.2361275088547816, "eval_loss": 0.4984550178050995, "eval_runtime": 22.074, "eval_samples_per_second": 90.604, "eval_steps_per_second": 90.604, "step": 400 }, { "epoch": 0.23907910271546634, "grad_norm": 20.125, "learning_rate": 5.339354393369962e-05, "loss": 0.4524, "step": 405 }, { "epoch": 0.24203069657615112, "grad_norm": 11.375, "learning_rate": 5.321302749902615e-05, "loss": 0.492, "step": 410 }, { "epoch": 0.2449822904368359, "grad_norm": 20.75, "learning_rate": 5.303039190014818e-05, "loss": 0.4989, "step": 415 }, { "epoch": 0.24793388429752067, "grad_norm": 14.0, "learning_rate": 5.284565381023873e-05, "loss": 0.6195, "step": 420 }, { "epoch": 0.25088547815820544, "grad_norm": 25.125, "learning_rate": 5.265883009441147e-05, "loss": 0.5687, "step": 425 }, { "epoch": 0.25088547815820544, "eval_loss": 0.4847257137298584, "eval_runtime": 22.029, "eval_samples_per_second": 90.789, "eval_steps_per_second": 90.789, "step": 425 }, { "epoch": 0.2538370720188902, "grad_norm": 16.75, "learning_rate": 5.2469937808181055e-05, "loss": 0.4048, "step": 430 }, { "epoch": 0.256788665879575, "grad_norm": 39.75, "learning_rate": 5.227899419590614e-05, "loss": 0.5483, "step": 435 }, { "epoch": 0.2597402597402597, "grad_norm": 15.125, "learning_rate": 5.208601668921508e-05, "loss": 0.4843, "step": 440 }, { "epoch": 0.2626918536009445, "grad_norm": 15.5, "learning_rate": 5.1891022905414546e-05, "loss": 0.6146, "step": 445 }, { "epoch": 0.26564344746162927, "grad_norm": 22.875, "learning_rate": 5.169403064588125e-05, "loss": 0.4279, "step": 450 }, { "epoch": 0.26564344746162927, "eval_loss": 0.48101305961608887, "eval_runtime": 22.0566, "eval_samples_per_second": 90.676, "eval_steps_per_second": 90.676, "step": 450 }, { "epoch": 0.26859504132231404, "grad_norm": 28.625, "learning_rate": 5.1495057894436757e-05, "loss": 0.5749, "step": 455 }, { "epoch": 0.2715466351829988, "grad_norm": 20.5, "learning_rate": 5.1294122815705773e-05, "loss": 0.4963, "step": 460 }, { "epoch": 0.2744982290436836, "grad_norm": 22.25, "learning_rate": 5.109124375345781e-05, "loss": 0.4213, "step": 465 }, { "epoch": 0.27744982290436837, "grad_norm": 18.125, "learning_rate": 5.0886439228932576e-05, "loss": 0.5002, "step": 470 }, { "epoch": 0.28040141676505315, "grad_norm": 34.0, "learning_rate": 5.067972793914911e-05, "loss": 0.5136, "step": 475 }, { "epoch": 0.28040141676505315, "eval_loss": 0.5137303471565247, "eval_runtime": 22.0698, "eval_samples_per_second": 90.621, "eval_steps_per_second": 90.621, "step": 475 }, { "epoch": 0.2833530106257379, "grad_norm": 15.0, "learning_rate": 5.047112875519892e-05, "loss": 0.5162, "step": 480 }, { "epoch": 0.28630460448642264, "grad_norm": 38.0, "learning_rate": 5.02606607205232e-05, "loss": 0.4831, "step": 485 }, { "epoch": 0.2892561983471074, "grad_norm": 26.5, "learning_rate": 5.004834304917425e-05, "loss": 0.5147, "step": 490 }, { "epoch": 0.2922077922077922, "grad_norm": 37.0, "learning_rate": 4.983419512406151e-05, "loss": 0.553, "step": 495 }, { "epoch": 0.29515938606847697, "grad_norm": 23.625, "learning_rate": 4.9618236495181936e-05, "loss": 0.3999, "step": 500 }, { "epoch": 0.29515938606847697, "eval_loss": 0.5061969757080078, "eval_runtime": 22.0382, "eval_samples_per_second": 90.752, "eval_steps_per_second": 90.752, "step": 500 }, { "epoch": 0.29811097992916175, "grad_norm": 13.75, "learning_rate": 4.9400486877835325e-05, "loss": 0.4205, "step": 505 }, { "epoch": 0.3010625737898465, "grad_norm": 22.5, "learning_rate": 4.91809661508244e-05, "loss": 0.5904, "step": 510 }, { "epoch": 0.3040141676505313, "grad_norm": 16.75, "learning_rate": 4.895969435464009e-05, "loss": 0.5749, "step": 515 }, { "epoch": 0.3069657615112161, "grad_norm": 10.8125, "learning_rate": 4.873669168963196e-05, "loss": 0.5841, "step": 520 }, { "epoch": 0.30991735537190085, "grad_norm": 14.25, "learning_rate": 4.851197851416409e-05, "loss": 0.5454, "step": 525 }, { "epoch": 0.30991735537190085, "eval_loss": 0.6045746207237244, "eval_runtime": 22.0653, "eval_samples_per_second": 90.64, "eval_steps_per_second": 90.64, "step": 525 }, { "epoch": 0.31286894923258557, "grad_norm": 12.4375, "learning_rate": 4.828557534275651e-05, "loss": 0.4654, "step": 530 }, { "epoch": 0.31582054309327035, "grad_norm": 28.125, "learning_rate": 4.8057502844212406e-05, "loss": 0.5669, "step": 535 }, { "epoch": 0.3187721369539551, "grad_norm": 17.625, "learning_rate": 4.78277818397312e-05, "loss": 0.3907, "step": 540 }, { "epoch": 0.3217237308146399, "grad_norm": 12.625, "learning_rate": 4.7596433301007775e-05, "loss": 0.508, "step": 545 }, { "epoch": 0.3246753246753247, "grad_norm": 26.75, "learning_rate": 4.736347834831789e-05, "loss": 0.408, "step": 550 }, { "epoch": 0.3246753246753247, "eval_loss": 0.48477041721343994, "eval_runtime": 22.0413, "eval_samples_per_second": 90.739, "eval_steps_per_second": 90.739, "step": 550 }, { "epoch": 0.32762691853600945, "grad_norm": 32.25, "learning_rate": 4.712893824859008e-05, "loss": 0.3498, "step": 555 }, { "epoch": 0.3305785123966942, "grad_norm": 21.25, "learning_rate": 4.6892834413464163e-05, "loss": 0.6738, "step": 560 }, { "epoch": 0.333530106257379, "grad_norm": 23.0, "learning_rate": 4.6655188397336515e-05, "loss": 0.4393, "step": 565 }, { "epoch": 0.3364817001180638, "grad_norm": 13.625, "learning_rate": 4.641602189539235e-05, "loss": 0.5021, "step": 570 }, { "epoch": 0.33943329397874855, "grad_norm": 16.5, "learning_rate": 4.617535674162509e-05, "loss": 0.5409, "step": 575 }, { "epoch": 0.33943329397874855, "eval_loss": 0.4802681505680084, "eval_runtime": 22.041, "eval_samples_per_second": 90.74, "eval_steps_per_second": 90.74, "step": 575 }, { "epoch": 0.34238488783943327, "grad_norm": 18.0, "learning_rate": 4.59332149068431e-05, "loss": 0.4619, "step": 580 }, { "epoch": 0.34533648170011805, "grad_norm": 19.25, "learning_rate": 4.5689618496664e-05, "loss": 0.5019, "step": 585 }, { "epoch": 0.3482880755608028, "grad_norm": 17.25, "learning_rate": 4.544458974949646e-05, "loss": 0.3654, "step": 590 }, { "epoch": 0.3512396694214876, "grad_norm": 25.625, "learning_rate": 4.519815103451012e-05, "loss": 0.6236, "step": 595 }, { "epoch": 0.3541912632821724, "grad_norm": 30.125, "learning_rate": 4.4950324849593455e-05, "loss": 0.4846, "step": 600 }, { "epoch": 0.3541912632821724, "eval_loss": 0.47454240918159485, "eval_runtime": 22.0763, "eval_samples_per_second": 90.595, "eval_steps_per_second": 90.595, "step": 600 }, { "epoch": 0.35714285714285715, "grad_norm": 17.75, "learning_rate": 4.470113381929984e-05, "loss": 0.5738, "step": 605 }, { "epoch": 0.3600944510035419, "grad_norm": 16.0, "learning_rate": 4.445060069278218e-05, "loss": 0.3647, "step": 610 }, { "epoch": 0.3630460448642267, "grad_norm": 12.3125, "learning_rate": 4.419874834171601e-05, "loss": 0.4772, "step": 615 }, { "epoch": 0.3659976387249115, "grad_norm": 12.4375, "learning_rate": 4.3945599758211594e-05, "loss": 0.4309, "step": 620 }, { "epoch": 0.3689492325855962, "grad_norm": 15.4375, "learning_rate": 4.369117805271482e-05, "loss": 0.3699, "step": 625 }, { "epoch": 0.3689492325855962, "eval_loss": 0.5046902298927307, "eval_runtime": 22.0518, "eval_samples_per_second": 90.696, "eval_steps_per_second": 90.696, "step": 625 }, { "epoch": 0.371900826446281, "grad_norm": 26.5, "learning_rate": 4.343550645189751e-05, "loss": 0.4449, "step": 630 }, { "epoch": 0.37485242030696575, "grad_norm": 21.125, "learning_rate": 4.317860829653692e-05, "loss": 0.2902, "step": 635 }, { "epoch": 0.3778040141676505, "grad_norm": 28.875, "learning_rate": 4.292050703938496e-05, "loss": 0.5582, "step": 640 }, { "epoch": 0.3807556080283353, "grad_norm": 13.75, "learning_rate": 4.266122624302714e-05, "loss": 0.5335, "step": 645 }, { "epoch": 0.3837072018890201, "grad_norm": 13.5, "learning_rate": 4.2400789577731485e-05, "loss": 0.3741, "step": 650 }, { "epoch": 0.3837072018890201, "eval_loss": 0.4519728720188141, "eval_runtime": 22.0605, "eval_samples_per_second": 90.66, "eval_steps_per_second": 90.66, "step": 650 }, { "epoch": 0.38665879574970485, "grad_norm": 19.5, "learning_rate": 4.213922081928763e-05, "loss": 0.5017, "step": 655 }, { "epoch": 0.38961038961038963, "grad_norm": 13.4375, "learning_rate": 4.187654384683628e-05, "loss": 0.359, "step": 660 }, { "epoch": 0.3925619834710744, "grad_norm": 11.1875, "learning_rate": 4.161278264068925e-05, "loss": 0.2609, "step": 665 }, { "epoch": 0.3955135773317591, "grad_norm": 13.0625, "learning_rate": 4.134796128014022e-05, "loss": 0.4038, "step": 670 }, { "epoch": 0.3984651711924439, "grad_norm": 17.625, "learning_rate": 4.108210394126652e-05, "loss": 0.3166, "step": 675 }, { "epoch": 0.3984651711924439, "eval_loss": 0.44906917214393616, "eval_runtime": 22.0309, "eval_samples_per_second": 90.782, "eval_steps_per_second": 90.782, "step": 675 }, { "epoch": 0.4014167650531287, "grad_norm": 32.25, "learning_rate": 4.0815234894722035e-05, "loss": 0.3372, "step": 680 }, { "epoch": 0.40436835891381345, "grad_norm": 27.25, "learning_rate": 4.05473785035215e-05, "loss": 0.3346, "step": 685 }, { "epoch": 0.40731995277449823, "grad_norm": 24.75, "learning_rate": 4.0278559220816304e-05, "loss": 0.3993, "step": 690 }, { "epoch": 0.410271546635183, "grad_norm": 20.0, "learning_rate": 4.0008801587662194e-05, "loss": 0.34, "step": 695 }, { "epoch": 0.4132231404958678, "grad_norm": 27.25, "learning_rate": 3.9738130230778796e-05, "loss": 0.4442, "step": 700 }, { "epoch": 0.4132231404958678, "eval_loss": 0.4622591435909271, "eval_runtime": 22.0728, "eval_samples_per_second": 90.609, "eval_steps_per_second": 90.609, "step": 700 }, { "epoch": 0.41617473435655256, "grad_norm": 22.625, "learning_rate": 3.946656986030142e-05, "loss": 0.3391, "step": 705 }, { "epoch": 0.41912632821723733, "grad_norm": 26.125, "learning_rate": 3.919414526752524e-05, "loss": 0.333, "step": 710 }, { "epoch": 0.42207792207792205, "grad_norm": 7.5, "learning_rate": 3.8920881322642036e-05, "loss": 0.3791, "step": 715 }, { "epoch": 0.42502951593860683, "grad_norm": 32.0, "learning_rate": 3.864680297246972e-05, "loss": 0.5102, "step": 720 }, { "epoch": 0.4279811097992916, "grad_norm": 15.5, "learning_rate": 3.8371935238174924e-05, "loss": 0.4723, "step": 725 }, { "epoch": 0.4279811097992916, "eval_loss": 0.44943633675575256, "eval_runtime": 22.0672, "eval_samples_per_second": 90.632, "eval_steps_per_second": 90.632, "step": 725 }, { "epoch": 0.4309327036599764, "grad_norm": 26.875, "learning_rate": 3.809630321298872e-05, "loss": 0.4499, "step": 730 }, { "epoch": 0.43388429752066116, "grad_norm": 14.3125, "learning_rate": 3.78199320599159e-05, "loss": 0.2574, "step": 735 }, { "epoch": 0.43683589138134593, "grad_norm": 10.0625, "learning_rate": 3.754284700943767e-05, "loss": 0.3981, "step": 740 }, { "epoch": 0.4397874852420307, "grad_norm": 20.5, "learning_rate": 3.726507335720842e-05, "loss": 0.3897, "step": 745 }, { "epoch": 0.4427390791027155, "grad_norm": 12.4375, "learning_rate": 3.6986636461746365e-05, "loss": 0.3282, "step": 750 }, { "epoch": 0.4427390791027155, "eval_loss": 0.46209147572517395, "eval_runtime": 22.0979, "eval_samples_per_second": 90.506, "eval_steps_per_second": 90.506, "step": 750 }, { "epoch": 0.44569067296340026, "grad_norm": 17.0, "learning_rate": 3.6707561742118546e-05, "loss": 0.3905, "step": 755 }, { "epoch": 0.448642266824085, "grad_norm": 16.875, "learning_rate": 3.642787467562024e-05, "loss": 0.2688, "step": 760 }, { "epoch": 0.45159386068476975, "grad_norm": 17.625, "learning_rate": 3.614760079544913e-05, "loss": 0.369, "step": 765 }, { "epoch": 0.45454545454545453, "grad_norm": 24.0, "learning_rate": 3.5866765688374296e-05, "loss": 0.4732, "step": 770 }, { "epoch": 0.4574970484061393, "grad_norm": 16.375, "learning_rate": 3.558539499240037e-05, "loss": 0.3928, "step": 775 }, { "epoch": 0.4574970484061393, "eval_loss": 0.4500948488712311, "eval_runtime": 22.071, "eval_samples_per_second": 90.617, "eval_steps_per_second": 90.617, "step": 775 }, { "epoch": 0.4604486422668241, "grad_norm": 11.375, "learning_rate": 3.530351439442696e-05, "loss": 0.498, "step": 780 }, { "epoch": 0.46340023612750886, "grad_norm": 9.125, "learning_rate": 3.502114962790366e-05, "loss": 0.3545, "step": 785 }, { "epoch": 0.46635182998819363, "grad_norm": 12.375, "learning_rate": 3.473832647048079e-05, "loss": 0.5654, "step": 790 }, { "epoch": 0.4693034238488784, "grad_norm": 10.9375, "learning_rate": 3.445507074165612e-05, "loss": 0.3995, "step": 795 }, { "epoch": 0.4722550177095632, "grad_norm": 8.375, "learning_rate": 3.41714083004177e-05, "loss": 0.4174, "step": 800 }, { "epoch": 0.4722550177095632, "eval_loss": 0.4428013265132904, "eval_runtime": 22.0412, "eval_samples_per_second": 90.739, "eval_steps_per_second": 90.739, "step": 800 }, { "epoch": 0.47520661157024796, "grad_norm": 14.3125, "learning_rate": 3.3887365042883226e-05, "loss": 0.487, "step": 805 }, { "epoch": 0.4781582054309327, "grad_norm": 12.4375, "learning_rate": 3.360296689993586e-05, "loss": 0.367, "step": 810 }, { "epoch": 0.48110979929161746, "grad_norm": 10.375, "learning_rate": 3.331823983485695e-05, "loss": 0.2972, "step": 815 }, { "epoch": 0.48406139315230223, "grad_norm": 14.0, "learning_rate": 3.303320984095584e-05, "loss": 0.417, "step": 820 }, { "epoch": 0.487012987012987, "grad_norm": 25.0, "learning_rate": 3.274790293919685e-05, "loss": 0.3814, "step": 825 }, { "epoch": 0.487012987012987, "eval_loss": 0.45124053955078125, "eval_runtime": 22.0625, "eval_samples_per_second": 90.652, "eval_steps_per_second": 90.652, "step": 825 }, { "epoch": 0.4899645808736718, "grad_norm": 19.125, "learning_rate": 3.246234517582378e-05, "loss": 0.4698, "step": 830 }, { "epoch": 0.49291617473435656, "grad_norm": 19.625, "learning_rate": 3.217656261998208e-05, "loss": 0.4826, "step": 835 }, { "epoch": 0.49586776859504134, "grad_norm": 9.25, "learning_rate": 3.189058136133898e-05, "loss": 0.4592, "step": 840 }, { "epoch": 0.4988193624557261, "grad_norm": 7.40625, "learning_rate": 3.1604427507701675e-05, "loss": 0.551, "step": 845 }, { "epoch": 0.5017709563164109, "grad_norm": 22.25, "learning_rate": 3.131812718263392e-05, "loss": 0.3845, "step": 850 }, { "epoch": 0.5017709563164109, "eval_loss": 0.42552411556243896, "eval_runtime": 22.0632, "eval_samples_per_second": 90.649, "eval_steps_per_second": 90.649, "step": 850 }, { "epoch": 0.5047225501770957, "grad_norm": 11.875, "learning_rate": 3.1031706523071115e-05, "loss": 0.5727, "step": 855 }, { "epoch": 0.5076741440377804, "grad_norm": 5.21875, "learning_rate": 3.0745191676934285e-05, "loss": 0.4759, "step": 860 }, { "epoch": 0.5106257378984652, "grad_norm": 20.75, "learning_rate": 3.0458608800742883e-05, "loss": 0.4727, "step": 865 }, { "epoch": 0.51357733175915, "grad_norm": 19.125, "learning_rate": 3.0171984057227008e-05, "loss": 0.4733, "step": 870 }, { "epoch": 0.5165289256198347, "grad_norm": 21.125, "learning_rate": 2.988534361293888e-05, "loss": 0.3789, "step": 875 }, { "epoch": 0.5165289256198347, "eval_loss": 0.41610679030418396, "eval_runtime": 22.0756, "eval_samples_per_second": 90.598, "eval_steps_per_second": 90.598, "step": 875 }, { "epoch": 0.5194805194805194, "grad_norm": 10.8125, "learning_rate": 2.959871363586411e-05, "loss": 0.4258, "step": 880 }, { "epoch": 0.5224321133412042, "grad_norm": 9.0625, "learning_rate": 2.9312120293032703e-05, "loss": 0.408, "step": 885 }, { "epoch": 0.525383707201889, "grad_norm": 29.375, "learning_rate": 2.902558974813026e-05, "loss": 0.5145, "step": 890 }, { "epoch": 0.5283353010625738, "grad_norm": 14.9375, "learning_rate": 2.873914815910944e-05, "loss": 0.4281, "step": 895 }, { "epoch": 0.5312868949232585, "grad_norm": 19.125, "learning_rate": 2.8452821675801944e-05, "loss": 0.3189, "step": 900 }, { "epoch": 0.5312868949232585, "eval_loss": 0.41652774810791016, "eval_runtime": 22.0235, "eval_samples_per_second": 90.812, "eval_steps_per_second": 90.812, "step": 900 }, { "epoch": 0.5342384887839433, "grad_norm": 19.375, "learning_rate": 2.81666364375312e-05, "loss": 0.4598, "step": 905 }, { "epoch": 0.5371900826446281, "grad_norm": 13.4375, "learning_rate": 2.7880618570726142e-05, "loss": 0.4811, "step": 910 }, { "epoch": 0.5401416765053129, "grad_norm": 23.0, "learning_rate": 2.7594794186535993e-05, "loss": 0.4931, "step": 915 }, { "epoch": 0.5430932703659976, "grad_norm": 24.125, "learning_rate": 2.7309189378446578e-05, "loss": 0.5955, "step": 920 }, { "epoch": 0.5460448642266824, "grad_norm": 19.875, "learning_rate": 2.702383021989817e-05, "loss": 0.3743, "step": 925 }, { "epoch": 0.5460448642266824, "eval_loss": 0.41767558455467224, "eval_runtime": 22.0719, "eval_samples_per_second": 90.613, "eval_steps_per_second": 90.613, "step": 925 } ], "logging_steps": 5, "max_steps": 1694, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 25, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.3649613399392256e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }