{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 100, "global_step": 4591, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0021784119376974187, "grad_norm": 583.7144775390625, "learning_rate": 1.956521739130435e-07, "loss": 8.685, "step": 10 }, { "epoch": 0.0043568238753948375, "grad_norm": 563.7721557617188, "learning_rate": 4.130434782608696e-07, "loss": 8.2663, "step": 20 }, { "epoch": 0.006535235813092255, "grad_norm": 223.9811553955078, "learning_rate": 6.304347826086957e-07, "loss": 5.5525, "step": 30 }, { "epoch": 0.008713647750789675, "grad_norm": 579.8836669921875, "learning_rate": 8.478260869565217e-07, "loss": 2.5749, "step": 40 }, { "epoch": 0.010892059688487093, "grad_norm": 234.20037841796875, "learning_rate": 1.065217391304348e-06, "loss": 1.9417, "step": 50 }, { "epoch": 0.01307047162618451, "grad_norm": 62.84681701660156, "learning_rate": 1.2826086956521742e-06, "loss": 1.3996, "step": 60 }, { "epoch": 0.01524888356388193, "grad_norm": 49.981231689453125, "learning_rate": 1.5e-06, "loss": 1.2478, "step": 70 }, { "epoch": 0.01742729550157935, "grad_norm": 80.07563018798828, "learning_rate": 1.7173913043478263e-06, "loss": 1.2043, "step": 80 }, { "epoch": 0.019605707439276768, "grad_norm": 64.25047302246094, "learning_rate": 1.934782608695652e-06, "loss": 1.1828, "step": 90 }, { "epoch": 0.021784119376974186, "grad_norm": 14.763628005981445, "learning_rate": 2.1521739130434787e-06, "loss": 1.1666, "step": 100 }, { "epoch": 0.021784119376974186, "eval_loss": 1.1604220867156982, "eval_runtime": 9.6262, "eval_samples_per_second": 518.585, "eval_steps_per_second": 2.078, "step": 100 }, { "epoch": 0.023962531314671603, "grad_norm": 52.79157257080078, "learning_rate": 2.3695652173913043e-06, "loss": 1.1818, "step": 110 }, { "epoch": 0.02614094325236902, "grad_norm": 74.7455825805664, "learning_rate": 2.586956521739131e-06, "loss": 1.1769, "step": 120 }, { "epoch": 0.028319355190066443, "grad_norm": 25.616329193115234, "learning_rate": 2.804347826086957e-06, "loss": 1.1554, "step": 130 }, { "epoch": 0.03049776712776386, "grad_norm": 47.48295211791992, "learning_rate": 3.0217391304347825e-06, "loss": 1.1475, "step": 140 }, { "epoch": 0.03267617906546128, "grad_norm": 11.702336311340332, "learning_rate": 3.239130434782609e-06, "loss": 1.1473, "step": 150 }, { "epoch": 0.0348545910031587, "grad_norm": 66.81829833984375, "learning_rate": 3.456521739130435e-06, "loss": 1.2304, "step": 160 }, { "epoch": 0.03703300294085612, "grad_norm": 33.29442596435547, "learning_rate": 3.673913043478261e-06, "loss": 1.1667, "step": 170 }, { "epoch": 0.039211414878553535, "grad_norm": 15.700839042663574, "learning_rate": 3.891304347826087e-06, "loss": 1.1553, "step": 180 }, { "epoch": 0.04138982681625095, "grad_norm": 5.018191814422607, "learning_rate": 4.108695652173914e-06, "loss": 1.1635, "step": 190 }, { "epoch": 0.04356823875394837, "grad_norm": 16.987112045288086, "learning_rate": 4.32608695652174e-06, "loss": 1.1794, "step": 200 }, { "epoch": 0.04356823875394837, "eval_loss": 1.1022545099258423, "eval_runtime": 9.5442, "eval_samples_per_second": 523.038, "eval_steps_per_second": 2.096, "step": 200 }, { "epoch": 0.04574665069164579, "grad_norm": 13.58285140991211, "learning_rate": 4.543478260869566e-06, "loss": 1.1459, "step": 210 }, { "epoch": 0.04792506262934321, "grad_norm": 17.155811309814453, "learning_rate": 4.760869565217392e-06, "loss": 1.1388, "step": 220 }, { "epoch": 0.050103474567040625, "grad_norm": 76.64653778076172, "learning_rate": 4.978260869565218e-06, "loss": 1.1497, "step": 230 }, { "epoch": 0.05228188650473804, "grad_norm": 25.240867614746094, "learning_rate": 5.195652173913044e-06, "loss": 1.1459, "step": 240 }, { "epoch": 0.05446029844243547, "grad_norm": 34.7780876159668, "learning_rate": 5.41304347826087e-06, "loss": 1.1401, "step": 250 }, { "epoch": 0.056638710380132885, "grad_norm": 54.864139556884766, "learning_rate": 5.630434782608696e-06, "loss": 1.1731, "step": 260 }, { "epoch": 0.0588171223178303, "grad_norm": 35.154762268066406, "learning_rate": 5.847826086956521e-06, "loss": 1.1628, "step": 270 }, { "epoch": 0.06099553425552772, "grad_norm": 15.862505912780762, "learning_rate": 6.065217391304349e-06, "loss": 1.1329, "step": 280 }, { "epoch": 0.06317394619322514, "grad_norm": 12.98637866973877, "learning_rate": 6.282608695652174e-06, "loss": 1.1319, "step": 290 }, { "epoch": 0.06535235813092256, "grad_norm": 25.900012969970703, "learning_rate": 6.5000000000000004e-06, "loss": 1.1542, "step": 300 }, { "epoch": 0.06535235813092256, "eval_loss": 1.096329927444458, "eval_runtime": 9.6101, "eval_samples_per_second": 519.452, "eval_steps_per_second": 2.081, "step": 300 }, { "epoch": 0.06753077006861997, "grad_norm": 13.545440673828125, "learning_rate": 6.7173913043478265e-06, "loss": 1.1598, "step": 310 }, { "epoch": 0.0697091820063174, "grad_norm": 29.305356979370117, "learning_rate": 6.934782608695653e-06, "loss": 1.1462, "step": 320 }, { "epoch": 0.07188759394401481, "grad_norm": 31.123804092407227, "learning_rate": 7.152173913043479e-06, "loss": 1.143, "step": 330 }, { "epoch": 0.07406600588171224, "grad_norm": 16.058088302612305, "learning_rate": 7.369565217391305e-06, "loss": 1.1237, "step": 340 }, { "epoch": 0.07624441781940965, "grad_norm": 53.67496109008789, "learning_rate": 7.586956521739131e-06, "loss": 1.1447, "step": 350 }, { "epoch": 0.07842282975710707, "grad_norm": 45.335391998291016, "learning_rate": 7.804347826086957e-06, "loss": 1.1486, "step": 360 }, { "epoch": 0.08060124169480448, "grad_norm": 1.5897856950759888, "learning_rate": 8.021739130434783e-06, "loss": 1.1459, "step": 370 }, { "epoch": 0.0827796536325019, "grad_norm": 21.029211044311523, "learning_rate": 8.239130434782609e-06, "loss": 1.1447, "step": 380 }, { "epoch": 0.08495806557019933, "grad_norm": 19.90410041809082, "learning_rate": 8.456521739130435e-06, "loss": 1.1186, "step": 390 }, { "epoch": 0.08713647750789674, "grad_norm": 11.852821350097656, "learning_rate": 8.673913043478261e-06, "loss": 1.1265, "step": 400 }, { "epoch": 0.08713647750789674, "eval_loss": 1.0634950399398804, "eval_runtime": 9.6061, "eval_samples_per_second": 519.668, "eval_steps_per_second": 2.082, "step": 400 }, { "epoch": 0.08931488944559417, "grad_norm": 14.51545238494873, "learning_rate": 8.891304347826087e-06, "loss": 1.1214, "step": 410 }, { "epoch": 0.09149330138329158, "grad_norm": 13.543514251708984, "learning_rate": 9.108695652173913e-06, "loss": 1.1109, "step": 420 }, { "epoch": 0.093671713320989, "grad_norm": 12.666316986083984, "learning_rate": 9.32608695652174e-06, "loss": 1.101, "step": 430 }, { "epoch": 0.09585012525868641, "grad_norm": 18.478477478027344, "learning_rate": 9.543478260869567e-06, "loss": 1.0728, "step": 440 }, { "epoch": 0.09802853719638384, "grad_norm": 21.53498077392578, "learning_rate": 9.760869565217393e-06, "loss": 1.0367, "step": 450 }, { "epoch": 0.10020694913408125, "grad_norm": 20.17803382873535, "learning_rate": 9.978260869565218e-06, "loss": 1.0203, "step": 460 }, { "epoch": 0.10238536107177867, "grad_norm": 11.546687126159668, "learning_rate": 9.999882884955554e-06, "loss": 0.9821, "step": 470 }, { "epoch": 0.10456377300947609, "grad_norm": 6.028177738189697, "learning_rate": 9.999478049870269e-06, "loss": 0.9369, "step": 480 }, { "epoch": 0.10674218494717351, "grad_norm": 28.20073699951172, "learning_rate": 9.998784072287277e-06, "loss": 0.9232, "step": 490 }, { "epoch": 0.10892059688487093, "grad_norm": 14.939946174621582, "learning_rate": 9.997800992342448e-06, "loss": 0.8963, "step": 500 }, { "epoch": 0.10892059688487093, "eval_loss": 0.8135343790054321, "eval_runtime": 9.689, "eval_samples_per_second": 515.224, "eval_steps_per_second": 2.064, "step": 500 }, { "epoch": 0.11109900882256835, "grad_norm": 14.081145286560059, "learning_rate": 9.99652886689176e-06, "loss": 0.8828, "step": 510 }, { "epoch": 0.11327742076026577, "grad_norm": 14.179576873779297, "learning_rate": 9.994967769507992e-06, "loss": 0.8611, "step": 520 }, { "epoch": 0.11545583269796318, "grad_norm": 14.495073318481445, "learning_rate": 9.993117790476494e-06, "loss": 0.8478, "step": 530 }, { "epoch": 0.1176342446356606, "grad_norm": 13.599958419799805, "learning_rate": 9.990979036789943e-06, "loss": 0.8425, "step": 540 }, { "epoch": 0.11981265657335802, "grad_norm": 11.789984703063965, "learning_rate": 9.988551632142171e-06, "loss": 0.8296, "step": 550 }, { "epoch": 0.12199106851105544, "grad_norm": 13.627972602844238, "learning_rate": 9.985835716921e-06, "loss": 0.8092, "step": 560 }, { "epoch": 0.12416948044875285, "grad_norm": 24.143400192260742, "learning_rate": 9.982831448200127e-06, "loss": 0.8161, "step": 570 }, { "epoch": 0.12634789238645028, "grad_norm": 7.413597106933594, "learning_rate": 9.979538999730047e-06, "loss": 0.8126, "step": 580 }, { "epoch": 0.1285263043241477, "grad_norm": 4.810773849487305, "learning_rate": 9.97595856192799e-06, "loss": 0.777, "step": 590 }, { "epoch": 0.13070471626184513, "grad_norm": 13.047112464904785, "learning_rate": 9.972090341866917e-06, "loss": 0.7587, "step": 600 }, { "epoch": 0.13070471626184513, "eval_loss": 0.670462429523468, "eval_runtime": 9.5938, "eval_samples_per_second": 520.335, "eval_steps_per_second": 2.085, "step": 600 }, { "epoch": 0.13288312819954254, "grad_norm": 7.8898539543151855, "learning_rate": 9.967934563263545e-06, "loss": 0.7472, "step": 610 }, { "epoch": 0.13506154013723995, "grad_norm": 12.270000457763672, "learning_rate": 9.963491466465405e-06, "loss": 0.7417, "step": 620 }, { "epoch": 0.13723995207493736, "grad_norm": 11.972742080688477, "learning_rate": 9.958761308436941e-06, "loss": 0.727, "step": 630 }, { "epoch": 0.1394183640126348, "grad_norm": 8.720413208007812, "learning_rate": 9.953744362744656e-06, "loss": 0.7179, "step": 640 }, { "epoch": 0.1415967759503322, "grad_norm": 10.330451965332031, "learning_rate": 9.948440919541277e-06, "loss": 0.7123, "step": 650 }, { "epoch": 0.14377518788802962, "grad_norm": 8.81712818145752, "learning_rate": 9.942851285548986e-06, "loss": 0.7052, "step": 660 }, { "epoch": 0.14595359982572703, "grad_norm": 12.095762252807617, "learning_rate": 9.936975784041678e-06, "loss": 0.6999, "step": 670 }, { "epoch": 0.14813201176342447, "grad_norm": 9.912598609924316, "learning_rate": 9.930814754826262e-06, "loss": 0.6873, "step": 680 }, { "epoch": 0.15031042370112188, "grad_norm": 4.136904716491699, "learning_rate": 9.924368554223007e-06, "loss": 0.6988, "step": 690 }, { "epoch": 0.1524888356388193, "grad_norm": 6.229986667633057, "learning_rate": 9.917637555044942e-06, "loss": 0.6947, "step": 700 }, { "epoch": 0.1524888356388193, "eval_loss": 0.606589138507843, "eval_runtime": 9.6122, "eval_samples_per_second": 519.339, "eval_steps_per_second": 2.081, "step": 700 }, { "epoch": 0.15466724757651673, "grad_norm": 8.056920051574707, "learning_rate": 9.910622146576285e-06, "loss": 0.6722, "step": 710 }, { "epoch": 0.15684565951421414, "grad_norm": 6.351996898651123, "learning_rate": 9.903322734549936e-06, "loss": 0.6715, "step": 720 }, { "epoch": 0.15902407145191155, "grad_norm": 4.359755992889404, "learning_rate": 9.895739741124012e-06, "loss": 0.6622, "step": 730 }, { "epoch": 0.16120248338960896, "grad_norm": 5.6291327476501465, "learning_rate": 9.887873604857424e-06, "loss": 0.6508, "step": 740 }, { "epoch": 0.1633808953273064, "grad_norm": 4.779679775238037, "learning_rate": 9.879724780684518e-06, "loss": 0.638, "step": 750 }, { "epoch": 0.1655593072650038, "grad_norm": 8.211064338684082, "learning_rate": 9.87129373988877e-06, "loss": 0.638, "step": 760 }, { "epoch": 0.16773771920270122, "grad_norm": 11.320868492126465, "learning_rate": 9.86258097007552e-06, "loss": 0.623, "step": 770 }, { "epoch": 0.16991613114039866, "grad_norm": 3.3580856323242188, "learning_rate": 9.853586975143772e-06, "loss": 0.5998, "step": 780 }, { "epoch": 0.17209454307809607, "grad_norm": 9.876014709472656, "learning_rate": 9.844312275257059e-06, "loss": 0.5885, "step": 790 }, { "epoch": 0.17427295501579348, "grad_norm": 7.176331520080566, "learning_rate": 9.834757406813353e-06, "loss": 0.5813, "step": 800 }, { "epoch": 0.17427295501579348, "eval_loss": 0.5262527465820312, "eval_runtime": 9.6798, "eval_samples_per_second": 515.712, "eval_steps_per_second": 2.066, "step": 800 }, { "epoch": 0.1764513669534909, "grad_norm": 8.124480247497559, "learning_rate": 9.824922922414048e-06, "loss": 0.5716, "step": 810 }, { "epoch": 0.17862977889118833, "grad_norm": 4.01912260055542, "learning_rate": 9.814809390831994e-06, "loss": 0.5618, "step": 820 }, { "epoch": 0.18080819082888575, "grad_norm": 8.250012397766113, "learning_rate": 9.804417396978605e-06, "loss": 0.5795, "step": 830 }, { "epoch": 0.18298660276658316, "grad_norm": 3.1755073070526123, "learning_rate": 9.793747541870034e-06, "loss": 0.5608, "step": 840 }, { "epoch": 0.18516501470428057, "grad_norm": 4.885599136352539, "learning_rate": 9.782800442592408e-06, "loss": 0.56, "step": 850 }, { "epoch": 0.187343426641978, "grad_norm": 5.620588302612305, "learning_rate": 9.771576732266147e-06, "loss": 0.5522, "step": 860 }, { "epoch": 0.18952183857967542, "grad_norm": 5.071799278259277, "learning_rate": 9.760077060009337e-06, "loss": 0.5444, "step": 870 }, { "epoch": 0.19170025051737283, "grad_norm": 3.8804523944854736, "learning_rate": 9.7483020909002e-06, "loss": 0.538, "step": 880 }, { "epoch": 0.19387866245507027, "grad_norm": 5.729625701904297, "learning_rate": 9.73625250593862e-06, "loss": 0.5356, "step": 890 }, { "epoch": 0.19605707439276768, "grad_norm": 5.823616027832031, "learning_rate": 9.723929002006761e-06, "loss": 0.5319, "step": 900 }, { "epoch": 0.19605707439276768, "eval_loss": 0.4962804913520813, "eval_runtime": 9.658, "eval_samples_per_second": 516.876, "eval_steps_per_second": 2.071, "step": 900 }, { "epoch": 0.1982354863304651, "grad_norm": 11.661290168762207, "learning_rate": 9.71133229182877e-06, "loss": 0.5237, "step": 910 }, { "epoch": 0.2004138982681625, "grad_norm": 8.839614868164062, "learning_rate": 9.698463103929542e-06, "loss": 0.538, "step": 920 }, { "epoch": 0.20259231020585994, "grad_norm": 5.292098522186279, "learning_rate": 9.685322182592604e-06, "loss": 0.5202, "step": 930 }, { "epoch": 0.20477072214355735, "grad_norm": 8.042999267578125, "learning_rate": 9.671910287817052e-06, "loss": 0.5218, "step": 940 }, { "epoch": 0.20694913408125476, "grad_norm": 5.952273368835449, "learning_rate": 9.658228195273612e-06, "loss": 0.5167, "step": 950 }, { "epoch": 0.20912754601895217, "grad_norm": 19.233489990234375, "learning_rate": 9.644276696259768e-06, "loss": 0.5153, "step": 960 }, { "epoch": 0.2113059579566496, "grad_norm": 4.609057426452637, "learning_rate": 9.630056597654002e-06, "loss": 0.5138, "step": 970 }, { "epoch": 0.21348436989434702, "grad_norm": 2.951719045639038, "learning_rate": 9.615568721869139e-06, "loss": 0.5077, "step": 980 }, { "epoch": 0.21566278183204443, "grad_norm": 6.947376251220703, "learning_rate": 9.600813906804757e-06, "loss": 0.5024, "step": 990 }, { "epoch": 0.21784119376974187, "grad_norm": 5.463963508605957, "learning_rate": 9.585793005798759e-06, "loss": 0.503, "step": 1000 }, { "epoch": 0.21784119376974187, "eval_loss": 0.49929723143577576, "eval_runtime": 9.6264, "eval_samples_per_second": 518.576, "eval_steps_per_second": 2.078, "step": 1000 }, { "epoch": 0.22001960570743928, "grad_norm": 3.5020666122436523, "learning_rate": 9.570506887577994e-06, "loss": 0.5113, "step": 1010 }, { "epoch": 0.2221980176451367, "grad_norm": 3.083266019821167, "learning_rate": 9.554956436208031e-06, "loss": 0.5043, "step": 1020 }, { "epoch": 0.2243764295828341, "grad_norm": 6.1208062171936035, "learning_rate": 9.539142551042024e-06, "loss": 0.4984, "step": 1030 }, { "epoch": 0.22655484152053154, "grad_norm": 2.8016839027404785, "learning_rate": 9.5230661466687e-06, "loss": 0.4957, "step": 1040 }, { "epoch": 0.22873325345822895, "grad_norm": 6.235162734985352, "learning_rate": 9.506728152859457e-06, "loss": 0.4936, "step": 1050 }, { "epoch": 0.23091166539592636, "grad_norm": 8.94963264465332, "learning_rate": 9.490129514514604e-06, "loss": 0.49, "step": 1060 }, { "epoch": 0.2330900773336238, "grad_norm": 5.008547306060791, "learning_rate": 9.473271191608699e-06, "loss": 0.4958, "step": 1070 }, { "epoch": 0.2352684892713212, "grad_norm": 8.403609275817871, "learning_rate": 9.456154159135043e-06, "loss": 0.4886, "step": 1080 }, { "epoch": 0.23744690120901862, "grad_norm": 12.323060035705566, "learning_rate": 9.438779407049282e-06, "loss": 0.4933, "step": 1090 }, { "epoch": 0.23962531314671603, "grad_norm": 6.123791217803955, "learning_rate": 9.421147940212152e-06, "loss": 0.4843, "step": 1100 }, { "epoch": 0.23962531314671603, "eval_loss": 0.4703376889228821, "eval_runtime": 9.6059, "eval_samples_per_second": 519.681, "eval_steps_per_second": 2.082, "step": 1100 }, { "epoch": 0.24180372508441347, "grad_norm": 3.6410789489746094, "learning_rate": 9.40326077833138e-06, "loss": 0.4797, "step": 1110 }, { "epoch": 0.24398213702211088, "grad_norm": 4.062999248504639, "learning_rate": 9.385118955902683e-06, "loss": 0.4844, "step": 1120 }, { "epoch": 0.2461605489598083, "grad_norm": 5.992898464202881, "learning_rate": 9.366723522149965e-06, "loss": 0.4801, "step": 1130 }, { "epoch": 0.2483389608975057, "grad_norm": 3.5720529556274414, "learning_rate": 9.348075540964611e-06, "loss": 0.4725, "step": 1140 }, { "epoch": 0.25051737283520314, "grad_norm": 3.079658269882202, "learning_rate": 9.329176090843985e-06, "loss": 0.4705, "step": 1150 }, { "epoch": 0.25269578477290056, "grad_norm": 4.961737632751465, "learning_rate": 9.310026264829027e-06, "loss": 0.4746, "step": 1160 }, { "epoch": 0.25487419671059797, "grad_norm": 4.140722274780273, "learning_rate": 9.290627170441055e-06, "loss": 0.465, "step": 1170 }, { "epoch": 0.2570526086482954, "grad_norm": 2.8426339626312256, "learning_rate": 9.270979929617712e-06, "loss": 0.4687, "step": 1180 }, { "epoch": 0.2592310205859928, "grad_norm": 6.142527103424072, "learning_rate": 9.251085678648072e-06, "loss": 0.4665, "step": 1190 }, { "epoch": 0.26140943252369025, "grad_norm": 6.779698371887207, "learning_rate": 9.230945568106925e-06, "loss": 0.4691, "step": 1200 }, { "epoch": 0.26140943252369025, "eval_loss": 0.4688015878200531, "eval_runtime": 9.5403, "eval_samples_per_second": 523.252, "eval_steps_per_second": 2.096, "step": 1200 }, { "epoch": 0.26358784446138767, "grad_norm": 4.152688980102539, "learning_rate": 9.210560762788238e-06, "loss": 0.4664, "step": 1210 }, { "epoch": 0.2657662563990851, "grad_norm": 6.0830817222595215, "learning_rate": 9.189932441637787e-06, "loss": 0.466, "step": 1220 }, { "epoch": 0.2679446683367825, "grad_norm": 3.1995880603790283, "learning_rate": 9.169061797684972e-06, "loss": 0.4615, "step": 1230 }, { "epoch": 0.2701230802744799, "grad_norm": 4.960101127624512, "learning_rate": 9.147950037973821e-06, "loss": 0.4645, "step": 1240 }, { "epoch": 0.2723014922121773, "grad_norm": 2.4431183338165283, "learning_rate": 9.126598383493179e-06, "loss": 0.4602, "step": 1250 }, { "epoch": 0.2744799041498747, "grad_norm": 3.671884298324585, "learning_rate": 9.105008069106093e-06, "loss": 0.457, "step": 1260 }, { "epoch": 0.2766583160875722, "grad_norm": 5.00780725479126, "learning_rate": 9.083180343478401e-06, "loss": 0.4622, "step": 1270 }, { "epoch": 0.2788367280252696, "grad_norm": 9.24307918548584, "learning_rate": 9.061116469006504e-06, "loss": 0.4601, "step": 1280 }, { "epoch": 0.281015139962967, "grad_norm": 3.016066789627075, "learning_rate": 9.038817721744362e-06, "loss": 0.4608, "step": 1290 }, { "epoch": 0.2831935519006644, "grad_norm": 4.565346717834473, "learning_rate": 9.0162853913297e-06, "loss": 0.4602, "step": 1300 }, { "epoch": 0.2831935519006644, "eval_loss": 0.45925456285476685, "eval_runtime": 9.5436, "eval_samples_per_second": 523.073, "eval_steps_per_second": 2.096, "step": 1300 }, { "epoch": 0.28537196383836183, "grad_norm": 2.7883694171905518, "learning_rate": 8.99352078090941e-06, "loss": 0.4573, "step": 1310 }, { "epoch": 0.28755037577605924, "grad_norm": 2.8337409496307373, "learning_rate": 8.970525207064194e-06, "loss": 0.4548, "step": 1320 }, { "epoch": 0.28972878771375665, "grad_norm": 7.179544448852539, "learning_rate": 8.947299999732415e-06, "loss": 0.4565, "step": 1330 }, { "epoch": 0.29190719965145406, "grad_norm": 5.800296783447266, "learning_rate": 8.923846502133186e-06, "loss": 0.4588, "step": 1340 }, { "epoch": 0.29408561158915153, "grad_norm": 5.762913227081299, "learning_rate": 8.900166070688676e-06, "loss": 0.4519, "step": 1350 }, { "epoch": 0.29626402352684894, "grad_norm": 10.822574615478516, "learning_rate": 8.87626007494567e-06, "loss": 0.456, "step": 1360 }, { "epoch": 0.29844243546454635, "grad_norm": 6.762709140777588, "learning_rate": 8.852129897496367e-06, "loss": 0.4638, "step": 1370 }, { "epoch": 0.30062084740224376, "grad_norm": 3.9109952449798584, "learning_rate": 8.827776933898397e-06, "loss": 0.4512, "step": 1380 }, { "epoch": 0.3027992593399412, "grad_norm": 4.887868881225586, "learning_rate": 8.803202592594136e-06, "loss": 0.4474, "step": 1390 }, { "epoch": 0.3049776712776386, "grad_norm": 8.851065635681152, "learning_rate": 8.778408294829232e-06, "loss": 0.4495, "step": 1400 }, { "epoch": 0.3049776712776386, "eval_loss": 0.45423299074172974, "eval_runtime": 9.5738, "eval_samples_per_second": 521.421, "eval_steps_per_second": 2.089, "step": 1400 }, { "epoch": 0.307156083215336, "grad_norm": 6.868368148803711, "learning_rate": 8.753395474570413e-06, "loss": 0.4507, "step": 1410 }, { "epoch": 0.30933449515303346, "grad_norm": 6.118240833282471, "learning_rate": 8.728165578422558e-06, "loss": 0.4476, "step": 1420 }, { "epoch": 0.3115129070907309, "grad_norm": 4.283236026763916, "learning_rate": 8.702720065545024e-06, "loss": 0.4516, "step": 1430 }, { "epoch": 0.3136913190284283, "grad_norm": 7.236689567565918, "learning_rate": 8.677060407567264e-06, "loss": 0.447, "step": 1440 }, { "epoch": 0.3158697309661257, "grad_norm": 1.7403960227966309, "learning_rate": 8.651188088503713e-06, "loss": 0.4424, "step": 1450 }, { "epoch": 0.3180481429038231, "grad_norm": 3.7462501525878906, "learning_rate": 8.625104604667965e-06, "loss": 0.4436, "step": 1460 }, { "epoch": 0.3202265548415205, "grad_norm": 2.083803653717041, "learning_rate": 8.598811464586225e-06, "loss": 0.4404, "step": 1470 }, { "epoch": 0.3224049667792179, "grad_norm": 3.62467360496521, "learning_rate": 8.572310188910075e-06, "loss": 0.4414, "step": 1480 }, { "epoch": 0.3245833787169154, "grad_norm": 3.4055850505828857, "learning_rate": 8.545602310328522e-06, "loss": 0.4489, "step": 1490 }, { "epoch": 0.3267617906546128, "grad_norm": 3.6146199703216553, "learning_rate": 8.518689373479354e-06, "loss": 0.4435, "step": 1500 }, { "epoch": 0.3267617906546128, "eval_loss": 0.4498658776283264, "eval_runtime": 9.6498, "eval_samples_per_second": 517.318, "eval_steps_per_second": 2.073, "step": 1500 }, { "epoch": 0.3289402025923102, "grad_norm": 4.079049587249756, "learning_rate": 8.491572934859808e-06, "loss": 0.4423, "step": 1510 }, { "epoch": 0.3311186145300076, "grad_norm": 4.84357213973999, "learning_rate": 8.464254562736556e-06, "loss": 0.4451, "step": 1520 }, { "epoch": 0.33329702646770504, "grad_norm": 2.70137619972229, "learning_rate": 8.436735837054998e-06, "loss": 0.4429, "step": 1530 }, { "epoch": 0.33547543840540245, "grad_norm": 3.2318954467773438, "learning_rate": 8.409018349347888e-06, "loss": 0.4394, "step": 1540 }, { "epoch": 0.33765385034309986, "grad_norm": 3.6141505241394043, "learning_rate": 8.381103702643295e-06, "loss": 0.4406, "step": 1550 }, { "epoch": 0.3398322622807973, "grad_norm": 4.4054484367370605, "learning_rate": 8.352993511371878e-06, "loss": 0.4397, "step": 1560 }, { "epoch": 0.34201067421849474, "grad_norm": 3.0935497283935547, "learning_rate": 8.324689401273538e-06, "loss": 0.4349, "step": 1570 }, { "epoch": 0.34418908615619215, "grad_norm": 2.5806384086608887, "learning_rate": 8.296193009303369e-06, "loss": 0.4366, "step": 1580 }, { "epoch": 0.34636749809388956, "grad_norm": 3.518735647201538, "learning_rate": 8.267505983537003e-06, "loss": 0.4356, "step": 1590 }, { "epoch": 0.34854591003158697, "grad_norm": 4.760187149047852, "learning_rate": 8.238629983075296e-06, "loss": 0.4351, "step": 1600 }, { "epoch": 0.34854591003158697, "eval_loss": 0.44457298517227173, "eval_runtime": 9.6018, "eval_samples_per_second": 519.901, "eval_steps_per_second": 2.083, "step": 1600 }, { "epoch": 0.3507243219692844, "grad_norm": 4.355133533477783, "learning_rate": 8.209566677948358e-06, "loss": 0.4352, "step": 1610 }, { "epoch": 0.3529027339069818, "grad_norm": 3.173840045928955, "learning_rate": 8.18031774901898e-06, "loss": 0.4359, "step": 1620 }, { "epoch": 0.3550811458446792, "grad_norm": 1.9210759401321411, "learning_rate": 8.150884887885424e-06, "loss": 0.4341, "step": 1630 }, { "epoch": 0.35725955778237667, "grad_norm": 2.729641914367676, "learning_rate": 8.121269796783585e-06, "loss": 0.4326, "step": 1640 }, { "epoch": 0.3594379697200741, "grad_norm": 2.5791547298431396, "learning_rate": 8.091474188488535e-06, "loss": 0.4313, "step": 1650 }, { "epoch": 0.3616163816577715, "grad_norm": 1.8370323181152344, "learning_rate": 8.061499786215488e-06, "loss": 0.4328, "step": 1660 }, { "epoch": 0.3637947935954689, "grad_norm": 3.280975818634033, "learning_rate": 8.031348323520113e-06, "loss": 0.4333, "step": 1670 }, { "epoch": 0.3659732055331663, "grad_norm": 3.277691125869751, "learning_rate": 8.001021544198295e-06, "loss": 0.4365, "step": 1680 }, { "epoch": 0.3681516174708637, "grad_norm": 1.8916176557540894, "learning_rate": 7.970521202185269e-06, "loss": 0.4311, "step": 1690 }, { "epoch": 0.37033002940856113, "grad_norm": 5.67519998550415, "learning_rate": 7.939849061454192e-06, "loss": 0.4335, "step": 1700 }, { "epoch": 0.37033002940856113, "eval_loss": 0.44085103273391724, "eval_runtime": 9.5393, "eval_samples_per_second": 523.31, "eval_steps_per_second": 2.097, "step": 1700 }, { "epoch": 0.3725084413462586, "grad_norm": 4.427841663360596, "learning_rate": 7.909006895914119e-06, "loss": 0.4328, "step": 1710 }, { "epoch": 0.374686853283956, "grad_norm": 2.4348649978637695, "learning_rate": 7.877996489307409e-06, "loss": 0.4273, "step": 1720 }, { "epoch": 0.3768652652216534, "grad_norm": 3.5188827514648438, "learning_rate": 7.846819635106569e-06, "loss": 0.4291, "step": 1730 }, { "epoch": 0.37904367715935083, "grad_norm": 4.911815643310547, "learning_rate": 7.81547813641052e-06, "loss": 0.4292, "step": 1740 }, { "epoch": 0.38122208909704824, "grad_norm": 5.043581962585449, "learning_rate": 7.783973805840329e-06, "loss": 0.4317, "step": 1750 }, { "epoch": 0.38340050103474566, "grad_norm": 5.273864269256592, "learning_rate": 7.75230846543436e-06, "loss": 0.4299, "step": 1760 }, { "epoch": 0.38557891297244307, "grad_norm": 2.7299153804779053, "learning_rate": 7.720483946542913e-06, "loss": 0.4255, "step": 1770 }, { "epoch": 0.38775732491014053, "grad_norm": 1.7716805934906006, "learning_rate": 7.688502089722302e-06, "loss": 0.4275, "step": 1780 }, { "epoch": 0.38993573684783794, "grad_norm": 3.8951661586761475, "learning_rate": 7.656364744628402e-06, "loss": 0.4247, "step": 1790 }, { "epoch": 0.39211414878553535, "grad_norm": 4.414074420928955, "learning_rate": 7.624073769909684e-06, "loss": 0.4259, "step": 1800 }, { "epoch": 0.39211414878553535, "eval_loss": 0.438446044921875, "eval_runtime": 9.5359, "eval_samples_per_second": 523.494, "eval_steps_per_second": 2.097, "step": 1800 }, { "epoch": 0.39429256072323277, "grad_norm": 2.8621246814727783, "learning_rate": 7.5916310330997175e-06, "loss": 0.426, "step": 1810 }, { "epoch": 0.3964709726609302, "grad_norm": 2.847648859024048, "learning_rate": 7.559038410509161e-06, "loss": 0.4263, "step": 1820 }, { "epoch": 0.3986493845986276, "grad_norm": 3.055957317352295, "learning_rate": 7.526297787117247e-06, "loss": 0.4249, "step": 1830 }, { "epoch": 0.400827796536325, "grad_norm": 5.001469612121582, "learning_rate": 7.493411056462768e-06, "loss": 0.4236, "step": 1840 }, { "epoch": 0.40300620847402246, "grad_norm": 1.1893895864486694, "learning_rate": 7.460380120534563e-06, "loss": 0.4238, "step": 1850 }, { "epoch": 0.4051846204117199, "grad_norm": 10.078166007995605, "learning_rate": 7.427206889661511e-06, "loss": 0.429, "step": 1860 }, { "epoch": 0.4073630323494173, "grad_norm": 2.957148551940918, "learning_rate": 7.393893282402057e-06, "loss": 0.4287, "step": 1870 }, { "epoch": 0.4095414442871147, "grad_norm": 2.5597410202026367, "learning_rate": 7.360441225433252e-06, "loss": 0.4264, "step": 1880 }, { "epoch": 0.4117198562248121, "grad_norm": 3.9550392627716064, "learning_rate": 7.326852653439317e-06, "loss": 0.4231, "step": 1890 }, { "epoch": 0.4138982681625095, "grad_norm": 3.9910600185394287, "learning_rate": 7.293129508999758e-06, "loss": 0.4254, "step": 1900 }, { "epoch": 0.4138982681625095, "eval_loss": 0.4347061216831207, "eval_runtime": 9.6662, "eval_samples_per_second": 516.439, "eval_steps_per_second": 2.069, "step": 1900 }, { "epoch": 0.41607668010020693, "grad_norm": 1.821865439414978, "learning_rate": 7.259273742477017e-06, "loss": 0.4201, "step": 1910 }, { "epoch": 0.41825509203790434, "grad_norm": 2.225336790084839, "learning_rate": 7.2252873119036684e-06, "loss": 0.4254, "step": 1920 }, { "epoch": 0.4204335039756018, "grad_norm": 3.4500694274902344, "learning_rate": 7.1911721828691904e-06, "loss": 0.4219, "step": 1930 }, { "epoch": 0.4226119159132992, "grad_norm": 2.226417064666748, "learning_rate": 7.156930328406268e-06, "loss": 0.4235, "step": 1940 }, { "epoch": 0.42479032785099663, "grad_norm": 4.276403427124023, "learning_rate": 7.122563728876698e-06, "loss": 0.4216, "step": 1950 }, { "epoch": 0.42696873978869404, "grad_norm": 4.915441513061523, "learning_rate": 7.088074371856847e-06, "loss": 0.4286, "step": 1960 }, { "epoch": 0.42914715172639145, "grad_norm": 4.776780605316162, "learning_rate": 7.053464252022708e-06, "loss": 0.4246, "step": 1970 }, { "epoch": 0.43132556366408886, "grad_norm": 2.5758256912231445, "learning_rate": 7.018735371034531e-06, "loss": 0.42, "step": 1980 }, { "epoch": 0.4335039756017863, "grad_norm": 2.303910255432129, "learning_rate": 6.9838897374210665e-06, "loss": 0.4183, "step": 1990 }, { "epoch": 0.43568238753948374, "grad_norm": 9.267624855041504, "learning_rate": 6.948929366463397e-06, "loss": 0.4193, "step": 2000 }, { "epoch": 0.43568238753948374, "eval_loss": 0.4357667863368988, "eval_runtime": 9.586, "eval_samples_per_second": 520.759, "eval_steps_per_second": 2.086, "step": 2000 }, { "epoch": 0.43786079947718115, "grad_norm": 2.5708439350128174, "learning_rate": 6.913856280078387e-06, "loss": 0.4255, "step": 2010 }, { "epoch": 0.44003921141487856, "grad_norm": 2.246631622314453, "learning_rate": 6.878672506701742e-06, "loss": 0.4179, "step": 2020 }, { "epoch": 0.442217623352576, "grad_norm": 2.604588270187378, "learning_rate": 6.843380081170704e-06, "loss": 0.4193, "step": 2030 }, { "epoch": 0.4443960352902734, "grad_norm": 2.7791225910186768, "learning_rate": 6.807981044606359e-06, "loss": 0.4199, "step": 2040 }, { "epoch": 0.4465744472279708, "grad_norm": 4.1223344802856445, "learning_rate": 6.772477444295592e-06, "loss": 0.4198, "step": 2050 }, { "epoch": 0.4487528591656682, "grad_norm": 1.6104360818862915, "learning_rate": 6.736871333572686e-06, "loss": 0.4166, "step": 2060 }, { "epoch": 0.45093127110336567, "grad_norm": 6.032946586608887, "learning_rate": 6.701164771700564e-06, "loss": 0.4196, "step": 2070 }, { "epoch": 0.4531096830410631, "grad_norm": 3.254844903945923, "learning_rate": 6.665359823751696e-06, "loss": 0.4205, "step": 2080 }, { "epoch": 0.4552880949787605, "grad_norm": 2.6492910385131836, "learning_rate": 6.629458560488664e-06, "loss": 0.4185, "step": 2090 }, { "epoch": 0.4574665069164579, "grad_norm": 7.052666664123535, "learning_rate": 6.593463058244407e-06, "loss": 0.4164, "step": 2100 }, { "epoch": 0.4574665069164579, "eval_loss": 0.4329228401184082, "eval_runtime": 9.6522, "eval_samples_per_second": 517.188, "eval_steps_per_second": 2.072, "step": 2100 }, { "epoch": 0.4596449188541553, "grad_norm": 3.9242568016052246, "learning_rate": 6.557375398802124e-06, "loss": 0.4188, "step": 2110 }, { "epoch": 0.4618233307918527, "grad_norm": 2.833991289138794, "learning_rate": 6.521197669274885e-06, "loss": 0.4166, "step": 2120 }, { "epoch": 0.46400174272955014, "grad_norm": 2.6651554107666016, "learning_rate": 6.484931961984927e-06, "loss": 0.4164, "step": 2130 }, { "epoch": 0.4661801546672476, "grad_norm": 1.856614112854004, "learning_rate": 6.448580374342632e-06, "loss": 0.4159, "step": 2140 }, { "epoch": 0.468358566604945, "grad_norm": 2.084019660949707, "learning_rate": 6.412145008725239e-06, "loss": 0.4155, "step": 2150 }, { "epoch": 0.4705369785426424, "grad_norm": 2.50307297706604, "learning_rate": 6.375627972355245e-06, "loss": 0.4155, "step": 2160 }, { "epoch": 0.47271539048033984, "grad_norm": 3.0096487998962402, "learning_rate": 6.339031377178536e-06, "loss": 0.4231, "step": 2170 }, { "epoch": 0.47489380241803725, "grad_norm": 2.756303071975708, "learning_rate": 6.302357339742245e-06, "loss": 0.418, "step": 2180 }, { "epoch": 0.47707221435573466, "grad_norm": 3.530158042907715, "learning_rate": 6.265607981072346e-06, "loss": 0.4165, "step": 2190 }, { "epoch": 0.47925062629343207, "grad_norm": 4.0776519775390625, "learning_rate": 6.228785426550976e-06, "loss": 0.4142, "step": 2200 }, { "epoch": 0.47925062629343207, "eval_loss": 0.4327417314052582, "eval_runtime": 9.6888, "eval_samples_per_second": 515.234, "eval_steps_per_second": 2.064, "step": 2200 }, { "epoch": 0.4814290382311295, "grad_norm": 2.6378114223480225, "learning_rate": 6.191891805793523e-06, "loss": 0.4127, "step": 2210 }, { "epoch": 0.48360745016882695, "grad_norm": 2.1185641288757324, "learning_rate": 6.1549292525254585e-06, "loss": 0.4145, "step": 2220 }, { "epoch": 0.48578586210652436, "grad_norm": 5.306119918823242, "learning_rate": 6.117899904458935e-06, "loss": 0.4153, "step": 2230 }, { "epoch": 0.48796427404422177, "grad_norm": 2.197924852371216, "learning_rate": 6.080805903169145e-06, "loss": 0.4156, "step": 2240 }, { "epoch": 0.4901426859819192, "grad_norm": 2.9446542263031006, "learning_rate": 6.043649393970478e-06, "loss": 0.414, "step": 2250 }, { "epoch": 0.4923210979196166, "grad_norm": 7.410131931304932, "learning_rate": 6.006432525792435e-06, "loss": 0.4185, "step": 2260 }, { "epoch": 0.494499509857314, "grad_norm": 3.202277898788452, "learning_rate": 5.9691574510553505e-06, "loss": 0.4155, "step": 2270 }, { "epoch": 0.4966779217950114, "grad_norm": 2.3827662467956543, "learning_rate": 5.931826325545912e-06, "loss": 0.412, "step": 2280 }, { "epoch": 0.4988563337327089, "grad_norm": 3.133537769317627, "learning_rate": 5.894441308292471e-06, "loss": 0.413, "step": 2290 }, { "epoch": 0.5010347456704063, "grad_norm": 3.3716533184051514, "learning_rate": 5.85700456144019e-06, "loss": 0.4119, "step": 2300 }, { "epoch": 0.5010347456704063, "eval_loss": 0.42866024374961853, "eval_runtime": 9.6451, "eval_samples_per_second": 517.568, "eval_steps_per_second": 2.074, "step": 2300 }, { "epoch": 0.5032131576081037, "grad_norm": 3.1281096935272217, "learning_rate": 5.819518250125986e-06, "loss": 0.4141, "step": 2310 }, { "epoch": 0.5053915695458011, "grad_norm": 2.40556263923645, "learning_rate": 5.781984542353313e-06, "loss": 0.412, "step": 2320 }, { "epoch": 0.5075699814834985, "grad_norm": 2.3655827045440674, "learning_rate": 5.744405608866775e-06, "loss": 0.4112, "step": 2330 }, { "epoch": 0.5097483934211959, "grad_norm": 3.645313024520874, "learning_rate": 5.7067836230265915e-06, "loss": 0.4102, "step": 2340 }, { "epoch": 0.5119268053588933, "grad_norm": 1.9526983499526978, "learning_rate": 5.669120760682888e-06, "loss": 0.414, "step": 2350 }, { "epoch": 0.5141052172965908, "grad_norm": 3.076853036880493, "learning_rate": 5.631419200049867e-06, "loss": 0.4122, "step": 2360 }, { "epoch": 0.5162836292342882, "grad_norm": 2.6010568141937256, "learning_rate": 5.593681121579828e-06, "loss": 0.4118, "step": 2370 }, { "epoch": 0.5184620411719856, "grad_norm": 1.341263771057129, "learning_rate": 5.555908707837061e-06, "loss": 0.4108, "step": 2380 }, { "epoch": 0.520640453109683, "grad_norm": 2.0360755920410156, "learning_rate": 5.518104143371623e-06, "loss": 0.4132, "step": 2390 }, { "epoch": 0.5228188650473805, "grad_norm": 1.9817912578582764, "learning_rate": 5.4802696145929925e-06, "loss": 0.4109, "step": 2400 }, { "epoch": 0.5228188650473805, "eval_loss": 0.42882996797561646, "eval_runtime": 9.7082, "eval_samples_per_second": 514.203, "eval_steps_per_second": 2.06, "step": 2400 }, { "epoch": 0.5249972769850779, "grad_norm": 1.8024429082870483, "learning_rate": 5.44240730964362e-06, "loss": 0.4126, "step": 2410 }, { "epoch": 0.5271756889227753, "grad_norm": 1.9342011213302612, "learning_rate": 5.404519418272376e-06, "loss": 0.4135, "step": 2420 }, { "epoch": 0.5293541008604727, "grad_norm": 1.8404808044433594, "learning_rate": 5.366608131707913e-06, "loss": 0.4089, "step": 2430 }, { "epoch": 0.5315325127981702, "grad_norm": 1.6226404905319214, "learning_rate": 5.328675642531934e-06, "loss": 0.4097, "step": 2440 }, { "epoch": 0.5337109247358676, "grad_norm": 3.6844136714935303, "learning_rate": 5.290724144552379e-06, "loss": 0.4098, "step": 2450 }, { "epoch": 0.535889336673565, "grad_norm": 4.781176567077637, "learning_rate": 5.252755832676562e-06, "loss": 0.4113, "step": 2460 }, { "epoch": 0.5380677486112624, "grad_norm": 3.025144338607788, "learning_rate": 5.214772902784218e-06, "loss": 0.4084, "step": 2470 }, { "epoch": 0.5402461605489598, "grad_norm": 1.8959859609603882, "learning_rate": 5.176777551600504e-06, "loss": 0.4085, "step": 2480 }, { "epoch": 0.5424245724866572, "grad_norm": 1.4723670482635498, "learning_rate": 5.13877197656896e-06, "loss": 0.4089, "step": 2490 }, { "epoch": 0.5446029844243546, "grad_norm": 1.6879379749298096, "learning_rate": 5.100758375724418e-06, "loss": 0.4117, "step": 2500 }, { "epoch": 0.5446029844243546, "eval_loss": 0.43057578802108765, "eval_runtime": 9.6671, "eval_samples_per_second": 516.389, "eval_steps_per_second": 2.069, "step": 2500 }, { "epoch": 0.546781396362052, "grad_norm": 2.652782440185547, "learning_rate": 5.062738947565879e-06, "loss": 0.4075, "step": 2510 }, { "epoch": 0.5489598082997494, "grad_norm": 3.549682140350342, "learning_rate": 5.0247158909293605e-06, "loss": 0.4061, "step": 2520 }, { "epoch": 0.5511382202374469, "grad_norm": 4.6838603019714355, "learning_rate": 4.986691404860738e-06, "loss": 0.4106, "step": 2530 }, { "epoch": 0.5533166321751444, "grad_norm": 2.4103927612304688, "learning_rate": 4.948667688488552e-06, "loss": 0.4093, "step": 2540 }, { "epoch": 0.5554950441128418, "grad_norm": 3.744180679321289, "learning_rate": 4.910646940896828e-06, "loss": 0.4082, "step": 2550 }, { "epoch": 0.5576734560505392, "grad_norm": 3.863417863845825, "learning_rate": 4.872631360997899e-06, "loss": 0.4061, "step": 2560 }, { "epoch": 0.5598518679882366, "grad_norm": 3.8889987468719482, "learning_rate": 4.83462314740522e-06, "loss": 0.4075, "step": 2570 }, { "epoch": 0.562030279925934, "grad_norm": 3.1181349754333496, "learning_rate": 4.796624498306224e-06, "loss": 0.4113, "step": 2580 }, { "epoch": 0.5642086918636314, "grad_norm": 2.6836493015289307, "learning_rate": 4.758637611335183e-06, "loss": 0.4088, "step": 2590 }, { "epoch": 0.5663871038013288, "grad_norm": 2.120168924331665, "learning_rate": 4.720664683446111e-06, "loss": 0.4073, "step": 2600 }, { "epoch": 0.5663871038013288, "eval_loss": 0.4350495934486389, "eval_runtime": 9.6231, "eval_samples_per_second": 518.752, "eval_steps_per_second": 2.078, "step": 2600 }, { "epoch": 0.5685655157390263, "grad_norm": 5.942187786102295, "learning_rate": 4.682707910785712e-06, "loss": 0.4149, "step": 2610 }, { "epoch": 0.5707439276767237, "grad_norm": 2.514582395553589, "learning_rate": 4.644769488566351e-06, "loss": 0.4083, "step": 2620 }, { "epoch": 0.5729223396144211, "grad_norm": 3.884115219116211, "learning_rate": 4.606851610939108e-06, "loss": 0.4052, "step": 2630 }, { "epoch": 0.5751007515521185, "grad_norm": 4.213008403778076, "learning_rate": 4.5689564708668735e-06, "loss": 0.4076, "step": 2640 }, { "epoch": 0.5772791634898159, "grad_norm": 2.6122772693634033, "learning_rate": 4.531086259997528e-06, "loss": 0.4088, "step": 2650 }, { "epoch": 0.5794575754275133, "grad_norm": 2.8703975677490234, "learning_rate": 4.493243168537178e-06, "loss": 0.4081, "step": 2660 }, { "epoch": 0.5816359873652107, "grad_norm": 2.920668840408325, "learning_rate": 4.455429385123489e-06, "loss": 0.4074, "step": 2670 }, { "epoch": 0.5838143993029081, "grad_norm": 2.7703404426574707, "learning_rate": 4.417647096699118e-06, "loss": 0.4037, "step": 2680 }, { "epoch": 0.5859928112406056, "grad_norm": 3.3332200050354004, "learning_rate": 4.379898488385215e-06, "loss": 0.4063, "step": 2690 }, { "epoch": 0.5881712231783031, "grad_norm": 1.8352432250976562, "learning_rate": 4.342185743355062e-06, "loss": 0.4062, "step": 2700 }, { "epoch": 0.5881712231783031, "eval_loss": 0.4279891550540924, "eval_runtime": 9.6691, "eval_samples_per_second": 516.286, "eval_steps_per_second": 2.068, "step": 2700 }, { "epoch": 0.5903496351160005, "grad_norm": 1.7296792268753052, "learning_rate": 4.304511042707802e-06, "loss": 0.4064, "step": 2710 }, { "epoch": 0.5925280470536979, "grad_norm": 3.613008499145508, "learning_rate": 4.266876565342298e-06, "loss": 0.4075, "step": 2720 }, { "epoch": 0.5947064589913953, "grad_norm": 3.5010597705841064, "learning_rate": 4.229284487831122e-06, "loss": 0.4091, "step": 2730 }, { "epoch": 0.5968848709290927, "grad_norm": 2.1212668418884277, "learning_rate": 4.191736984294667e-06, "loss": 0.4075, "step": 2740 }, { "epoch": 0.5990632828667901, "grad_norm": 2.979283571243286, "learning_rate": 4.154236226275411e-06, "loss": 0.404, "step": 2750 }, { "epoch": 0.6012416948044875, "grad_norm": 1.792733907699585, "learning_rate": 4.1167843826123265e-06, "loss": 0.4042, "step": 2760 }, { "epoch": 0.6034201067421849, "grad_norm": 3.041388750076294, "learning_rate": 4.079383619315447e-06, "loss": 0.4061, "step": 2770 }, { "epoch": 0.6055985186798823, "grad_norm": 2.6816020011901855, "learning_rate": 4.0420360994406e-06, "loss": 0.4071, "step": 2780 }, { "epoch": 0.6077769306175798, "grad_norm": 2.446981430053711, "learning_rate": 4.004743982964298e-06, "loss": 0.402, "step": 2790 }, { "epoch": 0.6099553425552772, "grad_norm": 2.502835273742676, "learning_rate": 3.96750942665883e-06, "loss": 0.4037, "step": 2800 }, { "epoch": 0.6099553425552772, "eval_loss": 0.4277077615261078, "eval_runtime": 9.743, "eval_samples_per_second": 512.368, "eval_steps_per_second": 2.053, "step": 2800 }, { "epoch": 0.6121337544929746, "grad_norm": 1.2970865964889526, "learning_rate": 3.930334583967514e-06, "loss": 0.404, "step": 2810 }, { "epoch": 0.614312166430672, "grad_norm": 3.3795993328094482, "learning_rate": 3.8932216048801615e-06, "loss": 0.4065, "step": 2820 }, { "epoch": 0.6164905783683695, "grad_norm": 2.872642755508423, "learning_rate": 3.856172635808732e-06, "loss": 0.4059, "step": 2830 }, { "epoch": 0.6186689903060669, "grad_norm": 2.930790662765503, "learning_rate": 3.8191898194631926e-06, "loss": 0.4053, "step": 2840 }, { "epoch": 0.6208474022437643, "grad_norm": 3.7308971881866455, "learning_rate": 3.782275294727601e-06, "loss": 0.4055, "step": 2850 }, { "epoch": 0.6230258141814617, "grad_norm": 3.462425470352173, "learning_rate": 3.745431196536398e-06, "loss": 0.4038, "step": 2860 }, { "epoch": 0.6252042261191592, "grad_norm": 2.3263561725616455, "learning_rate": 3.7086596557509396e-06, "loss": 0.4045, "step": 2870 }, { "epoch": 0.6273826380568566, "grad_norm": 1.485323190689087, "learning_rate": 3.6719627990362594e-06, "loss": 0.4025, "step": 2880 }, { "epoch": 0.629561049994554, "grad_norm": 3.1288955211639404, "learning_rate": 3.6353427487380687e-06, "loss": 0.402, "step": 2890 }, { "epoch": 0.6317394619322514, "grad_norm": 2.5085573196411133, "learning_rate": 3.598801622760021e-06, "loss": 0.4054, "step": 2900 }, { "epoch": 0.6317394619322514, "eval_loss": 0.42717447876930237, "eval_runtime": 9.6573, "eval_samples_per_second": 516.914, "eval_steps_per_second": 2.071, "step": 2900 }, { "epoch": 0.6339178738699488, "grad_norm": 4.0328049659729, "learning_rate": 3.5623415344412125e-06, "loss": 0.4006, "step": 2910 }, { "epoch": 0.6360962858076462, "grad_norm": 2.123347759246826, "learning_rate": 3.525964592433966e-06, "loss": 0.4046, "step": 2920 }, { "epoch": 0.6382746977453436, "grad_norm": 3.348815441131592, "learning_rate": 3.489672900581876e-06, "loss": 0.4025, "step": 2930 }, { "epoch": 0.640453109683041, "grad_norm": 4.057186603546143, "learning_rate": 3.453468557798133e-06, "loss": 0.4037, "step": 2940 }, { "epoch": 0.6426315216207384, "grad_norm": 3.547410249710083, "learning_rate": 3.417353657944137e-06, "loss": 0.403, "step": 2950 }, { "epoch": 0.6448099335584359, "grad_norm": 3.1567604541778564, "learning_rate": 3.3813302897083955e-06, "loss": 0.4027, "step": 2960 }, { "epoch": 0.6469883454961333, "grad_norm": 4.546987533569336, "learning_rate": 3.345400536485727e-06, "loss": 0.4029, "step": 2970 }, { "epoch": 0.6491667574338308, "grad_norm": 0.840340793132782, "learning_rate": 3.3095664762567687e-06, "loss": 0.4018, "step": 2980 }, { "epoch": 0.6513451693715282, "grad_norm": 1.2068535089492798, "learning_rate": 3.2738301814678015e-06, "loss": 0.4025, "step": 2990 }, { "epoch": 0.6535235813092256, "grad_norm": 2.895763635635376, "learning_rate": 3.238193718910883e-06, "loss": 0.4031, "step": 3000 }, { "epoch": 0.6535235813092256, "eval_loss": 0.42841657996177673, "eval_runtime": 9.6401, "eval_samples_per_second": 517.835, "eval_steps_per_second": 2.075, "step": 3000 }, { "epoch": 0.655701993246923, "grad_norm": 4.0693678855896, "learning_rate": 3.2026591496043226e-06, "loss": 0.4022, "step": 3010 }, { "epoch": 0.6578804051846204, "grad_norm": 3.907064914703369, "learning_rate": 3.167228528673478e-06, "loss": 0.4018, "step": 3020 }, { "epoch": 0.6600588171223178, "grad_norm": 6.200908660888672, "learning_rate": 3.1319039052319033e-06, "loss": 0.4035, "step": 3030 }, { "epoch": 0.6622372290600153, "grad_norm": 3.6325767040252686, "learning_rate": 3.0966873222628368e-06, "loss": 0.404, "step": 3040 }, { "epoch": 0.6644156409977127, "grad_norm": 2.363156795501709, "learning_rate": 3.0615808165010453e-06, "loss": 0.4034, "step": 3050 }, { "epoch": 0.6665940529354101, "grad_norm": 3.558595895767212, "learning_rate": 3.0265864183150306e-06, "loss": 0.4009, "step": 3060 }, { "epoch": 0.6687724648731075, "grad_norm": 3.482257604598999, "learning_rate": 2.991706151589606e-06, "loss": 0.4024, "step": 3070 }, { "epoch": 0.6709508768108049, "grad_norm": 2.492945909500122, "learning_rate": 2.956942033608843e-06, "loss": 0.4016, "step": 3080 }, { "epoch": 0.6731292887485023, "grad_norm": 2.730677366256714, "learning_rate": 2.9222960749394073e-06, "loss": 0.4028, "step": 3090 }, { "epoch": 0.6753077006861997, "grad_norm": 1.6951932907104492, "learning_rate": 2.887770279314273e-06, "loss": 0.3998, "step": 3100 }, { "epoch": 0.6753077006861997, "eval_loss": 0.42821940779685974, "eval_runtime": 9.5688, "eval_samples_per_second": 521.694, "eval_steps_per_second": 2.09, "step": 3100 }, { "epoch": 0.6774861126238971, "grad_norm": 3.0309555530548096, "learning_rate": 2.853366643516841e-06, "loss": 0.4005, "step": 3110 }, { "epoch": 0.6796645245615947, "grad_norm": 3.3345096111297607, "learning_rate": 2.8190871572654543e-06, "loss": 0.4013, "step": 3120 }, { "epoch": 0.6818429364992921, "grad_norm": 1.4350861310958862, "learning_rate": 2.784933803098326e-06, "loss": 0.4007, "step": 3130 }, { "epoch": 0.6840213484369895, "grad_norm": 1.698664665222168, "learning_rate": 2.7509085562588743e-06, "loss": 0.4002, "step": 3140 }, { "epoch": 0.6861997603746869, "grad_norm": 1.3729372024536133, "learning_rate": 2.717013384581498e-06, "loss": 0.4017, "step": 3150 }, { "epoch": 0.6883781723123843, "grad_norm": 1.2046958208084106, "learning_rate": 2.683250248377752e-06, "loss": 0.4025, "step": 3160 }, { "epoch": 0.6905565842500817, "grad_norm": 2.475562810897827, "learning_rate": 2.6496211003229795e-06, "loss": 0.3992, "step": 3170 }, { "epoch": 0.6927349961877791, "grad_norm": 3.6771671772003174, "learning_rate": 2.6161278853433885e-06, "loss": 0.402, "step": 3180 }, { "epoch": 0.6949134081254765, "grad_norm": 1.5112743377685547, "learning_rate": 2.582772540503555e-06, "loss": 0.4014, "step": 3190 }, { "epoch": 0.6970918200631739, "grad_norm": 2.9181530475616455, "learning_rate": 2.5495569948944e-06, "loss": 0.4003, "step": 3200 }, { "epoch": 0.6970918200631739, "eval_loss": 0.42963868379592896, "eval_runtime": 9.6232, "eval_samples_per_second": 518.744, "eval_steps_per_second": 2.078, "step": 3200 }, { "epoch": 0.6992702320008714, "grad_norm": 3.9118452072143555, "learning_rate": 2.5164831695216217e-06, "loss": 0.4006, "step": 3210 }, { "epoch": 0.7014486439385688, "grad_norm": 3.3592333793640137, "learning_rate": 2.4835529771945938e-06, "loss": 0.4019, "step": 3220 }, { "epoch": 0.7036270558762662, "grad_norm": 3.0435197353363037, "learning_rate": 2.450768322415739e-06, "loss": 0.3993, "step": 3230 }, { "epoch": 0.7058054678139636, "grad_norm": 1.2945152521133423, "learning_rate": 2.4181311012703833e-06, "loss": 0.4025, "step": 3240 }, { "epoch": 0.707983879751661, "grad_norm": 4.202663421630859, "learning_rate": 2.3856432013170978e-06, "loss": 0.4009, "step": 3250 }, { "epoch": 0.7101622916893584, "grad_norm": 1.378665566444397, "learning_rate": 2.35330650147853e-06, "loss": 0.3978, "step": 3260 }, { "epoch": 0.7123407036270559, "grad_norm": 1.7930536270141602, "learning_rate": 2.321122871932739e-06, "loss": 0.3982, "step": 3270 }, { "epoch": 0.7145191155647533, "grad_norm": 1.7397598028182983, "learning_rate": 2.289094174005037e-06, "loss": 0.4005, "step": 3280 }, { "epoch": 0.7166975275024507, "grad_norm": 2.92988920211792, "learning_rate": 2.257222260060335e-06, "loss": 0.4007, "step": 3290 }, { "epoch": 0.7188759394401482, "grad_norm": 4.326840877532959, "learning_rate": 2.2255089733960162e-06, "loss": 0.4021, "step": 3300 }, { "epoch": 0.7188759394401482, "eval_loss": 0.42818737030029297, "eval_runtime": 9.6384, "eval_samples_per_second": 517.929, "eval_steps_per_second": 2.075, "step": 3300 }, { "epoch": 0.7210543513778456, "grad_norm": 1.8383755683898926, "learning_rate": 2.1939561481353273e-06, "loss": 0.403, "step": 3310 }, { "epoch": 0.723232763315543, "grad_norm": 4.371707439422607, "learning_rate": 2.1625656091213056e-06, "loss": 0.4023, "step": 3320 }, { "epoch": 0.7254111752532404, "grad_norm": 5.149106502532959, "learning_rate": 2.131339171811236e-06, "loss": 0.3982, "step": 3330 }, { "epoch": 0.7275895871909378, "grad_norm": 2.2125470638275146, "learning_rate": 2.100278642171662e-06, "loss": 0.3994, "step": 3340 }, { "epoch": 0.7297679991286352, "grad_norm": 2.824982166290283, "learning_rate": 2.069385816573928e-06, "loss": 0.401, "step": 3350 }, { "epoch": 0.7319464110663326, "grad_norm": 3.157496213912964, "learning_rate": 2.0386624816902978e-06, "loss": 0.3996, "step": 3360 }, { "epoch": 0.73412482300403, "grad_norm": 2.6652979850769043, "learning_rate": 2.008110414390613e-06, "loss": 0.4014, "step": 3370 }, { "epoch": 0.7363032349417274, "grad_norm": 3.388657331466675, "learning_rate": 1.977731381639541e-06, "loss": 0.3998, "step": 3380 }, { "epoch": 0.7384816468794249, "grad_norm": 3.359178066253662, "learning_rate": 1.9475271403943686e-06, "loss": 0.3994, "step": 3390 }, { "epoch": 0.7406600588171223, "grad_norm": 1.4487273693084717, "learning_rate": 1.9174994375034034e-06, "loss": 0.3982, "step": 3400 }, { "epoch": 0.7406600588171223, "eval_loss": 0.429605096578598, "eval_runtime": 9.6367, "eval_samples_per_second": 518.021, "eval_steps_per_second": 2.075, "step": 3400 }, { "epoch": 0.7428384707548198, "grad_norm": 3.843029260635376, "learning_rate": 1.887650009604935e-06, "loss": 0.4, "step": 3410 }, { "epoch": 0.7450168826925172, "grad_norm": 4.428391456604004, "learning_rate": 1.8579805830268016e-06, "loss": 0.4005, "step": 3420 }, { "epoch": 0.7471952946302146, "grad_norm": 2.102658748626709, "learning_rate": 1.8284928736865499e-06, "loss": 0.3997, "step": 3430 }, { "epoch": 0.749373706567912, "grad_norm": 2.6355855464935303, "learning_rate": 1.7991885869921928e-06, "loss": 0.3983, "step": 3440 }, { "epoch": 0.7515521185056094, "grad_norm": 3.8726813793182373, "learning_rate": 1.7700694177435807e-06, "loss": 0.4002, "step": 3450 }, { "epoch": 0.7537305304433068, "grad_norm": 1.449800729751587, "learning_rate": 1.7411370500343812e-06, "loss": 0.3986, "step": 3460 }, { "epoch": 0.7559089423810043, "grad_norm": 2.0149781703948975, "learning_rate": 1.7123931571546826e-06, "loss": 0.3958, "step": 3470 }, { "epoch": 0.7580873543187017, "grad_norm": 1.6387373208999634, "learning_rate": 1.6838394014942179e-06, "loss": 0.3997, "step": 3480 }, { "epoch": 0.7602657662563991, "grad_norm": 1.898795485496521, "learning_rate": 1.6554774344462199e-06, "loss": 0.3986, "step": 3490 }, { "epoch": 0.7624441781940965, "grad_norm": 3.0024943351745605, "learning_rate": 1.6273088963119222e-06, "loss": 0.3988, "step": 3500 }, { "epoch": 0.7624441781940965, "eval_loss": 0.4297761917114258, "eval_runtime": 9.6746, "eval_samples_per_second": 515.992, "eval_steps_per_second": 2.067, "step": 3500 }, { "epoch": 0.7646225901317939, "grad_norm": 0.9623323082923889, "learning_rate": 1.5993354162056802e-06, "loss": 0.3997, "step": 3510 }, { "epoch": 0.7668010020694913, "grad_norm": 2.5859196186065674, "learning_rate": 1.5715586119607617e-06, "loss": 0.4001, "step": 3520 }, { "epoch": 0.7689794140071887, "grad_norm": 2.1984033584594727, "learning_rate": 1.5439800900357765e-06, "loss": 0.398, "step": 3530 }, { "epoch": 0.7711578259448861, "grad_norm": 1.5113025903701782, "learning_rate": 1.516601445421768e-06, "loss": 0.4008, "step": 3540 }, { "epoch": 0.7733362378825835, "grad_norm": 1.8336790800094604, "learning_rate": 1.4894242615499689e-06, "loss": 0.3986, "step": 3550 }, { "epoch": 0.7755146498202811, "grad_norm": 2.773827075958252, "learning_rate": 1.4624501102002209e-06, "loss": 0.3984, "step": 3560 }, { "epoch": 0.7776930617579785, "grad_norm": 2.334609270095825, "learning_rate": 1.4356805514100763e-06, "loss": 0.4002, "step": 3570 }, { "epoch": 0.7798714736956759, "grad_norm": 1.920845627784729, "learning_rate": 1.4091171333845692e-06, "loss": 0.3957, "step": 3580 }, { "epoch": 0.7820498856333733, "grad_norm": 1.4427253007888794, "learning_rate": 1.3827613924066817e-06, "loss": 0.396, "step": 3590 }, { "epoch": 0.7842282975710707, "grad_norm": 1.5900869369506836, "learning_rate": 1.3566148527484863e-06, "loss": 0.3988, "step": 3600 }, { "epoch": 0.7842282975710707, "eval_loss": 0.42993584275245667, "eval_runtime": 9.6537, "eval_samples_per_second": 517.108, "eval_steps_per_second": 2.072, "step": 3600 }, { "epoch": 0.7864067095087681, "grad_norm": 0.8249953389167786, "learning_rate": 1.330679026582996e-06, "loss": 0.3996, "step": 3610 }, { "epoch": 0.7885851214464655, "grad_norm": 0.8866672515869141, "learning_rate": 1.3049554138967052e-06, "loss": 0.3967, "step": 3620 }, { "epoch": 0.7907635333841629, "grad_norm": 1.9449257850646973, "learning_rate": 1.2794455024028396e-06, "loss": 0.3978, "step": 3630 }, { "epoch": 0.7929419453218604, "grad_norm": 2.8876960277557373, "learning_rate": 1.25415076745532e-06, "loss": 0.3974, "step": 3640 }, { "epoch": 0.7951203572595578, "grad_norm": 1.4642447233200073, "learning_rate": 1.2290726719634255e-06, "loss": 0.3963, "step": 3650 }, { "epoch": 0.7972987691972552, "grad_norm": 0.9553173184394836, "learning_rate": 1.2042126663071952e-06, "loss": 0.3967, "step": 3660 }, { "epoch": 0.7994771811349526, "grad_norm": 3.112318515777588, "learning_rate": 1.1795721882535439e-06, "loss": 0.3984, "step": 3670 }, { "epoch": 0.80165559307265, "grad_norm": 3.6302273273468018, "learning_rate": 1.1551526628731074e-06, "loss": 0.3967, "step": 3680 }, { "epoch": 0.8038340050103474, "grad_norm": 1.9708794355392456, "learning_rate": 1.130955502457829e-06, "loss": 0.3972, "step": 3690 }, { "epoch": 0.8060124169480449, "grad_norm": 2.380471706390381, "learning_rate": 1.1069821064392722e-06, "loss": 0.3949, "step": 3700 }, { "epoch": 0.8060124169480449, "eval_loss": 0.4309428334236145, "eval_runtime": 9.5957, "eval_samples_per_second": 520.233, "eval_steps_per_second": 2.084, "step": 3700 }, { "epoch": 0.8081908288857423, "grad_norm": 2.012770652770996, "learning_rate": 1.083233861307697e-06, "loss": 0.3977, "step": 3710 }, { "epoch": 0.8103692408234398, "grad_norm": 1.0971107482910156, "learning_rate": 1.059712140531861e-06, "loss": 0.3984, "step": 3720 }, { "epoch": 0.8125476527611372, "grad_norm": 0.9638540744781494, "learning_rate": 1.0364183044795929e-06, "loss": 0.3956, "step": 3730 }, { "epoch": 0.8147260646988346, "grad_norm": 0.7911339402198792, "learning_rate": 1.013353700339112e-06, "loss": 0.397, "step": 3740 }, { "epoch": 0.816904476636532, "grad_norm": 0.8662415146827698, "learning_rate": 9.905196620411178e-07, "loss": 0.3965, "step": 3750 }, { "epoch": 0.8190828885742294, "grad_norm": 1.298143744468689, "learning_rate": 9.679175101816408e-07, "loss": 0.395, "step": 3760 }, { "epoch": 0.8212613005119268, "grad_norm": 0.6701905131340027, "learning_rate": 9.455485519456658e-07, "loss": 0.3947, "step": 3770 }, { "epoch": 0.8234397124496242, "grad_norm": 0.8428287506103516, "learning_rate": 9.234140810315345e-07, "loss": 0.3962, "step": 3780 }, { "epoch": 0.8256181243873216, "grad_norm": 2.1166293621063232, "learning_rate": 9.015153775761199e-07, "loss": 0.3946, "step": 3790 }, { "epoch": 0.827796536325019, "grad_norm": 3.267383575439453, "learning_rate": 8.79853708080795e-07, "loss": 0.3961, "step": 3800 }, { "epoch": 0.827796536325019, "eval_loss": 0.4297568202018738, "eval_runtime": 9.5765, "eval_samples_per_second": 521.274, "eval_steps_per_second": 2.088, "step": 3800 }, { "epoch": 0.8299749482627164, "grad_norm": 1.0721772909164429, "learning_rate": 8.584303253381848e-07, "loss": 0.3958, "step": 3810 }, { "epoch": 0.8321533602004139, "grad_norm": 1.8855377435684204, "learning_rate": 8.372464683597081e-07, "loss": 0.3953, "step": 3820 }, { "epoch": 0.8343317721381113, "grad_norm": 0.7978574633598328, "learning_rate": 8.16303362303923e-07, "loss": 0.3931, "step": 3830 }, { "epoch": 0.8365101840758087, "grad_norm": 1.466511845588684, "learning_rate": 7.956022184056689e-07, "loss": 0.3969, "step": 3840 }, { "epoch": 0.8386885960135062, "grad_norm": 1.6769752502441406, "learning_rate": 7.751442339060161e-07, "loss": 0.3958, "step": 3850 }, { "epoch": 0.8408670079512036, "grad_norm": 1.4472180604934692, "learning_rate": 7.549305919830241e-07, "loss": 0.3969, "step": 3860 }, { "epoch": 0.843045419888901, "grad_norm": 1.6235806941986084, "learning_rate": 7.349624616833123e-07, "loss": 0.3966, "step": 3870 }, { "epoch": 0.8452238318265984, "grad_norm": 0.7935024499893188, "learning_rate": 7.152409978544483e-07, "loss": 0.3941, "step": 3880 }, { "epoch": 0.8474022437642958, "grad_norm": 0.7577910423278809, "learning_rate": 6.957673410781617e-07, "loss": 0.3946, "step": 3890 }, { "epoch": 0.8495806557019933, "grad_norm": 0.8020604252815247, "learning_rate": 6.765426176043721e-07, "loss": 0.3952, "step": 3900 }, { "epoch": 0.8495806557019933, "eval_loss": 0.4307285249233246, "eval_runtime": 9.5556, "eval_samples_per_second": 522.416, "eval_steps_per_second": 2.093, "step": 3900 }, { "epoch": 0.8517590676396907, "grad_norm": 1.1028465032577515, "learning_rate": 6.57567939286059e-07, "loss": 0.396, "step": 3910 }, { "epoch": 0.8539374795773881, "grad_norm": 1.3756065368652344, "learning_rate": 6.388444035149538e-07, "loss": 0.3975, "step": 3920 }, { "epoch": 0.8561158915150855, "grad_norm": 1.218600869178772, "learning_rate": 6.203730931580798e-07, "loss": 0.3936, "step": 3930 }, { "epoch": 0.8582943034527829, "grad_norm": 0.8649505972862244, "learning_rate": 6.021550764951157e-07, "loss": 0.3947, "step": 3940 }, { "epoch": 0.8604727153904803, "grad_norm": 3.2568881511688232, "learning_rate": 5.841914071566179e-07, "loss": 0.3956, "step": 3950 }, { "epoch": 0.8626511273281777, "grad_norm": 1.6868760585784912, "learning_rate": 5.664831240630841e-07, "loss": 0.3974, "step": 3960 }, { "epoch": 0.8648295392658751, "grad_norm": 0.7339500784873962, "learning_rate": 5.490312513648666e-07, "loss": 0.395, "step": 3970 }, { "epoch": 0.8670079512035725, "grad_norm": 0.862421989440918, "learning_rate": 5.318367983829393e-07, "loss": 0.3947, "step": 3980 }, { "epoch": 0.8691863631412701, "grad_norm": 1.4295686483383179, "learning_rate": 5.149007595505285e-07, "loss": 0.3943, "step": 3990 }, { "epoch": 0.8713647750789675, "grad_norm": 1.2400861978530884, "learning_rate": 4.982241143555966e-07, "loss": 0.397, "step": 4000 }, { "epoch": 0.8713647750789675, "eval_loss": 0.43102139234542847, "eval_runtime": 9.6775, "eval_samples_per_second": 515.838, "eval_steps_per_second": 2.067, "step": 4000 }, { "epoch": 0.8735431870166649, "grad_norm": 1.7299386262893677, "learning_rate": 4.818078272841942e-07, "loss": 0.3977, "step": 4010 }, { "epoch": 0.8757215989543623, "grad_norm": 0.814769446849823, "learning_rate": 4.6565284776468376e-07, "loss": 0.3936, "step": 4020 }, { "epoch": 0.8779000108920597, "grad_norm": 1.483109712600708, "learning_rate": 4.4976011011282294e-07, "loss": 0.398, "step": 4030 }, { "epoch": 0.8800784228297571, "grad_norm": 2.0057575702667236, "learning_rate": 4.3413053347773406e-07, "loss": 0.3951, "step": 4040 }, { "epoch": 0.8822568347674545, "grad_norm": 0.9725803136825562, "learning_rate": 4.1876502178874403e-07, "loss": 0.3962, "step": 4050 }, { "epoch": 0.884435246705152, "grad_norm": 1.1756798028945923, "learning_rate": 4.036644637031045e-07, "loss": 0.3953, "step": 4060 }, { "epoch": 0.8866136586428494, "grad_norm": 1.6209012269973755, "learning_rate": 3.8882973255459975e-07, "loss": 0.3947, "step": 4070 }, { "epoch": 0.8887920705805468, "grad_norm": 1.3361040353775024, "learning_rate": 3.742616863030352e-07, "loss": 0.3952, "step": 4080 }, { "epoch": 0.8909704825182442, "grad_norm": 0.8442427515983582, "learning_rate": 3.5996116748461984e-07, "loss": 0.3943, "step": 4090 }, { "epoch": 0.8931488944559416, "grad_norm": 1.1721961498260498, "learning_rate": 3.459290031632362e-07, "loss": 0.3935, "step": 4100 }, { "epoch": 0.8931488944559416, "eval_loss": 0.4306623339653015, "eval_runtime": 9.6333, "eval_samples_per_second": 518.202, "eval_steps_per_second": 2.076, "step": 4100 }, { "epoch": 0.895327306393639, "grad_norm": 0.9018988013267517, "learning_rate": 3.321660048826092e-07, "loss": 0.3938, "step": 4110 }, { "epoch": 0.8975057183313364, "grad_norm": 0.8927802443504333, "learning_rate": 3.186729686193718e-07, "loss": 0.3968, "step": 4120 }, { "epoch": 0.8996841302690338, "grad_norm": 1.160204291343689, "learning_rate": 3.0545067473702617e-07, "loss": 0.3959, "step": 4130 }, { "epoch": 0.9018625422067313, "grad_norm": 0.6945912837982178, "learning_rate": 2.92499887940817e-07, "loss": 0.3948, "step": 4140 }, { "epoch": 0.9040409541444288, "grad_norm": 0.9483731389045715, "learning_rate": 2.798213572335001e-07, "loss": 0.3955, "step": 4150 }, { "epoch": 0.9062193660821262, "grad_norm": 0.9849236607551575, "learning_rate": 2.6741581587202747e-07, "loss": 0.3946, "step": 4160 }, { "epoch": 0.9083977780198236, "grad_norm": 1.0742136240005493, "learning_rate": 2.55283981325139e-07, "loss": 0.3954, "step": 4170 }, { "epoch": 0.910576189957521, "grad_norm": 0.8884495496749878, "learning_rate": 2.4342655523186706e-07, "loss": 0.3922, "step": 4180 }, { "epoch": 0.9127546018952184, "grad_norm": 0.7058791518211365, "learning_rate": 2.3184422336095857e-07, "loss": 0.3929, "step": 4190 }, { "epoch": 0.9149330138329158, "grad_norm": 1.1460115909576416, "learning_rate": 2.2053765557121298e-07, "loss": 0.3931, "step": 4200 }, { "epoch": 0.9149330138329158, "eval_loss": 0.43217477202415466, "eval_runtime": 9.5759, "eval_samples_per_second": 521.307, "eval_steps_per_second": 2.089, "step": 4200 }, { "epoch": 0.9171114257706132, "grad_norm": 1.6075760126113892, "learning_rate": 2.0950750577274247e-07, "loss": 0.3943, "step": 4210 }, { "epoch": 0.9192898377083106, "grad_norm": 1.4762377738952637, "learning_rate": 1.9875441188915324e-07, "loss": 0.395, "step": 4220 }, { "epoch": 0.921468249646008, "grad_norm": 0.6985810995101929, "learning_rate": 1.8827899582064824e-07, "loss": 0.3941, "step": 4230 }, { "epoch": 0.9236466615837055, "grad_norm": 1.9295434951782227, "learning_rate": 1.7808186340806664e-07, "loss": 0.3929, "step": 4240 }, { "epoch": 0.9258250735214029, "grad_norm": 0.6554338335990906, "learning_rate": 1.6816360439783797e-07, "loss": 0.3932, "step": 4250 }, { "epoch": 0.9280034854591003, "grad_norm": 0.8134569525718689, "learning_rate": 1.5852479240787978e-07, "loss": 0.3929, "step": 4260 }, { "epoch": 0.9301818973967977, "grad_norm": 0.6693784594535828, "learning_rate": 1.4916598489441824e-07, "loss": 0.3932, "step": 4270 }, { "epoch": 0.9323603093344952, "grad_norm": 0.8736308217048645, "learning_rate": 1.400877231197534e-07, "loss": 0.3918, "step": 4280 }, { "epoch": 0.9345387212721926, "grad_norm": 0.8335554599761963, "learning_rate": 1.3129053212094956e-07, "loss": 0.3943, "step": 4290 }, { "epoch": 0.93671713320989, "grad_norm": 0.784169614315033, "learning_rate": 1.2277492067947428e-07, "loss": 0.3942, "step": 4300 }, { "epoch": 0.93671713320989, "eval_loss": 0.4312984347343445, "eval_runtime": 9.573, "eval_samples_per_second": 521.466, "eval_steps_per_second": 2.089, "step": 4300 }, { "epoch": 0.9388955451475874, "grad_norm": 0.7941382527351379, "learning_rate": 1.1454138129177172e-07, "loss": 0.3953, "step": 4310 }, { "epoch": 0.9410739570852849, "grad_norm": 0.8930506110191345, "learning_rate": 1.0659039014077943e-07, "loss": 0.3922, "step": 4320 }, { "epoch": 0.9432523690229823, "grad_norm": 0.682492733001709, "learning_rate": 9.892240706838808e-08, "loss": 0.3958, "step": 4330 }, { "epoch": 0.9454307809606797, "grad_norm": 0.6189770698547363, "learning_rate": 9.15378755488483e-08, "loss": 0.3945, "step": 4340 }, { "epoch": 0.9476091928983771, "grad_norm": 1.1786706447601318, "learning_rate": 8.443722266312015e-08, "loss": 0.3943, "step": 4350 }, { "epoch": 0.9497876048360745, "grad_norm": 0.7087965607643127, "learning_rate": 7.762085907417505e-08, "loss": 0.3931, "step": 4360 }, { "epoch": 0.9519660167737719, "grad_norm": 0.8723616600036621, "learning_rate": 7.108917900324419e-08, "loss": 0.3934, "step": 4370 }, { "epoch": 0.9541444287114693, "grad_norm": 0.8475690484046936, "learning_rate": 6.484256020701962e-08, "loss": 0.3937, "step": 4380 }, { "epoch": 0.9563228406491667, "grad_norm": 0.9160191416740417, "learning_rate": 5.8881363955807234e-08, "loss": 0.3966, "step": 4390 }, { "epoch": 0.9585012525868641, "grad_norm": 0.7470030784606934, "learning_rate": 5.320593501263071e-08, "loss": 0.3951, "step": 4400 }, { "epoch": 0.9585012525868641, "eval_loss": 0.4317244589328766, "eval_runtime": 9.5949, "eval_samples_per_second": 520.275, "eval_steps_per_second": 2.084, "step": 4400 }, { "epoch": 0.9606796645245615, "grad_norm": 0.7705516219139099, "learning_rate": 4.7816601613295263e-08, "loss": 0.3955, "step": 4410 }, { "epoch": 0.962858076462259, "grad_norm": 0.6091457605361938, "learning_rate": 4.271367544740113e-08, "loss": 0.3923, "step": 4420 }, { "epoch": 0.9650364883999565, "grad_norm": 0.6471231579780579, "learning_rate": 3.7897451640321326e-08, "loss": 0.3947, "step": 4430 }, { "epoch": 0.9672149003376539, "grad_norm": 0.8809778094291687, "learning_rate": 3.3368208736128116e-08, "loss": 0.393, "step": 4440 }, { "epoch": 0.9693933122753513, "grad_norm": 0.7096254825592041, "learning_rate": 2.9126208681488076e-08, "loss": 0.3916, "step": 4450 }, { "epoch": 0.9715717242130487, "grad_norm": 0.8418611288070679, "learning_rate": 2.5171696810509794e-08, "loss": 0.3942, "step": 4460 }, { "epoch": 0.9737501361507461, "grad_norm": 0.6566638350486755, "learning_rate": 2.1504901830555203e-08, "loss": 0.3929, "step": 4470 }, { "epoch": 0.9759285480884435, "grad_norm": 0.7813677787780762, "learning_rate": 1.812603580901462e-08, "loss": 0.3926, "step": 4480 }, { "epoch": 0.978106960026141, "grad_norm": 0.7559650540351868, "learning_rate": 1.5035294161039882e-08, "loss": 0.3955, "step": 4490 }, { "epoch": 0.9802853719638384, "grad_norm": 0.799744188785553, "learning_rate": 1.2232855638241726e-08, "loss": 0.3922, "step": 4500 }, { "epoch": 0.9802853719638384, "eval_loss": 0.43170347809791565, "eval_runtime": 9.6697, "eval_samples_per_second": 516.251, "eval_steps_per_second": 2.068, "step": 4500 }, { "epoch": 0.9824637839015358, "grad_norm": 0.6862944960594177, "learning_rate": 9.718882318355272e-09, "loss": 0.395, "step": 4510 }, { "epoch": 0.9846421958392332, "grad_norm": 0.7467820048332214, "learning_rate": 7.49351959586253e-09, "loss": 0.3943, "step": 4520 }, { "epoch": 0.9868206077769306, "grad_norm": 0.789197564125061, "learning_rate": 5.556896173586346e-09, "loss": 0.3942, "step": 4530 }, { "epoch": 0.988999019714628, "grad_norm": 0.7414279580116272, "learning_rate": 3.9091240552457945e-09, "loss": 0.3938, "step": 4540 }, { "epoch": 0.9911774316523254, "grad_norm": 0.6362519860267639, "learning_rate": 2.5502985389774806e-09, "loss": 0.3957, "step": 4550 }, { "epoch": 0.9933558435900228, "grad_norm": 0.7834025621414185, "learning_rate": 1.4804982118266087e-09, "loss": 0.3927, "step": 4560 }, { "epoch": 0.9955342555277203, "grad_norm": 0.6968185901641846, "learning_rate": 6.997849451989558e-10, "loss": 0.3935, "step": 4570 }, { "epoch": 0.9977126674654178, "grad_norm": 0.7227827906608582, "learning_rate": 2.0820389128539674e-10, "loss": 0.3946, "step": 4580 }, { "epoch": 0.9998910794031152, "grad_norm": 0.5763871669769287, "learning_rate": 5.7834804478851255e-12, "loss": 0.3973, "step": 4590 }, { "epoch": 1.0, "step": 4591, "total_flos": 5.052191529362166e+18, "train_loss": 0.5649781546881453, "train_runtime": 40899.9, "train_samples_per_second": 57.46, "train_steps_per_second": 0.112 } ], "logging_steps": 10, "max_steps": 4591, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 300, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.052191529362166e+18, "train_batch_size": 256, "trial_name": null, "trial_params": null }