| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.3107520198881293, | |
| "eval_steps": 500, | |
| "global_step": 4000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0015537600994406464, | |
| "grad_norm": 1.6171875, | |
| "learning_rate": 9.952500000000001e-05, | |
| "loss": 1.8667953491210938, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.003107520198881293, | |
| "grad_norm": 1.46875, | |
| "learning_rate": 9.9025e-05, | |
| "loss": 1.4211945533752441, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.0046612802983219395, | |
| "grad_norm": 1.46875, | |
| "learning_rate": 9.8525e-05, | |
| "loss": 1.3616345405578614, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.006215040397762586, | |
| "grad_norm": 1.375, | |
| "learning_rate": 9.8025e-05, | |
| "loss": 1.3699079513549806, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.007768800497203232, | |
| "grad_norm": 1.265625, | |
| "learning_rate": 9.7525e-05, | |
| "loss": 1.326680564880371, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.009322560596643879, | |
| "grad_norm": 1.2421875, | |
| "learning_rate": 9.7025e-05, | |
| "loss": 1.3107091903686523, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.010876320696084525, | |
| "grad_norm": 1.2421875, | |
| "learning_rate": 9.652500000000002e-05, | |
| "loss": 1.3247576713562013, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.012430080795525171, | |
| "grad_norm": 1.234375, | |
| "learning_rate": 9.6025e-05, | |
| "loss": 1.306673526763916, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.013983840894965818, | |
| "grad_norm": 1.2265625, | |
| "learning_rate": 9.5525e-05, | |
| "loss": 1.2789793014526367, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.015537600994406464, | |
| "grad_norm": 1.1796875, | |
| "learning_rate": 9.5025e-05, | |
| "loss": 1.2699722290039062, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.01709136109384711, | |
| "grad_norm": 1.1796875, | |
| "learning_rate": 9.452500000000001e-05, | |
| "loss": 1.2735892295837403, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.018645121193287758, | |
| "grad_norm": 1.15625, | |
| "learning_rate": 9.402500000000001e-05, | |
| "loss": 1.255952739715576, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.020198881292728402, | |
| "grad_norm": 1.0859375, | |
| "learning_rate": 9.352500000000001e-05, | |
| "loss": 1.2663206100463866, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.02175264139216905, | |
| "grad_norm": 1.09375, | |
| "learning_rate": 9.302500000000001e-05, | |
| "loss": 1.2674468994140624, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.023306401491609695, | |
| "grad_norm": 1.0703125, | |
| "learning_rate": 9.252500000000001e-05, | |
| "loss": 1.2581491470336914, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.024860161591050343, | |
| "grad_norm": 1.15625, | |
| "learning_rate": 9.2025e-05, | |
| "loss": 1.2293106079101563, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.026413921690490987, | |
| "grad_norm": 1.0703125, | |
| "learning_rate": 9.1525e-05, | |
| "loss": 1.2330900192260743, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.027967681789931635, | |
| "grad_norm": 1.0859375, | |
| "learning_rate": 9.1025e-05, | |
| "loss": 1.2186731338500976, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.02952144188937228, | |
| "grad_norm": 1.078125, | |
| "learning_rate": 9.0525e-05, | |
| "loss": 1.2420782089233398, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.031075201988812928, | |
| "grad_norm": 0.984375, | |
| "learning_rate": 9.0025e-05, | |
| "loss": 1.2236801147460938, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.032628962088253576, | |
| "grad_norm": 1.0390625, | |
| "learning_rate": 8.952500000000001e-05, | |
| "loss": 1.2241475105285644, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.03418272218769422, | |
| "grad_norm": 0.98828125, | |
| "learning_rate": 8.902500000000001e-05, | |
| "loss": 1.2227598190307618, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.035736482287134864, | |
| "grad_norm": 1.0390625, | |
| "learning_rate": 8.8525e-05, | |
| "loss": 1.2154598236083984, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.037290242386575516, | |
| "grad_norm": 1.0859375, | |
| "learning_rate": 8.8025e-05, | |
| "loss": 1.20736665725708, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.03884400248601616, | |
| "grad_norm": 1.0546875, | |
| "learning_rate": 8.7525e-05, | |
| "loss": 1.1936758041381836, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.040397762585456805, | |
| "grad_norm": 1.09375, | |
| "learning_rate": 8.7025e-05, | |
| "loss": 1.1817484855651856, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.04195152268489745, | |
| "grad_norm": 1.0703125, | |
| "learning_rate": 8.6525e-05, | |
| "loss": 1.2060453414916992, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.0435052827843381, | |
| "grad_norm": 0.99609375, | |
| "learning_rate": 8.6025e-05, | |
| "loss": 1.2340307235717773, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.045059042883778745, | |
| "grad_norm": 1.0, | |
| "learning_rate": 8.5525e-05, | |
| "loss": 1.2149415969848634, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.04661280298321939, | |
| "grad_norm": 0.97265625, | |
| "learning_rate": 8.502499999999999e-05, | |
| "loss": 1.2079537391662598, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.048166563082660034, | |
| "grad_norm": 1.015625, | |
| "learning_rate": 8.4525e-05, | |
| "loss": 1.1671560287475586, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.049720323182100686, | |
| "grad_norm": 1.0, | |
| "learning_rate": 8.402500000000001e-05, | |
| "loss": 1.1967806816101074, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.05127408328154133, | |
| "grad_norm": 1.0234375, | |
| "learning_rate": 8.352500000000001e-05, | |
| "loss": 1.178822135925293, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.052827843380981974, | |
| "grad_norm": 1.0625, | |
| "learning_rate": 8.302500000000001e-05, | |
| "loss": 1.1784405708312988, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.054381603480422626, | |
| "grad_norm": 0.9765625, | |
| "learning_rate": 8.252500000000001e-05, | |
| "loss": 1.1611870765686034, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.05593536357986327, | |
| "grad_norm": 1.0078125, | |
| "learning_rate": 8.2025e-05, | |
| "loss": 1.1876888275146484, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.057489123679303915, | |
| "grad_norm": 0.94921875, | |
| "learning_rate": 8.1525e-05, | |
| "loss": 1.18981876373291, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.05904288377874456, | |
| "grad_norm": 0.93359375, | |
| "learning_rate": 8.1025e-05, | |
| "loss": 1.1651229858398438, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.06059664387818521, | |
| "grad_norm": 0.953125, | |
| "learning_rate": 8.0525e-05, | |
| "loss": 1.1725165367126464, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.062150403977625855, | |
| "grad_norm": 0.94921875, | |
| "learning_rate": 8.002500000000001e-05, | |
| "loss": 1.1620153427124023, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.0637041640770665, | |
| "grad_norm": 0.94921875, | |
| "learning_rate": 7.952500000000001e-05, | |
| "loss": 1.180180549621582, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.06525792417650715, | |
| "grad_norm": 0.953125, | |
| "learning_rate": 7.902500000000001e-05, | |
| "loss": 1.169425106048584, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.06681168427594779, | |
| "grad_norm": 0.96875, | |
| "learning_rate": 7.8525e-05, | |
| "loss": 1.1701436996459962, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.06836544437538844, | |
| "grad_norm": 0.98828125, | |
| "learning_rate": 7.8025e-05, | |
| "loss": 1.1514653205871581, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.06991920447482909, | |
| "grad_norm": 0.9453125, | |
| "learning_rate": 7.7525e-05, | |
| "loss": 1.166794776916504, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.07147296457426973, | |
| "grad_norm": 0.98828125, | |
| "learning_rate": 7.7025e-05, | |
| "loss": 1.1640316009521485, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.07302672467371038, | |
| "grad_norm": 1.0, | |
| "learning_rate": 7.6525e-05, | |
| "loss": 1.1756441116333007, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.07458048477315103, | |
| "grad_norm": 0.921875, | |
| "learning_rate": 7.6025e-05, | |
| "loss": 1.1775863647460938, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.07613424487259167, | |
| "grad_norm": 0.94140625, | |
| "learning_rate": 7.5525e-05, | |
| "loss": 1.1620708465576173, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.07768800497203232, | |
| "grad_norm": 0.93359375, | |
| "learning_rate": 7.502500000000001e-05, | |
| "loss": 1.1424921989440917, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.07924176507147296, | |
| "grad_norm": 0.94140625, | |
| "learning_rate": 7.4525e-05, | |
| "loss": 1.1636884689331055, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.08079552517091361, | |
| "grad_norm": 0.96875, | |
| "learning_rate": 7.4025e-05, | |
| "loss": 1.1640792846679688, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.08234928527035426, | |
| "grad_norm": 0.96875, | |
| "learning_rate": 7.3525e-05, | |
| "loss": 1.1577949523925781, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.0839030453697949, | |
| "grad_norm": 0.9375, | |
| "learning_rate": 7.3025e-05, | |
| "loss": 1.1563040733337402, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.08545680546923555, | |
| "grad_norm": 0.9296875, | |
| "learning_rate": 7.2525e-05, | |
| "loss": 1.1314552307128907, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.0870105655686762, | |
| "grad_norm": 0.953125, | |
| "learning_rate": 7.2025e-05, | |
| "loss": 1.1430686950683593, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.08856432566811684, | |
| "grad_norm": 0.921875, | |
| "learning_rate": 7.1525e-05, | |
| "loss": 1.1624882698059082, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.09011808576755749, | |
| "grad_norm": 0.9375, | |
| "learning_rate": 7.1025e-05, | |
| "loss": 1.1404618263244628, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.09167184586699814, | |
| "grad_norm": 0.94921875, | |
| "learning_rate": 7.0525e-05, | |
| "loss": 1.142820453643799, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.09322560596643878, | |
| "grad_norm": 0.9296875, | |
| "learning_rate": 7.002500000000001e-05, | |
| "loss": 1.1361354827880858, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.09477936606587943, | |
| "grad_norm": 0.91796875, | |
| "learning_rate": 6.952500000000001e-05, | |
| "loss": 1.141007137298584, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.09633312616532007, | |
| "grad_norm": 0.93359375, | |
| "learning_rate": 6.902500000000001e-05, | |
| "loss": 1.140983772277832, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.09788688626476072, | |
| "grad_norm": 0.94921875, | |
| "learning_rate": 6.852500000000001e-05, | |
| "loss": 1.1491514205932618, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.09944064636420137, | |
| "grad_norm": 0.953125, | |
| "learning_rate": 6.8025e-05, | |
| "loss": 1.1248952865600585, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.10099440646364201, | |
| "grad_norm": 0.9296875, | |
| "learning_rate": 6.7525e-05, | |
| "loss": 1.1090205192565918, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.10254816656308266, | |
| "grad_norm": 0.9765625, | |
| "learning_rate": 6.7025e-05, | |
| "loss": 1.1208727836608887, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.10410192666252331, | |
| "grad_norm": 0.9375, | |
| "learning_rate": 6.6525e-05, | |
| "loss": 1.1511701583862304, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.10565568676196395, | |
| "grad_norm": 0.8828125, | |
| "learning_rate": 6.6025e-05, | |
| "loss": 1.1259963989257813, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.1072094468614046, | |
| "grad_norm": 0.921875, | |
| "learning_rate": 6.552500000000001e-05, | |
| "loss": 1.1311534881591796, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.10876320696084525, | |
| "grad_norm": 0.9375, | |
| "learning_rate": 6.502500000000001e-05, | |
| "loss": 1.1099111557006835, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.11031696706028589, | |
| "grad_norm": 0.953125, | |
| "learning_rate": 6.4525e-05, | |
| "loss": 1.126076602935791, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.11187072715972654, | |
| "grad_norm": 0.8984375, | |
| "learning_rate": 6.4025e-05, | |
| "loss": 1.1338358879089356, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.11342448725916718, | |
| "grad_norm": 0.91796875, | |
| "learning_rate": 6.3525e-05, | |
| "loss": 1.1149433135986329, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.11497824735860783, | |
| "grad_norm": 0.96484375, | |
| "learning_rate": 6.3025e-05, | |
| "loss": 1.1221566200256348, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.11653200745804848, | |
| "grad_norm": 0.921875, | |
| "learning_rate": 6.2525e-05, | |
| "loss": 1.1223237991333008, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.11808576755748912, | |
| "grad_norm": 0.90234375, | |
| "learning_rate": 6.2025e-05, | |
| "loss": 1.1033407211303712, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.11963952765692977, | |
| "grad_norm": 0.90234375, | |
| "learning_rate": 6.1525e-05, | |
| "loss": 1.1114818572998046, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.12119328775637042, | |
| "grad_norm": 0.91796875, | |
| "learning_rate": 6.1025e-05, | |
| "loss": 1.1105637550354004, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.12274704785581106, | |
| "grad_norm": 0.90625, | |
| "learning_rate": 6.0525e-05, | |
| "loss": 1.1189040184020995, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.12430080795525171, | |
| "grad_norm": 0.921875, | |
| "learning_rate": 6.0024999999999995e-05, | |
| "loss": 1.1234063148498534, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.12585456805469236, | |
| "grad_norm": 0.953125, | |
| "learning_rate": 5.9525e-05, | |
| "loss": 1.1020179748535157, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.127408328154133, | |
| "grad_norm": 0.921875, | |
| "learning_rate": 5.9025000000000005e-05, | |
| "loss": 1.1260129928588867, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.12896208825357364, | |
| "grad_norm": 0.87890625, | |
| "learning_rate": 5.8525000000000006e-05, | |
| "loss": 1.1067705154418945, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.1305158483530143, | |
| "grad_norm": 0.94140625, | |
| "learning_rate": 5.802500000000001e-05, | |
| "loss": 1.115440273284912, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.13206960845245494, | |
| "grad_norm": 0.91796875, | |
| "learning_rate": 5.752500000000001e-05, | |
| "loss": 1.1169689178466797, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.13362336855189558, | |
| "grad_norm": 0.921875, | |
| "learning_rate": 5.7025000000000004e-05, | |
| "loss": 1.1162803649902344, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.13517712865133624, | |
| "grad_norm": 0.8984375, | |
| "learning_rate": 5.6525000000000005e-05, | |
| "loss": 1.1266037940979003, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.13673088875077688, | |
| "grad_norm": 0.953125, | |
| "learning_rate": 5.6025000000000007e-05, | |
| "loss": 1.1090587615966796, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.13828464885021752, | |
| "grad_norm": 0.8828125, | |
| "learning_rate": 5.552500000000001e-05, | |
| "loss": 1.1141853332519531, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.13983840894965818, | |
| "grad_norm": 0.88671875, | |
| "learning_rate": 5.5025e-05, | |
| "loss": 1.1379024505615234, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.14139216904909882, | |
| "grad_norm": 0.91015625, | |
| "learning_rate": 5.4525000000000004e-05, | |
| "loss": 1.110361099243164, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.14294592914853946, | |
| "grad_norm": 0.8984375, | |
| "learning_rate": 5.4025000000000005e-05, | |
| "loss": 1.1104223251342773, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.14449968924798012, | |
| "grad_norm": 0.953125, | |
| "learning_rate": 5.3525e-05, | |
| "loss": 1.0993282318115234, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.14605344934742076, | |
| "grad_norm": 0.88671875, | |
| "learning_rate": 5.3025e-05, | |
| "loss": 1.1151838302612305, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.1476072094468614, | |
| "grad_norm": 0.91796875, | |
| "learning_rate": 5.2525e-05, | |
| "loss": 1.1184428215026856, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.14916096954630206, | |
| "grad_norm": 0.921875, | |
| "learning_rate": 5.2025000000000004e-05, | |
| "loss": 1.1321413040161132, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.1507147296457427, | |
| "grad_norm": 0.875, | |
| "learning_rate": 5.1525e-05, | |
| "loss": 1.1035722732543944, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.15226848974518334, | |
| "grad_norm": 0.92578125, | |
| "learning_rate": 5.1025e-05, | |
| "loss": 1.0843469619750976, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.153822249844624, | |
| "grad_norm": 0.89453125, | |
| "learning_rate": 5.0525e-05, | |
| "loss": 1.102191162109375, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.15537600994406464, | |
| "grad_norm": 0.93359375, | |
| "learning_rate": 5.0025e-05, | |
| "loss": 1.1058999061584474, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.15692977004350528, | |
| "grad_norm": 0.8984375, | |
| "learning_rate": 4.9525000000000004e-05, | |
| "loss": 1.1077130317687989, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.15848353014294592, | |
| "grad_norm": 0.91796875, | |
| "learning_rate": 4.9025000000000006e-05, | |
| "loss": 1.0922195434570312, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.16003729024238658, | |
| "grad_norm": 0.92578125, | |
| "learning_rate": 4.8525e-05, | |
| "loss": 1.130363941192627, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.16159105034182722, | |
| "grad_norm": 0.8984375, | |
| "learning_rate": 4.8025e-05, | |
| "loss": 1.0666452407836915, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.16314481044126786, | |
| "grad_norm": 0.91015625, | |
| "learning_rate": 4.7525e-05, | |
| "loss": 1.096773052215576, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.16469857054070852, | |
| "grad_norm": 0.9296875, | |
| "learning_rate": 4.7025000000000005e-05, | |
| "loss": 1.0890857696533203, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.16625233064014916, | |
| "grad_norm": 0.91015625, | |
| "learning_rate": 4.6525e-05, | |
| "loss": 1.1071590423583983, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.1678060907395898, | |
| "grad_norm": 0.87890625, | |
| "learning_rate": 4.6025e-05, | |
| "loss": 1.1148256301879882, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.16935985083903046, | |
| "grad_norm": 0.92578125, | |
| "learning_rate": 4.5525e-05, | |
| "loss": 1.090849018096924, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.1709136109384711, | |
| "grad_norm": 0.921875, | |
| "learning_rate": 4.5025000000000003e-05, | |
| "loss": 1.1214483261108399, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.17246737103791174, | |
| "grad_norm": 0.89453125, | |
| "learning_rate": 4.4525e-05, | |
| "loss": 1.099250030517578, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.1740211311373524, | |
| "grad_norm": 0.90234375, | |
| "learning_rate": 4.4025e-05, | |
| "loss": 1.0906652450561523, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.17557489123679304, | |
| "grad_norm": 0.875, | |
| "learning_rate": 4.352500000000001e-05, | |
| "loss": 1.0940235137939454, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.17712865133623368, | |
| "grad_norm": 0.875, | |
| "learning_rate": 4.3025e-05, | |
| "loss": 1.0771520614624024, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.17868241143567434, | |
| "grad_norm": 0.875, | |
| "learning_rate": 4.2525000000000004e-05, | |
| "loss": 1.0870559692382813, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.18023617153511498, | |
| "grad_norm": 0.90625, | |
| "learning_rate": 4.2025000000000005e-05, | |
| "loss": 1.0863205909729003, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.18178993163455562, | |
| "grad_norm": 0.9296875, | |
| "learning_rate": 4.1525e-05, | |
| "loss": 1.0996931076049805, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.18334369173399628, | |
| "grad_norm": 0.875, | |
| "learning_rate": 4.1025e-05, | |
| "loss": 1.0987712860107421, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.18489745183343692, | |
| "grad_norm": 0.9296875, | |
| "learning_rate": 4.0525e-05, | |
| "loss": 1.1171295166015625, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.18645121193287756, | |
| "grad_norm": 0.9296875, | |
| "learning_rate": 4.0025000000000004e-05, | |
| "loss": 1.103907012939453, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.18800497203231822, | |
| "grad_norm": 0.94140625, | |
| "learning_rate": 3.9525e-05, | |
| "loss": 1.090281867980957, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.18955873213175886, | |
| "grad_norm": 0.9140625, | |
| "learning_rate": 3.9025e-05, | |
| "loss": 1.0880367279052734, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.1911124922311995, | |
| "grad_norm": 0.90625, | |
| "learning_rate": 3.8525e-05, | |
| "loss": 1.0839120864868164, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.19266625233064014, | |
| "grad_norm": 0.9375, | |
| "learning_rate": 3.8025e-05, | |
| "loss": 1.1069268226623534, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.1942200124300808, | |
| "grad_norm": 0.87109375, | |
| "learning_rate": 3.7525e-05, | |
| "loss": 1.0701780319213867, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.19577377252952144, | |
| "grad_norm": 0.921875, | |
| "learning_rate": 3.7025000000000005e-05, | |
| "loss": 1.1017258644104004, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.19732753262896208, | |
| "grad_norm": 0.9140625, | |
| "learning_rate": 3.652500000000001e-05, | |
| "loss": 1.1134037017822265, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.19888129272840274, | |
| "grad_norm": 0.9453125, | |
| "learning_rate": 3.6025e-05, | |
| "loss": 1.0741355895996094, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.20043505282784338, | |
| "grad_norm": 0.890625, | |
| "learning_rate": 3.5525e-05, | |
| "loss": 1.0729135513305663, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.20198881292728402, | |
| "grad_norm": 0.921875, | |
| "learning_rate": 3.5025000000000004e-05, | |
| "loss": 1.0833280563354493, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.20354257302672468, | |
| "grad_norm": 0.8984375, | |
| "learning_rate": 3.4525e-05, | |
| "loss": 1.1003621101379395, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.20509633312616532, | |
| "grad_norm": 0.8984375, | |
| "learning_rate": 3.4025e-05, | |
| "loss": 1.0958992004394532, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.20665009322560596, | |
| "grad_norm": 0.90625, | |
| "learning_rate": 3.3525e-05, | |
| "loss": 1.098832893371582, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.20820385332504662, | |
| "grad_norm": 0.89453125, | |
| "learning_rate": 3.3025e-05, | |
| "loss": 1.0823541641235352, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.20975761342448726, | |
| "grad_norm": 0.91015625, | |
| "learning_rate": 3.2525e-05, | |
| "loss": 1.0849027633666992, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.2113113735239279, | |
| "grad_norm": 0.890625, | |
| "learning_rate": 3.2025e-05, | |
| "loss": 1.0959321975708007, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.21286513362336856, | |
| "grad_norm": 0.88671875, | |
| "learning_rate": 3.1525e-05, | |
| "loss": 1.083869743347168, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.2144188937228092, | |
| "grad_norm": 0.8984375, | |
| "learning_rate": 3.1025e-05, | |
| "loss": 1.0912399291992188, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.21597265382224984, | |
| "grad_norm": 0.9140625, | |
| "learning_rate": 3.0525e-05, | |
| "loss": 1.088266944885254, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.2175264139216905, | |
| "grad_norm": 0.86328125, | |
| "learning_rate": 3.0025000000000005e-05, | |
| "loss": 1.0917407035827638, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.21908017402113114, | |
| "grad_norm": 0.88671875, | |
| "learning_rate": 2.9525000000000003e-05, | |
| "loss": 1.088615608215332, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.22063393412057178, | |
| "grad_norm": 0.89453125, | |
| "learning_rate": 2.9025e-05, | |
| "loss": 1.0885720252990723, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.22218769422001244, | |
| "grad_norm": 0.8828125, | |
| "learning_rate": 2.8525000000000002e-05, | |
| "loss": 1.072688388824463, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.22374145431945308, | |
| "grad_norm": 0.89453125, | |
| "learning_rate": 2.8025e-05, | |
| "loss": 1.0892706871032716, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.22529521441889372, | |
| "grad_norm": 0.93359375, | |
| "learning_rate": 2.7525e-05, | |
| "loss": 1.083917236328125, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.22684897451833436, | |
| "grad_norm": 0.921875, | |
| "learning_rate": 2.7025e-05, | |
| "loss": 1.1079219818115233, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.22840273461777502, | |
| "grad_norm": 0.91796875, | |
| "learning_rate": 2.6525e-05, | |
| "loss": 1.0851898193359375, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.22995649471721566, | |
| "grad_norm": 0.91015625, | |
| "learning_rate": 2.6025e-05, | |
| "loss": 1.0698064804077148, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.2315102548166563, | |
| "grad_norm": 0.8828125, | |
| "learning_rate": 2.5525e-05, | |
| "loss": 1.0661302566528321, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.23306401491609696, | |
| "grad_norm": 0.9140625, | |
| "learning_rate": 2.5025e-05, | |
| "loss": 1.0647387504577637, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.2346177750155376, | |
| "grad_norm": 0.8984375, | |
| "learning_rate": 2.4525e-05, | |
| "loss": 1.074313259124756, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.23617153511497824, | |
| "grad_norm": 0.9140625, | |
| "learning_rate": 2.4025e-05, | |
| "loss": 1.0964820861816407, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.2377252952144189, | |
| "grad_norm": 0.89453125, | |
| "learning_rate": 2.3525e-05, | |
| "loss": 1.0810314178466798, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.23927905531385954, | |
| "grad_norm": 0.9296875, | |
| "learning_rate": 2.3025e-05, | |
| "loss": 1.078106689453125, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.24083281541330018, | |
| "grad_norm": 0.90234375, | |
| "learning_rate": 2.2525000000000002e-05, | |
| "loss": 1.0860605239868164, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.24238657551274084, | |
| "grad_norm": 0.88671875, | |
| "learning_rate": 2.2025e-05, | |
| "loss": 1.0743337631225587, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.24394033561218148, | |
| "grad_norm": 0.88671875, | |
| "learning_rate": 2.1525e-05, | |
| "loss": 1.0778118133544923, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 0.24549409571162212, | |
| "grad_norm": 0.9296875, | |
| "learning_rate": 2.1025e-05, | |
| "loss": 1.0981425285339355, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 0.24704785581106278, | |
| "grad_norm": 0.91015625, | |
| "learning_rate": 2.0525e-05, | |
| "loss": 1.0836584091186523, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.24860161591050342, | |
| "grad_norm": 0.8671875, | |
| "learning_rate": 2.0025000000000002e-05, | |
| "loss": 1.089561367034912, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.2501553760099441, | |
| "grad_norm": 0.9296875, | |
| "learning_rate": 1.9525e-05, | |
| "loss": 1.0788454055786132, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 0.2517091361093847, | |
| "grad_norm": 0.91015625, | |
| "learning_rate": 1.9025e-05, | |
| "loss": 1.078543758392334, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.25326289620882536, | |
| "grad_norm": 0.9140625, | |
| "learning_rate": 1.8525e-05, | |
| "loss": 1.1045896530151367, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 0.254816656308266, | |
| "grad_norm": 0.88671875, | |
| "learning_rate": 1.8025e-05, | |
| "loss": 1.0857179641723633, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 0.25637041640770664, | |
| "grad_norm": 0.87109375, | |
| "learning_rate": 1.7525e-05, | |
| "loss": 1.0980566024780274, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.2579241765071473, | |
| "grad_norm": 0.90234375, | |
| "learning_rate": 1.7025e-05, | |
| "loss": 1.0898794174194335, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 0.25947793660658797, | |
| "grad_norm": 0.93359375, | |
| "learning_rate": 1.6525000000000002e-05, | |
| "loss": 1.0813383102416991, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 0.2610316967060286, | |
| "grad_norm": 0.9140625, | |
| "learning_rate": 1.6025e-05, | |
| "loss": 1.0852994918823242, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.26258545680546924, | |
| "grad_norm": 0.9296875, | |
| "learning_rate": 1.5525e-05, | |
| "loss": 1.0736425399780274, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 0.2641392169049099, | |
| "grad_norm": 0.87890625, | |
| "learning_rate": 1.5025000000000001e-05, | |
| "loss": 1.0873353004455566, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.2656929770043505, | |
| "grad_norm": 0.90625, | |
| "learning_rate": 1.4524999999999999e-05, | |
| "loss": 1.0847795486450196, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.26724673710379115, | |
| "grad_norm": 0.90234375, | |
| "learning_rate": 1.4025000000000002e-05, | |
| "loss": 1.0983787536621095, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 0.26880049720323185, | |
| "grad_norm": 0.8828125, | |
| "learning_rate": 1.3525000000000002e-05, | |
| "loss": 1.0719210624694824, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 0.2703542573026725, | |
| "grad_norm": 0.87890625, | |
| "learning_rate": 1.3025000000000002e-05, | |
| "loss": 1.0850942611694336, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.2719080174021131, | |
| "grad_norm": 0.890625, | |
| "learning_rate": 1.2525000000000001e-05, | |
| "loss": 1.0767460823059083, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.27346177750155376, | |
| "grad_norm": 0.90234375, | |
| "learning_rate": 1.2025000000000001e-05, | |
| "loss": 1.0981364250183105, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 0.2750155376009944, | |
| "grad_norm": 0.90234375, | |
| "learning_rate": 1.1525e-05, | |
| "loss": 1.0713367462158203, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.27656929770043504, | |
| "grad_norm": 0.8984375, | |
| "learning_rate": 1.1025e-05, | |
| "loss": 1.074637222290039, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 0.2781230577998757, | |
| "grad_norm": 0.91015625, | |
| "learning_rate": 1.0525e-05, | |
| "loss": 1.0917478561401368, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 0.27967681789931637, | |
| "grad_norm": 0.8828125, | |
| "learning_rate": 1.0025000000000001e-05, | |
| "loss": 1.0934381484985352, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.281230577998757, | |
| "grad_norm": 0.953125, | |
| "learning_rate": 9.525000000000001e-06, | |
| "loss": 1.0998394966125489, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 0.28278433809819764, | |
| "grad_norm": 0.8828125, | |
| "learning_rate": 9.025e-06, | |
| "loss": 1.087621021270752, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 0.2843380981976383, | |
| "grad_norm": 0.91015625, | |
| "learning_rate": 8.525e-06, | |
| "loss": 1.091860866546631, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.2858918582970789, | |
| "grad_norm": 0.9375, | |
| "learning_rate": 8.025e-06, | |
| "loss": 1.072180938720703, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 0.28744561839651955, | |
| "grad_norm": 0.8828125, | |
| "learning_rate": 7.525e-06, | |
| "loss": 1.0778998374938964, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.28899937849596025, | |
| "grad_norm": 0.875, | |
| "learning_rate": 7.025000000000001e-06, | |
| "loss": 1.0709803581237793, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.2905531385954009, | |
| "grad_norm": 0.91015625, | |
| "learning_rate": 6.525e-06, | |
| "loss": 1.1010093688964844, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 0.2921068986948415, | |
| "grad_norm": 0.91015625, | |
| "learning_rate": 6.025e-06, | |
| "loss": 1.0703121185302735, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 0.29366065879428216, | |
| "grad_norm": 0.88671875, | |
| "learning_rate": 5.5250000000000005e-06, | |
| "loss": 1.0773000717163086, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 0.2952144188937228, | |
| "grad_norm": 0.91796875, | |
| "learning_rate": 5.025e-06, | |
| "loss": 1.0894964218139649, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.29676817899316343, | |
| "grad_norm": 0.87890625, | |
| "learning_rate": 4.525e-06, | |
| "loss": 1.060502815246582, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 0.2983219390926041, | |
| "grad_norm": 0.890625, | |
| "learning_rate": 4.0250000000000004e-06, | |
| "loss": 1.0844976425170898, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 0.29987569919204476, | |
| "grad_norm": 0.90625, | |
| "learning_rate": 3.5249999999999997e-06, | |
| "loss": 1.068390464782715, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 0.3014294592914854, | |
| "grad_norm": 0.8828125, | |
| "learning_rate": 3.0250000000000003e-06, | |
| "loss": 1.0733464241027832, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 0.30298321939092604, | |
| "grad_norm": 0.92578125, | |
| "learning_rate": 2.5250000000000004e-06, | |
| "loss": 1.0655000686645508, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.3045369794903667, | |
| "grad_norm": 0.88671875, | |
| "learning_rate": 2.025e-06, | |
| "loss": 1.0655412673950195, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 0.3060907395898073, | |
| "grad_norm": 0.90234375, | |
| "learning_rate": 1.525e-06, | |
| "loss": 1.0558664321899414, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 0.307644499689248, | |
| "grad_norm": 0.94140625, | |
| "learning_rate": 1.0250000000000001e-06, | |
| "loss": 1.1024866104125977, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 0.30919825978868865, | |
| "grad_norm": 0.91015625, | |
| "learning_rate": 5.250000000000001e-07, | |
| "loss": 1.0651491165161133, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 0.3107520198881293, | |
| "grad_norm": 0.9609375, | |
| "learning_rate": 2.5000000000000002e-08, | |
| "loss": 1.0959912300109864, | |
| "step": 4000 | |
| } | |
| ], | |
| "logging_steps": 20, | |
| "max_steps": 4000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 9.3750546137088e+17, | |
| "train_batch_size": 32, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |