{ "best_metric": null, "best_model_checkpoint": null, "epoch": 41.328125, "eval_steps": 500, "global_step": 2645, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0, "eval_runtime": 2.9591, "eval_samples_per_second": 0.338, "eval_steps_per_second": 0.338, "step": 0 }, { "epoch": 0.15625, "grad_norm": 8.098007202148438, "learning_rate": 4.000000000000001e-06, "loss": 8.5142, "step": 10 }, { "epoch": 0.3125, "grad_norm": 6.856651306152344, "learning_rate": 9e-06, "loss": 7.9708, "step": 20 }, { "epoch": 0.46875, "grad_norm": 5.80125093460083, "learning_rate": 1.4000000000000001e-05, "loss": 8.01, "step": 30 }, { "epoch": 0.625, "grad_norm": 5.687861442565918, "learning_rate": 1.9e-05, "loss": 7.6212, "step": 40 }, { "epoch": 0.78125, "grad_norm": 6.117136478424072, "learning_rate": 2.4e-05, "loss": 7.3311, "step": 50 }, { "epoch": 0.9375, "grad_norm": 5.861077785491943, "learning_rate": 2.9e-05, "loss": 7.5553, "step": 60 }, { "epoch": 1.09375, "grad_norm": 6.021880149841309, "learning_rate": 3.4000000000000007e-05, "loss": 7.05, "step": 70 }, { "epoch": 1.25, "grad_norm": 7.620683670043945, "learning_rate": 3.9000000000000006e-05, "loss": 6.7186, "step": 80 }, { "epoch": 1.40625, "grad_norm": 7.098018169403076, "learning_rate": 4.4000000000000006e-05, "loss": 6.5668, "step": 90 }, { "epoch": 1.5625, "grad_norm": 7.157684326171875, "learning_rate": 4.9e-05, "loss": 6.6148, "step": 100 }, { "epoch": 1.71875, "grad_norm": 7.009696006774902, "learning_rate": 4.9936507936507936e-05, "loss": 6.5017, "step": 110 }, { "epoch": 1.875, "grad_norm": 7.6622467041015625, "learning_rate": 4.985714285714286e-05, "loss": 6.53, "step": 120 }, { "epoch": 2.03125, "grad_norm": 8.221333503723145, "learning_rate": 4.977777777777778e-05, "loss": 6.306, "step": 130 }, { "epoch": 2.1875, "grad_norm": 8.832132339477539, "learning_rate": 4.96984126984127e-05, "loss": 4.8787, "step": 140 }, { "epoch": 2.34375, "grad_norm": 9.245682716369629, "learning_rate": 4.961904761904762e-05, "loss": 4.7796, "step": 150 }, { "epoch": 2.5, "grad_norm": 9.249746322631836, "learning_rate": 4.953968253968254e-05, "loss": 4.79, "step": 160 }, { "epoch": 2.65625, "grad_norm": 8.098647117614746, "learning_rate": 4.9460317460317465e-05, "loss": 4.8157, "step": 170 }, { "epoch": 2.8125, "grad_norm": 9.594801902770996, "learning_rate": 4.9380952380952386e-05, "loss": 4.5135, "step": 180 }, { "epoch": 2.96875, "grad_norm": 9.500739097595215, "learning_rate": 4.930158730158731e-05, "loss": 4.9346, "step": 190 }, { "epoch": 3.125, "grad_norm": 10.759628295898438, "learning_rate": 4.922222222222222e-05, "loss": 3.4195, "step": 200 }, { "epoch": 3.28125, "grad_norm": 12.300085067749023, "learning_rate": 4.9142857142857144e-05, "loss": 3.136, "step": 210 }, { "epoch": 3.4375, "grad_norm": 9.825932502746582, "learning_rate": 4.9063492063492065e-05, "loss": 3.0981, "step": 220 }, { "epoch": 3.59375, "grad_norm": 10.764175415039062, "learning_rate": 4.898412698412699e-05, "loss": 3.0372, "step": 230 }, { "epoch": 3.75, "grad_norm": 10.330010414123535, "learning_rate": 4.890476190476191e-05, "loss": 3.0934, "step": 240 }, { "epoch": 3.90625, "grad_norm": 10.405282020568848, "learning_rate": 4.882539682539683e-05, "loss": 3.135, "step": 250 }, { "epoch": 4.0625, "grad_norm": 9.842012405395508, "learning_rate": 4.874603174603175e-05, "loss": 2.7457, "step": 260 }, { "epoch": 4.21875, "grad_norm": 10.659170150756836, "learning_rate": 4.866666666666667e-05, "loss": 1.981, "step": 270 }, { "epoch": 4.375, "grad_norm": 9.820462226867676, "learning_rate": 4.858730158730159e-05, "loss": 2.0881, "step": 280 }, { "epoch": 4.53125, "grad_norm": 9.645284652709961, "learning_rate": 4.850793650793651e-05, "loss": 1.9438, "step": 290 }, { "epoch": 4.6875, "grad_norm": 9.881282806396484, "learning_rate": 4.842857142857143e-05, "loss": 1.8556, "step": 300 }, { "epoch": 4.84375, "grad_norm": 10.5478515625, "learning_rate": 4.834920634920635e-05, "loss": 2.0494, "step": 310 }, { "epoch": 5.0, "grad_norm": 19.77553939819336, "learning_rate": 4.8269841269841274e-05, "loss": 2.0695, "step": 320 }, { "epoch": 5.15625, "grad_norm": 8.124504089355469, "learning_rate": 4.819047619047619e-05, "loss": 1.1424, "step": 330 }, { "epoch": 5.3125, "grad_norm": 9.63291072845459, "learning_rate": 4.811111111111111e-05, "loss": 1.1688, "step": 340 }, { "epoch": 5.46875, "grad_norm": 8.620952606201172, "learning_rate": 4.803174603174603e-05, "loss": 1.2346, "step": 350 }, { "epoch": 5.625, "grad_norm": 8.060315132141113, "learning_rate": 4.795238095238095e-05, "loss": 1.176, "step": 360 }, { "epoch": 5.78125, "grad_norm": 7.583358287811279, "learning_rate": 4.7873015873015874e-05, "loss": 1.2118, "step": 370 }, { "epoch": 5.9375, "grad_norm": 10.535112380981445, "learning_rate": 4.7793650793650796e-05, "loss": 1.2446, "step": 380 }, { "epoch": 6.09375, "grad_norm": 7.170854568481445, "learning_rate": 4.771428571428572e-05, "loss": 1.0445, "step": 390 }, { "epoch": 6.25, "grad_norm": 6.0907392501831055, "learning_rate": 4.763492063492064e-05, "loss": 0.7688, "step": 400 }, { "epoch": 6.40625, "grad_norm": 7.953549385070801, "learning_rate": 4.755555555555556e-05, "loss": 0.7333, "step": 410 }, { "epoch": 6.5625, "grad_norm": 10.276406288146973, "learning_rate": 4.747619047619048e-05, "loss": 0.8645, "step": 420 }, { "epoch": 6.71875, "grad_norm": 7.596552848815918, "learning_rate": 4.73968253968254e-05, "loss": 0.7799, "step": 430 }, { "epoch": 6.875, "grad_norm": 6.478920936584473, "learning_rate": 4.7317460317460325e-05, "loss": 0.8262, "step": 440 }, { "epoch": 7.03125, "grad_norm": 4.837390422821045, "learning_rate": 4.723809523809524e-05, "loss": 0.7218, "step": 450 }, { "epoch": 7.1875, "grad_norm": 5.8777289390563965, "learning_rate": 4.715873015873016e-05, "loss": 0.5238, "step": 460 }, { "epoch": 7.34375, "grad_norm": 4.988452434539795, "learning_rate": 4.707936507936508e-05, "loss": 0.5285, "step": 470 }, { "epoch": 7.5, "grad_norm": 5.605051517486572, "learning_rate": 4.7e-05, "loss": 0.5368, "step": 480 }, { "epoch": 7.65625, "grad_norm": 7.234042644500732, "learning_rate": 4.692063492063492e-05, "loss": 0.5803, "step": 490 }, { "epoch": 7.8125, "grad_norm": 6.273004055023193, "learning_rate": 4.684126984126984e-05, "loss": 0.519, "step": 500 }, { "epoch": 7.8125, "eval_runtime": 2.3062, "eval_samples_per_second": 0.434, "eval_steps_per_second": 0.434, "step": 500 }, { "epoch": 7.96875, "grad_norm": 4.803182125091553, "learning_rate": 4.676190476190476e-05, "loss": 0.5611, "step": 510 }, { "epoch": 8.125, "grad_norm": 4.977607727050781, "learning_rate": 4.668253968253968e-05, "loss": 0.3616, "step": 520 }, { "epoch": 8.28125, "grad_norm": 4.4240336418151855, "learning_rate": 4.6603174603174605e-05, "loss": 0.4303, "step": 530 }, { "epoch": 8.4375, "grad_norm": 5.406126022338867, "learning_rate": 4.6523809523809526e-05, "loss": 0.384, "step": 540 }, { "epoch": 8.59375, "grad_norm": 4.34033203125, "learning_rate": 4.644444444444445e-05, "loss": 0.3784, "step": 550 }, { "epoch": 8.75, "grad_norm": 5.256556034088135, "learning_rate": 4.636507936507937e-05, "loss": 0.4345, "step": 560 }, { "epoch": 8.90625, "grad_norm": 5.054710388183594, "learning_rate": 4.628571428571429e-05, "loss": 0.4257, "step": 570 }, { "epoch": 9.0625, "grad_norm": 5.220774173736572, "learning_rate": 4.6206349206349205e-05, "loss": 0.3146, "step": 580 }, { "epoch": 9.21875, "grad_norm": 4.420045375823975, "learning_rate": 4.612698412698413e-05, "loss": 0.298, "step": 590 }, { "epoch": 9.375, "grad_norm": 4.474491596221924, "learning_rate": 4.604761904761905e-05, "loss": 0.3023, "step": 600 }, { "epoch": 9.53125, "grad_norm": 4.131107807159424, "learning_rate": 4.596825396825397e-05, "loss": 0.2846, "step": 610 }, { "epoch": 9.6875, "grad_norm": 4.535623550415039, "learning_rate": 4.588888888888889e-05, "loss": 0.3244, "step": 620 }, { "epoch": 9.84375, "grad_norm": 4.365594387054443, "learning_rate": 4.580952380952381e-05, "loss": 0.2925, "step": 630 }, { "epoch": 10.0, "grad_norm": 7.354922294616699, "learning_rate": 4.5730158730158734e-05, "loss": 0.3234, "step": 640 }, { "epoch": 10.15625, "grad_norm": 3.916835308074951, "learning_rate": 4.5650793650793656e-05, "loss": 0.236, "step": 650 }, { "epoch": 10.3125, "grad_norm": 4.322419166564941, "learning_rate": 4.557142857142858e-05, "loss": 0.2353, "step": 660 }, { "epoch": 10.46875, "grad_norm": 5.4846038818359375, "learning_rate": 4.54920634920635e-05, "loss": 0.2422, "step": 670 }, { "epoch": 10.625, "grad_norm": 4.341657638549805, "learning_rate": 4.5412698412698414e-05, "loss": 0.2592, "step": 680 }, { "epoch": 10.78125, "grad_norm": 4.786070346832275, "learning_rate": 4.5333333333333335e-05, "loss": 0.2501, "step": 690 }, { "epoch": 10.9375, "grad_norm": 3.9264745712280273, "learning_rate": 4.525396825396826e-05, "loss": 0.2683, "step": 700 }, { "epoch": 11.09375, "grad_norm": 3.851856231689453, "learning_rate": 4.517460317460318e-05, "loss": 0.1978, "step": 710 }, { "epoch": 11.25, "grad_norm": 4.643581390380859, "learning_rate": 4.509523809523809e-05, "loss": 0.1841, "step": 720 }, { "epoch": 11.40625, "grad_norm": 4.105888843536377, "learning_rate": 4.5015873015873014e-05, "loss": 0.1976, "step": 730 }, { "epoch": 11.5625, "grad_norm": 3.809528350830078, "learning_rate": 4.4936507936507936e-05, "loss": 0.1896, "step": 740 }, { "epoch": 11.71875, "grad_norm": 4.457248687744141, "learning_rate": 4.485714285714286e-05, "loss": 0.2125, "step": 750 }, { "epoch": 11.875, "grad_norm": 3.7699806690216064, "learning_rate": 4.477777777777778e-05, "loss": 0.2212, "step": 760 }, { "epoch": 12.03125, "grad_norm": 4.643237590789795, "learning_rate": 4.46984126984127e-05, "loss": 0.2235, "step": 770 }, { "epoch": 12.1875, "grad_norm": 4.064443588256836, "learning_rate": 4.461904761904762e-05, "loss": 0.1503, "step": 780 }, { "epoch": 12.34375, "grad_norm": 4.605109691619873, "learning_rate": 4.4539682539682543e-05, "loss": 0.172, "step": 790 }, { "epoch": 12.5, "grad_norm": 3.8005576133728027, "learning_rate": 4.4460317460317465e-05, "loss": 0.1713, "step": 800 }, { "epoch": 12.65625, "grad_norm": 4.829875469207764, "learning_rate": 4.4380952380952386e-05, "loss": 0.1783, "step": 810 }, { "epoch": 12.8125, "grad_norm": 4.428802490234375, "learning_rate": 4.430158730158731e-05, "loss": 0.1742, "step": 820 }, { "epoch": 12.96875, "grad_norm": 5.068906307220459, "learning_rate": 4.422222222222222e-05, "loss": 0.1797, "step": 830 }, { "epoch": 13.125, "grad_norm": 3.3268795013427734, "learning_rate": 4.4142857142857144e-05, "loss": 0.1389, "step": 840 }, { "epoch": 13.28125, "grad_norm": 3.312582015991211, "learning_rate": 4.4063492063492066e-05, "loss": 0.1191, "step": 850 }, { "epoch": 13.4375, "grad_norm": 3.7299697399139404, "learning_rate": 4.398412698412699e-05, "loss": 0.1432, "step": 860 }, { "epoch": 13.59375, "grad_norm": 4.627827167510986, "learning_rate": 4.39047619047619e-05, "loss": 0.1672, "step": 870 }, { "epoch": 13.75, "grad_norm": 4.645421981811523, "learning_rate": 4.3825396825396823e-05, "loss": 0.1503, "step": 880 }, { "epoch": 13.90625, "grad_norm": 5.321810245513916, "learning_rate": 4.3746031746031745e-05, "loss": 0.1486, "step": 890 }, { "epoch": 14.0625, "grad_norm": 2.487105369567871, "learning_rate": 4.3666666666666666e-05, "loss": 0.1193, "step": 900 }, { "epoch": 14.21875, "grad_norm": 4.098151683807373, "learning_rate": 4.358730158730159e-05, "loss": 0.1182, "step": 910 }, { "epoch": 14.375, "grad_norm": 3.2538163661956787, "learning_rate": 4.350793650793651e-05, "loss": 0.1145, "step": 920 }, { "epoch": 14.53125, "grad_norm": 3.7881181240081787, "learning_rate": 4.342857142857143e-05, "loss": 0.1255, "step": 930 }, { "epoch": 14.6875, "grad_norm": 2.861689567565918, "learning_rate": 4.334920634920635e-05, "loss": 0.144, "step": 940 }, { "epoch": 14.84375, "grad_norm": 3.1621310710906982, "learning_rate": 4.3269841269841274e-05, "loss": 0.1293, "step": 950 }, { "epoch": 15.0, "grad_norm": 11.439355850219727, "learning_rate": 4.3190476190476195e-05, "loss": 0.1553, "step": 960 }, { "epoch": 15.15625, "grad_norm": 3.6791841983795166, "learning_rate": 4.311111111111111e-05, "loss": 0.0949, "step": 970 }, { "epoch": 15.3125, "grad_norm": 2.71702241897583, "learning_rate": 4.303174603174603e-05, "loss": 0.1125, "step": 980 }, { "epoch": 15.46875, "grad_norm": 2.713571786880493, "learning_rate": 4.295238095238095e-05, "loss": 0.137, "step": 990 }, { "epoch": 15.625, "grad_norm": 2.975768804550171, "learning_rate": 4.2873015873015875e-05, "loss": 0.108, "step": 1000 }, { "epoch": 15.625, "eval_runtime": 2.2622, "eval_samples_per_second": 0.442, "eval_steps_per_second": 0.442, "step": 1000 }, { "epoch": 15.78125, "grad_norm": 3.5744853019714355, "learning_rate": 4.2793650793650796e-05, "loss": 0.1207, "step": 1010 }, { "epoch": 15.9375, "grad_norm": 1.8845595121383667, "learning_rate": 4.271428571428572e-05, "loss": 0.1013, "step": 1020 }, { "epoch": 16.09375, "grad_norm": 4.227961540222168, "learning_rate": 4.263492063492064e-05, "loss": 0.112, "step": 1030 }, { "epoch": 16.25, "grad_norm": 3.289071798324585, "learning_rate": 4.255555555555556e-05, "loss": 0.1054, "step": 1040 }, { "epoch": 16.40625, "grad_norm": 1.9135278463363647, "learning_rate": 4.247619047619048e-05, "loss": 0.0933, "step": 1050 }, { "epoch": 16.5625, "grad_norm": 2.586151599884033, "learning_rate": 4.2396825396825404e-05, "loss": 0.1205, "step": 1060 }, { "epoch": 16.71875, "grad_norm": 3.5681819915771484, "learning_rate": 4.231746031746032e-05, "loss": 0.1015, "step": 1070 }, { "epoch": 16.875, "grad_norm": 3.7675039768218994, "learning_rate": 4.223809523809524e-05, "loss": 0.128, "step": 1080 }, { "epoch": 17.03125, "grad_norm": 2.631277322769165, "learning_rate": 4.215873015873016e-05, "loss": 0.107, "step": 1090 }, { "epoch": 17.1875, "grad_norm": 3.59251070022583, "learning_rate": 4.2079365079365076e-05, "loss": 0.0918, "step": 1100 }, { "epoch": 17.34375, "grad_norm": 3.785374641418457, "learning_rate": 4.2e-05, "loss": 0.1031, "step": 1110 }, { "epoch": 17.5, "grad_norm": 3.6201488971710205, "learning_rate": 4.192063492063492e-05, "loss": 0.0965, "step": 1120 }, { "epoch": 17.65625, "grad_norm": 3.608516216278076, "learning_rate": 4.184126984126984e-05, "loss": 0.1046, "step": 1130 }, { "epoch": 17.8125, "grad_norm": 3.3631608486175537, "learning_rate": 4.176190476190476e-05, "loss": 0.1047, "step": 1140 }, { "epoch": 17.96875, "grad_norm": 2.2599599361419678, "learning_rate": 4.1682539682539684e-05, "loss": 0.0976, "step": 1150 }, { "epoch": 18.125, "grad_norm": 3.3925111293792725, "learning_rate": 4.1603174603174605e-05, "loss": 0.081, "step": 1160 }, { "epoch": 18.28125, "grad_norm": 2.5443062782287598, "learning_rate": 4.152380952380953e-05, "loss": 0.0917, "step": 1170 }, { "epoch": 18.4375, "grad_norm": 4.180214881896973, "learning_rate": 4.144444444444445e-05, "loss": 0.0913, "step": 1180 }, { "epoch": 18.59375, "grad_norm": 2.3229408264160156, "learning_rate": 4.136507936507937e-05, "loss": 0.081, "step": 1190 }, { "epoch": 18.75, "grad_norm": 3.885908603668213, "learning_rate": 4.128571428571429e-05, "loss": 0.1078, "step": 1200 }, { "epoch": 18.90625, "grad_norm": 3.1222493648529053, "learning_rate": 4.120634920634921e-05, "loss": 0.1177, "step": 1210 }, { "epoch": 19.0625, "grad_norm": 3.3118879795074463, "learning_rate": 4.112698412698413e-05, "loss": 0.0775, "step": 1220 }, { "epoch": 19.21875, "grad_norm": 1.8756282329559326, "learning_rate": 4.104761904761905e-05, "loss": 0.0906, "step": 1230 }, { "epoch": 19.375, "grad_norm": 3.6675071716308594, "learning_rate": 4.096825396825397e-05, "loss": 0.0938, "step": 1240 }, { "epoch": 19.53125, "grad_norm": 2.8696744441986084, "learning_rate": 4.088888888888889e-05, "loss": 0.0821, "step": 1250 }, { "epoch": 19.6875, "grad_norm": 3.246438980102539, "learning_rate": 4.0809523809523813e-05, "loss": 0.1216, "step": 1260 }, { "epoch": 19.84375, "grad_norm": 2.579602003097534, "learning_rate": 4.073015873015873e-05, "loss": 0.0705, "step": 1270 }, { "epoch": 20.0, "grad_norm": 14.062989234924316, "learning_rate": 4.065079365079365e-05, "loss": 0.1125, "step": 1280 }, { "epoch": 20.15625, "grad_norm": 2.5907585620880127, "learning_rate": 4.057142857142857e-05, "loss": 0.0764, "step": 1290 }, { "epoch": 20.3125, "grad_norm": 4.850607872009277, "learning_rate": 4.049206349206349e-05, "loss": 0.0882, "step": 1300 }, { "epoch": 20.46875, "grad_norm": 2.3619115352630615, "learning_rate": 4.0412698412698414e-05, "loss": 0.0833, "step": 1310 }, { "epoch": 20.625, "grad_norm": 3.1803815364837646, "learning_rate": 4.0333333333333336e-05, "loss": 0.0894, "step": 1320 }, { "epoch": 20.78125, "grad_norm": 15.570907592773438, "learning_rate": 4.025396825396826e-05, "loss": 0.2225, "step": 1330 }, { "epoch": 20.9375, "grad_norm": 2.721440076828003, "learning_rate": 4.018253968253968e-05, "loss": 0.1233, "step": 1340 }, { "epoch": 21.09375, "grad_norm": 1.798134446144104, "learning_rate": 4.01031746031746e-05, "loss": 0.067, "step": 1350 }, { "epoch": 21.25, "grad_norm": 3.387782335281372, "learning_rate": 4.0023809523809524e-05, "loss": 0.0692, "step": 1360 }, { "epoch": 21.40625, "grad_norm": 2.723196029663086, "learning_rate": 3.9944444444444446e-05, "loss": 0.0683, "step": 1370 }, { "epoch": 21.5625, "grad_norm": 3.666444778442383, "learning_rate": 3.986507936507937e-05, "loss": 0.0883, "step": 1380 }, { "epoch": 21.71875, "grad_norm": 1.9571526050567627, "learning_rate": 3.978571428571429e-05, "loss": 0.0802, "step": 1390 }, { "epoch": 21.875, "grad_norm": 2.5271904468536377, "learning_rate": 3.970634920634921e-05, "loss": 0.0737, "step": 1400 }, { "epoch": 22.03125, "grad_norm": 3.5426900386810303, "learning_rate": 3.962698412698413e-05, "loss": 0.0838, "step": 1410 }, { "epoch": 22.1875, "grad_norm": 1.7901580333709717, "learning_rate": 3.954761904761905e-05, "loss": 0.0671, "step": 1420 }, { "epoch": 22.34375, "grad_norm": 3.4260764122009277, "learning_rate": 3.946825396825397e-05, "loss": 0.0775, "step": 1430 }, { "epoch": 22.5, "grad_norm": 2.499107837677002, "learning_rate": 3.938888888888889e-05, "loss": 0.0826, "step": 1440 }, { "epoch": 22.65625, "grad_norm": 2.7331862449645996, "learning_rate": 3.930952380952381e-05, "loss": 0.0771, "step": 1450 }, { "epoch": 22.8125, "grad_norm": 3.2004685401916504, "learning_rate": 3.923015873015873e-05, "loss": 0.088, "step": 1460 }, { "epoch": 22.96875, "grad_norm": 3.175179958343506, "learning_rate": 3.9150793650793654e-05, "loss": 0.0861, "step": 1470 }, { "epoch": 23.125, "grad_norm": 2.032646417617798, "learning_rate": 3.9071428571428575e-05, "loss": 0.0682, "step": 1480 }, { "epoch": 23.28125, "grad_norm": 1.5635634660720825, "learning_rate": 3.89920634920635e-05, "loss": 0.0833, "step": 1490 }, { "epoch": 23.4375, "grad_norm": 1.8121321201324463, "learning_rate": 3.891269841269842e-05, "loss": 0.0965, "step": 1500 }, { "epoch": 23.4375, "eval_runtime": 2.0794, "eval_samples_per_second": 0.481, "eval_steps_per_second": 0.481, "step": 1500 }, { "epoch": 23.59375, "grad_norm": 1.8793394565582275, "learning_rate": 3.883333333333333e-05, "loss": 0.0887, "step": 1510 }, { "epoch": 23.75, "grad_norm": 2.1231632232666016, "learning_rate": 3.8753968253968255e-05, "loss": 0.0879, "step": 1520 }, { "epoch": 23.90625, "grad_norm": 3.5764803886413574, "learning_rate": 3.8674603174603176e-05, "loss": 0.0861, "step": 1530 }, { "epoch": 24.0625, "grad_norm": 2.105710029602051, "learning_rate": 3.85952380952381e-05, "loss": 0.07, "step": 1540 }, { "epoch": 24.21875, "grad_norm": 2.8722870349884033, "learning_rate": 3.851587301587302e-05, "loss": 0.076, "step": 1550 }, { "epoch": 24.375, "grad_norm": 2.6126277446746826, "learning_rate": 3.843650793650794e-05, "loss": 0.0983, "step": 1560 }, { "epoch": 24.53125, "grad_norm": 2.2635769844055176, "learning_rate": 3.8357142857142855e-05, "loss": 0.0864, "step": 1570 }, { "epoch": 24.6875, "grad_norm": 2.11098313331604, "learning_rate": 3.827777777777778e-05, "loss": 0.0923, "step": 1580 }, { "epoch": 24.84375, "grad_norm": 1.6163533926010132, "learning_rate": 3.81984126984127e-05, "loss": 0.0711, "step": 1590 }, { "epoch": 25.0, "grad_norm": 2.54638671875, "learning_rate": 3.811904761904762e-05, "loss": 0.0705, "step": 1600 }, { "epoch": 25.15625, "grad_norm": 2.592470407485962, "learning_rate": 3.803968253968254e-05, "loss": 0.0765, "step": 1610 }, { "epoch": 25.3125, "grad_norm": 2.1981208324432373, "learning_rate": 3.796031746031746e-05, "loss": 0.0856, "step": 1620 }, { "epoch": 25.46875, "grad_norm": 1.384098768234253, "learning_rate": 3.7880952380952384e-05, "loss": 0.0951, "step": 1630 }, { "epoch": 25.625, "grad_norm": 1.502350091934204, "learning_rate": 3.7801587301587306e-05, "loss": 0.0858, "step": 1640 }, { "epoch": 25.78125, "grad_norm": 1.4763522148132324, "learning_rate": 3.772222222222223e-05, "loss": 0.0647, "step": 1650 }, { "epoch": 25.9375, "grad_norm": 1.7555052042007446, "learning_rate": 3.764285714285715e-05, "loss": 0.0874, "step": 1660 }, { "epoch": 26.09375, "grad_norm": 1.0187015533447266, "learning_rate": 3.756349206349207e-05, "loss": 0.0635, "step": 1670 }, { "epoch": 26.25, "grad_norm": 1.7138936519622803, "learning_rate": 3.7484126984126985e-05, "loss": 0.0772, "step": 1680 }, { "epoch": 26.40625, "grad_norm": 2.3353724479675293, "learning_rate": 3.7404761904761907e-05, "loss": 0.0892, "step": 1690 }, { "epoch": 26.5625, "grad_norm": 2.6141700744628906, "learning_rate": 3.732539682539682e-05, "loss": 0.0897, "step": 1700 }, { "epoch": 26.71875, "grad_norm": 1.7785848379135132, "learning_rate": 3.724603174603174e-05, "loss": 0.0899, "step": 1710 }, { "epoch": 26.875, "grad_norm": 2.6693010330200195, "learning_rate": 3.7166666666666664e-05, "loss": 0.0901, "step": 1720 }, { "epoch": 27.03125, "grad_norm": 2.059981346130371, "learning_rate": 3.7087301587301586e-05, "loss": 0.0822, "step": 1730 }, { "epoch": 27.1875, "grad_norm": 1.6238901615142822, "learning_rate": 3.700793650793651e-05, "loss": 0.0777, "step": 1740 }, { "epoch": 27.34375, "grad_norm": 2.782425880432129, "learning_rate": 3.692857142857143e-05, "loss": 0.0782, "step": 1750 }, { "epoch": 27.5, "grad_norm": 1.8468166589736938, "learning_rate": 3.684920634920635e-05, "loss": 0.0835, "step": 1760 }, { "epoch": 27.65625, "grad_norm": 1.3156135082244873, "learning_rate": 3.676984126984127e-05, "loss": 0.1072, "step": 1770 }, { "epoch": 27.8125, "grad_norm": 3.260084390640259, "learning_rate": 3.669047619047619e-05, "loss": 0.0939, "step": 1780 }, { "epoch": 27.96875, "grad_norm": 2.518204689025879, "learning_rate": 3.6611111111111115e-05, "loss": 0.0884, "step": 1790 }, { "epoch": 28.125, "grad_norm": 2.598057985305786, "learning_rate": 3.6531746031746036e-05, "loss": 0.0683, "step": 1800 }, { "epoch": 28.28125, "grad_norm": 1.8533433675765991, "learning_rate": 3.645238095238096e-05, "loss": 0.0833, "step": 1810 }, { "epoch": 28.4375, "grad_norm": 1.2828975915908813, "learning_rate": 3.637301587301587e-05, "loss": 0.0832, "step": 1820 }, { "epoch": 28.59375, "grad_norm": 1.7714905738830566, "learning_rate": 3.6293650793650794e-05, "loss": 0.0901, "step": 1830 }, { "epoch": 28.75, "grad_norm": 2.098923921585083, "learning_rate": 3.6214285714285716e-05, "loss": 0.0954, "step": 1840 }, { "epoch": 28.90625, "grad_norm": 2.298226833343506, "learning_rate": 3.613492063492064e-05, "loss": 0.0888, "step": 1850 }, { "epoch": 29.0625, "grad_norm": 1.5519624948501587, "learning_rate": 3.605555555555556e-05, "loss": 0.0664, "step": 1860 }, { "epoch": 29.21875, "grad_norm": 2.015573501586914, "learning_rate": 3.597619047619048e-05, "loss": 0.0637, "step": 1870 }, { "epoch": 29.375, "grad_norm": 1.925529956817627, "learning_rate": 3.58968253968254e-05, "loss": 0.0803, "step": 1880 }, { "epoch": 29.53125, "grad_norm": 2.4342522621154785, "learning_rate": 3.581746031746032e-05, "loss": 0.0911, "step": 1890 }, { "epoch": 29.6875, "grad_norm": 1.8124195337295532, "learning_rate": 3.573809523809524e-05, "loss": 0.0803, "step": 1900 }, { "epoch": 29.84375, "grad_norm": 2.3409860134124756, "learning_rate": 3.565873015873016e-05, "loss": 0.0943, "step": 1910 }, { "epoch": 30.0, "grad_norm": 1.2545162439346313, "learning_rate": 3.557936507936508e-05, "loss": 0.0853, "step": 1920 }, { "epoch": 30.15625, "grad_norm": 1.9356091022491455, "learning_rate": 3.55e-05, "loss": 0.0585, "step": 1930 }, { "epoch": 30.3125, "grad_norm": 1.6294385194778442, "learning_rate": 3.5420634920634924e-05, "loss": 0.0876, "step": 1940 }, { "epoch": 30.46875, "grad_norm": 2.081688165664673, "learning_rate": 3.534126984126984e-05, "loss": 0.0827, "step": 1950 }, { "epoch": 30.625, "grad_norm": 1.7468382120132446, "learning_rate": 3.526190476190476e-05, "loss": 0.0732, "step": 1960 }, { "epoch": 30.78125, "grad_norm": 2.1742124557495117, "learning_rate": 3.518253968253968e-05, "loss": 0.084, "step": 1970 }, { "epoch": 30.9375, "grad_norm": 2.147754430770874, "learning_rate": 3.51031746031746e-05, "loss": 0.0856, "step": 1980 }, { "epoch": 31.09375, "grad_norm": 2.1251964569091797, "learning_rate": 3.5023809523809525e-05, "loss": 0.0762, "step": 1990 }, { "epoch": 31.25, "grad_norm": 2.6378941535949707, "learning_rate": 3.4944444444444446e-05, "loss": 0.071, "step": 2000 }, { "epoch": 31.25, "eval_runtime": 2.0743, "eval_samples_per_second": 0.482, "eval_steps_per_second": 0.482, "step": 2000 }, { "epoch": 31.40625, "grad_norm": 2.126807689666748, "learning_rate": 3.486507936507937e-05, "loss": 0.0785, "step": 2010 }, { "epoch": 31.5625, "grad_norm": 2.4585835933685303, "learning_rate": 3.478571428571429e-05, "loss": 0.1053, "step": 2020 }, { "epoch": 31.71875, "grad_norm": 1.9649542570114136, "learning_rate": 3.470634920634921e-05, "loss": 0.0734, "step": 2030 }, { "epoch": 31.875, "grad_norm": 2.3006460666656494, "learning_rate": 3.462698412698413e-05, "loss": 0.0858, "step": 2040 }, { "epoch": 32.03125, "grad_norm": 2.163447856903076, "learning_rate": 3.4547619047619054e-05, "loss": 0.0826, "step": 2050 }, { "epoch": 32.1875, "grad_norm": 2.0441508293151855, "learning_rate": 3.4468253968253975e-05, "loss": 0.0712, "step": 2060 }, { "epoch": 32.34375, "grad_norm": 1.5795445442199707, "learning_rate": 3.438888888888889e-05, "loss": 0.056, "step": 2070 }, { "epoch": 32.5, "grad_norm": 3.5661540031433105, "learning_rate": 3.430952380952381e-05, "loss": 0.069, "step": 2080 }, { "epoch": 32.65625, "grad_norm": 1.6884055137634277, "learning_rate": 3.423015873015873e-05, "loss": 0.0899, "step": 2090 }, { "epoch": 32.8125, "grad_norm": 2.421724557876587, "learning_rate": 3.415079365079365e-05, "loss": 0.0768, "step": 2100 }, { "epoch": 32.96875, "grad_norm": 1.1584899425506592, "learning_rate": 3.407142857142857e-05, "loss": 0.0763, "step": 2110 }, { "epoch": 33.125, "grad_norm": 2.069801092147827, "learning_rate": 3.399206349206349e-05, "loss": 0.0636, "step": 2120 }, { "epoch": 33.28125, "grad_norm": 2.5826573371887207, "learning_rate": 3.391269841269841e-05, "loss": 0.0676, "step": 2130 }, { "epoch": 33.4375, "grad_norm": 1.119449496269226, "learning_rate": 3.3833333333333334e-05, "loss": 0.0613, "step": 2140 }, { "epoch": 33.59375, "grad_norm": 1.6656103134155273, "learning_rate": 3.3753968253968255e-05, "loss": 0.0598, "step": 2150 }, { "epoch": 33.75, "grad_norm": 1.7956265211105347, "learning_rate": 3.3674603174603177e-05, "loss": 0.0778, "step": 2160 }, { "epoch": 33.90625, "grad_norm": 2.2106685638427734, "learning_rate": 3.35952380952381e-05, "loss": 0.0737, "step": 2170 }, { "epoch": 34.0625, "grad_norm": 2.4478724002838135, "learning_rate": 3.351587301587302e-05, "loss": 0.0642, "step": 2180 }, { "epoch": 34.21875, "grad_norm": 2.3334341049194336, "learning_rate": 3.343650793650794e-05, "loss": 0.0697, "step": 2190 }, { "epoch": 34.375, "grad_norm": 1.870275855064392, "learning_rate": 3.3357142857142856e-05, "loss": 0.0523, "step": 2200 }, { "epoch": 34.53125, "grad_norm": 1.283444881439209, "learning_rate": 3.327777777777778e-05, "loss": 0.0619, "step": 2210 }, { "epoch": 34.6875, "grad_norm": 1.7918671369552612, "learning_rate": 3.31984126984127e-05, "loss": 0.061, "step": 2220 }, { "epoch": 34.84375, "grad_norm": 1.6546680927276611, "learning_rate": 3.311904761904762e-05, "loss": 0.058, "step": 2230 }, { "epoch": 35.0, "grad_norm": 3.8063647747039795, "learning_rate": 3.303968253968254e-05, "loss": 0.0605, "step": 2240 }, { "epoch": 35.15625, "grad_norm": 1.8701483011245728, "learning_rate": 3.296031746031746e-05, "loss": 0.0652, "step": 2250 }, { "epoch": 35.3125, "grad_norm": 0.48190346360206604, "learning_rate": 3.2880952380952385e-05, "loss": 0.0443, "step": 2260 }, { "epoch": 35.46875, "grad_norm": 1.0227997303009033, "learning_rate": 3.2801587301587306e-05, "loss": 0.0397, "step": 2270 }, { "epoch": 35.625, "grad_norm": 1.4256937503814697, "learning_rate": 3.272222222222223e-05, "loss": 0.0541, "step": 2280 }, { "epoch": 35.78125, "grad_norm": 2.3451199531555176, "learning_rate": 3.264285714285714e-05, "loss": 0.0568, "step": 2290 }, { "epoch": 35.9375, "grad_norm": 1.3683526515960693, "learning_rate": 3.2563492063492064e-05, "loss": 0.0587, "step": 2300 }, { "epoch": 36.09375, "grad_norm": 1.4543867111206055, "learning_rate": 3.2484126984126986e-05, "loss": 0.0394, "step": 2310 }, { "epoch": 36.25, "grad_norm": 1.9606877565383911, "learning_rate": 3.240476190476191e-05, "loss": 0.0545, "step": 2320 }, { "epoch": 36.40625, "grad_norm": 1.6910959482192993, "learning_rate": 3.232539682539683e-05, "loss": 0.0638, "step": 2330 }, { "epoch": 36.5625, "grad_norm": 1.735841155052185, "learning_rate": 3.224603174603174e-05, "loss": 0.0569, "step": 2340 }, { "epoch": 36.71875, "grad_norm": 1.5598944425582886, "learning_rate": 3.2166666666666665e-05, "loss": 0.0649, "step": 2350 }, { "epoch": 36.875, "grad_norm": 1.5700335502624512, "learning_rate": 3.2087301587301586e-05, "loss": 0.0689, "step": 2360 }, { "epoch": 37.03125, "grad_norm": 1.0767812728881836, "learning_rate": 3.200793650793651e-05, "loss": 0.0535, "step": 2370 }, { "epoch": 37.1875, "grad_norm": 2.505707263946533, "learning_rate": 3.192857142857143e-05, "loss": 0.0486, "step": 2380 }, { "epoch": 37.34375, "grad_norm": 2.1414477825164795, "learning_rate": 3.184920634920635e-05, "loss": 0.0462, "step": 2390 }, { "epoch": 37.5, "grad_norm": 1.0493581295013428, "learning_rate": 3.176984126984127e-05, "loss": 0.0515, "step": 2400 }, { "epoch": 37.65625, "grad_norm": 1.5255446434020996, "learning_rate": 3.1690476190476194e-05, "loss": 0.0485, "step": 2410 }, { "epoch": 37.8125, "grad_norm": 1.3334754705429077, "learning_rate": 3.1611111111111115e-05, "loss": 0.056, "step": 2420 }, { "epoch": 37.96875, "grad_norm": 1.745110273361206, "learning_rate": 3.153174603174604e-05, "loss": 0.0473, "step": 2430 }, { "epoch": 38.125, "grad_norm": 1.4413131475448608, "learning_rate": 3.145238095238096e-05, "loss": 0.0317, "step": 2440 }, { "epoch": 38.28125, "grad_norm": 1.6811962127685547, "learning_rate": 3.137301587301587e-05, "loss": 0.0496, "step": 2450 }, { "epoch": 38.4375, "grad_norm": 1.2820957899093628, "learning_rate": 3.1293650793650795e-05, "loss": 0.0369, "step": 2460 }, { "epoch": 38.59375, "grad_norm": 1.7850005626678467, "learning_rate": 3.1214285714285716e-05, "loss": 0.0478, "step": 2470 }, { "epoch": 38.75, "grad_norm": 2.456017255783081, "learning_rate": 3.113492063492064e-05, "loss": 0.0507, "step": 2480 }, { "epoch": 38.90625, "grad_norm": 2.1933865547180176, "learning_rate": 3.105555555555555e-05, "loss": 0.0436, "step": 2490 }, { "epoch": 39.0625, "grad_norm": 0.9716876745223999, "learning_rate": 3.0976190476190474e-05, "loss": 0.0439, "step": 2500 }, { "epoch": 39.0625, "eval_runtime": 2.0072, "eval_samples_per_second": 0.498, "eval_steps_per_second": 0.498, "step": 2500 }, { "epoch": 39.21875, "grad_norm": 2.0628113746643066, "learning_rate": 3.0896825396825395e-05, "loss": 0.0408, "step": 2510 }, { "epoch": 39.375, "grad_norm": 1.5137745141983032, "learning_rate": 3.081746031746032e-05, "loss": 0.0445, "step": 2520 }, { "epoch": 39.53125, "grad_norm": 1.6775884628295898, "learning_rate": 3.073809523809524e-05, "loss": 0.051, "step": 2530 }, { "epoch": 39.6875, "grad_norm": 1.440619707107544, "learning_rate": 3.065873015873016e-05, "loss": 0.0387, "step": 2540 }, { "epoch": 39.84375, "grad_norm": 2.1149260997772217, "learning_rate": 3.057936507936508e-05, "loss": 0.0387, "step": 2550 }, { "epoch": 40.0, "grad_norm": 2.163191318511963, "learning_rate": 3.05e-05, "loss": 0.0363, "step": 2560 }, { "epoch": 40.15625, "grad_norm": 3.246634006500244, "learning_rate": 3.042063492063492e-05, "loss": 0.0429, "step": 2570 }, { "epoch": 40.3125, "grad_norm": 1.188644289970398, "learning_rate": 3.0341269841269842e-05, "loss": 0.0316, "step": 2580 }, { "epoch": 40.46875, "grad_norm": 2.2617383003234863, "learning_rate": 3.0261904761904764e-05, "loss": 0.0406, "step": 2590 }, { "epoch": 40.625, "grad_norm": 0.6037064790725708, "learning_rate": 3.0182539682539685e-05, "loss": 0.031, "step": 2600 }, { "epoch": 40.78125, "grad_norm": 1.1936763525009155, "learning_rate": 3.0103174603174607e-05, "loss": 0.0401, "step": 2610 }, { "epoch": 40.9375, "grad_norm": 3.0310215950012207, "learning_rate": 3.0023809523809525e-05, "loss": 0.0339, "step": 2620 }, { "epoch": 41.09375, "grad_norm": 0.6716585755348206, "learning_rate": 2.9944444444444446e-05, "loss": 0.0338, "step": 2630 }, { "epoch": 41.25, "grad_norm": 0.5010519623756409, "learning_rate": 2.9865079365079368e-05, "loss": 0.0315, "step": 2640 } ], "logging_steps": 10, "max_steps": 6400, "num_input_tokens_seen": 0, "num_train_epochs": 100, "save_steps": 115, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }