{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 5000, "global_step": 2250, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0044444444444444444, "grad_norm": 172.13922119140625, "learning_rate": 5.000000000000001e-07, "loss": 11.5329, "step": 10 }, { "epoch": 0.008888888888888889, "grad_norm": 179.96237182617188, "learning_rate": 1.5e-06, "loss": 11.3904, "step": 20 }, { "epoch": 0.013333333333333334, "grad_norm": 1364.69921875, "learning_rate": 2.5e-06, "loss": 10.7917, "step": 30 }, { "epoch": 0.017777777777777778, "grad_norm": 8219.5224609375, "learning_rate": 3.3e-06, "loss": 9.1512, "step": 40 }, { "epoch": 0.022222222222222223, "grad_norm": 12965.9892578125, "learning_rate": 4.2999999999999995e-06, "loss": 7.8213, "step": 50 }, { "epoch": 0.02666666666666667, "grad_norm": 31872.74609375, "learning_rate": 5.3e-06, "loss": 6.6272, "step": 60 }, { "epoch": 0.03111111111111111, "grad_norm": 788.8919677734375, "learning_rate": 6.300000000000001e-06, "loss": 6.3266, "step": 70 }, { "epoch": 0.035555555555555556, "grad_norm": 106.24903869628906, "learning_rate": 7.2999999999999996e-06, "loss": 5.7356, "step": 80 }, { "epoch": 0.04, "grad_norm": 409.2347412109375, "learning_rate": 8.3e-06, "loss": 5.018, "step": 90 }, { "epoch": 0.044444444444444446, "grad_norm": 817.6410522460938, "learning_rate": 9.3e-06, "loss": 5.422, "step": 100 }, { "epoch": 0.04888888888888889, "grad_norm": 589.453125, "learning_rate": 1.03e-05, "loss": 3.6104, "step": 110 }, { "epoch": 0.05333333333333334, "grad_norm": 299.5021667480469, "learning_rate": 1.13e-05, "loss": 3.3433, "step": 120 }, { "epoch": 0.057777777777777775, "grad_norm": 105.66629791259766, "learning_rate": 1.23e-05, "loss": 2.1184, "step": 130 }, { "epoch": 0.06222222222222222, "grad_norm": 94.36296844482422, "learning_rate": 1.3300000000000001e-05, "loss": 1.6441, "step": 140 }, { "epoch": 0.06666666666666667, "grad_norm": 99.46455383300781, "learning_rate": 1.43e-05, "loss": 1.0691, "step": 150 }, { "epoch": 0.07111111111111111, "grad_norm": 128.905517578125, "learning_rate": 1.53e-05, "loss": 1.347, "step": 160 }, { "epoch": 0.07555555555555556, "grad_norm": 10.669831275939941, "learning_rate": 1.63e-05, "loss": 0.8163, "step": 170 }, { "epoch": 0.08, "grad_norm": 17.684833526611328, "learning_rate": 1.73e-05, "loss": 1.0315, "step": 180 }, { "epoch": 0.08444444444444445, "grad_norm": 35.8447151184082, "learning_rate": 1.83e-05, "loss": 0.6946, "step": 190 }, { "epoch": 0.08888888888888889, "grad_norm": 88.678466796875, "learning_rate": 1.93e-05, "loss": 1.0834, "step": 200 }, { "epoch": 0.09333333333333334, "grad_norm": 50.5704231262207, "learning_rate": 2.0300000000000002e-05, "loss": 0.9062, "step": 210 }, { "epoch": 0.09777777777777778, "grad_norm": 5.56904935836792, "learning_rate": 2.13e-05, "loss": 0.9961, "step": 220 }, { "epoch": 0.10222222222222223, "grad_norm": 2.848574638366699, "learning_rate": 2.23e-05, "loss": 0.7547, "step": 230 }, { "epoch": 0.10666666666666667, "grad_norm": 51.646854400634766, "learning_rate": 2.3300000000000004e-05, "loss": 0.9434, "step": 240 }, { "epoch": 0.1111111111111111, "grad_norm": 4.968932151794434, "learning_rate": 2.43e-05, "loss": 0.9282, "step": 250 }, { "epoch": 0.11555555555555555, "grad_norm": 8.864336967468262, "learning_rate": 2.5300000000000002e-05, "loss": 1.1923, "step": 260 }, { "epoch": 0.12, "grad_norm": 6.225351333618164, "learning_rate": 2.6300000000000002e-05, "loss": 0.7519, "step": 270 }, { "epoch": 0.12444444444444444, "grad_norm": 16.043298721313477, "learning_rate": 2.7300000000000003e-05, "loss": 0.8965, "step": 280 }, { "epoch": 0.1288888888888889, "grad_norm": 13.324910163879395, "learning_rate": 2.83e-05, "loss": 0.8238, "step": 290 }, { "epoch": 0.13333333333333333, "grad_norm": 3.661782741546631, "learning_rate": 2.93e-05, "loss": 0.6619, "step": 300 }, { "epoch": 0.13777777777777778, "grad_norm": 24.01588249206543, "learning_rate": 3.03e-05, "loss": 0.9874, "step": 310 }, { "epoch": 0.14222222222222222, "grad_norm": 3.4116015434265137, "learning_rate": 3.13e-05, "loss": 0.7866, "step": 320 }, { "epoch": 0.14666666666666667, "grad_norm": 5.378733158111572, "learning_rate": 3.2300000000000006e-05, "loss": 0.784, "step": 330 }, { "epoch": 0.1511111111111111, "grad_norm": 3.0666747093200684, "learning_rate": 3.33e-05, "loss": 0.8068, "step": 340 }, { "epoch": 0.15555555555555556, "grad_norm": 5.233002662658691, "learning_rate": 3.430000000000001e-05, "loss": 0.8338, "step": 350 }, { "epoch": 0.16, "grad_norm": 12.038925170898438, "learning_rate": 3.53e-05, "loss": 0.9978, "step": 360 }, { "epoch": 0.16444444444444445, "grad_norm": 4.42678165435791, "learning_rate": 3.63e-05, "loss": 0.9186, "step": 370 }, { "epoch": 0.1688888888888889, "grad_norm": 3.2533979415893555, "learning_rate": 3.73e-05, "loss": 0.853, "step": 380 }, { "epoch": 0.17333333333333334, "grad_norm": 6.089888095855713, "learning_rate": 3.83e-05, "loss": 0.6885, "step": 390 }, { "epoch": 0.17777777777777778, "grad_norm": 2.185434341430664, "learning_rate": 3.9300000000000007e-05, "loss": 0.7051, "step": 400 }, { "epoch": 0.18222222222222223, "grad_norm": 3.463395118713379, "learning_rate": 4.0300000000000004e-05, "loss": 0.7594, "step": 410 }, { "epoch": 0.18666666666666668, "grad_norm": 111.5975341796875, "learning_rate": 4.13e-05, "loss": 0.7679, "step": 420 }, { "epoch": 0.19111111111111112, "grad_norm": 2.420258045196533, "learning_rate": 4.23e-05, "loss": 0.9897, "step": 430 }, { "epoch": 0.19555555555555557, "grad_norm": 3.507974863052368, "learning_rate": 4.33e-05, "loss": 0.7764, "step": 440 }, { "epoch": 0.2, "grad_norm": 1.6020970344543457, "learning_rate": 4.43e-05, "loss": 0.6163, "step": 450 }, { "epoch": 0.20444444444444446, "grad_norm": 1.9605637788772583, "learning_rate": 4.53e-05, "loss": 0.9071, "step": 460 }, { "epoch": 0.2088888888888889, "grad_norm": 6.756528377532959, "learning_rate": 4.630000000000001e-05, "loss": 0.939, "step": 470 }, { "epoch": 0.21333333333333335, "grad_norm": 6.34450101852417, "learning_rate": 4.73e-05, "loss": 0.832, "step": 480 }, { "epoch": 0.21777777777777776, "grad_norm": 3.8294358253479004, "learning_rate": 4.83e-05, "loss": 0.9205, "step": 490 }, { "epoch": 0.2222222222222222, "grad_norm": 2.1834263801574707, "learning_rate": 4.93e-05, "loss": 0.6335, "step": 500 }, { "epoch": 0.22666666666666666, "grad_norm": 2.837082624435425, "learning_rate": 4.9914285714285717e-05, "loss": 0.6933, "step": 510 }, { "epoch": 0.2311111111111111, "grad_norm": 1.8396414518356323, "learning_rate": 4.962857142857143e-05, "loss": 1.0322, "step": 520 }, { "epoch": 0.23555555555555555, "grad_norm": 4.330628871917725, "learning_rate": 4.934285714285715e-05, "loss": 0.7124, "step": 530 }, { "epoch": 0.24, "grad_norm": 2.490299701690674, "learning_rate": 4.905714285714286e-05, "loss": 0.7426, "step": 540 }, { "epoch": 0.24444444444444444, "grad_norm": 1.341838002204895, "learning_rate": 4.8771428571428574e-05, "loss": 0.7588, "step": 550 }, { "epoch": 0.24888888888888888, "grad_norm": 2.234884262084961, "learning_rate": 4.848571428571429e-05, "loss": 0.8314, "step": 560 }, { "epoch": 0.25333333333333335, "grad_norm": 2.8187923431396484, "learning_rate": 4.82e-05, "loss": 0.8109, "step": 570 }, { "epoch": 0.2577777777777778, "grad_norm": 0.9691933393478394, "learning_rate": 4.7914285714285715e-05, "loss": 0.6785, "step": 580 }, { "epoch": 0.26222222222222225, "grad_norm": 1.8897961378097534, "learning_rate": 4.762857142857143e-05, "loss": 0.9523, "step": 590 }, { "epoch": 0.26666666666666666, "grad_norm": 7.316945552825928, "learning_rate": 4.734285714285715e-05, "loss": 0.5996, "step": 600 }, { "epoch": 0.27111111111111114, "grad_norm": 2.491607427597046, "learning_rate": 4.7057142857142864e-05, "loss": 0.5426, "step": 610 }, { "epoch": 0.27555555555555555, "grad_norm": 2.7054076194763184, "learning_rate": 4.677142857142857e-05, "loss": 0.6294, "step": 620 }, { "epoch": 0.28, "grad_norm": 1.5034236907958984, "learning_rate": 4.648571428571429e-05, "loss": 0.4571, "step": 630 }, { "epoch": 0.28444444444444444, "grad_norm": 5.013509750366211, "learning_rate": 4.6200000000000005e-05, "loss": 0.9696, "step": 640 }, { "epoch": 0.28888888888888886, "grad_norm": 1.4784796237945557, "learning_rate": 4.5914285714285714e-05, "loss": 0.79, "step": 650 }, { "epoch": 0.29333333333333333, "grad_norm": 1.679485559463501, "learning_rate": 4.562857142857143e-05, "loss": 0.5445, "step": 660 }, { "epoch": 0.29777777777777775, "grad_norm": 11.93566608428955, "learning_rate": 4.534285714285714e-05, "loss": 0.8077, "step": 670 }, { "epoch": 0.3022222222222222, "grad_norm": 2.100651264190674, "learning_rate": 4.5057142857142856e-05, "loss": 0.5643, "step": 680 }, { "epoch": 0.30666666666666664, "grad_norm": 2.368565797805786, "learning_rate": 4.477142857142858e-05, "loss": 1.0012, "step": 690 }, { "epoch": 0.3111111111111111, "grad_norm": 2.4255175590515137, "learning_rate": 4.448571428571429e-05, "loss": 0.6344, "step": 700 }, { "epoch": 0.31555555555555553, "grad_norm": 3.4809205532073975, "learning_rate": 4.4200000000000004e-05, "loss": 0.87, "step": 710 }, { "epoch": 0.32, "grad_norm": 12.76645565032959, "learning_rate": 4.391428571428572e-05, "loss": 0.673, "step": 720 }, { "epoch": 0.3244444444444444, "grad_norm": 3.417945623397827, "learning_rate": 4.362857142857143e-05, "loss": 0.6753, "step": 730 }, { "epoch": 0.3288888888888889, "grad_norm": 2.9550516605377197, "learning_rate": 4.3342857142857145e-05, "loss": 0.6621, "step": 740 }, { "epoch": 0.3333333333333333, "grad_norm": 1.9887003898620605, "learning_rate": 4.3057142857142854e-05, "loss": 0.671, "step": 750 }, { "epoch": 0.3377777777777778, "grad_norm": 1.826278567314148, "learning_rate": 4.277142857142857e-05, "loss": 0.7536, "step": 760 }, { "epoch": 0.3422222222222222, "grad_norm": 4.41416597366333, "learning_rate": 4.2485714285714286e-05, "loss": 0.7106, "step": 770 }, { "epoch": 0.3466666666666667, "grad_norm": 1.8409727811813354, "learning_rate": 4.22e-05, "loss": 0.7142, "step": 780 }, { "epoch": 0.3511111111111111, "grad_norm": 2.143974781036377, "learning_rate": 4.191428571428572e-05, "loss": 0.5623, "step": 790 }, { "epoch": 0.35555555555555557, "grad_norm": 1.1921825408935547, "learning_rate": 4.162857142857143e-05, "loss": 0.6638, "step": 800 }, { "epoch": 0.36, "grad_norm": 2.4289398193359375, "learning_rate": 4.1342857142857144e-05, "loss": 0.503, "step": 810 }, { "epoch": 0.36444444444444446, "grad_norm": 1.718072533607483, "learning_rate": 4.105714285714286e-05, "loss": 0.4197, "step": 820 }, { "epoch": 0.3688888888888889, "grad_norm": 2.278563976287842, "learning_rate": 4.077142857142857e-05, "loss": 0.611, "step": 830 }, { "epoch": 0.37333333333333335, "grad_norm": 2.060332775115967, "learning_rate": 4.0485714285714285e-05, "loss": 0.5883, "step": 840 }, { "epoch": 0.37777777777777777, "grad_norm": 2.5066239833831787, "learning_rate": 4.02e-05, "loss": 0.6179, "step": 850 }, { "epoch": 0.38222222222222224, "grad_norm": 5.195372104644775, "learning_rate": 3.991428571428572e-05, "loss": 0.7709, "step": 860 }, { "epoch": 0.38666666666666666, "grad_norm": 3.0029399394989014, "learning_rate": 3.9628571428571433e-05, "loss": 0.584, "step": 870 }, { "epoch": 0.39111111111111113, "grad_norm": 6.307543754577637, "learning_rate": 3.934285714285714e-05, "loss": 0.6633, "step": 880 }, { "epoch": 0.39555555555555555, "grad_norm": 6.635744571685791, "learning_rate": 3.905714285714286e-05, "loss": 0.85, "step": 890 }, { "epoch": 0.4, "grad_norm": 4.6865434646606445, "learning_rate": 3.8771428571428575e-05, "loss": 0.6508, "step": 900 }, { "epoch": 0.40444444444444444, "grad_norm": 1.8700509071350098, "learning_rate": 3.8485714285714284e-05, "loss": 0.5405, "step": 910 }, { "epoch": 0.4088888888888889, "grad_norm": 3.749453067779541, "learning_rate": 3.82e-05, "loss": 0.6869, "step": 920 }, { "epoch": 0.41333333333333333, "grad_norm": 2.443854570388794, "learning_rate": 3.7914285714285716e-05, "loss": 0.5013, "step": 930 }, { "epoch": 0.4177777777777778, "grad_norm": 1.066603183746338, "learning_rate": 3.762857142857143e-05, "loss": 0.5993, "step": 940 }, { "epoch": 0.4222222222222222, "grad_norm": 5.453496932983398, "learning_rate": 3.734285714285715e-05, "loss": 0.633, "step": 950 }, { "epoch": 0.4266666666666667, "grad_norm": 3.3490309715270996, "learning_rate": 3.705714285714286e-05, "loss": 0.7452, "step": 960 }, { "epoch": 0.4311111111111111, "grad_norm": 2.070544719696045, "learning_rate": 3.6771428571428574e-05, "loss": 0.7698, "step": 970 }, { "epoch": 0.43555555555555553, "grad_norm": 3.1262307167053223, "learning_rate": 3.648571428571429e-05, "loss": 0.7202, "step": 980 }, { "epoch": 0.44, "grad_norm": 4.966836929321289, "learning_rate": 3.62e-05, "loss": 0.6881, "step": 990 }, { "epoch": 0.4444444444444444, "grad_norm": 0.9398611783981323, "learning_rate": 3.5914285714285715e-05, "loss": 0.3937, "step": 1000 }, { "epoch": 0.4488888888888889, "grad_norm": 2.2783327102661133, "learning_rate": 3.562857142857143e-05, "loss": 0.5421, "step": 1010 }, { "epoch": 0.4533333333333333, "grad_norm": 1.5350534915924072, "learning_rate": 3.534285714285715e-05, "loss": 1.1952, "step": 1020 }, { "epoch": 0.4577777777777778, "grad_norm": 2.5785741806030273, "learning_rate": 3.505714285714286e-05, "loss": 0.7421, "step": 1030 }, { "epoch": 0.4622222222222222, "grad_norm": 3.4625842571258545, "learning_rate": 3.477142857142857e-05, "loss": 0.6769, "step": 1040 }, { "epoch": 0.4666666666666667, "grad_norm": 1.2348237037658691, "learning_rate": 3.448571428571429e-05, "loss": 0.4613, "step": 1050 }, { "epoch": 0.4711111111111111, "grad_norm": 1.4216806888580322, "learning_rate": 3.4200000000000005e-05, "loss": 0.6116, "step": 1060 }, { "epoch": 0.47555555555555556, "grad_norm": 1.089032769203186, "learning_rate": 3.3914285714285714e-05, "loss": 0.5686, "step": 1070 }, { "epoch": 0.48, "grad_norm": 2.135321855545044, "learning_rate": 3.362857142857143e-05, "loss": 0.5668, "step": 1080 }, { "epoch": 0.48444444444444446, "grad_norm": 1.846582055091858, "learning_rate": 3.334285714285714e-05, "loss": 0.62, "step": 1090 }, { "epoch": 0.4888888888888889, "grad_norm": 2.3308606147766113, "learning_rate": 3.305714285714286e-05, "loss": 0.7323, "step": 1100 }, { "epoch": 0.49333333333333335, "grad_norm": 10.256339073181152, "learning_rate": 3.277142857142858e-05, "loss": 0.6122, "step": 1110 }, { "epoch": 0.49777777777777776, "grad_norm": 5.60221529006958, "learning_rate": 3.248571428571429e-05, "loss": 0.7023, "step": 1120 }, { "epoch": 0.5022222222222222, "grad_norm": 1.3172476291656494, "learning_rate": 3.2200000000000003e-05, "loss": 0.6019, "step": 1130 }, { "epoch": 0.5066666666666667, "grad_norm": 2.984369993209839, "learning_rate": 3.191428571428571e-05, "loss": 0.8876, "step": 1140 }, { "epoch": 0.5111111111111111, "grad_norm": 2.645580291748047, "learning_rate": 3.162857142857143e-05, "loss": 0.6198, "step": 1150 }, { "epoch": 0.5155555555555555, "grad_norm": 6.525484085083008, "learning_rate": 3.1342857142857145e-05, "loss": 0.7378, "step": 1160 }, { "epoch": 0.52, "grad_norm": 2.520108938217163, "learning_rate": 3.1057142857142854e-05, "loss": 0.6144, "step": 1170 }, { "epoch": 0.5244444444444445, "grad_norm": 1.4671144485473633, "learning_rate": 3.077142857142857e-05, "loss": 0.5624, "step": 1180 }, { "epoch": 0.5288888888888889, "grad_norm": 1.6521360874176025, "learning_rate": 3.048571428571429e-05, "loss": 0.8754, "step": 1190 }, { "epoch": 0.5333333333333333, "grad_norm": 1.7824636697769165, "learning_rate": 3.02e-05, "loss": 0.6161, "step": 1200 }, { "epoch": 0.5377777777777778, "grad_norm": 3.1892054080963135, "learning_rate": 2.9914285714285718e-05, "loss": 0.6746, "step": 1210 }, { "epoch": 0.5422222222222223, "grad_norm": 2.471684217453003, "learning_rate": 2.9628571428571428e-05, "loss": 0.7047, "step": 1220 }, { "epoch": 0.5466666666666666, "grad_norm": 1.1938486099243164, "learning_rate": 2.9342857142857144e-05, "loss": 0.7613, "step": 1230 }, { "epoch": 0.5511111111111111, "grad_norm": 4.988212585449219, "learning_rate": 2.905714285714286e-05, "loss": 0.7503, "step": 1240 }, { "epoch": 0.5555555555555556, "grad_norm": 6.780579566955566, "learning_rate": 2.8771428571428572e-05, "loss": 0.5375, "step": 1250 }, { "epoch": 0.56, "grad_norm": 2.5735678672790527, "learning_rate": 2.848571428571429e-05, "loss": 0.703, "step": 1260 }, { "epoch": 0.5644444444444444, "grad_norm": 3.509777069091797, "learning_rate": 2.8199999999999998e-05, "loss": 0.5273, "step": 1270 }, { "epoch": 0.5688888888888889, "grad_norm": 1.5865542888641357, "learning_rate": 2.7914285714285714e-05, "loss": 0.4821, "step": 1280 }, { "epoch": 0.5733333333333334, "grad_norm": 1.2859346866607666, "learning_rate": 2.762857142857143e-05, "loss": 0.6347, "step": 1290 }, { "epoch": 0.5777777777777777, "grad_norm": 2.0483293533325195, "learning_rate": 2.7342857142857142e-05, "loss": 0.4798, "step": 1300 }, { "epoch": 0.5822222222222222, "grad_norm": 1.2751628160476685, "learning_rate": 2.705714285714286e-05, "loss": 0.4283, "step": 1310 }, { "epoch": 0.5866666666666667, "grad_norm": 1.7461509704589844, "learning_rate": 2.6771428571428575e-05, "loss": 0.577, "step": 1320 }, { "epoch": 0.5911111111111111, "grad_norm": 2.4847190380096436, "learning_rate": 2.6485714285714287e-05, "loss": 0.6608, "step": 1330 }, { "epoch": 0.5955555555555555, "grad_norm": 7.046814441680908, "learning_rate": 2.6200000000000003e-05, "loss": 0.7501, "step": 1340 }, { "epoch": 0.6, "grad_norm": 1.8862992525100708, "learning_rate": 2.5914285714285713e-05, "loss": 0.5347, "step": 1350 }, { "epoch": 0.6044444444444445, "grad_norm": 1.941525936126709, "learning_rate": 2.562857142857143e-05, "loss": 0.6533, "step": 1360 }, { "epoch": 0.6088888888888889, "grad_norm": 1.9246207475662231, "learning_rate": 2.5342857142857145e-05, "loss": 0.4619, "step": 1370 }, { "epoch": 0.6133333333333333, "grad_norm": 2.3284366130828857, "learning_rate": 2.5057142857142857e-05, "loss": 0.5442, "step": 1380 }, { "epoch": 0.6177777777777778, "grad_norm": 3.2495412826538086, "learning_rate": 2.4771428571428573e-05, "loss": 0.5538, "step": 1390 }, { "epoch": 0.6222222222222222, "grad_norm": 1.9646495580673218, "learning_rate": 2.4485714285714286e-05, "loss": 0.6764, "step": 1400 }, { "epoch": 0.6266666666666667, "grad_norm": 2.3741135597229004, "learning_rate": 2.4200000000000002e-05, "loss": 0.4923, "step": 1410 }, { "epoch": 0.6311111111111111, "grad_norm": 3.175323486328125, "learning_rate": 2.3914285714285715e-05, "loss": 0.5318, "step": 1420 }, { "epoch": 0.6355555555555555, "grad_norm": 4.018131732940674, "learning_rate": 2.362857142857143e-05, "loss": 0.6042, "step": 1430 }, { "epoch": 0.64, "grad_norm": 5.065487384796143, "learning_rate": 2.3342857142857143e-05, "loss": 0.4944, "step": 1440 }, { "epoch": 0.6444444444444445, "grad_norm": 4.733597278594971, "learning_rate": 2.3057142857142856e-05, "loss": 0.5209, "step": 1450 }, { "epoch": 0.6488888888888888, "grad_norm": 2.8453125953674316, "learning_rate": 2.2771428571428572e-05, "loss": 0.6981, "step": 1460 }, { "epoch": 0.6533333333333333, "grad_norm": 5.391908168792725, "learning_rate": 2.2485714285714288e-05, "loss": 0.6173, "step": 1470 }, { "epoch": 0.6577777777777778, "grad_norm": 3.4642038345336914, "learning_rate": 2.22e-05, "loss": 0.5644, "step": 1480 }, { "epoch": 0.6622222222222223, "grad_norm": 5.350740909576416, "learning_rate": 2.1914285714285714e-05, "loss": 0.6647, "step": 1490 }, { "epoch": 0.6666666666666666, "grad_norm": 1.7532992362976074, "learning_rate": 2.162857142857143e-05, "loss": 0.5771, "step": 1500 }, { "epoch": 0.6711111111111111, "grad_norm": 1.6923083066940308, "learning_rate": 2.1342857142857146e-05, "loss": 0.4541, "step": 1510 }, { "epoch": 0.6755555555555556, "grad_norm": 3.4142796993255615, "learning_rate": 2.105714285714286e-05, "loss": 0.6353, "step": 1520 }, { "epoch": 0.68, "grad_norm": 1.2692922353744507, "learning_rate": 2.077142857142857e-05, "loss": 0.6471, "step": 1530 }, { "epoch": 0.6844444444444444, "grad_norm": 6.867006778717041, "learning_rate": 2.0485714285714287e-05, "loss": 0.5668, "step": 1540 }, { "epoch": 0.6888888888888889, "grad_norm": 1.1940428018569946, "learning_rate": 2.0200000000000003e-05, "loss": 0.7113, "step": 1550 }, { "epoch": 0.6933333333333334, "grad_norm": 0.9745553731918335, "learning_rate": 1.9914285714285716e-05, "loss": 0.6865, "step": 1560 }, { "epoch": 0.6977777777777778, "grad_norm": 1.4725788831710815, "learning_rate": 1.962857142857143e-05, "loss": 0.7809, "step": 1570 }, { "epoch": 0.7022222222222222, "grad_norm": 2.4500861167907715, "learning_rate": 1.9342857142857144e-05, "loss": 0.5468, "step": 1580 }, { "epoch": 0.7066666666666667, "grad_norm": 13.177860260009766, "learning_rate": 1.9057142857142857e-05, "loss": 0.5427, "step": 1590 }, { "epoch": 0.7111111111111111, "grad_norm": 5.718009948730469, "learning_rate": 1.8771428571428573e-05, "loss": 0.6003, "step": 1600 }, { "epoch": 0.7155555555555555, "grad_norm": 2.4186694622039795, "learning_rate": 1.8485714285714286e-05, "loss": 0.4843, "step": 1610 }, { "epoch": 0.72, "grad_norm": 1.150259256362915, "learning_rate": 1.8200000000000002e-05, "loss": 0.4339, "step": 1620 }, { "epoch": 0.7244444444444444, "grad_norm": 2.5581448078155518, "learning_rate": 1.7914285714285715e-05, "loss": 0.7765, "step": 1630 }, { "epoch": 0.7288888888888889, "grad_norm": 2.4859399795532227, "learning_rate": 1.762857142857143e-05, "loss": 0.5869, "step": 1640 }, { "epoch": 0.7333333333333333, "grad_norm": 1.6887503862380981, "learning_rate": 1.7342857142857143e-05, "loss": 0.4651, "step": 1650 }, { "epoch": 0.7377777777777778, "grad_norm": 2.0939149856567383, "learning_rate": 1.7057142857142856e-05, "loss": 0.7603, "step": 1660 }, { "epoch": 0.7422222222222222, "grad_norm": 1.9502744674682617, "learning_rate": 1.6771428571428572e-05, "loss": 0.5475, "step": 1670 }, { "epoch": 0.7466666666666667, "grad_norm": 0.8679394721984863, "learning_rate": 1.6485714285714288e-05, "loss": 0.4752, "step": 1680 }, { "epoch": 0.7511111111111111, "grad_norm": 1.2583119869232178, "learning_rate": 1.62e-05, "loss": 0.5677, "step": 1690 }, { "epoch": 0.7555555555555555, "grad_norm": 2.252943754196167, "learning_rate": 1.5914285714285713e-05, "loss": 0.4604, "step": 1700 }, { "epoch": 0.76, "grad_norm": 1.150745153427124, "learning_rate": 1.562857142857143e-05, "loss": 0.5713, "step": 1710 }, { "epoch": 0.7644444444444445, "grad_norm": 1.7483420372009277, "learning_rate": 1.5342857142857146e-05, "loss": 0.5989, "step": 1720 }, { "epoch": 0.7688888888888888, "grad_norm": 0.8522917628288269, "learning_rate": 1.5057142857142858e-05, "loss": 0.5211, "step": 1730 }, { "epoch": 0.7733333333333333, "grad_norm": 2.067723035812378, "learning_rate": 1.4771428571428573e-05, "loss": 0.6904, "step": 1740 }, { "epoch": 0.7777777777777778, "grad_norm": 1.3494257926940918, "learning_rate": 1.4485714285714285e-05, "loss": 0.6667, "step": 1750 }, { "epoch": 0.7822222222222223, "grad_norm": 1.3629013299942017, "learning_rate": 1.42e-05, "loss": 0.3367, "step": 1760 }, { "epoch": 0.7866666666666666, "grad_norm": 3.2202422618865967, "learning_rate": 1.3914285714285716e-05, "loss": 0.6033, "step": 1770 }, { "epoch": 0.7911111111111111, "grad_norm": 1.038692593574524, "learning_rate": 1.362857142857143e-05, "loss": 0.5305, "step": 1780 }, { "epoch": 0.7955555555555556, "grad_norm": 2.57888126373291, "learning_rate": 1.3342857142857143e-05, "loss": 0.5989, "step": 1790 }, { "epoch": 0.8, "grad_norm": 1.118715524673462, "learning_rate": 1.3057142857142857e-05, "loss": 0.4963, "step": 1800 }, { "epoch": 0.8044444444444444, "grad_norm": 2.586069107055664, "learning_rate": 1.2771428571428573e-05, "loss": 0.6829, "step": 1810 }, { "epoch": 0.8088888888888889, "grad_norm": 2.27518630027771, "learning_rate": 1.2485714285714287e-05, "loss": 0.7121, "step": 1820 }, { "epoch": 0.8133333333333334, "grad_norm": 1.1266392469406128, "learning_rate": 1.22e-05, "loss": 0.5293, "step": 1830 }, { "epoch": 0.8177777777777778, "grad_norm": 1.4908055067062378, "learning_rate": 1.1914285714285716e-05, "loss": 0.5562, "step": 1840 }, { "epoch": 0.8222222222222222, "grad_norm": 1.4283602237701416, "learning_rate": 1.1628571428571429e-05, "loss": 0.4831, "step": 1850 }, { "epoch": 0.8266666666666667, "grad_norm": 1.0552829504013062, "learning_rate": 1.1342857142857143e-05, "loss": 0.5761, "step": 1860 }, { "epoch": 0.8311111111111111, "grad_norm": 6.45361852645874, "learning_rate": 1.1057142857142858e-05, "loss": 0.6874, "step": 1870 }, { "epoch": 0.8355555555555556, "grad_norm": 1.207909345626831, "learning_rate": 1.0771428571428572e-05, "loss": 0.6289, "step": 1880 }, { "epoch": 0.84, "grad_norm": 3.6151106357574463, "learning_rate": 1.0485714285714286e-05, "loss": 0.78, "step": 1890 }, { "epoch": 0.8444444444444444, "grad_norm": 2.560347318649292, "learning_rate": 1.02e-05, "loss": 0.7657, "step": 1900 }, { "epoch": 0.8488888888888889, "grad_norm": 1.1826066970825195, "learning_rate": 9.914285714285715e-06, "loss": 0.5419, "step": 1910 }, { "epoch": 0.8533333333333334, "grad_norm": 2.4970545768737793, "learning_rate": 9.628571428571428e-06, "loss": 0.5137, "step": 1920 }, { "epoch": 0.8577777777777778, "grad_norm": 1.6849817037582397, "learning_rate": 9.342857142857144e-06, "loss": 0.4476, "step": 1930 }, { "epoch": 0.8622222222222222, "grad_norm": 3.5557351112365723, "learning_rate": 9.057142857142856e-06, "loss": 0.6422, "step": 1940 }, { "epoch": 0.8666666666666667, "grad_norm": 1.7327756881713867, "learning_rate": 8.771428571428572e-06, "loss": 0.656, "step": 1950 }, { "epoch": 0.8711111111111111, "grad_norm": 1.3222990036010742, "learning_rate": 8.485714285714285e-06, "loss": 0.6862, "step": 1960 }, { "epoch": 0.8755555555555555, "grad_norm": 1.3732144832611084, "learning_rate": 8.200000000000001e-06, "loss": 0.5173, "step": 1970 }, { "epoch": 0.88, "grad_norm": 5.491960048675537, "learning_rate": 7.914285714285714e-06, "loss": 0.5735, "step": 1980 }, { "epoch": 0.8844444444444445, "grad_norm": 1.9830958843231201, "learning_rate": 7.628571428571429e-06, "loss": 0.4936, "step": 1990 }, { "epoch": 0.8888888888888888, "grad_norm": 3.4833927154541016, "learning_rate": 7.342857142857143e-06, "loss": 0.5099, "step": 2000 }, { "epoch": 0.8933333333333333, "grad_norm": 1.197811484336853, "learning_rate": 7.057142857142858e-06, "loss": 0.4711, "step": 2010 }, { "epoch": 0.8977777777777778, "grad_norm": 2.15141224861145, "learning_rate": 6.771428571428571e-06, "loss": 0.5555, "step": 2020 }, { "epoch": 0.9022222222222223, "grad_norm": 1.7254974842071533, "learning_rate": 6.485714285714286e-06, "loss": 0.5582, "step": 2030 }, { "epoch": 0.9066666666666666, "grad_norm": 9.579049110412598, "learning_rate": 6.2e-06, "loss": 0.4571, "step": 2040 }, { "epoch": 0.9111111111111111, "grad_norm": 2.882373094558716, "learning_rate": 5.914285714285714e-06, "loss": 0.6287, "step": 2050 }, { "epoch": 0.9155555555555556, "grad_norm": 5.913229942321777, "learning_rate": 5.628571428571429e-06, "loss": 0.5779, "step": 2060 }, { "epoch": 0.92, "grad_norm": 4.352343559265137, "learning_rate": 5.342857142857143e-06, "loss": 0.5565, "step": 2070 }, { "epoch": 0.9244444444444444, "grad_norm": 2.8510098457336426, "learning_rate": 5.057142857142857e-06, "loss": 0.5999, "step": 2080 }, { "epoch": 0.9288888888888889, "grad_norm": 2.46824049949646, "learning_rate": 4.771428571428572e-06, "loss": 0.4779, "step": 2090 }, { "epoch": 0.9333333333333333, "grad_norm": 1.5886437892913818, "learning_rate": 4.485714285714286e-06, "loss": 0.5774, "step": 2100 }, { "epoch": 0.9377777777777778, "grad_norm": 1.5184223651885986, "learning_rate": 4.2000000000000004e-06, "loss": 0.4278, "step": 2110 }, { "epoch": 0.9422222222222222, "grad_norm": 4.225131511688232, "learning_rate": 3.914285714285715e-06, "loss": 0.6479, "step": 2120 }, { "epoch": 0.9466666666666667, "grad_norm": 2.665025234222412, "learning_rate": 3.6285714285714283e-06, "loss": 0.4907, "step": 2130 }, { "epoch": 0.9511111111111111, "grad_norm": 1.4222490787506104, "learning_rate": 3.3428571428571427e-06, "loss": 0.4263, "step": 2140 }, { "epoch": 0.9555555555555556, "grad_norm": 1.8098088502883911, "learning_rate": 3.0571428571428575e-06, "loss": 0.512, "step": 2150 }, { "epoch": 0.96, "grad_norm": 1.5185695886611938, "learning_rate": 2.771428571428572e-06, "loss": 0.4707, "step": 2160 }, { "epoch": 0.9644444444444444, "grad_norm": 3.510244846343994, "learning_rate": 2.4857142857142858e-06, "loss": 0.5985, "step": 2170 }, { "epoch": 0.9688888888888889, "grad_norm": 2.4858155250549316, "learning_rate": 2.2e-06, "loss": 0.5148, "step": 2180 }, { "epoch": 0.9733333333333334, "grad_norm": 2.5503909587860107, "learning_rate": 1.9142857142857145e-06, "loss": 0.5905, "step": 2190 }, { "epoch": 0.9777777777777777, "grad_norm": 1.640555739402771, "learning_rate": 1.6285714285714286e-06, "loss": 0.4983, "step": 2200 }, { "epoch": 0.9822222222222222, "grad_norm": 3.124906539916992, "learning_rate": 1.342857142857143e-06, "loss": 0.4201, "step": 2210 }, { "epoch": 0.9866666666666667, "grad_norm": 2.2014262676239014, "learning_rate": 1.0571428571428573e-06, "loss": 0.5991, "step": 2220 }, { "epoch": 0.9911111111111112, "grad_norm": 1.412158727645874, "learning_rate": 7.714285714285715e-07, "loss": 0.5062, "step": 2230 }, { "epoch": 0.9955555555555555, "grad_norm": 2.74906325340271, "learning_rate": 4.857142857142857e-07, "loss": 0.6499, "step": 2240 }, { "epoch": 1.0, "grad_norm": 1.1791956424713135, "learning_rate": 2.0000000000000002e-07, "loss": 0.521, "step": 2250 } ], "logging_steps": 10, "max_steps": 2250, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 5000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.20270921367552e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }