{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0, "eval_steps": 500, "global_step": 2120, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.009433962264150943, "grad_norm": 2.1125893592834473, "learning_rate": 1.9915094339622643e-05, "loss": 0.6661, "step": 10 }, { "epoch": 0.018867924528301886, "grad_norm": 4.568980693817139, "learning_rate": 1.9820754716981134e-05, "loss": 0.6476, "step": 20 }, { "epoch": 0.02830188679245283, "grad_norm": 4.55064582824707, "learning_rate": 1.9726415094339625e-05, "loss": 0.609, "step": 30 }, { "epoch": 0.03773584905660377, "grad_norm": 2.458791732788086, "learning_rate": 1.9632075471698116e-05, "loss": 0.5465, "step": 40 }, { "epoch": 0.04716981132075472, "grad_norm": 5.235497951507568, "learning_rate": 1.9537735849056604e-05, "loss": 0.4277, "step": 50 }, { "epoch": 0.05660377358490566, "grad_norm": 2.81006121635437, "learning_rate": 1.9443396226415095e-05, "loss": 0.3367, "step": 60 }, { "epoch": 0.0660377358490566, "grad_norm": 5.580306529998779, "learning_rate": 1.9349056603773586e-05, "loss": 0.232, "step": 70 }, { "epoch": 0.07547169811320754, "grad_norm": 6.972874641418457, "learning_rate": 1.9254716981132077e-05, "loss": 0.1852, "step": 80 }, { "epoch": 0.08490566037735849, "grad_norm": 2.1650376319885254, "learning_rate": 1.9160377358490568e-05, "loss": 0.1581, "step": 90 }, { "epoch": 0.09433962264150944, "grad_norm": 0.42251577973365784, "learning_rate": 1.906603773584906e-05, "loss": 0.1821, "step": 100 }, { "epoch": 0.10377358490566038, "grad_norm": 0.361060231924057, "learning_rate": 1.8971698113207547e-05, "loss": 0.3396, "step": 110 }, { "epoch": 0.11320754716981132, "grad_norm": 1.1767176389694214, "learning_rate": 1.887735849056604e-05, "loss": 0.0724, "step": 120 }, { "epoch": 0.12264150943396226, "grad_norm": 0.310159295797348, "learning_rate": 1.878301886792453e-05, "loss": 0.1826, "step": 130 }, { "epoch": 0.1320754716981132, "grad_norm": 0.29479867219924927, "learning_rate": 1.868867924528302e-05, "loss": 0.1407, "step": 140 }, { "epoch": 0.14150943396226415, "grad_norm": 0.25081831216812134, "learning_rate": 1.859433962264151e-05, "loss": 0.0804, "step": 150 }, { "epoch": 0.1509433962264151, "grad_norm": 7.194790840148926, "learning_rate": 1.8500000000000002e-05, "loss": 0.1394, "step": 160 }, { "epoch": 0.16037735849056603, "grad_norm": 0.19942767918109894, "learning_rate": 1.8405660377358494e-05, "loss": 0.0876, "step": 170 }, { "epoch": 0.16981132075471697, "grad_norm": 12.822174072265625, "learning_rate": 1.8311320754716985e-05, "loss": 0.1606, "step": 180 }, { "epoch": 0.1792452830188679, "grad_norm": 1.3328683376312256, "learning_rate": 1.8216981132075472e-05, "loss": 0.1835, "step": 190 }, { "epoch": 0.18867924528301888, "grad_norm": 0.17138229310512543, "learning_rate": 1.8122641509433963e-05, "loss": 0.2355, "step": 200 }, { "epoch": 0.19811320754716982, "grad_norm": 0.21127229928970337, "learning_rate": 1.8028301886792454e-05, "loss": 0.2031, "step": 210 }, { "epoch": 0.20754716981132076, "grad_norm": 2.211583137512207, "learning_rate": 1.7933962264150946e-05, "loss": 0.0223, "step": 220 }, { "epoch": 0.2169811320754717, "grad_norm": 0.17399153113365173, "learning_rate": 1.7839622641509437e-05, "loss": 0.2204, "step": 230 }, { "epoch": 0.22641509433962265, "grad_norm": 0.26820316910743713, "learning_rate": 1.7745283018867928e-05, "loss": 0.1022, "step": 240 }, { "epoch": 0.2358490566037736, "grad_norm": 0.15803369879722595, "learning_rate": 1.7650943396226415e-05, "loss": 0.0638, "step": 250 }, { "epoch": 0.24528301886792453, "grad_norm": 0.15400823950767517, "learning_rate": 1.7556603773584907e-05, "loss": 0.1257, "step": 260 }, { "epoch": 0.25471698113207547, "grad_norm": 0.7688778638839722, "learning_rate": 1.7462264150943398e-05, "loss": 0.2443, "step": 270 }, { "epoch": 0.2641509433962264, "grad_norm": 0.243111252784729, "learning_rate": 1.736792452830189e-05, "loss": 0.2897, "step": 280 }, { "epoch": 0.27358490566037735, "grad_norm": 5.870262145996094, "learning_rate": 1.727358490566038e-05, "loss": 0.2198, "step": 290 }, { "epoch": 0.2830188679245283, "grad_norm": 0.2840751111507416, "learning_rate": 1.7179245283018867e-05, "loss": 0.3108, "step": 300 }, { "epoch": 0.29245283018867924, "grad_norm": 0.1476099044084549, "learning_rate": 1.708490566037736e-05, "loss": 0.1688, "step": 310 }, { "epoch": 0.3018867924528302, "grad_norm": 0.3752176761627197, "learning_rate": 1.699056603773585e-05, "loss": 0.2876, "step": 320 }, { "epoch": 0.3113207547169811, "grad_norm": 30.54532241821289, "learning_rate": 1.689622641509434e-05, "loss": 0.1215, "step": 330 }, { "epoch": 0.32075471698113206, "grad_norm": 0.15333189070224762, "learning_rate": 1.6801886792452832e-05, "loss": 0.2964, "step": 340 }, { "epoch": 0.330188679245283, "grad_norm": 0.23220475018024445, "learning_rate": 1.6707547169811323e-05, "loss": 0.1506, "step": 350 }, { "epoch": 0.33962264150943394, "grad_norm": 121.28533172607422, "learning_rate": 1.661320754716981e-05, "loss": 0.2107, "step": 360 }, { "epoch": 0.3490566037735849, "grad_norm": 1.4383134841918945, "learning_rate": 1.6518867924528305e-05, "loss": 0.2117, "step": 370 }, { "epoch": 0.3584905660377358, "grad_norm": 0.14638499915599823, "learning_rate": 1.6424528301886793e-05, "loss": 0.1013, "step": 380 }, { "epoch": 0.36792452830188677, "grad_norm": 0.124303437769413, "learning_rate": 1.6330188679245284e-05, "loss": 0.0478, "step": 390 }, { "epoch": 0.37735849056603776, "grad_norm": 0.11256200820207596, "learning_rate": 1.6235849056603775e-05, "loss": 0.0689, "step": 400 }, { "epoch": 0.3867924528301887, "grad_norm": 0.10450290888547897, "learning_rate": 1.6141509433962266e-05, "loss": 0.1293, "step": 410 }, { "epoch": 0.39622641509433965, "grad_norm": 0.09205685555934906, "learning_rate": 1.6047169811320754e-05, "loss": 0.0672, "step": 420 }, { "epoch": 0.4056603773584906, "grad_norm": 1.0851916074752808, "learning_rate": 1.5952830188679248e-05, "loss": 0.1918, "step": 430 }, { "epoch": 0.41509433962264153, "grad_norm": 0.1460585743188858, "learning_rate": 1.5858490566037736e-05, "loss": 0.0866, "step": 440 }, { "epoch": 0.42452830188679247, "grad_norm": 0.10592811554670334, "learning_rate": 1.5764150943396227e-05, "loss": 0.0327, "step": 450 }, { "epoch": 0.4339622641509434, "grad_norm": 0.09917836636304855, "learning_rate": 1.5669811320754718e-05, "loss": 0.268, "step": 460 }, { "epoch": 0.44339622641509435, "grad_norm": 0.2683676481246948, "learning_rate": 1.557547169811321e-05, "loss": 0.2433, "step": 470 }, { "epoch": 0.4528301886792453, "grad_norm": 0.3096301853656769, "learning_rate": 1.54811320754717e-05, "loss": 0.0106, "step": 480 }, { "epoch": 0.46226415094339623, "grad_norm": 0.14542903006076813, "learning_rate": 1.538679245283019e-05, "loss": 0.1084, "step": 490 }, { "epoch": 0.4716981132075472, "grad_norm": 0.09798438102006912, "learning_rate": 1.529245283018868e-05, "loss": 0.1217, "step": 500 }, { "epoch": 0.4811320754716981, "grad_norm": 0.5679267644882202, "learning_rate": 1.5198113207547172e-05, "loss": 0.1084, "step": 510 }, { "epoch": 0.49056603773584906, "grad_norm": 0.07084541022777557, "learning_rate": 1.5103773584905661e-05, "loss": 0.1496, "step": 520 }, { "epoch": 0.5, "grad_norm": 0.09527801722288132, "learning_rate": 1.5009433962264152e-05, "loss": 0.0397, "step": 530 }, { "epoch": 0.5094339622641509, "grad_norm": 2.7925026416778564, "learning_rate": 1.4915094339622642e-05, "loss": 0.0082, "step": 540 }, { "epoch": 0.5188679245283019, "grad_norm": 61.901344299316406, "learning_rate": 1.4820754716981134e-05, "loss": 0.2448, "step": 550 }, { "epoch": 0.5283018867924528, "grad_norm": 0.10737080127000809, "learning_rate": 1.4726415094339624e-05, "loss": 0.2122, "step": 560 }, { "epoch": 0.5377358490566038, "grad_norm": 0.08159242570400238, "learning_rate": 1.4632075471698115e-05, "loss": 0.2056, "step": 570 }, { "epoch": 0.5471698113207547, "grad_norm": 26.218284606933594, "learning_rate": 1.4537735849056604e-05, "loss": 0.2137, "step": 580 }, { "epoch": 0.5566037735849056, "grad_norm": 0.09104019403457642, "learning_rate": 1.4443396226415095e-05, "loss": 0.1235, "step": 590 }, { "epoch": 0.5660377358490566, "grad_norm": 44.143577575683594, "learning_rate": 1.4349056603773586e-05, "loss": 0.1708, "step": 600 }, { "epoch": 0.5754716981132075, "grad_norm": 0.09825887531042099, "learning_rate": 1.4254716981132078e-05, "loss": 0.2829, "step": 610 }, { "epoch": 0.5849056603773585, "grad_norm": 0.08215257525444031, "learning_rate": 1.4160377358490567e-05, "loss": 0.0978, "step": 620 }, { "epoch": 0.5943396226415094, "grad_norm": 0.0858820378780365, "learning_rate": 1.4066037735849058e-05, "loss": 0.1308, "step": 630 }, { "epoch": 0.6037735849056604, "grad_norm": 0.07820013910531998, "learning_rate": 1.3971698113207547e-05, "loss": 0.0083, "step": 640 }, { "epoch": 0.6132075471698113, "grad_norm": 0.07088133692741394, "learning_rate": 1.387735849056604e-05, "loss": 0.0343, "step": 650 }, { "epoch": 0.6226415094339622, "grad_norm": 0.07562941312789917, "learning_rate": 1.378301886792453e-05, "loss": 0.1503, "step": 660 }, { "epoch": 0.6320754716981132, "grad_norm": 0.10470731556415558, "learning_rate": 1.368867924528302e-05, "loss": 0.0474, "step": 670 }, { "epoch": 0.6415094339622641, "grad_norm": 0.06047428399324417, "learning_rate": 1.359433962264151e-05, "loss": 0.3286, "step": 680 }, { "epoch": 0.6509433962264151, "grad_norm": 0.12241201102733612, "learning_rate": 1.3500000000000001e-05, "loss": 0.163, "step": 690 }, { "epoch": 0.660377358490566, "grad_norm": 0.07547236233949661, "learning_rate": 1.3405660377358492e-05, "loss": 0.1182, "step": 700 }, { "epoch": 0.6698113207547169, "grad_norm": 0.06518665701150894, "learning_rate": 1.3311320754716983e-05, "loss": 0.1177, "step": 710 }, { "epoch": 0.6792452830188679, "grad_norm": 0.05459102615714073, "learning_rate": 1.3216981132075473e-05, "loss": 0.0056, "step": 720 }, { "epoch": 0.6886792452830188, "grad_norm": 0.06507308781147003, "learning_rate": 1.3122641509433964e-05, "loss": 0.194, "step": 730 }, { "epoch": 0.6981132075471698, "grad_norm": 0.06847833096981049, "learning_rate": 1.3028301886792453e-05, "loss": 0.0334, "step": 740 }, { "epoch": 0.7075471698113207, "grad_norm": 0.05916672945022583, "learning_rate": 1.2933962264150946e-05, "loss": 0.1326, "step": 750 }, { "epoch": 0.7169811320754716, "grad_norm": 0.05806839466094971, "learning_rate": 1.2839622641509435e-05, "loss": 0.0734, "step": 760 }, { "epoch": 0.7264150943396226, "grad_norm": 0.058394916355609894, "learning_rate": 1.2745283018867926e-05, "loss": 0.3154, "step": 770 }, { "epoch": 0.7358490566037735, "grad_norm": 0.08128712326288223, "learning_rate": 1.2650943396226416e-05, "loss": 0.1004, "step": 780 }, { "epoch": 0.7452830188679245, "grad_norm": 0.1772662252187729, "learning_rate": 1.2556603773584907e-05, "loss": 0.1408, "step": 790 }, { "epoch": 0.7547169811320755, "grad_norm": 0.07657451182603836, "learning_rate": 1.2462264150943398e-05, "loss": 0.1138, "step": 800 }, { "epoch": 0.7641509433962265, "grad_norm": 0.0920625552535057, "learning_rate": 1.2367924528301887e-05, "loss": 0.0064, "step": 810 }, { "epoch": 0.7735849056603774, "grad_norm": 3.893165111541748, "learning_rate": 1.2273584905660379e-05, "loss": 0.1015, "step": 820 }, { "epoch": 0.7830188679245284, "grad_norm": 0.06635797768831253, "learning_rate": 1.2179245283018868e-05, "loss": 0.1405, "step": 830 }, { "epoch": 0.7924528301886793, "grad_norm": 1.1072912216186523, "learning_rate": 1.2084905660377359e-05, "loss": 0.1225, "step": 840 }, { "epoch": 0.8018867924528302, "grad_norm": 0.04783385619521141, "learning_rate": 1.1990566037735848e-05, "loss": 0.0895, "step": 850 }, { "epoch": 0.8113207547169812, "grad_norm": 0.05416665971279144, "learning_rate": 1.1896226415094341e-05, "loss": 0.1095, "step": 860 }, { "epoch": 0.8207547169811321, "grad_norm": 0.10190918296575546, "learning_rate": 1.180188679245283e-05, "loss": 0.1404, "step": 870 }, { "epoch": 0.8301886792452831, "grad_norm": 92.75904083251953, "learning_rate": 1.1707547169811322e-05, "loss": 0.0269, "step": 880 }, { "epoch": 0.839622641509434, "grad_norm": 0.06948844343423843, "learning_rate": 1.1613207547169811e-05, "loss": 0.0046, "step": 890 }, { "epoch": 0.8490566037735849, "grad_norm": 0.0549422986805439, "learning_rate": 1.1518867924528304e-05, "loss": 0.1588, "step": 900 }, { "epoch": 0.8584905660377359, "grad_norm": 0.09215264767408371, "learning_rate": 1.1424528301886793e-05, "loss": 0.0639, "step": 910 }, { "epoch": 0.8679245283018868, "grad_norm": 0.0980132520198822, "learning_rate": 1.1330188679245284e-05, "loss": 0.21, "step": 920 }, { "epoch": 0.8773584905660378, "grad_norm": 0.06740464270114899, "learning_rate": 1.1235849056603774e-05, "loss": 0.0059, "step": 930 }, { "epoch": 0.8867924528301887, "grad_norm": 0.42118075489997864, "learning_rate": 1.1141509433962265e-05, "loss": 0.1012, "step": 940 }, { "epoch": 0.8962264150943396, "grad_norm": 0.09071939438581467, "learning_rate": 1.1047169811320754e-05, "loss": 0.0636, "step": 950 }, { "epoch": 0.9056603773584906, "grad_norm": 0.06605428457260132, "learning_rate": 1.0952830188679247e-05, "loss": 0.1424, "step": 960 }, { "epoch": 0.9150943396226415, "grad_norm": 0.054869458079338074, "learning_rate": 1.0858490566037736e-05, "loss": 0.1506, "step": 970 }, { "epoch": 0.9245283018867925, "grad_norm": 8.528816223144531, "learning_rate": 1.0764150943396227e-05, "loss": 0.0766, "step": 980 }, { "epoch": 0.9339622641509434, "grad_norm": 0.04189067333936691, "learning_rate": 1.0669811320754717e-05, "loss": 0.0383, "step": 990 }, { "epoch": 0.9433962264150944, "grad_norm": 0.19604367017745972, "learning_rate": 1.057547169811321e-05, "loss": 0.0474, "step": 1000 }, { "epoch": 0.9528301886792453, "grad_norm": 0.04203261435031891, "learning_rate": 1.0481132075471699e-05, "loss": 0.0571, "step": 1010 }, { "epoch": 0.9622641509433962, "grad_norm": 20.811012268066406, "learning_rate": 1.038679245283019e-05, "loss": 0.2019, "step": 1020 }, { "epoch": 0.9716981132075472, "grad_norm": 0.10877203941345215, "learning_rate": 1.029245283018868e-05, "loss": 0.1336, "step": 1030 }, { "epoch": 0.9811320754716981, "grad_norm": 0.10228094458580017, "learning_rate": 1.019811320754717e-05, "loss": 0.0273, "step": 1040 }, { "epoch": 0.9905660377358491, "grad_norm": 1.2278637886047363, "learning_rate": 1.010377358490566e-05, "loss": 0.0452, "step": 1050 }, { "epoch": 1.0, "grad_norm": 0.03946210443973541, "learning_rate": 1.0009433962264153e-05, "loss": 0.0109, "step": 1060 }, { "epoch": 1.0, "eval_loss": 0.08500511199235916, "eval_runtime": 1.6971, "eval_samples_per_second": 555.67, "eval_steps_per_second": 69.532, "step": 1060 }, { "epoch": 1.009433962264151, "grad_norm": 0.03779453784227371, "learning_rate": 9.915094339622642e-06, "loss": 0.1822, "step": 1070 }, { "epoch": 1.0188679245283019, "grad_norm": 0.040553513914346695, "learning_rate": 9.820754716981133e-06, "loss": 0.054, "step": 1080 }, { "epoch": 1.028301886792453, "grad_norm": 0.02908136695623398, "learning_rate": 9.726415094339623e-06, "loss": 0.0062, "step": 1090 }, { "epoch": 1.0377358490566038, "grad_norm": 0.04287223890423775, "learning_rate": 9.632075471698114e-06, "loss": 0.0039, "step": 1100 }, { "epoch": 1.0471698113207548, "grad_norm": 0.03865998610854149, "learning_rate": 9.537735849056605e-06, "loss": 0.0446, "step": 1110 }, { "epoch": 1.0566037735849056, "grad_norm": 0.04521052539348602, "learning_rate": 9.443396226415094e-06, "loss": 0.0029, "step": 1120 }, { "epoch": 1.0660377358490567, "grad_norm": 0.03594714403152466, "learning_rate": 9.349056603773585e-06, "loss": 0.148, "step": 1130 }, { "epoch": 1.0754716981132075, "grad_norm": 0.029539430513978004, "learning_rate": 9.254716981132076e-06, "loss": 0.0271, "step": 1140 }, { "epoch": 1.0849056603773586, "grad_norm": 4.038683891296387, "learning_rate": 9.160377358490566e-06, "loss": 0.165, "step": 1150 }, { "epoch": 1.0943396226415094, "grad_norm": 0.027737464755773544, "learning_rate": 9.066037735849057e-06, "loss": 0.0028, "step": 1160 }, { "epoch": 1.1037735849056605, "grad_norm": 0.03591258078813553, "learning_rate": 8.971698113207548e-06, "loss": 0.1371, "step": 1170 }, { "epoch": 1.1132075471698113, "grad_norm": 12.058833122253418, "learning_rate": 8.877358490566039e-06, "loss": 0.236, "step": 1180 }, { "epoch": 1.1226415094339623, "grad_norm": 0.04443644732236862, "learning_rate": 8.783018867924528e-06, "loss": 0.003, "step": 1190 }, { "epoch": 1.1320754716981132, "grad_norm": 40.05979919433594, "learning_rate": 8.68867924528302e-06, "loss": 0.0157, "step": 1200 }, { "epoch": 1.1415094339622642, "grad_norm": 0.04003468528389931, "learning_rate": 8.59433962264151e-06, "loss": 0.0029, "step": 1210 }, { "epoch": 1.150943396226415, "grad_norm": 2.571934223175049, "learning_rate": 8.5e-06, "loss": 0.1204, "step": 1220 }, { "epoch": 1.1603773584905661, "grad_norm": 0.06405159831047058, "learning_rate": 8.405660377358491e-06, "loss": 0.0146, "step": 1230 }, { "epoch": 1.169811320754717, "grad_norm": 8.398702621459961, "learning_rate": 8.311320754716982e-06, "loss": 0.0048, "step": 1240 }, { "epoch": 1.179245283018868, "grad_norm": 19.97818946838379, "learning_rate": 8.216981132075471e-06, "loss": 0.0688, "step": 1250 }, { "epoch": 1.1886792452830188, "grad_norm": 0.03282918781042099, "learning_rate": 8.122641509433963e-06, "loss": 0.0035, "step": 1260 }, { "epoch": 1.1981132075471699, "grad_norm": 0.03513512760400772, "learning_rate": 8.028301886792454e-06, "loss": 0.0355, "step": 1270 }, { "epoch": 1.2075471698113207, "grad_norm": 0.03274456784129143, "learning_rate": 7.933962264150945e-06, "loss": 0.188, "step": 1280 }, { "epoch": 1.2169811320754718, "grad_norm": 0.03626378998160362, "learning_rate": 7.839622641509434e-06, "loss": 0.014, "step": 1290 }, { "epoch": 1.2264150943396226, "grad_norm": 0.8501767516136169, "learning_rate": 7.745283018867925e-06, "loss": 0.1076, "step": 1300 }, { "epoch": 1.2358490566037736, "grad_norm": 0.0267633106559515, "learning_rate": 7.650943396226416e-06, "loss": 0.0025, "step": 1310 }, { "epoch": 1.2452830188679245, "grad_norm": 0.04418088495731354, "learning_rate": 7.5566037735849066e-06, "loss": 0.1331, "step": 1320 }, { "epoch": 1.2547169811320755, "grad_norm": 6.747396469116211, "learning_rate": 7.462264150943397e-06, "loss": 0.3259, "step": 1330 }, { "epoch": 1.2641509433962264, "grad_norm": 0.032822128385305405, "learning_rate": 7.367924528301887e-06, "loss": 0.1229, "step": 1340 }, { "epoch": 1.2735849056603774, "grad_norm": 0.06315261125564575, "learning_rate": 7.273584905660378e-06, "loss": 0.0039, "step": 1350 }, { "epoch": 1.2830188679245282, "grad_norm": 0.03962039574980736, "learning_rate": 7.179245283018868e-06, "loss": 0.0072, "step": 1360 }, { "epoch": 1.2924528301886793, "grad_norm": 0.03726816922426224, "learning_rate": 7.0849056603773594e-06, "loss": 0.0138, "step": 1370 }, { "epoch": 1.3018867924528301, "grad_norm": 0.04036370664834976, "learning_rate": 6.99056603773585e-06, "loss": 0.1472, "step": 1380 }, { "epoch": 1.3113207547169812, "grad_norm": 0.03037635050714016, "learning_rate": 6.89622641509434e-06, "loss": 0.0069, "step": 1390 }, { "epoch": 1.320754716981132, "grad_norm": 0.02670077607035637, "learning_rate": 6.801886792452831e-06, "loss": 0.0021, "step": 1400 }, { "epoch": 1.330188679245283, "grad_norm": 0.05846339091658592, "learning_rate": 6.707547169811321e-06, "loss": 0.0022, "step": 1410 }, { "epoch": 1.3396226415094339, "grad_norm": 9.76332950592041, "learning_rate": 6.613207547169812e-06, "loss": 0.1445, "step": 1420 }, { "epoch": 1.349056603773585, "grad_norm": 0.048967599868774414, "learning_rate": 6.5188679245283026e-06, "loss": 0.0441, "step": 1430 }, { "epoch": 1.3584905660377358, "grad_norm": 0.025179484859108925, "learning_rate": 6.424528301886793e-06, "loss": 0.0188, "step": 1440 }, { "epoch": 1.3679245283018868, "grad_norm": 0.02890361286699772, "learning_rate": 6.330188679245284e-06, "loss": 0.0021, "step": 1450 }, { "epoch": 1.3773584905660377, "grad_norm": 0.02526584453880787, "learning_rate": 6.235849056603774e-06, "loss": 0.0022, "step": 1460 }, { "epoch": 1.3867924528301887, "grad_norm": 2.650209903717041, "learning_rate": 6.141509433962265e-06, "loss": 0.0482, "step": 1470 }, { "epoch": 1.3962264150943398, "grad_norm": 0.0368034765124321, "learning_rate": 6.0471698113207555e-06, "loss": 0.0034, "step": 1480 }, { "epoch": 1.4056603773584906, "grad_norm": 0.021781697869300842, "learning_rate": 5.952830188679246e-06, "loss": 0.0022, "step": 1490 }, { "epoch": 1.4150943396226414, "grad_norm": 0.019001835957169533, "learning_rate": 5.858490566037737e-06, "loss": 0.2058, "step": 1500 }, { "epoch": 1.4245283018867925, "grad_norm": 0.047584012150764465, "learning_rate": 5.764150943396227e-06, "loss": 0.1632, "step": 1510 }, { "epoch": 1.4339622641509435, "grad_norm": 0.02781420387327671, "learning_rate": 5.669811320754718e-06, "loss": 0.0018, "step": 1520 }, { "epoch": 1.4433962264150944, "grad_norm": 0.03572264313697815, "learning_rate": 5.575471698113208e-06, "loss": 0.0184, "step": 1530 }, { "epoch": 1.4528301886792452, "grad_norm": 0.021896662190556526, "learning_rate": 5.4811320754716994e-06, "loss": 0.155, "step": 1540 }, { "epoch": 1.4622641509433962, "grad_norm": 0.024771416559815407, "learning_rate": 5.38679245283019e-06, "loss": 0.0018, "step": 1550 }, { "epoch": 1.4716981132075473, "grad_norm": 3.849034547805786, "learning_rate": 5.29245283018868e-06, "loss": 0.1006, "step": 1560 }, { "epoch": 1.4811320754716981, "grad_norm": 0.12136558443307877, "learning_rate": 5.198113207547171e-06, "loss": 0.0024, "step": 1570 }, { "epoch": 1.490566037735849, "grad_norm": 0.029468955472111702, "learning_rate": 5.103773584905661e-06, "loss": 0.0021, "step": 1580 }, { "epoch": 1.5, "grad_norm": 2.4301178455352783, "learning_rate": 5.009433962264152e-06, "loss": 0.0024, "step": 1590 }, { "epoch": 1.509433962264151, "grad_norm": 0.024639485403895378, "learning_rate": 4.915094339622642e-06, "loss": 0.0019, "step": 1600 }, { "epoch": 1.5188679245283019, "grad_norm": 5.811135768890381, "learning_rate": 4.820754716981133e-06, "loss": 0.0282, "step": 1610 }, { "epoch": 1.5283018867924527, "grad_norm": 0.02190409228205681, "learning_rate": 4.726415094339623e-06, "loss": 0.0016, "step": 1620 }, { "epoch": 1.5377358490566038, "grad_norm": 0.025405917316675186, "learning_rate": 4.632075471698113e-06, "loss": 0.0505, "step": 1630 }, { "epoch": 1.5471698113207548, "grad_norm": 16.779964447021484, "learning_rate": 4.537735849056604e-06, "loss": 0.1169, "step": 1640 }, { "epoch": 1.5566037735849056, "grad_norm": 43.401329040527344, "learning_rate": 4.443396226415095e-06, "loss": 0.0245, "step": 1650 }, { "epoch": 1.5660377358490565, "grad_norm": 0.9246163964271545, "learning_rate": 4.349056603773586e-06, "loss": 0.0354, "step": 1660 }, { "epoch": 1.5754716981132075, "grad_norm": 0.02293049544095993, "learning_rate": 4.254716981132076e-06, "loss": 0.0036, "step": 1670 }, { "epoch": 1.5849056603773586, "grad_norm": 0.019884012639522552, "learning_rate": 4.160377358490566e-06, "loss": 0.0017, "step": 1680 }, { "epoch": 1.5943396226415094, "grad_norm": 0.0297381728887558, "learning_rate": 4.066037735849057e-06, "loss": 0.002, "step": 1690 }, { "epoch": 1.6037735849056602, "grad_norm": 0.01664115861058235, "learning_rate": 3.9716981132075475e-06, "loss": 0.0029, "step": 1700 }, { "epoch": 1.6132075471698113, "grad_norm": 0.021111104637384415, "learning_rate": 3.8773584905660386e-06, "loss": 0.0019, "step": 1710 }, { "epoch": 1.6226415094339623, "grad_norm": 5.916421890258789, "learning_rate": 3.783018867924529e-06, "loss": 0.0597, "step": 1720 }, { "epoch": 1.6320754716981132, "grad_norm": 0.11948911845684052, "learning_rate": 3.688679245283019e-06, "loss": 0.0288, "step": 1730 }, { "epoch": 1.641509433962264, "grad_norm": 0.029013920575380325, "learning_rate": 3.5943396226415093e-06, "loss": 0.0022, "step": 1740 }, { "epoch": 1.650943396226415, "grad_norm": 0.021010151132941246, "learning_rate": 3.5e-06, "loss": 0.0021, "step": 1750 }, { "epoch": 1.6603773584905661, "grad_norm": 0.020002691075205803, "learning_rate": 3.4056603773584906e-06, "loss": 0.0016, "step": 1760 }, { "epoch": 1.669811320754717, "grad_norm": 0.02407264895737171, "learning_rate": 3.3113207547169813e-06, "loss": 0.0928, "step": 1770 }, { "epoch": 1.6792452830188678, "grad_norm": 0.024595679715275764, "learning_rate": 3.216981132075472e-06, "loss": 0.0018, "step": 1780 }, { "epoch": 1.6886792452830188, "grad_norm": 0.017810633406043053, "learning_rate": 3.122641509433962e-06, "loss": 0.1722, "step": 1790 }, { "epoch": 1.6981132075471699, "grad_norm": 0.01612844318151474, "learning_rate": 3.028301886792453e-06, "loss": 0.1158, "step": 1800 }, { "epoch": 1.7075471698113207, "grad_norm": 0.023075569421052933, "learning_rate": 2.9339622641509435e-06, "loss": 0.0016, "step": 1810 }, { "epoch": 1.7169811320754715, "grad_norm": 0.01716834492981434, "learning_rate": 2.839622641509434e-06, "loss": 0.0015, "step": 1820 }, { "epoch": 1.7264150943396226, "grad_norm": 0.0205362718552351, "learning_rate": 2.745283018867925e-06, "loss": 0.0017, "step": 1830 }, { "epoch": 1.7358490566037736, "grad_norm": 0.023218706250190735, "learning_rate": 2.6509433962264155e-06, "loss": 0.0014, "step": 1840 }, { "epoch": 1.7452830188679245, "grad_norm": 0.018329111859202385, "learning_rate": 2.5566037735849057e-06, "loss": 0.1204, "step": 1850 }, { "epoch": 1.7547169811320755, "grad_norm": 0.019258936867117882, "learning_rate": 2.4622641509433964e-06, "loss": 0.0018, "step": 1860 }, { "epoch": 1.7641509433962264, "grad_norm": 0.023526819422841072, "learning_rate": 2.367924528301887e-06, "loss": 0.0074, "step": 1870 }, { "epoch": 1.7735849056603774, "grad_norm": 0.01679695025086403, "learning_rate": 2.2735849056603777e-06, "loss": 0.0014, "step": 1880 }, { "epoch": 1.7830188679245285, "grad_norm": 0.014311902225017548, "learning_rate": 2.1792452830188684e-06, "loss": 0.0015, "step": 1890 }, { "epoch": 1.7924528301886793, "grad_norm": 1.5082017183303833, "learning_rate": 2.0849056603773586e-06, "loss": 0.0708, "step": 1900 }, { "epoch": 1.8018867924528301, "grad_norm": 52.86256408691406, "learning_rate": 1.9905660377358493e-06, "loss": 0.0666, "step": 1910 }, { "epoch": 1.8113207547169812, "grad_norm": 0.020178191363811493, "learning_rate": 1.89622641509434e-06, "loss": 0.0694, "step": 1920 }, { "epoch": 1.8207547169811322, "grad_norm": 0.01885703019797802, "learning_rate": 1.8018867924528302e-06, "loss": 0.0416, "step": 1930 }, { "epoch": 1.830188679245283, "grad_norm": 0.021212713792920113, "learning_rate": 1.7075471698113208e-06, "loss": 0.0015, "step": 1940 }, { "epoch": 1.8396226415094339, "grad_norm": 0.020716039463877678, "learning_rate": 1.6132075471698113e-06, "loss": 0.0446, "step": 1950 }, { "epoch": 1.849056603773585, "grad_norm": 0.01979793794453144, "learning_rate": 1.518867924528302e-06, "loss": 0.0017, "step": 1960 }, { "epoch": 1.858490566037736, "grad_norm": 0.029191287234425545, "learning_rate": 1.4245283018867926e-06, "loss": 0.0016, "step": 1970 }, { "epoch": 1.8679245283018868, "grad_norm": 0.02217436581850052, "learning_rate": 1.330188679245283e-06, "loss": 0.0441, "step": 1980 }, { "epoch": 1.8773584905660377, "grad_norm": 0.0228409580886364, "learning_rate": 1.2358490566037737e-06, "loss": 0.0016, "step": 1990 }, { "epoch": 1.8867924528301887, "grad_norm": 0.022314254194498062, "learning_rate": 1.1415094339622642e-06, "loss": 0.1457, "step": 2000 }, { "epoch": 1.8962264150943398, "grad_norm": 0.017337264493107796, "learning_rate": 1.0471698113207548e-06, "loss": 0.1144, "step": 2010 }, { "epoch": 1.9056603773584906, "grad_norm": 0.021478446200489998, "learning_rate": 9.528301886792454e-07, "loss": 0.0014, "step": 2020 }, { "epoch": 1.9150943396226414, "grad_norm": 0.021643251180648804, "learning_rate": 8.584905660377358e-07, "loss": 0.0284, "step": 2030 }, { "epoch": 1.9245283018867925, "grad_norm": 0.02016671560704708, "learning_rate": 7.641509433962264e-07, "loss": 0.0482, "step": 2040 }, { "epoch": 1.9339622641509435, "grad_norm": 0.019274510443210602, "learning_rate": 6.698113207547171e-07, "loss": 0.0016, "step": 2050 }, { "epoch": 1.9433962264150944, "grad_norm": 0.014301498420536518, "learning_rate": 5.754716981132076e-07, "loss": 0.0045, "step": 2060 }, { "epoch": 1.9528301886792452, "grad_norm": 0.01988300494849682, "learning_rate": 4.811320754716982e-07, "loss": 0.0015, "step": 2070 }, { "epoch": 1.9622641509433962, "grad_norm": 0.02060701698064804, "learning_rate": 3.867924528301887e-07, "loss": 0.0044, "step": 2080 }, { "epoch": 1.9716981132075473, "grad_norm": 41.62403106689453, "learning_rate": 2.924528301886793e-07, "loss": 0.1818, "step": 2090 }, { "epoch": 1.9811320754716981, "grad_norm": 0.0649556890130043, "learning_rate": 1.9811320754716982e-07, "loss": 0.0016, "step": 2100 }, { "epoch": 1.990566037735849, "grad_norm": 0.02280914969742298, "learning_rate": 1.0377358490566039e-07, "loss": 0.0414, "step": 2110 }, { "epoch": 2.0, "grad_norm": 0.01854662224650383, "learning_rate": 9.433962264150945e-09, "loss": 0.0015, "step": 2120 }, { "epoch": 2.0, "eval_loss": 0.06546525657176971, "eval_runtime": 1.7009, "eval_samples_per_second": 554.403, "eval_steps_per_second": 69.374, "step": 2120 } ], "logging_steps": 10, "max_steps": 2120, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4461837276794880.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }