{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 300, "global_step": 2250, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0044444444444444444, "grad_norm": 240.13470458984375, "learning_rate": 2.7272727272727273e-05, "loss": 11.9773, "step": 10 }, { "epoch": 0.008888888888888889, "grad_norm": 34.34966278076172, "learning_rate": 7.272727272727273e-05, "loss": 9.6995, "step": 20 }, { "epoch": 0.013333333333333334, "grad_norm": 22.52939224243164, "learning_rate": 0.0001181818181818182, "loss": 6.6389, "step": 30 }, { "epoch": 0.017777777777777778, "grad_norm": 19.9622745513916, "learning_rate": 0.00016363636363636366, "loss": 3.7452, "step": 40 }, { "epoch": 0.022222222222222223, "grad_norm": 16.391326904296875, "learning_rate": 0.0001999995943808681, "loss": 3.7348, "step": 50 }, { "epoch": 0.02666666666666667, "grad_norm": 13.753044128417969, "learning_rate": 0.00019998539805675495, "loss": 3.3315, "step": 60 }, { "epoch": 0.03111111111111111, "grad_norm": 12.171226501464844, "learning_rate": 0.00019995092406646422, "loss": 3.2473, "step": 70 }, { "epoch": 0.035555555555555556, "grad_norm": 14.022675514221191, "learning_rate": 0.00019989617940153752, "loss": 3.1642, "step": 80 }, { "epoch": 0.04, "grad_norm": 10.995219230651855, "learning_rate": 0.0001998211751645364, "loss": 2.9791, "step": 90 }, { "epoch": 0.044444444444444446, "grad_norm": 13.217508316040039, "learning_rate": 0.00019972592656679084, "loss": 3.0747, "step": 100 }, { "epoch": 0.04888888888888889, "grad_norm": 13.215702056884766, "learning_rate": 0.0001996104529253142, "loss": 2.8731, "step": 110 }, { "epoch": 0.05333333333333334, "grad_norm": 12.901390075683594, "learning_rate": 0.00019947477765888558, "loss": 2.6857, "step": 120 }, { "epoch": 0.057777777777777775, "grad_norm": 10.362764358520508, "learning_rate": 0.00019931892828330047, "loss": 2.6396, "step": 130 }, { "epoch": 0.06222222222222222, "grad_norm": 17.187591552734375, "learning_rate": 0.00019914293640579023, "loss": 2.4217, "step": 140 }, { "epoch": 0.06666666666666667, "grad_norm": 11.482383728027344, "learning_rate": 0.00019894683771861206, "loss": 2.2934, "step": 150 }, { "epoch": 0.07111111111111111, "grad_norm": 9.790170669555664, "learning_rate": 0.00019873067199181033, "loss": 2.372, "step": 160 }, { "epoch": 0.07555555555555556, "grad_norm": 12.26194953918457, "learning_rate": 0.00019849448306515094, "loss": 2.2495, "step": 170 }, { "epoch": 0.08, "grad_norm": 10.313677787780762, "learning_rate": 0.00019823831883923046, "loss": 2.2395, "step": 180 }, { "epoch": 0.08444444444444445, "grad_norm": 12.778143882751465, "learning_rate": 0.00019796223126576138, "loss": 2.2095, "step": 190 }, { "epoch": 0.08888888888888889, "grad_norm": 10.047595024108887, "learning_rate": 0.00019766627633703622, "loss": 2.1296, "step": 200 }, { "epoch": 0.09333333333333334, "grad_norm": 13.92696475982666, "learning_rate": 0.00019735051407457176, "loss": 2.0027, "step": 210 }, { "epoch": 0.09777777777777778, "grad_norm": 13.361489295959473, "learning_rate": 0.00019701500851693635, "loss": 1.8503, "step": 220 }, { "epoch": 0.10222222222222223, "grad_norm": 11.596494674682617, "learning_rate": 0.00019665982770676255, "loss": 2.0362, "step": 230 }, { "epoch": 0.10666666666666667, "grad_norm": 15.200052261352539, "learning_rate": 0.00019628504367694754, "loss": 1.9158, "step": 240 }, { "epoch": 0.1111111111111111, "grad_norm": 14.349342346191406, "learning_rate": 0.00019589073243604454, "loss": 1.8403, "step": 250 }, { "epoch": 0.11555555555555555, "grad_norm": 14.814216613769531, "learning_rate": 0.00019547697395284766, "loss": 1.9161, "step": 260 }, { "epoch": 0.12, "grad_norm": 8.870625495910645, "learning_rate": 0.0001950438521401738, "loss": 1.7764, "step": 270 }, { "epoch": 0.12444444444444444, "grad_norm": 10.462891578674316, "learning_rate": 0.0001945914548378446, "loss": 1.9073, "step": 280 }, { "epoch": 0.1288888888888889, "grad_norm": 10.39869499206543, "learning_rate": 0.00019411987379487196, "loss": 1.7828, "step": 290 }, { "epoch": 0.13333333333333333, "grad_norm": 10.365347862243652, "learning_rate": 0.0001936292046508508, "loss": 1.7332, "step": 300 }, { "epoch": 0.13333333333333333, "eval_code_acc": 0.0, "eval_exact_match": 0.0, "eval_family_acc": 0.0, "eval_loss": 1.4483628273010254, "eval_runtime": 275.4346, "eval_samples_per_second": 1.815, "eval_steps_per_second": 1.815, "eval_super_acc": 0.0, "step": 300 }, { "epoch": 0.13777777777777778, "grad_norm": 12.327225685119629, "learning_rate": 0.00019311954691656264, "loss": 1.6749, "step": 310 }, { "epoch": 0.14222222222222222, "grad_norm": 11.838521003723145, "learning_rate": 0.00019259100395379434, "loss": 1.6202, "step": 320 }, { "epoch": 0.14666666666666667, "grad_norm": 14.03673267364502, "learning_rate": 0.0001920436829543756, "loss": 1.6491, "step": 330 }, { "epoch": 0.1511111111111111, "grad_norm": 15.616336822509766, "learning_rate": 0.00019147769491843978, "loss": 1.683, "step": 340 }, { "epoch": 0.15555555555555556, "grad_norm": 12.534092903137207, "learning_rate": 0.00019089315463191234, "loss": 1.5952, "step": 350 }, { "epoch": 0.16, "grad_norm": 11.09432315826416, "learning_rate": 0.00019029018064323165, "loss": 1.622, "step": 360 }, { "epoch": 0.16444444444444445, "grad_norm": 12.834360122680664, "learning_rate": 0.00018966889523930656, "loss": 1.5646, "step": 370 }, { "epoch": 0.1688888888888889, "grad_norm": 25.459949493408203, "learning_rate": 0.0001890294244207158, "loss": 1.5177, "step": 380 }, { "epoch": 0.17333333333333334, "grad_norm": 19.265300750732422, "learning_rate": 0.0001883718978761544, "loss": 1.5986, "step": 390 }, { "epoch": 0.17777777777777778, "grad_norm": 11.887044906616211, "learning_rate": 0.00018769644895613174, "loss": 1.613, "step": 400 }, { "epoch": 0.18222222222222223, "grad_norm": 21.538328170776367, "learning_rate": 0.00018700321464592764, "loss": 1.5335, "step": 410 }, { "epoch": 0.18666666666666668, "grad_norm": 18.00100326538086, "learning_rate": 0.00018629233553781052, "loss": 1.4608, "step": 420 }, { "epoch": 0.19111111111111112, "grad_norm": 14.892653465270996, "learning_rate": 0.00018556395580252458, "loss": 1.5059, "step": 430 }, { "epoch": 0.19555555555555557, "grad_norm": 10.61614990234375, "learning_rate": 0.000184818223160051, "loss": 1.2732, "step": 440 }, { "epoch": 0.2, "grad_norm": 11.014739990234375, "learning_rate": 0.00018405528884964952, "loss": 1.3921, "step": 450 }, { "epoch": 0.20444444444444446, "grad_norm": 9.74208927154541, "learning_rate": 0.00018327530759918597, "loss": 1.3523, "step": 460 }, { "epoch": 0.2088888888888889, "grad_norm": 15.92335319519043, "learning_rate": 0.0001824784375937528, "loss": 1.4222, "step": 470 }, { "epoch": 0.21333333333333335, "grad_norm": 15.541021347045898, "learning_rate": 0.00018166484044358764, "loss": 1.3772, "step": 480 }, { "epoch": 0.21777777777777776, "grad_norm": 13.139538764953613, "learning_rate": 0.00018083468115129834, "loss": 1.4395, "step": 490 }, { "epoch": 0.2222222222222222, "grad_norm": 21.101469039916992, "learning_rate": 0.00017998812807839892, "loss": 1.3891, "step": 500 }, { "epoch": 0.22666666666666666, "grad_norm": 14.130121231079102, "learning_rate": 0.00017912535291116508, "loss": 1.3276, "step": 510 }, { "epoch": 0.2311111111111111, "grad_norm": 25.728923797607422, "learning_rate": 0.00017824653062581503, "loss": 1.4297, "step": 520 }, { "epoch": 0.23555555555555555, "grad_norm": 14.234143257141113, "learning_rate": 0.00017735183945302322, "loss": 1.4121, "step": 530 }, { "epoch": 0.24, "grad_norm": 14.950554847717285, "learning_rate": 0.00017644146084177406, "loss": 1.2576, "step": 540 }, { "epoch": 0.24444444444444444, "grad_norm": 9.834321022033691, "learning_rate": 0.00017551557942256294, "loss": 1.3496, "step": 550 }, { "epoch": 0.24888888888888888, "grad_norm": 20.81060218811035, "learning_rate": 0.00017457438296995196, "loss": 1.2409, "step": 560 }, { "epoch": 0.25333333333333335, "grad_norm": 16.057870864868164, "learning_rate": 0.00017361806236448817, "loss": 1.369, "step": 570 }, { "epoch": 0.2577777777777778, "grad_norm": 9.714545249938965, "learning_rate": 0.00017264681155399164, "loss": 1.2626, "step": 580 }, { "epoch": 0.26222222222222225, "grad_norm": 16.824140548706055, "learning_rate": 0.00017166082751422177, "loss": 1.2978, "step": 590 }, { "epoch": 0.26666666666666666, "grad_norm": 13.657912254333496, "learning_rate": 0.00017066031020892934, "loss": 1.1963, "step": 600 }, { "epoch": 0.27111111111111114, "grad_norm": 22.651229858398438, "learning_rate": 0.00016964546254930247, "loss": 1.1826, "step": 610 }, { "epoch": 0.27555555555555555, "grad_norm": 28.09796905517578, "learning_rate": 0.0001686164903528152, "loss": 1.2849, "step": 620 }, { "epoch": 0.28, "grad_norm": 22.29288673400879, "learning_rate": 0.00016757360230148618, "loss": 1.2473, "step": 630 }, { "epoch": 0.28444444444444444, "grad_norm": 29.369836807250977, "learning_rate": 0.00016651700989955682, "loss": 1.2133, "step": 640 }, { "epoch": 0.28888888888888886, "grad_norm": 10.829903602600098, "learning_rate": 0.00016544692743059684, "loss": 1.2379, "step": 650 }, { "epoch": 0.29333333333333333, "grad_norm": 10.857136726379395, "learning_rate": 0.0001643635719140461, "loss": 1.1889, "step": 660 }, { "epoch": 0.29777777777777775, "grad_norm": 10.762494087219238, "learning_rate": 0.00016326716306120195, "loss": 1.2591, "step": 670 }, { "epoch": 0.3022222222222222, "grad_norm": 13.926369667053223, "learning_rate": 0.00016215792323066012, "loss": 1.2075, "step": 680 }, { "epoch": 0.30666666666666664, "grad_norm": 10.43800163269043, "learning_rate": 0.00016103607738321925, "loss": 1.2076, "step": 690 }, { "epoch": 0.3111111111111111, "grad_norm": 18.675508499145508, "learning_rate": 0.0001599018530362573, "loss": 1.2339, "step": 700 }, { "epoch": 0.31555555555555553, "grad_norm": 10.020101547241211, "learning_rate": 0.0001587554802175895, "loss": 1.14, "step": 710 }, { "epoch": 0.32, "grad_norm": 8.47937297821045, "learning_rate": 0.0001575971914188175, "loss": 1.145, "step": 720 }, { "epoch": 0.3244444444444444, "grad_norm": 16.2773380279541, "learning_rate": 0.00015642722154817848, "loss": 1.1076, "step": 730 }, { "epoch": 0.3288888888888889, "grad_norm": 10.47890853881836, "learning_rate": 0.00015524580788290425, "loss": 1.1414, "step": 740 }, { "epoch": 0.3333333333333333, "grad_norm": 9.702156066894531, "learning_rate": 0.0001540531900211, "loss": 1.22, "step": 750 }, { "epoch": 0.3377777777777778, "grad_norm": 10.579848289489746, "learning_rate": 0.0001528496098331523, "loss": 1.1548, "step": 760 }, { "epoch": 0.3422222222222222, "grad_norm": 24.226659774780273, "learning_rate": 0.00015163531141267628, "loss": 1.1407, "step": 770 }, { "epoch": 0.3466666666666667, "grad_norm": 11.10332202911377, "learning_rate": 0.00015041054102701184, "loss": 1.1642, "step": 780 }, { "epoch": 0.3511111111111111, "grad_norm": 14.13973331451416, "learning_rate": 0.00014917554706727915, "loss": 1.1726, "step": 790 }, { "epoch": 0.35555555555555557, "grad_norm": 21.75472640991211, "learning_rate": 0.00014793057999800335, "loss": 1.1478, "step": 800 }, { "epoch": 0.36, "grad_norm": 12.123833656311035, "learning_rate": 0.0001466758923063189, "loss": 1.1939, "step": 810 }, { "epoch": 0.36444444444444446, "grad_norm": 9.510560035705566, "learning_rate": 0.00014541173845076323, "loss": 1.0843, "step": 820 }, { "epoch": 0.3688888888888889, "grad_norm": 17.031314849853516, "learning_rate": 0.00014413837480967145, "loss": 1.1181, "step": 830 }, { "epoch": 0.37333333333333335, "grad_norm": 16.022037506103516, "learning_rate": 0.00014285605962918084, "loss": 1.1542, "step": 840 }, { "epoch": 0.37777777777777777, "grad_norm": 12.77236270904541, "learning_rate": 0.00014156505297085713, "loss": 1.114, "step": 850 }, { "epoch": 0.38222222222222224, "grad_norm": 38.819454193115234, "learning_rate": 0.00014026561665895224, "loss": 1.0932, "step": 860 }, { "epoch": 0.38666666666666666, "grad_norm": 10.85486125946045, "learning_rate": 0.00013895801422730473, "loss": 1.11, "step": 870 }, { "epoch": 0.39111111111111113, "grad_norm": 10.986682891845703, "learning_rate": 0.00013764251086589353, "loss": 1.0752, "step": 880 }, { "epoch": 0.39555555555555555, "grad_norm": 7.950289726257324, "learning_rate": 0.00013631937336705568, "loss": 1.1817, "step": 890 }, { "epoch": 0.4, "grad_norm": 7.103327751159668, "learning_rate": 0.00013498887007137918, "loss": 1.0813, "step": 900 }, { "epoch": 0.40444444444444444, "grad_norm": 7.365835189819336, "learning_rate": 0.0001336512708132819, "loss": 1.0193, "step": 910 }, { "epoch": 0.4088888888888889, "grad_norm": 11.938828468322754, "learning_rate": 0.00013230684686628744, "loss": 1.1339, "step": 920 }, { "epoch": 0.41333333333333333, "grad_norm": 12.521608352661133, "learning_rate": 0.00013095587088800902, "loss": 1.0743, "step": 930 }, { "epoch": 0.4177777777777778, "grad_norm": 9.13135051727295, "learning_rate": 0.00012959861686485304, "loss": 1.0734, "step": 940 }, { "epoch": 0.4222222222222222, "grad_norm": 16.696514129638672, "learning_rate": 0.0001282353600564527, "loss": 1.1145, "step": 950 }, { "epoch": 0.4266666666666667, "grad_norm": 12.382914543151855, "learning_rate": 0.00012686637693984384, "loss": 0.9964, "step": 960 }, { "epoch": 0.4311111111111111, "grad_norm": 10.711663246154785, "learning_rate": 0.00012549194515339344, "loss": 1.0572, "step": 970 }, { "epoch": 0.43555555555555553, "grad_norm": 13.973264694213867, "learning_rate": 0.00012411234344049293, "loss": 1.0616, "step": 980 }, { "epoch": 0.44, "grad_norm": 10.161416053771973, "learning_rate": 0.0001227278515930273, "loss": 1.0561, "step": 990 }, { "epoch": 0.4444444444444444, "grad_norm": 27.74120330810547, "learning_rate": 0.00012133875039463148, "loss": 1.1011, "step": 1000 }, { "epoch": 0.4488888888888889, "grad_norm": 9.277678489685059, "learning_rate": 0.00011994532156374574, "loss": 1.0957, "step": 1010 }, { "epoch": 0.4533333333333333, "grad_norm": 9.599855422973633, "learning_rate": 0.00011854784769648137, "loss": 1.0394, "step": 1020 }, { "epoch": 0.4577777777777778, "grad_norm": 9.483017921447754, "learning_rate": 0.00011714661220930833, "loss": 0.9773, "step": 1030 }, { "epoch": 0.4622222222222222, "grad_norm": 11.467011451721191, "learning_rate": 0.00011574189928157689, "loss": 1.0346, "step": 1040 }, { "epoch": 0.4666666666666667, "grad_norm": 8.953259468078613, "learning_rate": 0.00011433399379788387, "loss": 1.0622, "step": 1050 }, { "epoch": 0.4711111111111111, "grad_norm": 6.459799289703369, "learning_rate": 0.00011292318129029665, "loss": 0.9814, "step": 1060 }, { "epoch": 0.47555555555555556, "grad_norm": 8.728630065917969, "learning_rate": 0.00011150974788044521, "loss": 1.0526, "step": 1070 }, { "epoch": 0.48, "grad_norm": 7.729814529418945, "learning_rate": 0.00011009398022149495, "loss": 0.9997, "step": 1080 }, { "epoch": 0.48444444444444446, "grad_norm": 9.42880916595459, "learning_rate": 0.00010867616544001164, "loss": 0.9999, "step": 1090 }, { "epoch": 0.4888888888888889, "grad_norm": 8.457280158996582, "learning_rate": 0.00010725659107773045, "loss": 1.0464, "step": 1100 }, { "epoch": 0.49333333333333335, "grad_norm": 8.715860366821289, "learning_rate": 0.00010583554503324044, "loss": 1.0088, "step": 1110 }, { "epoch": 0.49777777777777776, "grad_norm": 6.529873847961426, "learning_rate": 0.00010441331550359712, "loss": 1.0749, "step": 1120 }, { "epoch": 0.5022222222222222, "grad_norm": 7.668039798736572, "learning_rate": 0.0001029901909258742, "loss": 1.026, "step": 1130 }, { "epoch": 0.5066666666666667, "grad_norm": 8.81876277923584, "learning_rate": 0.00010156645991866677, "loss": 1.0293, "step": 1140 }, { "epoch": 0.5111111111111111, "grad_norm": 8.55112075805664, "learning_rate": 0.00010014241122355762, "loss": 1.0282, "step": 1150 }, { "epoch": 0.5155555555555555, "grad_norm": 9.484146118164062, "learning_rate": 9.871833364655865e-05, "loss": 0.9964, "step": 1160 }, { "epoch": 0.52, "grad_norm": 10.939757347106934, "learning_rate": 9.729451599953917e-05, "loss": 1.0519, "step": 1170 }, { "epoch": 0.5244444444444445, "grad_norm": 9.229081153869629, "learning_rate": 9.587124704165302e-05, "loss": 1.0511, "step": 1180 }, { "epoch": 0.5288888888888889, "grad_norm": 12.294286727905273, "learning_rate": 9.44488154207766e-05, "loss": 1.0302, "step": 1190 }, { "epoch": 0.5333333333333333, "grad_norm": 11.527563095092773, "learning_rate": 9.302750961496888e-05, "loss": 1.0333, "step": 1200 }, { "epoch": 0.5377777777777778, "grad_norm": 10.483113288879395, "learning_rate": 9.160761787396665e-05, "loss": 0.9749, "step": 1210 }, { "epoch": 0.5422222222222223, "grad_norm": 9.348003387451172, "learning_rate": 9.018942816072545e-05, "loss": 0.9837, "step": 1220 }, { "epoch": 0.5466666666666666, "grad_norm": 9.57206916809082, "learning_rate": 8.87732280930188e-05, "loss": 1.0002, "step": 1230 }, { "epoch": 0.5511111111111111, "grad_norm": 9.370091438293457, "learning_rate": 8.735930488510774e-05, "loss": 1.0049, "step": 1240 }, { "epoch": 0.5555555555555556, "grad_norm": 9.066927909851074, "learning_rate": 8.594794528949183e-05, "loss": 0.9549, "step": 1250 }, { "epoch": 0.56, "grad_norm": 13.580326080322266, "learning_rate": 8.453943553875392e-05, "loss": 1.0505, "step": 1260 }, { "epoch": 0.5644444444444444, "grad_norm": 9.729880332946777, "learning_rate": 8.313406128751049e-05, "loss": 1.0413, "step": 1270 }, { "epoch": 0.5688888888888889, "grad_norm": 10.354995727539062, "learning_rate": 8.173210755447905e-05, "loss": 1.033, "step": 1280 }, { "epoch": 0.5733333333333334, "grad_norm": 10.784231185913086, "learning_rate": 8.033385866467444e-05, "loss": 1.0747, "step": 1290 }, { "epoch": 0.5777777777777777, "grad_norm": 8.267210006713867, "learning_rate": 7.893959819174619e-05, "loss": 0.9777, "step": 1300 }, { "epoch": 0.5822222222222222, "grad_norm": 8.181448936462402, "learning_rate": 7.754960890046785e-05, "loss": 0.9738, "step": 1310 }, { "epoch": 0.5866666666666667, "grad_norm": 6.555637836456299, "learning_rate": 7.616417268939037e-05, "loss": 0.9659, "step": 1320 }, { "epoch": 0.5911111111111111, "grad_norm": 8.430340766906738, "learning_rate": 7.47835705336716e-05, "loss": 0.999, "step": 1330 }, { "epoch": 0.5955555555555555, "grad_norm": 7.698472023010254, "learning_rate": 7.340808242809264e-05, "loss": 0.9666, "step": 1340 }, { "epoch": 0.6, "grad_norm": 6.320609092712402, "learning_rate": 7.203798733027304e-05, "loss": 0.9954, "step": 1350 }, { "epoch": 0.6044444444444445, "grad_norm": 7.057352542877197, "learning_rate": 7.067356310409659e-05, "loss": 0.9971, "step": 1360 }, { "epoch": 0.6088888888888889, "grad_norm": 10.81286334991455, "learning_rate": 6.931508646335874e-05, "loss": 0.9931, "step": 1370 }, { "epoch": 0.6133333333333333, "grad_norm": 7.427656173706055, "learning_rate": 6.796283291564722e-05, "loss": 0.9491, "step": 1380 }, { "epoch": 0.6177777777777778, "grad_norm": 7.356409072875977, "learning_rate": 6.66170767064675e-05, "loss": 1.0202, "step": 1390 }, { "epoch": 0.6222222222222222, "grad_norm": 8.578875541687012, "learning_rate": 6.527809076362399e-05, "loss": 1.0542, "step": 1400 }, { "epoch": 0.6266666666666667, "grad_norm": 8.644619941711426, "learning_rate": 6.394614664186862e-05, "loss": 1.0267, "step": 1410 }, { "epoch": 0.6311111111111111, "grad_norm": 9.160662651062012, "learning_rate": 6.262151446782785e-05, "loss": 0.9914, "step": 1420 }, { "epoch": 0.6355555555555555, "grad_norm": 7.767285346984863, "learning_rate": 6.130446288521915e-05, "loss": 0.987, "step": 1430 }, { "epoch": 0.64, "grad_norm": 12.88818073272705, "learning_rate": 5.999525900036855e-05, "loss": 0.9676, "step": 1440 }, { "epoch": 0.6444444444444445, "grad_norm": 10.068846702575684, "learning_rate": 5.86941683280398e-05, "loss": 0.9942, "step": 1450 }, { "epoch": 0.6488888888888888, "grad_norm": 8.70479679107666, "learning_rate": 5.7401454737586055e-05, "loss": 0.9848, "step": 1460 }, { "epoch": 0.6533333333333333, "grad_norm": 10.635972023010254, "learning_rate": 5.6117380399435826e-05, "loss": 0.9892, "step": 1470 }, { "epoch": 0.6577777777777778, "grad_norm": 6.84842586517334, "learning_rate": 5.484220573192307e-05, "loss": 0.961, "step": 1480 }, { "epoch": 0.6622222222222223, "grad_norm": 6.793154716491699, "learning_rate": 5.3576189348472526e-05, "loss": 0.9772, "step": 1490 }, { "epoch": 0.6666666666666666, "grad_norm": 10.692822456359863, "learning_rate": 5.231958800515164e-05, "loss": 1.0044, "step": 1500 }, { "epoch": 0.6711111111111111, "grad_norm": 7.949609279632568, "learning_rate": 5.107265654859855e-05, "loss": 1.0194, "step": 1510 }, { "epoch": 0.6755555555555556, "grad_norm": 8.028242111206055, "learning_rate": 4.983564786433763e-05, "loss": 0.9705, "step": 1520 }, { "epoch": 0.68, "grad_norm": 8.18526840209961, "learning_rate": 4.860881282549285e-05, "loss": 0.9802, "step": 1530 }, { "epoch": 0.6844444444444444, "grad_norm": 9.321311950683594, "learning_rate": 4.739240024190904e-05, "loss": 0.9649, "step": 1540 }, { "epoch": 0.6888888888888889, "grad_norm": 10.959417343139648, "learning_rate": 4.618665680969163e-05, "loss": 0.9957, "step": 1550 }, { "epoch": 0.6933333333333334, "grad_norm": 9.302586555480957, "learning_rate": 4.49918270611752e-05, "loss": 0.9833, "step": 1560 }, { "epoch": 0.6977777777777778, "grad_norm": 7.047448635101318, "learning_rate": 4.380815331533088e-05, "loss": 1.0179, "step": 1570 }, { "epoch": 0.7022222222222222, "grad_norm": 9.307101249694824, "learning_rate": 4.2635875628622345e-05, "loss": 0.9883, "step": 1580 }, { "epoch": 0.7066666666666667, "grad_norm": 8.306827545166016, "learning_rate": 4.147523174632103e-05, "loss": 0.984, "step": 1590 }, { "epoch": 0.7111111111111111, "grad_norm": 9.073155403137207, "learning_rate": 4.032645705428985e-05, "loss": 0.9916, "step": 1600 }, { "epoch": 0.7155555555555555, "grad_norm": 11.148294448852539, "learning_rate": 3.9189784531245334e-05, "loss": 0.993, "step": 1610 }, { "epoch": 0.72, "grad_norm": 7.878681659698486, "learning_rate": 3.806544470150831e-05, "loss": 0.9733, "step": 1620 }, { "epoch": 0.7244444444444444, "grad_norm": 9.204869270324707, "learning_rate": 3.6953665588251984e-05, "loss": 0.9689, "step": 1630 }, { "epoch": 0.7288888888888889, "grad_norm": 8.391727447509766, "learning_rate": 3.585467266725737e-05, "loss": 0.9782, "step": 1640 }, { "epoch": 0.7333333333333333, "grad_norm": 6.572085857391357, "learning_rate": 3.4768688821185566e-05, "loss": 0.9548, "step": 1650 }, { "epoch": 0.7377777777777778, "grad_norm": 9.943083763122559, "learning_rate": 3.3695934294375544e-05, "loss": 0.9904, "step": 1660 }, { "epoch": 0.7422222222222222, "grad_norm": 8.165312767028809, "learning_rate": 3.263662664817728e-05, "loss": 0.9728, "step": 1670 }, { "epoch": 0.7466666666666667, "grad_norm": 9.635257720947266, "learning_rate": 3.15909807168291e-05, "loss": 0.961, "step": 1680 }, { "epoch": 0.7511111111111111, "grad_norm": 7.636417865753174, "learning_rate": 3.055920856388779e-05, "loss": 0.9403, "step": 1690 }, { "epoch": 0.7555555555555555, "grad_norm": 6.770568370819092, "learning_rate": 2.95415194392207e-05, "loss": 0.9484, "step": 1700 }, { "epoch": 0.76, "grad_norm": 7.254674434661865, "learning_rate": 2.8538119736568845e-05, "loss": 0.9701, "step": 1710 }, { "epoch": 0.7644444444444445, "grad_norm": 8.287463188171387, "learning_rate": 2.7549212951688598e-05, "loss": 0.9591, "step": 1720 }, { "epoch": 0.7688888888888888, "grad_norm": 8.489920616149902, "learning_rate": 2.6574999641081812e-05, "loss": 0.9285, "step": 1730 }, { "epoch": 0.7733333333333333, "grad_norm": 7.725697994232178, "learning_rate": 2.561567738132149e-05, "loss": 0.8912, "step": 1740 }, { "epoch": 0.7777777777777778, "grad_norm": 8.986964225769043, "learning_rate": 2.467144072898202e-05, "loss": 0.9386, "step": 1750 }, { "epoch": 0.7822222222222223, "grad_norm": 8.926631927490234, "learning_rate": 2.3742481181182065e-05, "loss": 0.9224, "step": 1760 }, { "epoch": 0.7866666666666666, "grad_norm": 7.921815395355225, "learning_rate": 2.2828987136747505e-05, "loss": 0.9393, "step": 1770 }, { "epoch": 0.7911111111111111, "grad_norm": 6.680901050567627, "learning_rate": 2.193114385800309e-05, "loss": 0.9359, "step": 1780 }, { "epoch": 0.7955555555555556, "grad_norm": 6.957186698913574, "learning_rate": 2.104913343320013e-05, "loss": 0.9285, "step": 1790 }, { "epoch": 0.8, "grad_norm": 7.6232008934021, "learning_rate": 2.0183134739587807e-05, "loss": 0.9083, "step": 1800 }, { "epoch": 0.8044444444444444, "grad_norm": 6.6202898025512695, "learning_rate": 1.9333323407135652e-05, "loss": 0.9497, "step": 1810 }, { "epoch": 0.8088888888888889, "grad_norm": 8.43086051940918, "learning_rate": 1.8499871782914823e-05, "loss": 0.8944, "step": 1820 }, { "epoch": 0.8133333333333334, "grad_norm": 7.180028915405273, "learning_rate": 1.7682948896145037e-05, "loss": 0.9319, "step": 1830 }, { "epoch": 0.8177777777777778, "grad_norm": 9.3683500289917, "learning_rate": 1.688272042391421e-05, "loss": 0.9467, "step": 1840 }, { "epoch": 0.8222222222222222, "grad_norm": 7.685975551605225, "learning_rate": 1.609934865757835e-05, "loss": 0.8837, "step": 1850 }, { "epoch": 0.8266666666666667, "grad_norm": 6.820009231567383, "learning_rate": 1.5332992469847595e-05, "loss": 0.8969, "step": 1860 }, { "epoch": 0.8311111111111111, "grad_norm": 6.4122161865234375, "learning_rate": 1.4583807282566109e-05, "loss": 0.892, "step": 1870 }, { "epoch": 0.8355555555555556, "grad_norm": 11.386307716369629, "learning_rate": 1.3851945035191271e-05, "loss": 0.9494, "step": 1880 }, { "epoch": 0.84, "grad_norm": 7.07219934463501, "learning_rate": 1.3137554153979648e-05, "loss": 0.9254, "step": 1890 }, { "epoch": 0.8444444444444444, "grad_norm": 8.010754585266113, "learning_rate": 1.2440779521885026e-05, "loss": 0.8829, "step": 1900 }, { "epoch": 0.8488888888888889, "grad_norm": 7.019892692565918, "learning_rate": 1.1761762449175362e-05, "loss": 0.9155, "step": 1910 }, { "epoch": 0.8533333333333334, "grad_norm": 9.830785751342773, "learning_rate": 1.1100640644774174e-05, "loss": 0.9572, "step": 1920 }, { "epoch": 0.8577777777777778, "grad_norm": 8.173066139221191, "learning_rate": 1.0457548188332156e-05, "loss": 0.9465, "step": 1930 }, { "epoch": 0.8622222222222222, "grad_norm": 7.015248775482178, "learning_rate": 9.83261550303518e-06, "loss": 0.8799, "step": 1940 }, { "epoch": 0.8666666666666667, "grad_norm": 7.390592575073242, "learning_rate": 9.225969329153572e-06, "loss": 0.9065, "step": 1950 }, { "epoch": 0.8711111111111111, "grad_norm": 8.248661994934082, "learning_rate": 8.637732698338353e-06, "loss": 0.8821, "step": 1960 }, { "epoch": 0.8755555555555555, "grad_norm": 7.508656024932861, "learning_rate": 8.068024908669658e-06, "loss": 0.9156, "step": 1970 }, { "epoch": 0.88, "grad_norm": 11.526093482971191, "learning_rate": 7.516961500462438e-06, "loss": 0.9091, "step": 1980 }, { "epoch": 0.8844444444444445, "grad_norm": 7.947896480560303, "learning_rate": 6.984654232833998e-06, "loss": 0.9023, "step": 1990 }, { "epoch": 0.8888888888888888, "grad_norm": 8.093976974487305, "learning_rate": 6.471211061038695e-06, "loss": 0.8878, "step": 2000 }, { "epoch": 0.8933333333333333, "grad_norm": 11.323365211486816, "learning_rate": 5.976736114573867e-06, "loss": 0.9262, "step": 2010 }, { "epoch": 0.8977777777777778, "grad_norm": 8.999126434326172, "learning_rate": 5.501329676061662e-06, "loss": 0.878, "step": 2020 }, { "epoch": 0.9022222222222223, "grad_norm": 8.701828956604004, "learning_rate": 5.045088160911227e-06, "loss": 0.8847, "step": 2030 }, { "epoch": 0.9066666666666666, "grad_norm": 6.862762451171875, "learning_rate": 4.60810409776491e-06, "loss": 0.8999, "step": 2040 }, { "epoch": 0.9111111111111111, "grad_norm": 7.2732672691345215, "learning_rate": 4.190466109733004e-06, "loss": 0.9271, "step": 2050 }, { "epoch": 0.9155555555555556, "grad_norm": 6.5058979988098145, "learning_rate": 3.7922588964203533e-06, "loss": 0.8799, "step": 2060 }, { "epoch": 0.92, "grad_norm": 7.603868007659912, "learning_rate": 3.4135632167487274e-06, "loss": 0.9081, "step": 2070 }, { "epoch": 0.9244444444444444, "grad_norm": 7.810312271118164, "learning_rate": 3.054455872578421e-06, "loss": 0.9159, "step": 2080 }, { "epoch": 0.9288888888888889, "grad_norm": 8.112349510192871, "learning_rate": 2.71500969313242e-06, "loss": 0.8864, "step": 2090 }, { "epoch": 0.9333333333333333, "grad_norm": 8.802750587463379, "learning_rate": 2.3952935202260608e-06, "loss": 0.8618, "step": 2100 }, { "epoch": 0.9377777777777778, "grad_norm": 7.140369892120361, "learning_rate": 2.095372194305578e-06, "loss": 0.8829, "step": 2110 }, { "epoch": 0.9422222222222222, "grad_norm": 8.534485816955566, "learning_rate": 1.815306541298023e-06, "loss": 0.8757, "step": 2120 }, { "epoch": 0.9466666666666667, "grad_norm": 10.03408145904541, "learning_rate": 1.55515336027533e-06, "loss": 0.8699, "step": 2130 }, { "epoch": 0.9511111111111111, "grad_norm": 10.744848251342773, "learning_rate": 1.3149654119351874e-06, "loss": 0.9025, "step": 2140 }, { "epoch": 0.9555555555555556, "grad_norm": 12.541043281555176, "learning_rate": 1.0947914079008193e-06, "loss": 0.9111, "step": 2150 }, { "epoch": 0.96, "grad_norm": 9.694021224975586, "learning_rate": 8.946760008419208e-07, "loss": 0.8739, "step": 2160 }, { "epoch": 0.9644444444444444, "grad_norm": 8.234989166259766, "learning_rate": 7.146597754188578e-07, "loss": 0.8864, "step": 2170 }, { "epoch": 0.9688888888888889, "grad_norm": 8.109750747680664, "learning_rate": 5.547792400518171e-07, "loss": 0.8951, "step": 2180 }, { "epoch": 0.9733333333333334, "grad_norm": 5.8855180740356445, "learning_rate": 4.150668195166962e-07, "loss": 0.9224, "step": 2190 }, { "epoch": 0.9777777777777777, "grad_norm": 6.274494171142578, "learning_rate": 2.955508483691416e-07, "loss": 0.867, "step": 2200 }, { "epoch": 0.9822222222222222, "grad_norm": 9.63640022277832, "learning_rate": 1.9625556519811128e-07, "loss": 0.8781, "step": 2210 }, { "epoch": 0.9866666666666667, "grad_norm": 8.013216018676758, "learning_rate": 1.1720110771015202e-07, "loss": 0.9555, "step": 2220 }, { "epoch": 0.9911111111111112, "grad_norm": 11.773499488830566, "learning_rate": 5.840350864533273e-08, "loss": 0.8768, "step": 2230 }, { "epoch": 0.9955555555555555, "grad_norm": 6.226689338684082, "learning_rate": 1.9874692525745365e-08, "loss": 0.8743, "step": 2240 }, { "epoch": 1.0, "grad_norm": 6.99469518661499, "learning_rate": 1.6224732370728746e-09, "loss": 0.8964, "step": 2250 } ], "logging_steps": 10, "max_steps": 2250, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 300, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 6.137653654388736e+19, "train_batch_size": 4, "trial_name": null, "trial_params": null }