{ "best_global_step": 6000, "best_metric": 0.6719915270805359, "best_model_checkpoint": "/workspace/rails-finetune/adapters-qwen3-coder-30b/checkpoint-6000", "epoch": 0.9581794590278471, "eval_steps": 500, "global_step": 6000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0015969657650464117, "grad_norm": 0.15089154243469238, "learning_rate": 6.000000000000001e-07, "loss": 1.5709858894348145, "step": 10 }, { "epoch": 0.0031939315300928235, "grad_norm": 0.1435190588235855, "learning_rate": 1.2666666666666669e-06, "loss": 1.6188209533691407, "step": 20 }, { "epoch": 0.004790897295139236, "grad_norm": 0.14615978300571442, "learning_rate": 1.9333333333333336e-06, "loss": 1.6381675720214843, "step": 30 }, { "epoch": 0.006387863060185647, "grad_norm": 0.1281592696905136, "learning_rate": 2.6e-06, "loss": 1.5563851356506349, "step": 40 }, { "epoch": 0.00798482882523206, "grad_norm": 0.16032098233699799, "learning_rate": 3.266666666666667e-06, "loss": 1.6382104873657226, "step": 50 }, { "epoch": 0.009581794590278471, "grad_norm": 0.14609354734420776, "learning_rate": 3.9333333333333335e-06, "loss": 1.6256795883178712, "step": 60 }, { "epoch": 0.011178760355324883, "grad_norm": 0.13070641458034515, "learning_rate": 4.600000000000001e-06, "loss": 1.5222463607788086, "step": 70 }, { "epoch": 0.012775726120371294, "grad_norm": 0.1499311476945877, "learning_rate": 5.2666666666666665e-06, "loss": 1.518197727203369, "step": 80 }, { "epoch": 0.014372691885417706, "grad_norm": 0.12919697165489197, "learning_rate": 5.933333333333335e-06, "loss": 1.4710905075073242, "step": 90 }, { "epoch": 0.01596965765046412, "grad_norm": 0.11437301337718964, "learning_rate": 6.600000000000001e-06, "loss": 1.4049152374267577, "step": 100 }, { "epoch": 0.01756662341551053, "grad_norm": 0.13434389233589172, "learning_rate": 7.266666666666668e-06, "loss": 1.3853497505187988, "step": 110 }, { "epoch": 0.019163589180556943, "grad_norm": 0.08636850863695145, "learning_rate": 7.933333333333334e-06, "loss": 1.3351438522338868, "step": 120 }, { "epoch": 0.020760554945603353, "grad_norm": 0.08471965044736862, "learning_rate": 8.6e-06, "loss": 1.2937309265136718, "step": 130 }, { "epoch": 0.022357520710649767, "grad_norm": 0.058983951807022095, "learning_rate": 9.266666666666667e-06, "loss": 1.2116679191589355, "step": 140 }, { "epoch": 0.023954486475696177, "grad_norm": 0.06733114272356033, "learning_rate": 9.933333333333334e-06, "loss": 1.1680314064025878, "step": 150 }, { "epoch": 0.025551452240742588, "grad_norm": 0.045156873762607574, "learning_rate": 1.0600000000000002e-05, "loss": 1.1280742645263673, "step": 160 }, { "epoch": 0.027148418005789002, "grad_norm": 0.0415460579097271, "learning_rate": 1.1266666666666668e-05, "loss": 1.0879782676696776, "step": 170 }, { "epoch": 0.028745383770835412, "grad_norm": 0.05221620574593544, "learning_rate": 1.1933333333333335e-05, "loss": 1.0641801834106446, "step": 180 }, { "epoch": 0.030342349535881826, "grad_norm": 0.036492571234703064, "learning_rate": 1.2600000000000001e-05, "loss": 1.0208759307861328, "step": 190 }, { "epoch": 0.03193931530092824, "grad_norm": 0.03153146430850029, "learning_rate": 1.3266666666666668e-05, "loss": 1.0204400062561034, "step": 200 }, { "epoch": 0.03353628106597465, "grad_norm": 0.03343074768781662, "learning_rate": 1.3933333333333334e-05, "loss": 0.9532640457153321, "step": 210 }, { "epoch": 0.03513324683102106, "grad_norm": 0.034289468079805374, "learning_rate": 1.46e-05, "loss": 0.9065071105957031, "step": 220 }, { "epoch": 0.03673021259606747, "grad_norm": 0.040245093405246735, "learning_rate": 1.5266666666666667e-05, "loss": 0.895290470123291, "step": 230 }, { "epoch": 0.038327178361113885, "grad_norm": 0.03090902790427208, "learning_rate": 1.5933333333333336e-05, "loss": 0.89828462600708, "step": 240 }, { "epoch": 0.03992414412616029, "grad_norm": 0.023231355473399162, "learning_rate": 1.66e-05, "loss": 0.867582893371582, "step": 250 }, { "epoch": 0.041521109891206706, "grad_norm": 0.022748615592718124, "learning_rate": 1.726666666666667e-05, "loss": 0.8476214408874512, "step": 260 }, { "epoch": 0.04311807565625312, "grad_norm": 0.025588329881429672, "learning_rate": 1.7933333333333333e-05, "loss": 0.8639037132263183, "step": 270 }, { "epoch": 0.044715041421299534, "grad_norm": 0.029987547546625137, "learning_rate": 1.86e-05, "loss": 0.8353777885437011, "step": 280 }, { "epoch": 0.04631200718634594, "grad_norm": 0.020994428545236588, "learning_rate": 1.926666666666667e-05, "loss": 0.8189143180847168, "step": 290 }, { "epoch": 0.047908972951392355, "grad_norm": 0.023545585572719574, "learning_rate": 1.9933333333333334e-05, "loss": 0.7846664905548095, "step": 300 }, { "epoch": 0.04950593871643877, "grad_norm": 0.025134805589914322, "learning_rate": 1.999988754726792e-05, "loss": 0.7751604557037354, "step": 310 }, { "epoch": 0.051102904481485176, "grad_norm": 0.021275486797094345, "learning_rate": 1.9999498825021314e-05, "loss": 0.7967638969421387, "step": 320 }, { "epoch": 0.05269987024653159, "grad_norm": 0.019746815785765648, "learning_rate": 1.9998832455745586e-05, "loss": 0.8192337036132813, "step": 330 }, { "epoch": 0.054296836011578004, "grad_norm": 0.02567191608250141, "learning_rate": 1.9997888457943197e-05, "loss": 0.7604565620422363, "step": 340 }, { "epoch": 0.05589380177662441, "grad_norm": 0.024295911192893982, "learning_rate": 1.9996666857825287e-05, "loss": 0.7550592422485352, "step": 350 }, { "epoch": 0.057490767541670824, "grad_norm": 0.02220323495566845, "learning_rate": 1.9995167689310917e-05, "loss": 0.7514551639556885, "step": 360 }, { "epoch": 0.05908773330671724, "grad_norm": 0.024802017956972122, "learning_rate": 1.9993390994026144e-05, "loss": 0.7904987812042237, "step": 370 }, { "epoch": 0.06068469907176365, "grad_norm": 0.027685435488820076, "learning_rate": 1.9991336821302856e-05, "loss": 0.7770297050476074, "step": 380 }, { "epoch": 0.06228166483681006, "grad_norm": 0.019939841702580452, "learning_rate": 1.9989005228177406e-05, "loss": 0.7634893417358398, "step": 390 }, { "epoch": 0.06387863060185647, "grad_norm": 0.028217511251568794, "learning_rate": 1.9986396279389028e-05, "loss": 0.8059945106506348, "step": 400 }, { "epoch": 0.06547559636690288, "grad_norm": 0.019439322873950005, "learning_rate": 1.9983510047378046e-05, "loss": 0.7787755489349365, "step": 410 }, { "epoch": 0.0670725621319493, "grad_norm": 0.02309611812233925, "learning_rate": 1.9980346612283842e-05, "loss": 0.7745323657989502, "step": 420 }, { "epoch": 0.06866952789699571, "grad_norm": 0.020378535613417625, "learning_rate": 1.997690606194266e-05, "loss": 0.7243520736694335, "step": 430 }, { "epoch": 0.07026649366204211, "grad_norm": 0.021776653826236725, "learning_rate": 1.9973188491885146e-05, "loss": 0.7424108028411865, "step": 440 }, { "epoch": 0.07186345942708854, "grad_norm": 0.03435182571411133, "learning_rate": 1.9969194005333697e-05, "loss": 0.7804229736328125, "step": 450 }, { "epoch": 0.07346042519213494, "grad_norm": 0.026880595833063126, "learning_rate": 1.9964922713199613e-05, "loss": 0.7963083267211915, "step": 460 }, { "epoch": 0.07505739095718135, "grad_norm": 0.02643360197544098, "learning_rate": 1.9960374734079984e-05, "loss": 0.7517959117889405, "step": 470 }, { "epoch": 0.07665435672222777, "grad_norm": 0.036808405071496964, "learning_rate": 1.9955550194254436e-05, "loss": 0.7511186122894287, "step": 480 }, { "epoch": 0.07825132248727418, "grad_norm": 0.03115280158817768, "learning_rate": 1.995044922768159e-05, "loss": 0.765751314163208, "step": 490 }, { "epoch": 0.07984828825232058, "grad_norm": 0.025536708533763885, "learning_rate": 1.994507197599537e-05, "loss": 0.7346895217895508, "step": 500 }, { "epoch": 0.07984828825232058, "eval_loss": 0.7535409331321716, "eval_runtime": 1705.9703, "eval_samples_per_second": 3.263, "eval_steps_per_second": 1.631, "step": 500 }, { "epoch": 0.081445254017367, "grad_norm": 0.03875249624252319, "learning_rate": 1.993941858850106e-05, "loss": 0.7961405277252197, "step": 510 }, { "epoch": 0.08304221978241341, "grad_norm": 0.027557438239455223, "learning_rate": 1.993348922217114e-05, "loss": 0.7667338371276855, "step": 520 }, { "epoch": 0.08463918554745983, "grad_norm": 0.026045992970466614, "learning_rate": 1.9927284041640967e-05, "loss": 0.7275139808654785, "step": 530 }, { "epoch": 0.08623615131250624, "grad_norm": 0.026213862001895905, "learning_rate": 1.992080321920416e-05, "loss": 0.7866159439086914, "step": 540 }, { "epoch": 0.08783311707755265, "grad_norm": 0.028538137674331665, "learning_rate": 1.9914046934807853e-05, "loss": 0.7298189640045166, "step": 550 }, { "epoch": 0.08943008284259907, "grad_norm": 0.028126219287514687, "learning_rate": 1.9907015376047675e-05, "loss": 0.749655294418335, "step": 560 }, { "epoch": 0.09102704860764547, "grad_norm": 0.027085473760962486, "learning_rate": 1.9899708738162553e-05, "loss": 0.8086295127868652, "step": 570 }, { "epoch": 0.09262401437269188, "grad_norm": 0.034640420228242874, "learning_rate": 1.989212722402928e-05, "loss": 0.7653923511505127, "step": 580 }, { "epoch": 0.0942209801377383, "grad_norm": 0.03858828917145729, "learning_rate": 1.98842710441569e-05, "loss": 0.8174265861511231, "step": 590 }, { "epoch": 0.09581794590278471, "grad_norm": 0.04248794540762901, "learning_rate": 1.987614041668084e-05, "loss": 0.7357111930847168, "step": 600 }, { "epoch": 0.09741491166783112, "grad_norm": 0.03530848026275635, "learning_rate": 1.9867735567356876e-05, "loss": 0.7747371196746826, "step": 610 }, { "epoch": 0.09901187743287754, "grad_norm": 0.027734428644180298, "learning_rate": 1.9859056729554845e-05, "loss": 0.73280348777771, "step": 620 }, { "epoch": 0.10060884319792394, "grad_norm": 0.02657356671988964, "learning_rate": 1.9850104144252177e-05, "loss": 0.749216365814209, "step": 630 }, { "epoch": 0.10220580896297035, "grad_norm": 0.021906374022364616, "learning_rate": 1.98408780600272e-05, "loss": 0.7534349441528321, "step": 640 }, { "epoch": 0.10380277472801677, "grad_norm": 0.025684406980872154, "learning_rate": 1.9831378733052244e-05, "loss": 0.7199561595916748, "step": 650 }, { "epoch": 0.10539974049306318, "grad_norm": 0.025311095640063286, "learning_rate": 1.982160642708652e-05, "loss": 0.7382417678833008, "step": 660 }, { "epoch": 0.10699670625810959, "grad_norm": 0.035563357174396515, "learning_rate": 1.9811561413468794e-05, "loss": 0.7394683837890625, "step": 670 }, { "epoch": 0.10859367202315601, "grad_norm": 0.023977380245923996, "learning_rate": 1.9801243971109868e-05, "loss": 0.7267738819122315, "step": 680 }, { "epoch": 0.11019063778820241, "grad_norm": 0.02275015600025654, "learning_rate": 1.9790654386484818e-05, "loss": 0.7240358829498291, "step": 690 }, { "epoch": 0.11178760355324882, "grad_norm": 0.036198344081640244, "learning_rate": 1.9779792953625052e-05, "loss": 0.7299670696258544, "step": 700 }, { "epoch": 0.11338456931829524, "grad_norm": 0.027184955775737762, "learning_rate": 1.976865997411014e-05, "loss": 0.695775318145752, "step": 710 }, { "epoch": 0.11498153508334165, "grad_norm": 0.023749997839331627, "learning_rate": 1.9757255757059446e-05, "loss": 0.7071991920471191, "step": 720 }, { "epoch": 0.11657850084838806, "grad_norm": 0.027117466554045677, "learning_rate": 1.9745580619123535e-05, "loss": 0.7466438293457032, "step": 730 }, { "epoch": 0.11817546661343448, "grad_norm": 0.027798349037766457, "learning_rate": 1.9733634884475395e-05, "loss": 0.7503840923309326, "step": 740 }, { "epoch": 0.11977243237848088, "grad_norm": 0.026327304542064667, "learning_rate": 1.9721418884801414e-05, "loss": 0.724392032623291, "step": 750 }, { "epoch": 0.1213693981435273, "grad_norm": 0.029995381832122803, "learning_rate": 1.97089329592922e-05, "loss": 0.7100958824157715, "step": 760 }, { "epoch": 0.12296636390857371, "grad_norm": 0.03063913807272911, "learning_rate": 1.969617745463314e-05, "loss": 0.7130911827087403, "step": 770 }, { "epoch": 0.12456332967362012, "grad_norm": 0.021976860240101814, "learning_rate": 1.968315272499478e-05, "loss": 0.7231676578521729, "step": 780 }, { "epoch": 0.12616029543866653, "grad_norm": 0.02538118325173855, "learning_rate": 1.9669859132022994e-05, "loss": 0.7253612995147705, "step": 790 }, { "epoch": 0.12775726120371295, "grad_norm": 0.02788228541612625, "learning_rate": 1.9656297044828943e-05, "loss": 0.7297886848449707, "step": 800 }, { "epoch": 0.12935422696875937, "grad_norm": 0.028525004163384438, "learning_rate": 1.9642466839978814e-05, "loss": 0.7109212875366211, "step": 810 }, { "epoch": 0.13095119273380576, "grad_norm": 0.028368208557367325, "learning_rate": 1.962836890148339e-05, "loss": 0.7552286624908447, "step": 820 }, { "epoch": 0.13254815849885218, "grad_norm": 0.03732667118310928, "learning_rate": 1.9614003620787358e-05, "loss": 0.74849853515625, "step": 830 }, { "epoch": 0.1341451242638986, "grad_norm": 0.026267215609550476, "learning_rate": 1.9599371396758457e-05, "loss": 0.7323933124542237, "step": 840 }, { "epoch": 0.135742090028945, "grad_norm": 0.031789544969797134, "learning_rate": 1.958447263567641e-05, "loss": 0.7394798755645752, "step": 850 }, { "epoch": 0.13733905579399142, "grad_norm": 0.03209487721323967, "learning_rate": 1.956930775122162e-05, "loss": 0.7479897499084472, "step": 860 }, { "epoch": 0.13893602155903784, "grad_norm": 0.026272661983966827, "learning_rate": 1.9553877164463698e-05, "loss": 0.7194801807403565, "step": 870 }, { "epoch": 0.14053298732408423, "grad_norm": 0.02558542974293232, "learning_rate": 1.953818130384978e-05, "loss": 0.6967973709106445, "step": 880 }, { "epoch": 0.14212995308913065, "grad_norm": 0.03761237859725952, "learning_rate": 1.9522220605192615e-05, "loss": 0.6678271770477295, "step": 890 }, { "epoch": 0.14372691885417707, "grad_norm": 0.029605882242321968, "learning_rate": 1.9505995511658464e-05, "loss": 0.6922338008880615, "step": 900 }, { "epoch": 0.14532388461922346, "grad_norm": 0.03535737469792366, "learning_rate": 1.948950647375481e-05, "loss": 0.7159334182739258, "step": 910 }, { "epoch": 0.14692085038426989, "grad_norm": 0.030855044722557068, "learning_rate": 1.9472753949317843e-05, "loss": 0.7333884716033936, "step": 920 }, { "epoch": 0.1485178161493163, "grad_norm": 0.05045896768569946, "learning_rate": 1.9455738403499728e-05, "loss": 0.6935329914093018, "step": 930 }, { "epoch": 0.1501147819143627, "grad_norm": 0.033187806606292725, "learning_rate": 1.9438460308755724e-05, "loss": 0.7046589851379395, "step": 940 }, { "epoch": 0.15171174767940912, "grad_norm": 0.03167016804218292, "learning_rate": 1.9420920144831044e-05, "loss": 0.710863447189331, "step": 950 }, { "epoch": 0.15330871344445554, "grad_norm": 0.025539802387356758, "learning_rate": 1.9403118398747533e-05, "loss": 0.6936647891998291, "step": 960 }, { "epoch": 0.15490567920950193, "grad_norm": 0.04223870858550072, "learning_rate": 1.9385055564790157e-05, "loss": 0.6985628128051757, "step": 970 }, { "epoch": 0.15650264497454835, "grad_norm": 0.030605314299464226, "learning_rate": 1.9366732144493266e-05, "loss": 0.7294198989868164, "step": 980 }, { "epoch": 0.15809961073959478, "grad_norm": 0.03095085918903351, "learning_rate": 1.9348148646626687e-05, "loss": 0.7009531021118164, "step": 990 }, { "epoch": 0.15969657650464117, "grad_norm": 0.02993757091462612, "learning_rate": 1.9329305587181574e-05, "loss": 0.7266313552856445, "step": 1000 }, { "epoch": 0.15969657650464117, "eval_loss": 0.7159722447395325, "eval_runtime": 1703.1621, "eval_samples_per_second": 3.268, "eval_steps_per_second": 1.634, "step": 1000 }, { "epoch": 0.1612935422696876, "grad_norm": 0.035543542355298996, "learning_rate": 1.9310203489356092e-05, "loss": 0.74082350730896, "step": 1010 }, { "epoch": 0.162890508034734, "grad_norm": 0.03636915981769562, "learning_rate": 1.9290842883540897e-05, "loss": 0.717669153213501, "step": 1020 }, { "epoch": 0.1644874737997804, "grad_norm": 0.04510757327079773, "learning_rate": 1.92712243073044e-05, "loss": 0.7365827560424805, "step": 1030 }, { "epoch": 0.16608443956482682, "grad_norm": 0.03563699871301651, "learning_rate": 1.925134830537784e-05, "loss": 0.6981801986694336, "step": 1040 }, { "epoch": 0.16768140532987325, "grad_norm": 0.04541337490081787, "learning_rate": 1.9231215429640167e-05, "loss": 0.754232931137085, "step": 1050 }, { "epoch": 0.16927837109491967, "grad_norm": 0.033323634415864944, "learning_rate": 1.921082623910271e-05, "loss": 0.691849946975708, "step": 1060 }, { "epoch": 0.17087533685996606, "grad_norm": 0.03559419885277748, "learning_rate": 1.919018129989366e-05, "loss": 0.6994197845458985, "step": 1070 }, { "epoch": 0.17247230262501248, "grad_norm": 0.03182852268218994, "learning_rate": 1.916928118524235e-05, "loss": 0.6645867824554443, "step": 1080 }, { "epoch": 0.1740692683900589, "grad_norm": 0.02616371586918831, "learning_rate": 1.9148126475463336e-05, "loss": 0.7270137786865234, "step": 1090 }, { "epoch": 0.1756662341551053, "grad_norm": 0.03250862658023834, "learning_rate": 1.9126717757940288e-05, "loss": 0.7272531509399414, "step": 1100 }, { "epoch": 0.17726319992015172, "grad_norm": 0.036741774529218674, "learning_rate": 1.9105055627109683e-05, "loss": 0.7251851558685303, "step": 1110 }, { "epoch": 0.17886016568519814, "grad_norm": 0.03240974619984627, "learning_rate": 1.908314068444429e-05, "loss": 0.7145021915435791, "step": 1120 }, { "epoch": 0.18045713145024453, "grad_norm": 0.0324835442006588, "learning_rate": 1.9060973538436478e-05, "loss": 0.7045553684234619, "step": 1130 }, { "epoch": 0.18205409721529095, "grad_norm": 0.029804598540067673, "learning_rate": 1.9038554804581318e-05, "loss": 0.7342820644378663, "step": 1140 }, { "epoch": 0.18365106298033737, "grad_norm": 0.03631270304322243, "learning_rate": 1.9015885105359492e-05, "loss": 0.7435333728790283, "step": 1150 }, { "epoch": 0.18524802874538376, "grad_norm": 0.031312599778175354, "learning_rate": 1.8992965070220007e-05, "loss": 0.7185348033905029, "step": 1160 }, { "epoch": 0.18684499451043018, "grad_norm": 0.03553950414061546, "learning_rate": 1.896979533556273e-05, "loss": 0.7191666603088379, "step": 1170 }, { "epoch": 0.1884419602754766, "grad_norm": 0.024733861908316612, "learning_rate": 1.8946376544720698e-05, "loss": 0.7352997779846191, "step": 1180 }, { "epoch": 0.190038926040523, "grad_norm": 0.030409259721636772, "learning_rate": 1.8922709347942275e-05, "loss": 0.7265621185302734, "step": 1190 }, { "epoch": 0.19163589180556942, "grad_norm": 0.030736226588487625, "learning_rate": 1.8898794402373077e-05, "loss": 0.7150910377502442, "step": 1200 }, { "epoch": 0.19323285757061584, "grad_norm": 0.038398947566747665, "learning_rate": 1.887463237203775e-05, "loss": 0.7055376052856446, "step": 1210 }, { "epoch": 0.19482982333566223, "grad_norm": 0.03485625982284546, "learning_rate": 1.88502239278215e-05, "loss": 0.6555115699768066, "step": 1220 }, { "epoch": 0.19642678910070865, "grad_norm": 0.03220400586724281, "learning_rate": 1.8825569747451505e-05, "loss": 0.6946470737457275, "step": 1230 }, { "epoch": 0.19802375486575508, "grad_norm": 0.03333678096532822, "learning_rate": 1.880067051547806e-05, "loss": 0.7297664642333984, "step": 1240 }, { "epoch": 0.19962072063080147, "grad_norm": 0.0367787703871727, "learning_rate": 1.8775526923255597e-05, "loss": 0.7051557064056396, "step": 1250 }, { "epoch": 0.2012176863958479, "grad_norm": 0.03103001043200493, "learning_rate": 1.8750139668923472e-05, "loss": 0.7219597339630127, "step": 1260 }, { "epoch": 0.2028146521608943, "grad_norm": 0.04533790051937103, "learning_rate": 1.872450945738659e-05, "loss": 0.6916751384735107, "step": 1270 }, { "epoch": 0.2044116179259407, "grad_norm": 0.03510045260190964, "learning_rate": 1.8698637000295816e-05, "loss": 0.7119457721710205, "step": 1280 }, { "epoch": 0.20600858369098712, "grad_norm": 0.029134295880794525, "learning_rate": 1.867252301602825e-05, "loss": 0.6870355129241943, "step": 1290 }, { "epoch": 0.20760554945603354, "grad_norm": 0.03194071725010872, "learning_rate": 1.8646168229667238e-05, "loss": 0.734464693069458, "step": 1300 }, { "epoch": 0.20920251522107994, "grad_norm": 0.026471436023712158, "learning_rate": 1.861957337298227e-05, "loss": 0.7240866184234619, "step": 1310 }, { "epoch": 0.21079948098612636, "grad_norm": 0.0347883440554142, "learning_rate": 1.8592739184408657e-05, "loss": 0.6553101062774658, "step": 1320 }, { "epoch": 0.21239644675117278, "grad_norm": 0.026658106595277786, "learning_rate": 1.8565666409027004e-05, "loss": 0.7519384384155273, "step": 1330 }, { "epoch": 0.21399341251621917, "grad_norm": 0.030675504356622696, "learning_rate": 1.8538355798542556e-05, "loss": 0.6963082790374756, "step": 1340 }, { "epoch": 0.2155903782812656, "grad_norm": 0.04593832045793533, "learning_rate": 1.85108081112643e-05, "loss": 0.7145741939544678, "step": 1350 }, { "epoch": 0.21718734404631201, "grad_norm": 0.04163511469960213, "learning_rate": 1.8483024112083928e-05, "loss": 0.7342512130737304, "step": 1360 }, { "epoch": 0.2187843098113584, "grad_norm": 0.025683345273137093, "learning_rate": 1.8455004572454583e-05, "loss": 0.7134137630462647, "step": 1370 }, { "epoch": 0.22038127557640483, "grad_norm": 0.026864832267165184, "learning_rate": 1.8426750270369452e-05, "loss": 0.6854844570159913, "step": 1380 }, { "epoch": 0.22197824134145125, "grad_norm": 0.03378361091017723, "learning_rate": 1.839826199034015e-05, "loss": 0.6904460430145264, "step": 1390 }, { "epoch": 0.22357520710649764, "grad_norm": 0.03360961750149727, "learning_rate": 1.8369540523374963e-05, "loss": 0.7089653491973877, "step": 1400 }, { "epoch": 0.22517217287154406, "grad_norm": 0.033883776515722275, "learning_rate": 1.8340586666956846e-05, "loss": 0.7053616523742676, "step": 1410 }, { "epoch": 0.22676913863659048, "grad_norm": 0.03643488511443138, "learning_rate": 1.8311401225021318e-05, "loss": 0.7411230087280274, "step": 1420 }, { "epoch": 0.22836610440163688, "grad_norm": 0.030240802094340324, "learning_rate": 1.8281985007934115e-05, "loss": 0.7020374298095703, "step": 1430 }, { "epoch": 0.2299630701666833, "grad_norm": 0.03315526619553566, "learning_rate": 1.8252338832468702e-05, "loss": 0.7227590084075928, "step": 1440 }, { "epoch": 0.23156003593172972, "grad_norm": 0.04955840855836868, "learning_rate": 1.8222463521783584e-05, "loss": 0.7004672527313233, "step": 1450 }, { "epoch": 0.2331570016967761, "grad_norm": 0.03338323533535004, "learning_rate": 1.819235990539946e-05, "loss": 0.7032230377197266, "step": 1460 }, { "epoch": 0.23475396746182253, "grad_norm": 0.03659350797533989, "learning_rate": 1.8162028819176192e-05, "loss": 0.7022134780883789, "step": 1470 }, { "epoch": 0.23635093322686895, "grad_norm": 0.029659852385520935, "learning_rate": 1.813147110528958e-05, "loss": 0.7326688289642334, "step": 1480 }, { "epoch": 0.23794789899191537, "grad_norm": 0.03314507007598877, "learning_rate": 1.8100687612208e-05, "loss": 0.7101527690887451, "step": 1490 }, { "epoch": 0.23954486475696177, "grad_norm": 0.048770975321531296, "learning_rate": 1.806967919466883e-05, "loss": 0.7013855457305909, "step": 1500 }, { "epoch": 0.23954486475696177, "eval_loss": 0.7003746032714844, "eval_runtime": 1694.4125, "eval_samples_per_second": 3.285, "eval_steps_per_second": 1.642, "step": 1500 }, { "epoch": 0.2411418305220082, "grad_norm": 0.031342763453722, "learning_rate": 1.803844671365471e-05, "loss": 0.6915247917175293, "step": 1510 }, { "epoch": 0.2427387962870546, "grad_norm": 0.029329324141144753, "learning_rate": 1.800699103636967e-05, "loss": 0.721204423904419, "step": 1520 }, { "epoch": 0.244335762052101, "grad_norm": 0.03263983875513077, "learning_rate": 1.7975313036215015e-05, "loss": 0.7468688011169433, "step": 1530 }, { "epoch": 0.24593272781714742, "grad_norm": 0.035935308784246445, "learning_rate": 1.794341359276509e-05, "loss": 0.6986902713775635, "step": 1540 }, { "epoch": 0.24752969358219384, "grad_norm": 0.03797990456223488, "learning_rate": 1.7911293591742855e-05, "loss": 0.7199019908905029, "step": 1550 }, { "epoch": 0.24912665934724024, "grad_norm": 0.024390801787376404, "learning_rate": 1.787895392499529e-05, "loss": 0.6761057853698731, "step": 1560 }, { "epoch": 0.25072362511228663, "grad_norm": 0.0313715860247612, "learning_rate": 1.7846395490468643e-05, "loss": 0.6731560707092286, "step": 1570 }, { "epoch": 0.25232059087733305, "grad_norm": 0.0433608703315258, "learning_rate": 1.781361919218348e-05, "loss": 0.685046911239624, "step": 1580 }, { "epoch": 0.25391755664237947, "grad_norm": 0.050231028348207474, "learning_rate": 1.7780625940209596e-05, "loss": 0.7280925273895263, "step": 1590 }, { "epoch": 0.2555145224074259, "grad_norm": 0.0476137213408947, "learning_rate": 1.774741665064074e-05, "loss": 0.728914451599121, "step": 1600 }, { "epoch": 0.2571114881724723, "grad_norm": 0.03540361300110817, "learning_rate": 1.771399224556919e-05, "loss": 0.6937174320220947, "step": 1610 }, { "epoch": 0.25870845393751873, "grad_norm": 0.038516197353601456, "learning_rate": 1.7680353653060135e-05, "loss": 0.6788946151733398, "step": 1620 }, { "epoch": 0.2603054197025651, "grad_norm": 0.0472414456307888, "learning_rate": 1.7646501807125905e-05, "loss": 0.7246061325073242, "step": 1630 }, { "epoch": 0.2619023854676115, "grad_norm": 0.026418814435601234, "learning_rate": 1.7612437647700056e-05, "loss": 0.6792353630065918, "step": 1640 }, { "epoch": 0.26349935123265794, "grad_norm": 0.025900904089212418, "learning_rate": 1.757816212061126e-05, "loss": 0.6991750240325928, "step": 1650 }, { "epoch": 0.26509631699770436, "grad_norm": 0.036637961864471436, "learning_rate": 1.7543676177557042e-05, "loss": 0.7213243007659912, "step": 1660 }, { "epoch": 0.2666932827627508, "grad_norm": 0.031415775418281555, "learning_rate": 1.750898077607735e-05, "loss": 0.6988609790802002, "step": 1670 }, { "epoch": 0.2682902485277972, "grad_norm": 0.036175280809402466, "learning_rate": 1.7474076879527977e-05, "loss": 0.6812397003173828, "step": 1680 }, { "epoch": 0.2698872142928436, "grad_norm": 0.03462570160627365, "learning_rate": 1.743896545705382e-05, "loss": 0.6791361331939697, "step": 1690 }, { "epoch": 0.27148418005789, "grad_norm": 0.03334948793053627, "learning_rate": 1.740364748356195e-05, "loss": 0.6668567180633544, "step": 1700 }, { "epoch": 0.2730811458229364, "grad_norm": 0.03270823508501053, "learning_rate": 1.7368123939694554e-05, "loss": 0.6887404441833496, "step": 1710 }, { "epoch": 0.27467811158798283, "grad_norm": 0.03120650351047516, "learning_rate": 1.7332395811801706e-05, "loss": 0.6659372806549072, "step": 1720 }, { "epoch": 0.27627507735302925, "grad_norm": 0.042766936123371124, "learning_rate": 1.7296464091913986e-05, "loss": 0.7515771389007568, "step": 1730 }, { "epoch": 0.2778720431180757, "grad_norm": 0.027079230174422264, "learning_rate": 1.7260329777714923e-05, "loss": 0.7227123260498047, "step": 1740 }, { "epoch": 0.2794690088831221, "grad_norm": 0.03032068908214569, "learning_rate": 1.722399387251329e-05, "loss": 0.695145559310913, "step": 1750 }, { "epoch": 0.28106597464816846, "grad_norm": 0.03420598804950714, "learning_rate": 1.7187457385215274e-05, "loss": 0.6715566158294678, "step": 1760 }, { "epoch": 0.2826629404132149, "grad_norm": 0.037743836641311646, "learning_rate": 1.7150721330296428e-05, "loss": 0.6833428382873535, "step": 1770 }, { "epoch": 0.2842599061782613, "grad_norm": 0.035041823983192444, "learning_rate": 1.7113786727773528e-05, "loss": 0.7029437065124512, "step": 1780 }, { "epoch": 0.2858568719433077, "grad_norm": 0.03538225591182709, "learning_rate": 1.7076654603176234e-05, "loss": 0.7168496131896973, "step": 1790 }, { "epoch": 0.28745383770835414, "grad_norm": 0.03331568092107773, "learning_rate": 1.7039325987518623e-05, "loss": 0.7009388923645019, "step": 1800 }, { "epoch": 0.28905080347340056, "grad_norm": 0.033853549510240555, "learning_rate": 1.700180191727057e-05, "loss": 0.7036603450775146, "step": 1810 }, { "epoch": 0.29064776923844693, "grad_norm": 0.03764677420258522, "learning_rate": 1.696408343432895e-05, "loss": 0.6595663547515869, "step": 1820 }, { "epoch": 0.29224473500349335, "grad_norm": 0.028797749429941177, "learning_rate": 1.6926171585988728e-05, "loss": 0.6624193668365479, "step": 1830 }, { "epoch": 0.29384170076853977, "grad_norm": 0.03950833901762962, "learning_rate": 1.6888067424913863e-05, "loss": 0.7282033920288086, "step": 1840 }, { "epoch": 0.2954386665335862, "grad_norm": 0.032418642193078995, "learning_rate": 1.6849772009108094e-05, "loss": 0.6922025680541992, "step": 1850 }, { "epoch": 0.2970356322986326, "grad_norm": 0.04753576219081879, "learning_rate": 1.6811286401885554e-05, "loss": 0.6543939590454102, "step": 1860 }, { "epoch": 0.29863259806367903, "grad_norm": 0.03233656659722328, "learning_rate": 1.677261167184125e-05, "loss": 0.7344549655914306, "step": 1870 }, { "epoch": 0.3002295638287254, "grad_norm": 0.030969245359301567, "learning_rate": 1.673374889282139e-05, "loss": 0.6734820365905761, "step": 1880 }, { "epoch": 0.3018265295937718, "grad_norm": 0.03292040154337883, "learning_rate": 1.6694699143893566e-05, "loss": 0.6728582382202148, "step": 1890 }, { "epoch": 0.30342349535881824, "grad_norm": 0.05033507198095322, "learning_rate": 1.6655463509316797e-05, "loss": 0.7132375240325928, "step": 1900 }, { "epoch": 0.30502046112386466, "grad_norm": 0.03168455511331558, "learning_rate": 1.6616043078511425e-05, "loss": 0.7131926536560058, "step": 1910 }, { "epoch": 0.3066174268889111, "grad_norm": 0.03114555962383747, "learning_rate": 1.657643894602885e-05, "loss": 0.7184583187103272, "step": 1920 }, { "epoch": 0.3082143926539575, "grad_norm": 0.031201248988509178, "learning_rate": 1.6536652211521155e-05, "loss": 0.6606316566467285, "step": 1930 }, { "epoch": 0.30981135841900387, "grad_norm": 0.035202883183956146, "learning_rate": 1.6496683979710576e-05, "loss": 0.7029526233673096, "step": 1940 }, { "epoch": 0.3114083241840503, "grad_norm": 0.03515475615859032, "learning_rate": 1.6456535360358807e-05, "loss": 0.6629764080047608, "step": 1950 }, { "epoch": 0.3130052899490967, "grad_norm": 0.03606207296252251, "learning_rate": 1.6416207468236208e-05, "loss": 0.6876577377319336, "step": 1960 }, { "epoch": 0.31460225571414313, "grad_norm": 0.03580459579825401, "learning_rate": 1.6375701423090846e-05, "loss": 0.6879127979278564, "step": 1970 }, { "epoch": 0.31619922147918955, "grad_norm": 0.048258859664201736, "learning_rate": 1.6335018349617394e-05, "loss": 0.7255002498626709, "step": 1980 }, { "epoch": 0.317796187244236, "grad_norm": 0.046948857605457306, "learning_rate": 1.629415937742591e-05, "loss": 0.6919849395751954, "step": 1990 }, { "epoch": 0.31939315300928234, "grad_norm": 0.03351585939526558, "learning_rate": 1.625312564101049e-05, "loss": 0.6886905193328857, "step": 2000 }, { "epoch": 0.31939315300928234, "eval_loss": 0.6917140483856201, "eval_runtime": 1706.5142, "eval_samples_per_second": 3.262, "eval_steps_per_second": 1.631, "step": 2000 }, { "epoch": 0.32099011877432876, "grad_norm": 0.028385179117321968, "learning_rate": 1.621191827971772e-05, "loss": 0.7103724956512452, "step": 2010 }, { "epoch": 0.3225870845393752, "grad_norm": 0.02977878600358963, "learning_rate": 1.617053843771509e-05, "loss": 0.7012845516204834, "step": 2020 }, { "epoch": 0.3241840503044216, "grad_norm": 0.03950609639286995, "learning_rate": 1.6128987263959212e-05, "loss": 0.670585298538208, "step": 2030 }, { "epoch": 0.325781016069468, "grad_norm": 0.03531676530838013, "learning_rate": 1.6087265912163898e-05, "loss": 0.6601489543914795, "step": 2040 }, { "epoch": 0.32737798183451444, "grad_norm": 0.030362443998456, "learning_rate": 1.6045375540768136e-05, "loss": 0.7151473522186279, "step": 2050 }, { "epoch": 0.3289749475995608, "grad_norm": 0.03205285966396332, "learning_rate": 1.600331731290395e-05, "loss": 0.6645094394683838, "step": 2060 }, { "epoch": 0.33057191336460723, "grad_norm": 0.028163282200694084, "learning_rate": 1.5961092396364053e-05, "loss": 0.6868838310241699, "step": 2070 }, { "epoch": 0.33216887912965365, "grad_norm": 0.03369533643126488, "learning_rate": 1.5918701963569475e-05, "loss": 0.7029123783111573, "step": 2080 }, { "epoch": 0.33376584489470007, "grad_norm": 0.0379788838326931, "learning_rate": 1.5876147191536983e-05, "loss": 0.7173356056213379, "step": 2090 }, { "epoch": 0.3353628106597465, "grad_norm": 0.0344555526971817, "learning_rate": 1.583342926184639e-05, "loss": 0.7149466037750244, "step": 2100 }, { "epoch": 0.3369597764247929, "grad_norm": 0.03237446770071983, "learning_rate": 1.5790549360607772e-05, "loss": 0.6467396259307862, "step": 2110 }, { "epoch": 0.33855674218983933, "grad_norm": 0.032042454928159714, "learning_rate": 1.5747508678428515e-05, "loss": 0.6929362297058106, "step": 2120 }, { "epoch": 0.3401537079548857, "grad_norm": 0.042478688061237335, "learning_rate": 1.570430841038027e-05, "loss": 0.6702353000640869, "step": 2130 }, { "epoch": 0.3417506737199321, "grad_norm": 0.04559747502207756, "learning_rate": 1.5660949755965757e-05, "loss": 0.6600518226623535, "step": 2140 }, { "epoch": 0.34334763948497854, "grad_norm": 0.03681457042694092, "learning_rate": 1.5617433919085474e-05, "loss": 0.6836549758911132, "step": 2150 }, { "epoch": 0.34494460525002496, "grad_norm": 0.02465015836060047, "learning_rate": 1.5573762108004262e-05, "loss": 0.7304904460906982, "step": 2160 }, { "epoch": 0.3465415710150714, "grad_norm": 0.03969291225075722, "learning_rate": 1.5529935535317746e-05, "loss": 0.6897444725036621, "step": 2170 }, { "epoch": 0.3481385367801178, "grad_norm": 0.03233686834573746, "learning_rate": 1.548595541791869e-05, "loss": 0.684897518157959, "step": 2180 }, { "epoch": 0.34973550254516417, "grad_norm": 0.03138101100921631, "learning_rate": 1.544182297696319e-05, "loss": 0.7246172428131104, "step": 2190 }, { "epoch": 0.3513324683102106, "grad_norm": 0.03732126206159592, "learning_rate": 1.5397539437836765e-05, "loss": 0.6686253070831298, "step": 2200 }, { "epoch": 0.352929434075257, "grad_norm": 0.028484191745519638, "learning_rate": 1.535310603012035e-05, "loss": 0.6723896503448487, "step": 2210 }, { "epoch": 0.35452639984030343, "grad_norm": 0.04804018512368202, "learning_rate": 1.5308523987556144e-05, "loss": 0.7210356712341308, "step": 2220 }, { "epoch": 0.35612336560534985, "grad_norm": 0.035007379949092865, "learning_rate": 1.5263794548013347e-05, "loss": 0.6699332237243653, "step": 2230 }, { "epoch": 0.35772033137039627, "grad_norm": 0.034309230744838715, "learning_rate": 1.5218918953453808e-05, "loss": 0.6913293838500977, "step": 2240 }, { "epoch": 0.35931729713544264, "grad_norm": 0.05658178776502609, "learning_rate": 1.5173898449897527e-05, "loss": 0.6920335292816162, "step": 2250 }, { "epoch": 0.36091426290048906, "grad_norm": 0.03614016994833946, "learning_rate": 1.5128734287388051e-05, "loss": 0.6855293273925781, "step": 2260 }, { "epoch": 0.3625112286655355, "grad_norm": 0.0348992682993412, "learning_rate": 1.5083427719957792e-05, "loss": 0.7068172454833984, "step": 2270 }, { "epoch": 0.3641081944305819, "grad_norm": 0.037384193390607834, "learning_rate": 1.5037980005593178e-05, "loss": 0.7026480197906494, "step": 2280 }, { "epoch": 0.3657051601956283, "grad_norm": 0.03162362426519394, "learning_rate": 1.4992392406199741e-05, "loss": 0.7369038581848144, "step": 2290 }, { "epoch": 0.36730212596067474, "grad_norm": 0.03644828870892525, "learning_rate": 1.4946666187567073e-05, "loss": 0.7133200168609619, "step": 2300 }, { "epoch": 0.3688990917257211, "grad_norm": 0.05075901374220848, "learning_rate": 1.490080261933368e-05, "loss": 0.7146055221557617, "step": 2310 }, { "epoch": 0.3704960574907675, "grad_norm": 0.029153743758797646, "learning_rate": 1.4854802974951732e-05, "loss": 0.6756022453308106, "step": 2320 }, { "epoch": 0.37209302325581395, "grad_norm": 0.030742282047867775, "learning_rate": 1.4808668531651698e-05, "loss": 0.6855404376983643, "step": 2330 }, { "epoch": 0.37368998902086037, "grad_norm": 0.03525373339653015, "learning_rate": 1.4762400570406889e-05, "loss": 0.6626357078552246, "step": 2340 }, { "epoch": 0.3752869547859068, "grad_norm": 0.036246221512556076, "learning_rate": 1.4716000375897897e-05, "loss": 0.714638090133667, "step": 2350 }, { "epoch": 0.3768839205509532, "grad_norm": 0.03430463373661041, "learning_rate": 1.4669469236476897e-05, "loss": 0.6736037731170654, "step": 2360 }, { "epoch": 0.3784808863159996, "grad_norm": 0.035327013581991196, "learning_rate": 1.4622808444131904e-05, "loss": 0.6907155036926269, "step": 2370 }, { "epoch": 0.380077852081046, "grad_norm": 0.039804793894290924, "learning_rate": 1.457601929445089e-05, "loss": 0.6726024627685547, "step": 2380 }, { "epoch": 0.3816748178460924, "grad_norm": 0.03689780458807945, "learning_rate": 1.4529103086585803e-05, "loss": 0.6932793617248535, "step": 2390 }, { "epoch": 0.38327178361113884, "grad_norm": 0.03002229705452919, "learning_rate": 1.4482061123216507e-05, "loss": 0.6448677062988282, "step": 2400 }, { "epoch": 0.38486874937618526, "grad_norm": 0.04287157952785492, "learning_rate": 1.4434894710514596e-05, "loss": 0.6573331356048584, "step": 2410 }, { "epoch": 0.3864657151412317, "grad_norm": 0.033608682453632355, "learning_rate": 1.4387605158107146e-05, "loss": 0.6761981010437011, "step": 2420 }, { "epoch": 0.38806268090627805, "grad_norm": 0.0340590700507164, "learning_rate": 1.4340193779040335e-05, "loss": 0.7119544982910156, "step": 2430 }, { "epoch": 0.38965964667132447, "grad_norm": 0.04576168581843376, "learning_rate": 1.429266188974299e-05, "loss": 0.722407054901123, "step": 2440 }, { "epoch": 0.3912566124363709, "grad_norm": 0.030462034046649933, "learning_rate": 1.4245010809990052e-05, "loss": 0.6741881847381592, "step": 2450 }, { "epoch": 0.3928535782014173, "grad_norm": 0.024982597678899765, "learning_rate": 1.419724186286589e-05, "loss": 0.6923768997192383, "step": 2460 }, { "epoch": 0.39445054396646373, "grad_norm": 0.028493596240878105, "learning_rate": 1.414935637472761e-05, "loss": 0.7135504722595215, "step": 2470 }, { "epoch": 0.39604750973151015, "grad_norm": 0.03309721499681473, "learning_rate": 1.4101355675168197e-05, "loss": 0.6870269298553466, "step": 2480 }, { "epoch": 0.3976444754965565, "grad_norm": 0.033349502831697464, "learning_rate": 1.4053241096979611e-05, "loss": 0.7310259819030762, "step": 2490 }, { "epoch": 0.39924144126160294, "grad_norm": 0.0456295944750309, "learning_rate": 1.4005013976115777e-05, "loss": 0.6850334644317627, "step": 2500 }, { "epoch": 0.39924144126160294, "eval_loss": 0.6847082376480103, "eval_runtime": 1727.6141, "eval_samples_per_second": 3.222, "eval_steps_per_second": 1.611, "step": 2500 }, { "epoch": 0.40083840702664936, "grad_norm": 0.04150133952498436, "learning_rate": 1.3956675651655489e-05, "loss": 0.6748110771179199, "step": 2510 }, { "epoch": 0.4024353727916958, "grad_norm": 0.05328802019357681, "learning_rate": 1.390822746576523e-05, "loss": 0.6966916561126709, "step": 2520 }, { "epoch": 0.4040323385567422, "grad_norm": 0.04510056599974632, "learning_rate": 1.3859670763661908e-05, "loss": 0.6688936710357666, "step": 2530 }, { "epoch": 0.4056293043217886, "grad_norm": 0.03648371994495392, "learning_rate": 1.38110068935755e-05, "loss": 0.6503877162933349, "step": 2540 }, { "epoch": 0.40722627008683504, "grad_norm": 0.03958805277943611, "learning_rate": 1.3762237206711624e-05, "loss": 0.6709461212158203, "step": 2550 }, { "epoch": 0.4088232358518814, "grad_norm": 0.03620138764381409, "learning_rate": 1.3713363057214006e-05, "loss": 0.6775450706481934, "step": 2560 }, { "epoch": 0.4104202016169278, "grad_norm": 0.036233801394701004, "learning_rate": 1.3664385802126904e-05, "loss": 0.7039198875427246, "step": 2570 }, { "epoch": 0.41201716738197425, "grad_norm": 0.034267205744981766, "learning_rate": 1.3615306801357413e-05, "loss": 0.6893480777740478, "step": 2580 }, { "epoch": 0.41361413314702067, "grad_norm": 0.04056892171502113, "learning_rate": 1.3566127417637698e-05, "loss": 0.6852858543395997, "step": 2590 }, { "epoch": 0.4152110989120671, "grad_norm": 0.04654235392808914, "learning_rate": 1.351684901648718e-05, "loss": 0.6932817459106445, "step": 2600 }, { "epoch": 0.4168080646771135, "grad_norm": 0.05177690088748932, "learning_rate": 1.3467472966174595e-05, "loss": 0.663325309753418, "step": 2610 }, { "epoch": 0.4184050304421599, "grad_norm": 0.04320165514945984, "learning_rate": 1.3418000637680026e-05, "loss": 0.6278695106506348, "step": 2620 }, { "epoch": 0.4200019962072063, "grad_norm": 0.0365922674536705, "learning_rate": 1.336843340465682e-05, "loss": 0.6674781799316406, "step": 2630 }, { "epoch": 0.4215989619722527, "grad_norm": 0.03406934812664986, "learning_rate": 1.3318772643393447e-05, "loss": 0.6559438705444336, "step": 2640 }, { "epoch": 0.42319592773729914, "grad_norm": 0.030784646049141884, "learning_rate": 1.3269019732775304e-05, "loss": 0.7225832462310791, "step": 2650 }, { "epoch": 0.42479289350234556, "grad_norm": 0.0360528789460659, "learning_rate": 1.3219176054246404e-05, "loss": 0.7170959949493408, "step": 2660 }, { "epoch": 0.426389859267392, "grad_norm": 0.0355147160589695, "learning_rate": 1.316924299177103e-05, "loss": 0.683486795425415, "step": 2670 }, { "epoch": 0.42798682503243834, "grad_norm": 0.046759720891714096, "learning_rate": 1.3119221931795321e-05, "loss": 0.7131591320037842, "step": 2680 }, { "epoch": 0.42958379079748477, "grad_norm": 0.03554374352097511, "learning_rate": 1.3069114263208746e-05, "loss": 0.6715425491333008, "step": 2690 }, { "epoch": 0.4311807565625312, "grad_norm": 0.04263368248939514, "learning_rate": 1.3018921377305574e-05, "loss": 0.6871604442596435, "step": 2700 }, { "epoch": 0.4327777223275776, "grad_norm": 0.04496333748102188, "learning_rate": 1.2968644667746207e-05, "loss": 0.6588253021240235, "step": 2710 }, { "epoch": 0.43437468809262403, "grad_norm": 0.036086685955524445, "learning_rate": 1.2918285530518519e-05, "loss": 0.6757514953613282, "step": 2720 }, { "epoch": 0.43597165385767045, "grad_norm": 0.03308691456913948, "learning_rate": 1.2867845363899075e-05, "loss": 0.6658079624176025, "step": 2730 }, { "epoch": 0.4375686196227168, "grad_norm": 0.04500787332653999, "learning_rate": 1.2817325568414299e-05, "loss": 0.6812032222747803, "step": 2740 }, { "epoch": 0.43916558538776324, "grad_norm": 0.03841916844248772, "learning_rate": 1.2766727546801612e-05, "loss": 0.6864322662353516, "step": 2750 }, { "epoch": 0.44076255115280966, "grad_norm": 0.037914253771305084, "learning_rate": 1.2716052703970459e-05, "loss": 0.7053091526031494, "step": 2760 }, { "epoch": 0.4423595169178561, "grad_norm": 0.03686266764998436, "learning_rate": 1.2665302446963312e-05, "loss": 0.7154440879821777, "step": 2770 }, { "epoch": 0.4439564826829025, "grad_norm": 0.04227980226278305, "learning_rate": 1.261447818491661e-05, "loss": 0.716032886505127, "step": 2780 }, { "epoch": 0.4455534484479489, "grad_norm": 0.04683419317007065, "learning_rate": 1.2563581329021608e-05, "loss": 0.6738180637359619, "step": 2790 }, { "epoch": 0.4471504142129953, "grad_norm": 0.05264711007475853, "learning_rate": 1.2512613292485218e-05, "loss": 0.7129797458648681, "step": 2800 }, { "epoch": 0.4487473799780417, "grad_norm": 0.04030587151646614, "learning_rate": 1.2461575490490753e-05, "loss": 0.727331018447876, "step": 2810 }, { "epoch": 0.4503443457430881, "grad_norm": 0.04141924902796745, "learning_rate": 1.2410469340158655e-05, "loss": 0.676334810256958, "step": 2820 }, { "epoch": 0.45194131150813455, "grad_norm": 0.04476945474743843, "learning_rate": 1.2359296260507117e-05, "loss": 0.6823254108428956, "step": 2830 }, { "epoch": 0.45353827727318097, "grad_norm": 0.03869614750146866, "learning_rate": 1.23080576724127e-05, "loss": 0.6882941722869873, "step": 2840 }, { "epoch": 0.4551352430382274, "grad_norm": 0.031092161312699318, "learning_rate": 1.2256754998570887e-05, "loss": 0.6744899272918701, "step": 2850 }, { "epoch": 0.45673220880327375, "grad_norm": 0.03507015109062195, "learning_rate": 1.2205389663456566e-05, "loss": 0.7103989124298096, "step": 2860 }, { "epoch": 0.4583291745683202, "grad_norm": 0.032851576805114746, "learning_rate": 1.2153963093284483e-05, "loss": 0.7116181373596191, "step": 2870 }, { "epoch": 0.4599261403333666, "grad_norm": 0.05045435577630997, "learning_rate": 1.210247671596965e-05, "loss": 0.6489015579223633, "step": 2880 }, { "epoch": 0.461523106098413, "grad_norm": 0.032028377056121826, "learning_rate": 1.2050931961087673e-05, "loss": 0.6941986560821534, "step": 2890 }, { "epoch": 0.46312007186345944, "grad_norm": 0.038908179849386215, "learning_rate": 1.1999330259835096e-05, "loss": 0.6908363342285156, "step": 2900 }, { "epoch": 0.46471703762850586, "grad_norm": 0.03464297205209732, "learning_rate": 1.1947673044989627e-05, "loss": 0.7050665378570556, "step": 2910 }, { "epoch": 0.4663140033935522, "grad_norm": 0.028774775564670563, "learning_rate": 1.1895961750870375e-05, "loss": 0.6700276851654052, "step": 2920 }, { "epoch": 0.46791096915859864, "grad_norm": 0.03465382754802704, "learning_rate": 1.1844197813298018e-05, "loss": 0.6821481227874756, "step": 2930 }, { "epoch": 0.46950793492364506, "grad_norm": 0.03162831813097, "learning_rate": 1.1792382669554938e-05, "loss": 0.6927790164947509, "step": 2940 }, { "epoch": 0.4711049006886915, "grad_norm": 0.03354581072926521, "learning_rate": 1.1740517758345312e-05, "loss": 0.7100222110748291, "step": 2950 }, { "epoch": 0.4727018664537379, "grad_norm": 0.033408552408218384, "learning_rate": 1.168860451975516e-05, "loss": 0.6876598358154297, "step": 2960 }, { "epoch": 0.4742988322187843, "grad_norm": 0.04334475100040436, "learning_rate": 1.1636644395212375e-05, "loss": 0.6866564273834228, "step": 2970 }, { "epoch": 0.47589579798383075, "grad_norm": 0.040464501827955246, "learning_rate": 1.1584638827446684e-05, "loss": 0.6588007926940918, "step": 2980 }, { "epoch": 0.4774927637488771, "grad_norm": 0.04957546666264534, "learning_rate": 1.1532589260449589e-05, "loss": 0.7070925235748291, "step": 2990 }, { "epoch": 0.47908972951392353, "grad_norm": 0.038624707609415054, "learning_rate": 1.1480497139434294e-05, "loss": 0.6919936180114746, "step": 3000 }, { "epoch": 0.47908972951392353, "eval_loss": 0.6803271770477295, "eval_runtime": 1741.9445, "eval_samples_per_second": 3.195, "eval_steps_per_second": 1.598, "step": 3000 }, { "epoch": 0.48068669527896996, "grad_norm": 0.04100416228175163, "learning_rate": 1.1428363910795547e-05, "loss": 0.6852492809295654, "step": 3010 }, { "epoch": 0.4822836610440164, "grad_norm": 0.033158283680677414, "learning_rate": 1.1376191022069501e-05, "loss": 0.667516565322876, "step": 3020 }, { "epoch": 0.4838806268090628, "grad_norm": 0.03429755941033363, "learning_rate": 1.1323979921893514e-05, "loss": 0.6956333160400391, "step": 3030 }, { "epoch": 0.4854775925741092, "grad_norm": 0.038232944905757904, "learning_rate": 1.1271732059965925e-05, "loss": 0.6716940402984619, "step": 3040 }, { "epoch": 0.4870745583391556, "grad_norm": 0.039248283952474594, "learning_rate": 1.1219448887005805e-05, "loss": 0.6984889507293701, "step": 3050 }, { "epoch": 0.488671524104202, "grad_norm": 0.03670873865485191, "learning_rate": 1.1167131854712676e-05, "loss": 0.6709868907928467, "step": 3060 }, { "epoch": 0.4902684898692484, "grad_norm": 0.038310978561639786, "learning_rate": 1.11147824157262e-05, "loss": 0.6672838211059571, "step": 3070 }, { "epoch": 0.49186545563429485, "grad_norm": 0.049189358949661255, "learning_rate": 1.1062402023585846e-05, "loss": 0.6705825805664063, "step": 3080 }, { "epoch": 0.49346242139934127, "grad_norm": 0.0355689600110054, "learning_rate": 1.1009992132690523e-05, "loss": 0.6964725017547607, "step": 3090 }, { "epoch": 0.4950593871643877, "grad_norm": 0.03348139300942421, "learning_rate": 1.0957554198258225e-05, "loss": 0.6852751255035401, "step": 3100 }, { "epoch": 0.49665635292943405, "grad_norm": 0.03575340285897255, "learning_rate": 1.0905089676285588e-05, "loss": 0.670794153213501, "step": 3110 }, { "epoch": 0.4982533186944805, "grad_norm": 0.04119575023651123, "learning_rate": 1.0852600023507482e-05, "loss": 0.6651177406311035, "step": 3120 }, { "epoch": 0.4998502844595269, "grad_norm": 0.04122074693441391, "learning_rate": 1.0800086697356576e-05, "loss": 0.6976897239685058, "step": 3130 }, { "epoch": 0.5014472502245733, "grad_norm": 0.034223176538944244, "learning_rate": 1.0747551155922836e-05, "loss": 0.6890422821044921, "step": 3140 }, { "epoch": 0.5030442159896197, "grad_norm": 0.03588424623012543, "learning_rate": 1.069499485791307e-05, "loss": 0.702085018157959, "step": 3150 }, { "epoch": 0.5046411817546661, "grad_norm": 0.03808250650763512, "learning_rate": 1.0642419262610417e-05, "loss": 0.6910345554351807, "step": 3160 }, { "epoch": 0.5062381475197125, "grad_norm": 0.03852913901209831, "learning_rate": 1.0589825829833825e-05, "loss": 0.6685952186584473, "step": 3170 }, { "epoch": 0.5078351132847589, "grad_norm": 0.04528075084090233, "learning_rate": 1.0537216019897514e-05, "loss": 0.6971333980560303, "step": 3180 }, { "epoch": 0.5094320790498054, "grad_norm": 0.048296891152858734, "learning_rate": 1.0484591293570437e-05, "loss": 0.6506364345550537, "step": 3190 }, { "epoch": 0.5110290448148518, "grad_norm": 0.03280564025044441, "learning_rate": 1.0431953112035722e-05, "loss": 0.7106664657592774, "step": 3200 }, { "epoch": 0.5126260105798982, "grad_norm": 0.03674250841140747, "learning_rate": 1.0379302936850083e-05, "loss": 0.6798049449920655, "step": 3210 }, { "epoch": 0.5142229763449446, "grad_norm": 0.02851015329360962, "learning_rate": 1.0326642229903265e-05, "loss": 0.6674058914184571, "step": 3220 }, { "epoch": 0.515819942109991, "grad_norm": 0.04876931011676788, "learning_rate": 1.0273972453377433e-05, "loss": 0.6956850051879883, "step": 3230 }, { "epoch": 0.5174169078750375, "grad_norm": 0.04633413255214691, "learning_rate": 1.022129506970657e-05, "loss": 0.6859352111816406, "step": 3240 }, { "epoch": 0.5190138736400839, "grad_norm": 0.03311692923307419, "learning_rate": 1.0168611541535898e-05, "loss": 0.7103584289550782, "step": 3250 }, { "epoch": 0.5206108394051302, "grad_norm": 0.047599345445632935, "learning_rate": 1.0115923331681231e-05, "loss": 0.6435458183288574, "step": 3260 }, { "epoch": 0.5222078051701766, "grad_norm": 0.033676277846097946, "learning_rate": 1.0063231903088384e-05, "loss": 0.6581607341766358, "step": 3270 }, { "epoch": 0.523804770935223, "grad_norm": 0.04461989924311638, "learning_rate": 1.0010538718792544e-05, "loss": 0.7152835369110108, "step": 3280 }, { "epoch": 0.5254017367002695, "grad_norm": 0.04492766410112381, "learning_rate": 9.957845241877639e-06, "loss": 0.7056149482727051, "step": 3290 }, { "epoch": 0.5269987024653159, "grad_norm": 0.02856948785483837, "learning_rate": 9.905152935435733e-06, "loss": 0.6655986785888672, "step": 3300 }, { "epoch": 0.5285956682303623, "grad_norm": 0.0438731387257576, "learning_rate": 9.852463262526383e-06, "loss": 0.7045872688293457, "step": 3310 }, { "epoch": 0.5301926339954087, "grad_norm": 0.04025017097592354, "learning_rate": 9.799777686136025e-06, "loss": 0.6844919204711915, "step": 3320 }, { "epoch": 0.5317895997604551, "grad_norm": 0.04453453794121742, "learning_rate": 9.747097669137352e-06, "loss": 0.6857856750488281, "step": 3330 }, { "epoch": 0.5333865655255016, "grad_norm": 0.034065935760736465, "learning_rate": 9.694424674248687e-06, "loss": 0.6782355785369873, "step": 3340 }, { "epoch": 0.534983531290548, "grad_norm": 0.03808142989873886, "learning_rate": 9.641760163993394e-06, "loss": 0.7040555000305175, "step": 3350 }, { "epoch": 0.5365804970555944, "grad_norm": 0.03482348844408989, "learning_rate": 9.58910560065923e-06, "loss": 0.6730469226837158, "step": 3360 }, { "epoch": 0.5381774628206408, "grad_norm": 0.04541369527578354, "learning_rate": 9.536462446257777e-06, "loss": 0.7102465152740478, "step": 3370 }, { "epoch": 0.5397744285856872, "grad_norm": 0.043802615255117416, "learning_rate": 9.483832162483841e-06, "loss": 0.6941490173339844, "step": 3380 }, { "epoch": 0.5413713943507336, "grad_norm": 0.03652375563979149, "learning_rate": 9.431216210674849e-06, "loss": 0.704205322265625, "step": 3390 }, { "epoch": 0.54296836011578, "grad_norm": 0.03886988013982773, "learning_rate": 9.378616051770287e-06, "loss": 0.6582465171813965, "step": 3400 }, { "epoch": 0.5445653258808264, "grad_norm": 0.044551681727170944, "learning_rate": 9.326033146271144e-06, "loss": 0.7078223705291748, "step": 3410 }, { "epoch": 0.5461622916458728, "grad_norm": 0.03762778267264366, "learning_rate": 9.273468954199333e-06, "loss": 0.6956658840179444, "step": 3420 }, { "epoch": 0.5477592574109192, "grad_norm": 0.03794670104980469, "learning_rate": 9.220924935057184e-06, "loss": 0.6778861522674561, "step": 3430 }, { "epoch": 0.5493562231759657, "grad_norm": 0.03766616806387901, "learning_rate": 9.168402547786897e-06, "loss": 0.6759885311126709, "step": 3440 }, { "epoch": 0.5509531889410121, "grad_norm": 0.04172630235552788, "learning_rate": 9.115903250730035e-06, "loss": 0.6739662170410157, "step": 3450 }, { "epoch": 0.5525501547060585, "grad_norm": 0.04936975613236427, "learning_rate": 9.063428501587043e-06, "loss": 0.6581857681274415, "step": 3460 }, { "epoch": 0.5541471204711049, "grad_norm": 0.029507510364055634, "learning_rate": 9.010979757376766e-06, "loss": 0.6508080005645752, "step": 3470 }, { "epoch": 0.5557440862361513, "grad_norm": 0.03853528946638107, "learning_rate": 8.958558474395987e-06, "loss": 0.6520394325256348, "step": 3480 }, { "epoch": 0.5573410520011978, "grad_norm": 0.04168470576405525, "learning_rate": 8.906166108178999e-06, "loss": 0.6652052402496338, "step": 3490 }, { "epoch": 0.5589380177662442, "grad_norm": 0.05398240312933922, "learning_rate": 8.853804113457203e-06, "loss": 0.6777332782745361, "step": 3500 }, { "epoch": 0.5589380177662442, "eval_loss": 0.677163302898407, "eval_runtime": 1733.3138, "eval_samples_per_second": 3.211, "eval_steps_per_second": 1.606, "step": 3500 }, { "epoch": 0.5605349835312905, "grad_norm": 0.03181201219558716, "learning_rate": 8.801473944118683e-06, "loss": 0.687204122543335, "step": 3510 }, { "epoch": 0.5621319492963369, "grad_norm": 0.03610098734498024, "learning_rate": 8.749177053167865e-06, "loss": 0.7028386116027832, "step": 3520 }, { "epoch": 0.5637289150613833, "grad_norm": 0.0450638048350811, "learning_rate": 8.696914892685172e-06, "loss": 0.6494070529937744, "step": 3530 }, { "epoch": 0.5653258808264298, "grad_norm": 0.053478166460990906, "learning_rate": 8.644688913786678e-06, "loss": 0.6962610721588135, "step": 3540 }, { "epoch": 0.5669228465914762, "grad_norm": 0.03924020007252693, "learning_rate": 8.592500566583856e-06, "loss": 0.6941705703735351, "step": 3550 }, { "epoch": 0.5685198123565226, "grad_norm": 0.037506818771362305, "learning_rate": 8.540351300143284e-06, "loss": 0.6824637413024902, "step": 3560 }, { "epoch": 0.570116778121569, "grad_norm": 0.03883660212159157, "learning_rate": 8.488242562446416e-06, "loss": 0.7028771877288819, "step": 3570 }, { "epoch": 0.5717137438866154, "grad_norm": 0.047728635370731354, "learning_rate": 8.436175800349393e-06, "loss": 0.7102296829223633, "step": 3580 }, { "epoch": 0.5733107096516619, "grad_norm": 0.0423489511013031, "learning_rate": 8.384152459542849e-06, "loss": 0.7113842487335205, "step": 3590 }, { "epoch": 0.5749076754167083, "grad_norm": 0.034012071788311005, "learning_rate": 8.332173984511774e-06, "loss": 0.6995905876159668, "step": 3600 }, { "epoch": 0.5765046411817547, "grad_norm": 0.02976168505847454, "learning_rate": 8.280241818495423e-06, "loss": 0.7066914558410644, "step": 3610 }, { "epoch": 0.5781016069468011, "grad_norm": 0.04333082213997841, "learning_rate": 8.228357403447225e-06, "loss": 0.6559367656707764, "step": 3620 }, { "epoch": 0.5796985727118474, "grad_norm": 0.0726407989859581, "learning_rate": 8.176522179994744e-06, "loss": 0.691002082824707, "step": 3630 }, { "epoch": 0.5812955384768939, "grad_norm": 0.035537876188755035, "learning_rate": 8.124737587399697e-06, "loss": 0.7173010349273682, "step": 3640 }, { "epoch": 0.5828925042419403, "grad_norm": 0.036615390330553055, "learning_rate": 8.073005063517973e-06, "loss": 0.712274169921875, "step": 3650 }, { "epoch": 0.5844894700069867, "grad_norm": 0.02762191742658615, "learning_rate": 8.021326044759718e-06, "loss": 0.7043851852416992, "step": 3660 }, { "epoch": 0.5860864357720331, "grad_norm": 0.033354733139276505, "learning_rate": 7.969701966049447e-06, "loss": 0.6736726284027099, "step": 3670 }, { "epoch": 0.5876834015370795, "grad_norm": 0.04251250997185707, "learning_rate": 7.918134260786214e-06, "loss": 0.679281759262085, "step": 3680 }, { "epoch": 0.589280367302126, "grad_norm": 0.03284266218543053, "learning_rate": 7.866624360803786e-06, "loss": 0.6851751804351807, "step": 3690 }, { "epoch": 0.5908773330671724, "grad_norm": 0.050126783549785614, "learning_rate": 7.81517369633092e-06, "loss": 0.6820148468017578, "step": 3700 }, { "epoch": 0.5924742988322188, "grad_norm": 0.042198847979307175, "learning_rate": 7.763783695951623e-06, "loss": 0.6824104309082031, "step": 3710 }, { "epoch": 0.5940712645972652, "grad_norm": 0.03443041443824768, "learning_rate": 7.7124557865655e-06, "loss": 0.670203685760498, "step": 3720 }, { "epoch": 0.5956682303623116, "grad_norm": 0.036490943282842636, "learning_rate": 7.661191393348136e-06, "loss": 0.7197192192077637, "step": 3730 }, { "epoch": 0.5972651961273581, "grad_norm": 0.03636472299695015, "learning_rate": 7.609991939711517e-06, "loss": 0.6888130187988282, "step": 3740 }, { "epoch": 0.5988621618924044, "grad_norm": 0.040862370282411575, "learning_rate": 7.558858847264502e-06, "loss": 0.6817411422729492, "step": 3750 }, { "epoch": 0.6004591276574508, "grad_norm": 0.036771345883607864, "learning_rate": 7.507793535773377e-06, "loss": 0.7020951271057129, "step": 3760 }, { "epoch": 0.6020560934224972, "grad_norm": 0.036549679934978485, "learning_rate": 7.456797423122401e-06, "loss": 0.6926814079284668, "step": 3770 }, { "epoch": 0.6036530591875436, "grad_norm": 0.03661128878593445, "learning_rate": 7.405871925274457e-06, "loss": 0.7070647716522217, "step": 3780 }, { "epoch": 0.6052500249525901, "grad_norm": 0.05932987853884697, "learning_rate": 7.3550184562317295e-06, "loss": 0.7125431060791015, "step": 3790 }, { "epoch": 0.6068469907176365, "grad_norm": 0.03799287602305412, "learning_rate": 7.3042384279964394e-06, "loss": 0.6607150554656982, "step": 3800 }, { "epoch": 0.6084439564826829, "grad_norm": 0.031429167836904526, "learning_rate": 7.253533250531656e-06, "loss": 0.6806385517120361, "step": 3810 }, { "epoch": 0.6100409222477293, "grad_norm": 0.05243317410349846, "learning_rate": 7.202904331722127e-06, "loss": 0.6834061145782471, "step": 3820 }, { "epoch": 0.6116378880127757, "grad_norm": 0.0379512719810009, "learning_rate": 7.152353077335189e-06, "loss": 0.665333604812622, "step": 3830 }, { "epoch": 0.6132348537778222, "grad_norm": 0.03895451873540878, "learning_rate": 7.101880890981751e-06, "loss": 0.6566737651824951, "step": 3840 }, { "epoch": 0.6148318195428686, "grad_norm": 0.03916362673044205, "learning_rate": 7.0514891740773105e-06, "loss": 0.7027400970458985, "step": 3850 }, { "epoch": 0.616428785307915, "grad_norm": 0.03747180104255676, "learning_rate": 7.001179325803037e-06, "loss": 0.6675007820129395, "step": 3860 }, { "epoch": 0.6180257510729614, "grad_norm": 0.04694506898522377, "learning_rate": 6.95095274306693e-06, "loss": 0.6993120193481446, "step": 3870 }, { "epoch": 0.6196227168380077, "grad_norm": 0.035580579191446304, "learning_rate": 6.900810820465036e-06, "loss": 0.7056044578552246, "step": 3880 }, { "epoch": 0.6212196826030542, "grad_norm": 0.03560471907258034, "learning_rate": 6.8507549502427105e-06, "loss": 0.7067145824432373, "step": 3890 }, { "epoch": 0.6228166483681006, "grad_norm": 0.03395050764083862, "learning_rate": 6.800786522255974e-06, "loss": 0.6799113750457764, "step": 3900 }, { "epoch": 0.624413614133147, "grad_norm": 0.04874083772301674, "learning_rate": 6.7509069239329295e-06, "loss": 0.6687004089355468, "step": 3910 }, { "epoch": 0.6260105798981934, "grad_norm": 0.03185804560780525, "learning_rate": 6.701117540235204e-06, "loss": 0.6309501171112061, "step": 3920 }, { "epoch": 0.6276075456632398, "grad_norm": 0.04211097210645676, "learning_rate": 6.651419753619535e-06, "loss": 0.7135450839996338, "step": 3930 }, { "epoch": 0.6292045114282863, "grad_norm": 0.03911832720041275, "learning_rate": 6.601814943999363e-06, "loss": 0.7103962421417236, "step": 3940 }, { "epoch": 0.6308014771933327, "grad_norm": 0.033491455018520355, "learning_rate": 6.552304488706512e-06, "loss": 0.7135549068450928, "step": 3950 }, { "epoch": 0.6323984429583791, "grad_norm": 0.031158218160271645, "learning_rate": 6.502889762452969e-06, "loss": 0.6667858600616455, "step": 3960 }, { "epoch": 0.6339954087234255, "grad_norm": 0.037267692387104034, "learning_rate": 6.453572137292689e-06, "loss": 0.7077006816864013, "step": 3970 }, { "epoch": 0.635592374488472, "grad_norm": 0.03646966442465782, "learning_rate": 6.404352982583511e-06, "loss": 0.663360595703125, "step": 3980 }, { "epoch": 0.6371893402535184, "grad_norm": 0.0357484444975853, "learning_rate": 6.355233664949132e-06, "loss": 0.6508517742156983, "step": 3990 }, { "epoch": 0.6387863060185647, "grad_norm": 0.03508693352341652, "learning_rate": 6.306215548241174e-06, "loss": 0.669168758392334, "step": 4000 }, { "epoch": 0.6387863060185647, "eval_loss": 0.6750220060348511, "eval_runtime": 1753.5267, "eval_samples_per_second": 3.174, "eval_steps_per_second": 1.587, "step": 4000 }, { "epoch": 0.6403832717836111, "grad_norm": 0.03497602418065071, "learning_rate": 6.257299993501289e-06, "loss": 0.6791082859039307, "step": 4010 }, { "epoch": 0.6419802375486575, "grad_norm": 0.04195614159107208, "learning_rate": 6.208488358923393e-06, "loss": 0.6831603050231934, "step": 4020 }, { "epoch": 0.6435772033137039, "grad_norm": 0.049627698957920074, "learning_rate": 6.1597819998159506e-06, "loss": 0.7127910137176514, "step": 4030 }, { "epoch": 0.6451741690787504, "grad_norm": 0.03807014226913452, "learning_rate": 6.111182268564322e-06, "loss": 0.6687079429626465, "step": 4040 }, { "epoch": 0.6467711348437968, "grad_norm": 0.0430082343518734, "learning_rate": 6.062690514593237e-06, "loss": 0.6742999076843261, "step": 4050 }, { "epoch": 0.6483681006088432, "grad_norm": 0.032011304050683975, "learning_rate": 6.014308084329327e-06, "loss": 0.693959903717041, "step": 4060 }, { "epoch": 0.6499650663738896, "grad_norm": 0.0382765531539917, "learning_rate": 5.966036321163709e-06, "loss": 0.6869026184082031, "step": 4070 }, { "epoch": 0.651562032138936, "grad_norm": 0.038456711918115616, "learning_rate": 5.917876565414729e-06, "loss": 0.6802726268768311, "step": 4080 }, { "epoch": 0.6531589979039825, "grad_norm": 0.05505223199725151, "learning_rate": 5.8698301542907144e-06, "loss": 0.6540178775787353, "step": 4090 }, { "epoch": 0.6547559636690289, "grad_norm": 0.06022993102669716, "learning_rate": 5.8218984218528514e-06, "loss": 0.6731237888336181, "step": 4100 }, { "epoch": 0.6563529294340753, "grad_norm": 0.04032977297902107, "learning_rate": 5.774082698978154e-06, "loss": 0.7120064258575439, "step": 4110 }, { "epoch": 0.6579498951991216, "grad_norm": 0.04511059820652008, "learning_rate": 5.726384313322504e-06, "loss": 0.6774604797363282, "step": 4120 }, { "epoch": 0.659546860964168, "grad_norm": 0.04290354251861572, "learning_rate": 5.678804589283777e-06, "loss": 0.6660148143768311, "step": 4130 }, { "epoch": 0.6611438267292145, "grad_norm": 0.03994971513748169, "learning_rate": 5.6313448479650944e-06, "loss": 0.6876091957092285, "step": 4140 }, { "epoch": 0.6627407924942609, "grad_norm": 0.03699030727148056, "learning_rate": 5.584006407138114e-06, "loss": 0.6590198516845703, "step": 4150 }, { "epoch": 0.6643377582593073, "grad_norm": 0.041105858981609344, "learning_rate": 5.536790581206451e-06, "loss": 0.6922510147094727, "step": 4160 }, { "epoch": 0.6659347240243537, "grad_norm": 0.036437924951314926, "learning_rate": 5.489698681169196e-06, "loss": 0.7056612968444824, "step": 4170 }, { "epoch": 0.6675316897894001, "grad_norm": 0.03376461938023567, "learning_rate": 5.442732014584495e-06, "loss": 0.6548415184020996, "step": 4180 }, { "epoch": 0.6691286555544466, "grad_norm": 0.042223311960697174, "learning_rate": 5.395891885533244e-06, "loss": 0.6739361763000489, "step": 4190 }, { "epoch": 0.670725621319493, "grad_norm": 0.03633170202374458, "learning_rate": 5.349179594582898e-06, "loss": 0.7088988304138184, "step": 4200 }, { "epoch": 0.6723225870845394, "grad_norm": 0.037932299077510834, "learning_rate": 5.302596438751339e-06, "loss": 0.693552827835083, "step": 4210 }, { "epoch": 0.6739195528495858, "grad_norm": 0.03990362584590912, "learning_rate": 5.256143711470869e-06, "loss": 0.6921378135681152, "step": 4220 }, { "epoch": 0.6755165186146322, "grad_norm": 0.037905026227235794, "learning_rate": 5.20982270255231e-06, "loss": 0.6850985050201416, "step": 4230 }, { "epoch": 0.6771134843796787, "grad_norm": 0.0379810705780983, "learning_rate": 5.163634698149163e-06, "loss": 0.6650650978088379, "step": 4240 }, { "epoch": 0.678710450144725, "grad_norm": 0.02918444201350212, "learning_rate": 5.11758098072193e-06, "loss": 0.6373987674713135, "step": 4250 }, { "epoch": 0.6803074159097714, "grad_norm": 0.03856576979160309, "learning_rate": 5.071662829002477e-06, "loss": 0.7125391483306884, "step": 4260 }, { "epoch": 0.6819043816748178, "grad_norm": 0.03352683037519455, "learning_rate": 5.02588151795854e-06, "loss": 0.6727629661560058, "step": 4270 }, { "epoch": 0.6835013474398642, "grad_norm": 0.04891199618577957, "learning_rate": 4.98023831875833e-06, "loss": 0.722879695892334, "step": 4280 }, { "epoch": 0.6850983132049107, "grad_norm": 0.04167689383029938, "learning_rate": 4.9347344987352305e-06, "loss": 0.6740260124206543, "step": 4290 }, { "epoch": 0.6866952789699571, "grad_norm": 0.03634532913565636, "learning_rate": 4.889371321352607e-06, "loss": 0.649868106842041, "step": 4300 }, { "epoch": 0.6882922447350035, "grad_norm": 0.04563901200890541, "learning_rate": 4.844150046168722e-06, "loss": 0.6745593547821045, "step": 4310 }, { "epoch": 0.6898892105000499, "grad_norm": 0.04015154391527176, "learning_rate": 4.799071928801784e-06, "loss": 0.7058079719543457, "step": 4320 }, { "epoch": 0.6914861762650963, "grad_norm": 0.03595186024904251, "learning_rate": 4.7541382208950505e-06, "loss": 0.6532906532287598, "step": 4330 }, { "epoch": 0.6930831420301428, "grad_norm": 0.030780136585235596, "learning_rate": 4.709350170082103e-06, "loss": 0.6842369079589844, "step": 4340 }, { "epoch": 0.6946801077951892, "grad_norm": 0.04692146182060242, "learning_rate": 4.6647090199522025e-06, "loss": 0.7000434398651123, "step": 4350 }, { "epoch": 0.6962770735602356, "grad_norm": 0.0310053788125515, "learning_rate": 4.620216010015725e-06, "loss": 0.6727443218231202, "step": 4360 }, { "epoch": 0.6978740393252819, "grad_norm": 0.03317731246352196, "learning_rate": 4.575872375669793e-06, "loss": 0.6653543949127197, "step": 4370 }, { "epoch": 0.6994710050903283, "grad_norm": 0.05431032553315163, "learning_rate": 4.531679348163955e-06, "loss": 0.691961145401001, "step": 4380 }, { "epoch": 0.7010679708553748, "grad_norm": 0.037718210369348526, "learning_rate": 4.487638154565978e-06, "loss": 0.6595858573913574, "step": 4390 }, { "epoch": 0.7026649366204212, "grad_norm": 0.03125704079866409, "learning_rate": 4.443750017727814e-06, "loss": 0.7066734790802002, "step": 4400 }, { "epoch": 0.7042619023854676, "grad_norm": 0.050476983189582825, "learning_rate": 4.400016156251611e-06, "loss": 0.6906495094299316, "step": 4410 }, { "epoch": 0.705858868150514, "grad_norm": 0.034239206463098526, "learning_rate": 4.356437784455896e-06, "loss": 0.698778486251831, "step": 4420 }, { "epoch": 0.7074558339155604, "grad_norm": 0.03537076339125633, "learning_rate": 4.313016112341861e-06, "loss": 0.6730549335479736, "step": 4430 }, { "epoch": 0.7090527996806069, "grad_norm": 0.0366840697824955, "learning_rate": 4.269752345559761e-06, "loss": 0.6689012527465821, "step": 4440 }, { "epoch": 0.7106497654456533, "grad_norm": 0.04745423048734665, "learning_rate": 4.226647685375428e-06, "loss": 0.6668778419494629, "step": 4450 }, { "epoch": 0.7122467312106997, "grad_norm": 0.043418820947408676, "learning_rate": 4.183703328636924e-06, "loss": 0.6641871929168701, "step": 4460 }, { "epoch": 0.7138436969757461, "grad_norm": 0.0369136743247509, "learning_rate": 4.140920467741325e-06, "loss": 0.6592461585998535, "step": 4470 }, { "epoch": 0.7154406627407925, "grad_norm": 0.034368593245744705, "learning_rate": 4.098300290601581e-06, "loss": 0.6683683395385742, "step": 4480 }, { "epoch": 0.7170376285058389, "grad_norm": 0.042594823986291885, "learning_rate": 4.055843980613561e-06, "loss": 0.6916409015655518, "step": 4490 }, { "epoch": 0.7186345942708853, "grad_norm": 0.042385537177324295, "learning_rate": 4.013552716623185e-06, "loss": 0.6435957431793213, "step": 4500 }, { "epoch": 0.7186345942708853, "eval_loss": 0.6736165881156921, "eval_runtime": 1722.8539, "eval_samples_per_second": 3.231, "eval_steps_per_second": 1.615, "step": 4500 }, { "epoch": 0.7202315600359317, "grad_norm": 0.04700292646884918, "learning_rate": 3.971427672893672e-06, "loss": 0.688172435760498, "step": 4510 }, { "epoch": 0.7218285258009781, "grad_norm": 0.040195874869823456, "learning_rate": 3.929470019072972e-06, "loss": 0.6703804016113282, "step": 4520 }, { "epoch": 0.7234254915660245, "grad_norm": 0.036519281566143036, "learning_rate": 3.8876809201612695e-06, "loss": 0.6743984222412109, "step": 4530 }, { "epoch": 0.725022457331071, "grad_norm": 0.03687750920653343, "learning_rate": 3.846061536478626e-06, "loss": 0.6919309139251709, "step": 4540 }, { "epoch": 0.7266194230961174, "grad_norm": 0.034661829471588135, "learning_rate": 3.804613023632788e-06, "loss": 0.6852948665618896, "step": 4550 }, { "epoch": 0.7282163888611638, "grad_norm": 0.04090991988778114, "learning_rate": 3.763336532487076e-06, "loss": 0.6639271259307862, "step": 4560 }, { "epoch": 0.7298133546262102, "grad_norm": 0.053740449249744415, "learning_rate": 3.7222332091284384e-06, "loss": 0.7086254119873047, "step": 4570 }, { "epoch": 0.7314103203912566, "grad_norm": 0.05317610129714012, "learning_rate": 3.6813041948356408e-06, "loss": 0.688640546798706, "step": 4580 }, { "epoch": 0.7330072861563031, "grad_norm": 0.04328848421573639, "learning_rate": 3.640550626047562e-06, "loss": 0.6996944904327392, "step": 4590 }, { "epoch": 0.7346042519213495, "grad_norm": 0.03426145389676094, "learning_rate": 3.5999736343316406e-06, "loss": 0.7036723613739013, "step": 4600 }, { "epoch": 0.7362012176863958, "grad_norm": 0.04445657134056091, "learning_rate": 3.559574346352459e-06, "loss": 0.6727498054504395, "step": 4610 }, { "epoch": 0.7377981834514422, "grad_norm": 0.034136440604925156, "learning_rate": 3.5193538838404716e-06, "loss": 0.6709301471710205, "step": 4620 }, { "epoch": 0.7393951492164886, "grad_norm": 0.03670594096183777, "learning_rate": 3.4793133635608334e-06, "loss": 0.7098387241363525, "step": 4630 }, { "epoch": 0.740992114981535, "grad_norm": 0.04140735790133476, "learning_rate": 3.4394538972824167e-06, "loss": 0.6899067401885987, "step": 4640 }, { "epoch": 0.7425890807465815, "grad_norm": 0.03637846186757088, "learning_rate": 3.3997765917469348e-06, "loss": 0.6745570659637451, "step": 4650 }, { "epoch": 0.7441860465116279, "grad_norm": 0.03964189440011978, "learning_rate": 3.3602825486381886e-06, "loss": 0.6715593338012695, "step": 4660 }, { "epoch": 0.7457830122766743, "grad_norm": 0.036266524344682693, "learning_rate": 3.3209728645515204e-06, "loss": 0.6992376804351806, "step": 4670 }, { "epoch": 0.7473799780417207, "grad_norm": 0.050657473504543304, "learning_rate": 3.281848630963327e-06, "loss": 0.689194917678833, "step": 4680 }, { "epoch": 0.7489769438067672, "grad_norm": 0.03800279274582863, "learning_rate": 3.2429109342007747e-06, "loss": 0.7068089962005615, "step": 4690 }, { "epoch": 0.7505739095718136, "grad_norm": 0.034397054463624954, "learning_rate": 3.204160855411638e-06, "loss": 0.7055879592895508, "step": 4700 }, { "epoch": 0.75217087533686, "grad_norm": 0.03663492202758789, "learning_rate": 3.1655994705342595e-06, "loss": 0.6516903400421142, "step": 4710 }, { "epoch": 0.7537678411019064, "grad_norm": 0.10431419312953949, "learning_rate": 3.1272278502676932e-06, "loss": 0.6833277702331543, "step": 4720 }, { "epoch": 0.7553648068669528, "grad_norm": 0.04803668335080147, "learning_rate": 3.0890470600419785e-06, "loss": 0.7252533435821533, "step": 4730 }, { "epoch": 0.7569617726319992, "grad_norm": 0.052451301366090775, "learning_rate": 3.051058159988539e-06, "loss": 0.6383802413940429, "step": 4740 }, { "epoch": 0.7585587383970456, "grad_norm": 0.04214588552713394, "learning_rate": 3.0132622049107672e-06, "loss": 0.6385763645172119, "step": 4750 }, { "epoch": 0.760155704162092, "grad_norm": 0.04353920742869377, "learning_rate": 2.975660244254721e-06, "loss": 0.6771360874176026, "step": 4760 }, { "epoch": 0.7617526699271384, "grad_norm": 0.034283965826034546, "learning_rate": 2.9382533220799893e-06, "loss": 0.6694541931152344, "step": 4770 }, { "epoch": 0.7633496356921848, "grad_norm": 0.04156842082738876, "learning_rate": 2.901042477030709e-06, "loss": 0.719163179397583, "step": 4780 }, { "epoch": 0.7649466014572313, "grad_norm": 0.038868360221385956, "learning_rate": 2.864028742306725e-06, "loss": 0.6658177852630616, "step": 4790 }, { "epoch": 0.7665435672222777, "grad_norm": 0.03701276332139969, "learning_rate": 2.827213145634887e-06, "loss": 0.6825634956359863, "step": 4800 }, { "epoch": 0.7681405329873241, "grad_norm": 0.043009012937545776, "learning_rate": 2.7905967092405305e-06, "loss": 0.6813971519470214, "step": 4810 }, { "epoch": 0.7697374987523705, "grad_norm": 0.03427146002650261, "learning_rate": 2.754180449819092e-06, "loss": 0.6715961933135987, "step": 4820 }, { "epoch": 0.7713344645174169, "grad_norm": 0.04296226054430008, "learning_rate": 2.717965378507864e-06, "loss": 0.6857268810272217, "step": 4830 }, { "epoch": 0.7729314302824634, "grad_norm": 0.0429498665034771, "learning_rate": 2.681952500857944e-06, "loss": 0.6708025932312012, "step": 4840 }, { "epoch": 0.7745283960475098, "grad_norm": 0.03906711935997009, "learning_rate": 2.6461428168062973e-06, "loss": 0.6925381660461426, "step": 4850 }, { "epoch": 0.7761253618125561, "grad_norm": 0.04651563614606857, "learning_rate": 2.6105373206479913e-06, "loss": 0.6757769107818603, "step": 4860 }, { "epoch": 0.7777223275776025, "grad_norm": 0.04317731410264969, "learning_rate": 2.575137001008592e-06, "loss": 0.665839433670044, "step": 4870 }, { "epoch": 0.7793192933426489, "grad_norm": 0.03744199126958847, "learning_rate": 2.539942840816727e-06, "loss": 0.7045305252075196, "step": 4880 }, { "epoch": 0.7809162591076954, "grad_norm": 0.04245394840836525, "learning_rate": 2.5049558172767675e-06, "loss": 0.6957573890686035, "step": 4890 }, { "epoch": 0.7825132248727418, "grad_norm": 0.03942755609750748, "learning_rate": 2.47017690184172e-06, "loss": 0.6565013408660889, "step": 4900 }, { "epoch": 0.7841101906377882, "grad_norm": 0.04654339328408241, "learning_rate": 2.4356070601862327e-06, "loss": 0.6680932998657226, "step": 4910 }, { "epoch": 0.7857071564028346, "grad_norm": 0.05248269438743591, "learning_rate": 2.4012472521797923e-06, "loss": 0.6607920169830322, "step": 4920 }, { "epoch": 0.787304122167881, "grad_norm": 0.04624265059828758, "learning_rate": 2.3670984318600786e-06, "loss": 0.6683568954467773, "step": 4930 }, { "epoch": 0.7889010879329275, "grad_norm": 0.04104055091738701, "learning_rate": 2.333161547406464e-06, "loss": 0.6837501049041748, "step": 4940 }, { "epoch": 0.7904980536979739, "grad_norm": 0.039029769599437714, "learning_rate": 2.2994375411136884e-06, "loss": 0.6702389240264892, "step": 4950 }, { "epoch": 0.7920950194630203, "grad_norm": 0.03849470987915993, "learning_rate": 2.265927349365691e-06, "loss": 0.6521383285522461, "step": 4960 }, { "epoch": 0.7936919852280667, "grad_norm": 0.05400224030017853, "learning_rate": 2.232631902609629e-06, "loss": 0.6533032894134522, "step": 4970 }, { "epoch": 0.795288950993113, "grad_norm": 0.043887361884117126, "learning_rate": 2.1995521253300167e-06, "loss": 0.6794767379760742, "step": 4980 }, { "epoch": 0.7968859167581595, "grad_norm": 0.04501941800117493, "learning_rate": 2.166688936023078e-06, "loss": 0.6698923587799073, "step": 4990 }, { "epoch": 0.7984828825232059, "grad_norm": 0.045139558613300323, "learning_rate": 2.1340432471712367e-06, "loss": 0.661639404296875, "step": 5000 }, { "epoch": 0.7984828825232059, "eval_loss": 0.6726189255714417, "eval_runtime": 1709.319, "eval_samples_per_second": 3.256, "eval_steps_per_second": 1.628, "step": 5000 }, { "epoch": 0.8000798482882523, "grad_norm": 0.04251272603869438, "learning_rate": 2.101615965217768e-06, "loss": 0.7006937503814697, "step": 5010 }, { "epoch": 0.8016768140532987, "grad_norm": 0.040616679936647415, "learning_rate": 2.0694079905416475e-06, "loss": 0.6587623119354248, "step": 5020 }, { "epoch": 0.8032737798183451, "grad_norm": 0.04379381984472275, "learning_rate": 2.0374202174325443e-06, "loss": 0.6809040069580078, "step": 5030 }, { "epoch": 0.8048707455833916, "grad_norm": 0.042080674320459366, "learning_rate": 2.0056535340659867e-06, "loss": 0.6683236598968506, "step": 5040 }, { "epoch": 0.806467711348438, "grad_norm": 0.038132891058921814, "learning_rate": 1.97410882247871e-06, "loss": 0.6637454032897949, "step": 5050 }, { "epoch": 0.8080646771134844, "grad_norm": 0.029925232753157616, "learning_rate": 1.942786958544154e-06, "loss": 0.6917686939239502, "step": 5060 }, { "epoch": 0.8096616428785308, "grad_norm": 0.041264407336711884, "learning_rate": 1.91168881194815e-06, "loss": 0.7052114486694336, "step": 5070 }, { "epoch": 0.8112586086435772, "grad_norm": 0.03905145451426506, "learning_rate": 1.8808152461647766e-06, "loss": 0.667028284072876, "step": 5080 }, { "epoch": 0.8128555744086237, "grad_norm": 0.047729961574077606, "learning_rate": 1.8501671184323811e-06, "loss": 0.691020917892456, "step": 5090 }, { "epoch": 0.8144525401736701, "grad_norm": 0.034812819212675095, "learning_rate": 1.8197452797297699e-06, "loss": 0.7124451637268067, "step": 5100 }, { "epoch": 0.8160495059387164, "grad_norm": 0.044796060770750046, "learning_rate": 1.7895505747525933e-06, "loss": 0.7617973804473877, "step": 5110 }, { "epoch": 0.8176464717037628, "grad_norm": 0.039166904985904694, "learning_rate": 1.759583841889878e-06, "loss": 0.711750841140747, "step": 5120 }, { "epoch": 0.8192434374688092, "grad_norm": 0.03871138393878937, "learning_rate": 1.7298459132007628e-06, "loss": 0.6736746788024902, "step": 5130 }, { "epoch": 0.8208404032338557, "grad_norm": 0.046535786241292953, "learning_rate": 1.7003376143913886e-06, "loss": 0.7094714164733886, "step": 5140 }, { "epoch": 0.8224373689989021, "grad_norm": 0.04319629445672035, "learning_rate": 1.6710597647919613e-06, "loss": 0.6693609237670899, "step": 5150 }, { "epoch": 0.8240343347639485, "grad_norm": 0.04204268380999565, "learning_rate": 1.6420131773340263e-06, "loss": 0.6906074047088623, "step": 5160 }, { "epoch": 0.8256313005289949, "grad_norm": 0.041262589395046234, "learning_rate": 1.6131986585278714e-06, "loss": 0.6685310363769531, "step": 5170 }, { "epoch": 0.8272282662940413, "grad_norm": 0.03515459597110748, "learning_rate": 1.5846170084401458e-06, "loss": 0.6744651317596435, "step": 5180 }, { "epoch": 0.8288252320590878, "grad_norm": 0.044764138758182526, "learning_rate": 1.5562690206716479e-06, "loss": 0.6720765113830567, "step": 5190 }, { "epoch": 0.8304221978241342, "grad_norm": 0.03811672702431679, "learning_rate": 1.5281554823352896e-06, "loss": 0.6512242317199707, "step": 5200 }, { "epoch": 0.8320191635891806, "grad_norm": 0.03850555047392845, "learning_rate": 1.500277174034227e-06, "loss": 0.6925489425659179, "step": 5210 }, { "epoch": 0.833616129354227, "grad_norm": 0.039821285754442215, "learning_rate": 1.4726348698402015e-06, "loss": 0.6698274612426758, "step": 5220 }, { "epoch": 0.8352130951192733, "grad_norm": 0.042530301958322525, "learning_rate": 1.4452293372720483e-06, "loss": 0.657553243637085, "step": 5230 }, { "epoch": 0.8368100608843198, "grad_norm": 0.03285897895693779, "learning_rate": 1.4180613372743678e-06, "loss": 0.6513785362243653, "step": 5240 }, { "epoch": 0.8384070266493662, "grad_norm": 0.0418255515396595, "learning_rate": 1.39113162419642e-06, "loss": 0.6803596019744873, "step": 5250 }, { "epoch": 0.8400039924144126, "grad_norm": 0.03861439600586891, "learning_rate": 1.3644409457711604e-06, "loss": 0.7275567531585694, "step": 5260 }, { "epoch": 0.841600958179459, "grad_norm": 0.035479992628097534, "learning_rate": 1.3379900430944858e-06, "loss": 0.7078885555267334, "step": 5270 }, { "epoch": 0.8431979239445054, "grad_norm": 0.036275558173656464, "learning_rate": 1.31177965060466e-06, "loss": 0.6733812808990478, "step": 5280 }, { "epoch": 0.8447948897095519, "grad_norm": 0.037248894572257996, "learning_rate": 1.285810496061921e-06, "loss": 0.6728686809539794, "step": 5290 }, { "epoch": 0.8463918554745983, "grad_norm": 0.03526445850729942, "learning_rate": 1.2600833005282631e-06, "loss": 0.6712977409362793, "step": 5300 }, { "epoch": 0.8479888212396447, "grad_norm": 0.03622904419898987, "learning_rate": 1.2345987783474311e-06, "loss": 0.6881800651550293, "step": 5310 }, { "epoch": 0.8495857870046911, "grad_norm": 0.05652482062578201, "learning_rate": 1.2093576371250737e-06, "loss": 0.709586763381958, "step": 5320 }, { "epoch": 0.8511827527697375, "grad_norm": 0.05090609937906265, "learning_rate": 1.1843605777091028e-06, "loss": 0.7400435447692871, "step": 5330 }, { "epoch": 0.852779718534784, "grad_norm": 0.038989149034023285, "learning_rate": 1.1596082941702313e-06, "loss": 0.695259952545166, "step": 5340 }, { "epoch": 0.8543766842998303, "grad_norm": 0.04729596897959709, "learning_rate": 1.135101473782706e-06, "loss": 0.681358528137207, "step": 5350 }, { "epoch": 0.8559736500648767, "grad_norm": 0.0379403792321682, "learning_rate": 1.1108407970052149e-06, "loss": 0.6542623043060303, "step": 5360 }, { "epoch": 0.8575706158299231, "grad_norm": 0.044038884341716766, "learning_rate": 1.086826937462e-06, "loss": 0.6564295291900635, "step": 5370 }, { "epoch": 0.8591675815949695, "grad_norm": 0.04039445519447327, "learning_rate": 1.0630605619241564e-06, "loss": 0.6783324241638183, "step": 5380 }, { "epoch": 0.860764547360016, "grad_norm": 0.034845978021621704, "learning_rate": 1.0395423302911112e-06, "loss": 0.6549112796783447, "step": 5390 }, { "epoch": 0.8623615131250624, "grad_norm": 0.048201173543930054, "learning_rate": 1.016272895572309e-06, "loss": 0.6641885757446289, "step": 5400 }, { "epoch": 0.8639584788901088, "grad_norm": 0.037778809666633606, "learning_rate": 9.932529038690696e-07, "loss": 0.7034367561340332, "step": 5410 }, { "epoch": 0.8655554446551552, "grad_norm": 0.0427137166261673, "learning_rate": 9.704829943566574e-07, "loss": 0.6777073383331299, "step": 5420 }, { "epoch": 0.8671524104202016, "grad_norm": 0.037193115800619125, "learning_rate": 9.479637992665313e-07, "loss": 0.6748072147369385, "step": 5430 }, { "epoch": 0.8687493761852481, "grad_norm": 0.0337526798248291, "learning_rate": 9.256959438687929e-07, "loss": 0.6617064952850342, "step": 5440 }, { "epoch": 0.8703463419502945, "grad_norm": 0.039619434624910355, "learning_rate": 9.036800464548157e-07, "loss": 0.6612413883209228, "step": 5450 }, { "epoch": 0.8719433077153409, "grad_norm": 0.03671206533908844, "learning_rate": 8.819167183200905e-07, "loss": 0.6929868698120117, "step": 5460 }, { "epoch": 0.8735402734803873, "grad_norm": 0.034324243664741516, "learning_rate": 8.604065637472392e-07, "loss": 0.685936689376831, "step": 5470 }, { "epoch": 0.8751372392454336, "grad_norm": 0.035913076251745224, "learning_rate": 8.391501799892466e-07, "loss": 0.7082894802093506, "step": 5480 }, { "epoch": 0.87673420501048, "grad_norm": 0.04239356890320778, "learning_rate": 8.181481572528726e-07, "loss": 0.7099148750305175, "step": 5490 }, { "epoch": 0.8783311707755265, "grad_norm": 0.04144662618637085, "learning_rate": 7.974010786822684e-07, "loss": 0.6640007972717286, "step": 5500 }, { "epoch": 0.8783311707755265, "eval_loss": 0.672135055065155, "eval_runtime": 1721.4336, "eval_samples_per_second": 3.233, "eval_steps_per_second": 1.617, "step": 5500 }, { "epoch": 0.8799281365405729, "grad_norm": 0.030981754884123802, "learning_rate": 7.769095203427734e-07, "loss": 0.6549492359161377, "step": 5510 }, { "epoch": 0.8815251023056193, "grad_norm": 0.04086959362030029, "learning_rate": 7.56674051204932e-07, "loss": 0.6875146389007568, "step": 5520 }, { "epoch": 0.8831220680706657, "grad_norm": 0.0377834253013134, "learning_rate": 7.366952331286958e-07, "loss": 0.7493057250976562, "step": 5530 }, { "epoch": 0.8847190338357122, "grad_norm": 0.03834075480699539, "learning_rate": 7.169736208478118e-07, "loss": 0.6828866481781006, "step": 5540 }, { "epoch": 0.8863159996007586, "grad_norm": 0.03951543942093849, "learning_rate": 6.975097619544324e-07, "loss": 0.68464674949646, "step": 5550 }, { "epoch": 0.887912965365805, "grad_norm": 0.044246163219213486, "learning_rate": 6.78304196883901e-07, "loss": 0.6737260341644287, "step": 5560 }, { "epoch": 0.8895099311308514, "grad_norm": 0.04266819730401039, "learning_rate": 6.59357458899752e-07, "loss": 0.7164577960968017, "step": 5570 }, { "epoch": 0.8911068968958978, "grad_norm": 0.0311746709048748, "learning_rate": 6.40670074078904e-07, "loss": 0.680039358139038, "step": 5580 }, { "epoch": 0.8927038626609443, "grad_norm": 0.035248082131147385, "learning_rate": 6.222425612970462e-07, "loss": 0.6851954936981202, "step": 5590 }, { "epoch": 0.8943008284259906, "grad_norm": 0.03986518457531929, "learning_rate": 6.040754322142417e-07, "loss": 0.6760111331939698, "step": 5600 }, { "epoch": 0.895897794191037, "grad_norm": 0.03682507202029228, "learning_rate": 5.861691912607159e-07, "loss": 0.6883102416992187, "step": 5610 }, { "epoch": 0.8974947599560834, "grad_norm": 0.03000503219664097, "learning_rate": 5.685243356228409e-07, "loss": 0.6542916774749756, "step": 5620 }, { "epoch": 0.8990917257211298, "grad_norm": 0.044614020735025406, "learning_rate": 5.511413552293476e-07, "loss": 0.6680744647979736, "step": 5630 }, { "epoch": 0.9006886914861763, "grad_norm": 0.03846392408013344, "learning_rate": 5.340207327377156e-07, "loss": 0.7028224468231201, "step": 5640 }, { "epoch": 0.9022856572512227, "grad_norm": 0.038470469415187836, "learning_rate": 5.171629435207615e-07, "loss": 0.6939930438995361, "step": 5650 }, { "epoch": 0.9038826230162691, "grad_norm": 0.03757188841700554, "learning_rate": 5.005684556534574e-07, "loss": 0.656277084350586, "step": 5660 }, { "epoch": 0.9054795887813155, "grad_norm": 0.04011685401201248, "learning_rate": 4.842377298999191e-07, "loss": 0.638321828842163, "step": 5670 }, { "epoch": 0.9070765545463619, "grad_norm": 0.039341870695352554, "learning_rate": 4.681712197006205e-07, "loss": 0.6716189384460449, "step": 5680 }, { "epoch": 0.9086735203114084, "grad_norm": 0.046330180019140244, "learning_rate": 4.5236937115980204e-07, "loss": 0.6909313678741456, "step": 5690 }, { "epoch": 0.9102704860764548, "grad_norm": 0.041009679436683655, "learning_rate": 4.368326230330855e-07, "loss": 0.6272893428802491, "step": 5700 }, { "epoch": 0.9118674518415012, "grad_norm": 0.03455239534378052, "learning_rate": 4.215614067152851e-07, "loss": 0.6927034854888916, "step": 5710 }, { "epoch": 0.9134644176065475, "grad_norm": 0.03742032125592232, "learning_rate": 4.065561462284351e-07, "loss": 0.664287805557251, "step": 5720 }, { "epoch": 0.9150613833715939, "grad_norm": 0.03536292910575867, "learning_rate": 3.918172582100177e-07, "loss": 0.664261770248413, "step": 5730 }, { "epoch": 0.9166583491366403, "grad_norm": 0.044113483279943466, "learning_rate": 3.7734515190138934e-07, "loss": 0.7117574691772461, "step": 5740 }, { "epoch": 0.9182553149016868, "grad_norm": 0.04183537885546684, "learning_rate": 3.631402291364206e-07, "loss": 0.6857941627502442, "step": 5750 }, { "epoch": 0.9198522806667332, "grad_norm": 0.02898181416094303, "learning_rate": 3.492028843303419e-07, "loss": 0.6839983463287354, "step": 5760 }, { "epoch": 0.9214492464317796, "grad_norm": 0.040342144668102264, "learning_rate": 3.3553350446878244e-07, "loss": 0.6938966751098633, "step": 5770 }, { "epoch": 0.923046212196826, "grad_norm": 0.05028856173157692, "learning_rate": 3.221324690970351e-07, "loss": 0.6682254791259765, "step": 5780 }, { "epoch": 0.9246431779618725, "grad_norm": 0.033745042979717255, "learning_rate": 3.090001503095175e-07, "loss": 0.6571940422058106, "step": 5790 }, { "epoch": 0.9262401437269189, "grad_norm": 0.036207910627126694, "learning_rate": 2.9613691273942977e-07, "loss": 0.7331238269805909, "step": 5800 }, { "epoch": 0.9278371094919653, "grad_norm": 0.04358496144413948, "learning_rate": 2.835431135486455e-07, "loss": 0.6768048763275146, "step": 5810 }, { "epoch": 0.9294340752570117, "grad_norm": 0.051529187709093094, "learning_rate": 2.7121910241778014e-07, "loss": 0.6914044857025147, "step": 5820 }, { "epoch": 0.9310310410220581, "grad_norm": 0.03387345373630524, "learning_rate": 2.591652215364904e-07, "loss": 0.67225022315979, "step": 5830 }, { "epoch": 0.9326280067871044, "grad_norm": 0.043374814093112946, "learning_rate": 2.473818055939747e-07, "loss": 0.6870133876800537, "step": 5840 }, { "epoch": 0.9342249725521509, "grad_norm": 0.04162263497710228, "learning_rate": 2.358691817696712e-07, "loss": 0.6274053573608398, "step": 5850 }, { "epoch": 0.9358219383171973, "grad_norm": 0.03655523061752319, "learning_rate": 2.2462766972418226e-07, "loss": 0.6559162139892578, "step": 5860 }, { "epoch": 0.9374189040822437, "grad_norm": 0.036134812980890274, "learning_rate": 2.1365758159039073e-07, "loss": 0.6783801078796386, "step": 5870 }, { "epoch": 0.9390158698472901, "grad_norm": 0.03370088338851929, "learning_rate": 2.0295922196480355e-07, "loss": 0.679423189163208, "step": 5880 }, { "epoch": 0.9406128356123366, "grad_norm": 0.044190868735313416, "learning_rate": 1.9253288789908287e-07, "loss": 0.6669793128967285, "step": 5890 }, { "epoch": 0.942209801377383, "grad_norm": 0.03765581548213959, "learning_rate": 1.823788688918049e-07, "loss": 0.6280838966369628, "step": 5900 }, { "epoch": 0.9438067671424294, "grad_norm": 0.05461205169558525, "learning_rate": 1.7249744688042414e-07, "loss": 0.6410412788391113, "step": 5910 }, { "epoch": 0.9454037329074758, "grad_norm": 0.046921294182538986, "learning_rate": 1.6288889623343295e-07, "loss": 0.6527444839477539, "step": 5920 }, { "epoch": 0.9470006986725222, "grad_norm": 0.03945396840572357, "learning_rate": 1.535534837427588e-07, "loss": 0.694933271408081, "step": 5930 }, { "epoch": 0.9485976644375687, "grad_norm": 0.03962003439664841, "learning_rate": 1.44491468616349e-07, "loss": 0.6714414596557617, "step": 5940 }, { "epoch": 0.9501946302026151, "grad_norm": 0.07099556922912598, "learning_rate": 1.3570310247096764e-07, "loss": 0.7129894733428955, "step": 5950 }, { "epoch": 0.9517915959676615, "grad_norm": 0.03700670972466469, "learning_rate": 1.2718862932522224e-07, "loss": 0.6442203044891357, "step": 5960 }, { "epoch": 0.9533885617327078, "grad_norm": 0.04066568240523338, "learning_rate": 1.1894828559277816e-07, "loss": 0.6612192153930664, "step": 5970 }, { "epoch": 0.9549855274977542, "grad_norm": 0.04404400289058685, "learning_rate": 1.1098230007579702e-07, "loss": 0.6766380310058594, "step": 5980 }, { "epoch": 0.9565824932628006, "grad_norm": 0.04611429572105408, "learning_rate": 1.0329089395858527e-07, "loss": 0.6801462650299073, "step": 5990 }, { "epoch": 0.9581794590278471, "grad_norm": 0.03326278179883957, "learning_rate": 9.587428080145344e-08, "loss": 0.6539367198944092, "step": 6000 }, { "epoch": 0.9581794590278471, "eval_loss": 0.6719915270805359, "eval_runtime": 1743.0303, "eval_samples_per_second": 3.193, "eval_steps_per_second": 1.597, "step": 6000 } ], "logging_steps": 10, "max_steps": 6262, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.3174132430797578e+19, "train_batch_size": 2, "trial_name": null, "trial_params": null }