MNLP_M3_mcqa_model_trial / trainer_state.json
ellendagher's picture
Upload folder using huggingface_hub
e9d8b24 verified
{
"best_global_step": 14058,
"best_metric": 0.6273267865180969,
"best_model_checkpoint": "./mcqa_qwen3_letter_m3/checkpoint-14058",
"epoch": 2.0,
"eval_steps": 500,
"global_step": 14058,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.007113640405477503,
"grad_norm": 20.266469955444336,
"learning_rate": 4.551920341394026e-07,
"loss": 2.1439,
"step": 50
},
{
"epoch": 0.014227280810955006,
"grad_norm": 21.066892623901367,
"learning_rate": 9.293504030346136e-07,
"loss": 1.4471,
"step": 100
},
{
"epoch": 0.02134092121643251,
"grad_norm": 18.918508529663086,
"learning_rate": 1.4035087719298246e-06,
"loss": 1.3067,
"step": 150
},
{
"epoch": 0.028454561621910013,
"grad_norm": 9.563344955444336,
"learning_rate": 1.8681839734471315e-06,
"loss": 1.2745,
"step": 200
},
{
"epoch": 0.035568202027387516,
"grad_norm": 20.408035278320312,
"learning_rate": 2.3423423423423424e-06,
"loss": 1.2323,
"step": 250
},
{
"epoch": 0.04268184243286502,
"grad_norm": 17.020856857299805,
"learning_rate": 2.8165007112375536e-06,
"loss": 1.2476,
"step": 300
},
{
"epoch": 0.04979548283834252,
"grad_norm": 10.942744255065918,
"learning_rate": 3.2906590801327647e-06,
"loss": 1.2084,
"step": 350
},
{
"epoch": 0.056909123243820026,
"grad_norm": 11.239660263061523,
"learning_rate": 3.7648174490279754e-06,
"loss": 1.1453,
"step": 400
},
{
"epoch": 0.06402276364929753,
"grad_norm": 8.421320915222168,
"learning_rate": 4.2389758179231865e-06,
"loss": 1.1257,
"step": 450
},
{
"epoch": 0.07113640405477503,
"grad_norm": 7.429409027099609,
"learning_rate": 4.713134186818398e-06,
"loss": 1.0896,
"step": 500
},
{
"epoch": 0.07825004446025254,
"grad_norm": 8.3453950881958,
"learning_rate": 5.187292555713609e-06,
"loss": 1.0589,
"step": 550
},
{
"epoch": 0.08536368486573004,
"grad_norm": 9.157875061035156,
"learning_rate": 5.66145092460882e-06,
"loss": 1.0628,
"step": 600
},
{
"epoch": 0.09247732527120754,
"grad_norm": 9.339110374450684,
"learning_rate": 6.135609293504031e-06,
"loss": 1.0211,
"step": 650
},
{
"epoch": 0.09959096567668504,
"grad_norm": 7.010004043579102,
"learning_rate": 6.609767662399243e-06,
"loss": 1.0684,
"step": 700
},
{
"epoch": 0.10670460608216255,
"grad_norm": 8.618551254272461,
"learning_rate": 7.0839260312944525e-06,
"loss": 1.0467,
"step": 750
},
{
"epoch": 0.11381824648764005,
"grad_norm": 9.206182479858398,
"learning_rate": 7.558084400189663e-06,
"loss": 1.0855,
"step": 800
},
{
"epoch": 0.12093188689311755,
"grad_norm": 11.882657051086426,
"learning_rate": 8.032242769084875e-06,
"loss": 0.9897,
"step": 850
},
{
"epoch": 0.12804552729859506,
"grad_norm": 6.381322860717773,
"learning_rate": 8.506401137980086e-06,
"loss": 1.0227,
"step": 900
},
{
"epoch": 0.13515916770407255,
"grad_norm": 7.592447280883789,
"learning_rate": 8.980559506875298e-06,
"loss": 0.9508,
"step": 950
},
{
"epoch": 0.14227280810955006,
"grad_norm": 10.622550010681152,
"learning_rate": 9.454717875770508e-06,
"loss": 0.9916,
"step": 1000
},
{
"epoch": 0.14938644851502755,
"grad_norm": 16.836315155029297,
"learning_rate": 9.92887624466572e-06,
"loss": 0.9838,
"step": 1050
},
{
"epoch": 0.15650008892050507,
"grad_norm": 8.732917785644531,
"learning_rate": 1.040303461356093e-05,
"loss": 0.9678,
"step": 1100
},
{
"epoch": 0.16361372932598256,
"grad_norm": 10.75668716430664,
"learning_rate": 1.0877192982456142e-05,
"loss": 0.9535,
"step": 1150
},
{
"epoch": 0.17072736973146008,
"grad_norm": 10.604599952697754,
"learning_rate": 1.1351351351351352e-05,
"loss": 1.0164,
"step": 1200
},
{
"epoch": 0.17784101013693757,
"grad_norm": 11.192643165588379,
"learning_rate": 1.1825509720246564e-05,
"loss": 0.9811,
"step": 1250
},
{
"epoch": 0.18495465054241508,
"grad_norm": 11.667268753051758,
"learning_rate": 1.2299668089141774e-05,
"loss": 0.9329,
"step": 1300
},
{
"epoch": 0.19206829094789257,
"grad_norm": 12.122210502624512,
"learning_rate": 1.2773826458036987e-05,
"loss": 0.9693,
"step": 1350
},
{
"epoch": 0.1991819313533701,
"grad_norm": 10.696669578552246,
"learning_rate": 1.3247984826932197e-05,
"loss": 0.9118,
"step": 1400
},
{
"epoch": 0.20629557175884758,
"grad_norm": 9.4218168258667,
"learning_rate": 1.3722143195827407e-05,
"loss": 0.9242,
"step": 1450
},
{
"epoch": 0.2134092121643251,
"grad_norm": 15.12156867980957,
"learning_rate": 1.4196301564722618e-05,
"loss": 0.9161,
"step": 1500
},
{
"epoch": 0.22052285256980259,
"grad_norm": 10.341136932373047,
"learning_rate": 1.4670459933617828e-05,
"loss": 0.9112,
"step": 1550
},
{
"epoch": 0.2276364929752801,
"grad_norm": 8.621809005737305,
"learning_rate": 1.5144618302513041e-05,
"loss": 0.923,
"step": 1600
},
{
"epoch": 0.2347501333807576,
"grad_norm": 8.037166595458984,
"learning_rate": 1.561877667140825e-05,
"loss": 0.8646,
"step": 1650
},
{
"epoch": 0.2418637737862351,
"grad_norm": 12.160765647888184,
"learning_rate": 1.6092935040303463e-05,
"loss": 0.8811,
"step": 1700
},
{
"epoch": 0.2489774141917126,
"grad_norm": 9.850648880004883,
"learning_rate": 1.6567093409198673e-05,
"loss": 0.9158,
"step": 1750
},
{
"epoch": 0.2560910545971901,
"grad_norm": 6.849952220916748,
"learning_rate": 1.7041251778093886e-05,
"loss": 0.8903,
"step": 1800
},
{
"epoch": 0.26320469500266763,
"grad_norm": 7.7248334884643555,
"learning_rate": 1.7515410146989096e-05,
"loss": 0.8853,
"step": 1850
},
{
"epoch": 0.2703183354081451,
"grad_norm": 5.612997531890869,
"learning_rate": 1.7989568515884306e-05,
"loss": 0.8407,
"step": 1900
},
{
"epoch": 0.2774319758136226,
"grad_norm": 6.731159210205078,
"learning_rate": 1.846372688477952e-05,
"loss": 0.8447,
"step": 1950
},
{
"epoch": 0.28454561621910013,
"grad_norm": 8.783773422241211,
"learning_rate": 1.893788525367473e-05,
"loss": 0.8201,
"step": 2000
},
{
"epoch": 0.29165925662457765,
"grad_norm": 8.748090744018555,
"learning_rate": 1.9412043622569942e-05,
"loss": 0.8202,
"step": 2050
},
{
"epoch": 0.2987728970300551,
"grad_norm": 6.985095977783203,
"learning_rate": 1.9886201991465152e-05,
"loss": 0.8686,
"step": 2100
},
{
"epoch": 0.3058865374355326,
"grad_norm": 9.010405540466309,
"learning_rate": 1.9959947299077735e-05,
"loss": 0.8541,
"step": 2150
},
{
"epoch": 0.31300017784101014,
"grad_norm": 10.884967803955078,
"learning_rate": 1.9907246376811596e-05,
"loss": 0.9021,
"step": 2200
},
{
"epoch": 0.32011381824648766,
"grad_norm": 6.696805953979492,
"learning_rate": 1.9854545454545456e-05,
"loss": 0.8352,
"step": 2250
},
{
"epoch": 0.3272274586519651,
"grad_norm": 6.158838272094727,
"learning_rate": 1.9801844532279317e-05,
"loss": 0.8078,
"step": 2300
},
{
"epoch": 0.33434109905744264,
"grad_norm": 10.230331420898438,
"learning_rate": 1.9749143610013177e-05,
"loss": 0.8659,
"step": 2350
},
{
"epoch": 0.34145473946292015,
"grad_norm": 7.0724945068359375,
"learning_rate": 1.9696442687747038e-05,
"loss": 0.862,
"step": 2400
},
{
"epoch": 0.34856837986839767,
"grad_norm": 9.820209503173828,
"learning_rate": 1.9643741765480898e-05,
"loss": 0.8036,
"step": 2450
},
{
"epoch": 0.35568202027387513,
"grad_norm": 6.326645374298096,
"learning_rate": 1.959104084321476e-05,
"loss": 0.8255,
"step": 2500
},
{
"epoch": 0.36279566067935265,
"grad_norm": 4.881910800933838,
"learning_rate": 1.953939393939394e-05,
"loss": 0.861,
"step": 2550
},
{
"epoch": 0.36990930108483017,
"grad_norm": 6.977218151092529,
"learning_rate": 1.94866930171278e-05,
"loss": 0.8487,
"step": 2600
},
{
"epoch": 0.3770229414903077,
"grad_norm": 6.354854106903076,
"learning_rate": 1.943399209486166e-05,
"loss": 0.8122,
"step": 2650
},
{
"epoch": 0.38413658189578515,
"grad_norm": 6.679897308349609,
"learning_rate": 1.9381291172595522e-05,
"loss": 0.8239,
"step": 2700
},
{
"epoch": 0.39125022230126266,
"grad_norm": 7.958705902099609,
"learning_rate": 1.9328590250329382e-05,
"loss": 0.7808,
"step": 2750
},
{
"epoch": 0.3983638627067402,
"grad_norm": 6.450643062591553,
"learning_rate": 1.9275889328063243e-05,
"loss": 0.8359,
"step": 2800
},
{
"epoch": 0.4054775031122177,
"grad_norm": 5.986859321594238,
"learning_rate": 1.9223188405797103e-05,
"loss": 0.8278,
"step": 2850
},
{
"epoch": 0.41259114351769516,
"grad_norm": 8.40491771697998,
"learning_rate": 1.9170487483530964e-05,
"loss": 0.7932,
"step": 2900
},
{
"epoch": 0.4197047839231727,
"grad_norm": 7.491307735443115,
"learning_rate": 1.9117786561264824e-05,
"loss": 0.7932,
"step": 2950
},
{
"epoch": 0.4268184243286502,
"grad_norm": 5.475683689117432,
"learning_rate": 1.9065085638998684e-05,
"loss": 0.8292,
"step": 3000
},
{
"epoch": 0.4339320647341277,
"grad_norm": 6.489156723022461,
"learning_rate": 1.9012384716732545e-05,
"loss": 0.7585,
"step": 3050
},
{
"epoch": 0.44104570513960517,
"grad_norm": 6.417465686798096,
"learning_rate": 1.8959683794466402e-05,
"loss": 0.7628,
"step": 3100
},
{
"epoch": 0.4481593455450827,
"grad_norm": 4.215211868286133,
"learning_rate": 1.8906982872200266e-05,
"loss": 0.7418,
"step": 3150
},
{
"epoch": 0.4552729859505602,
"grad_norm": 5.503742218017578,
"learning_rate": 1.8854281949934126e-05,
"loss": 0.8346,
"step": 3200
},
{
"epoch": 0.4623866263560377,
"grad_norm": 5.3524651527404785,
"learning_rate": 1.8801581027667987e-05,
"loss": 0.8285,
"step": 3250
},
{
"epoch": 0.4695002667615152,
"grad_norm": 5.5221781730651855,
"learning_rate": 1.8748880105401844e-05,
"loss": 0.7363,
"step": 3300
},
{
"epoch": 0.4766139071669927,
"grad_norm": 7.9843268394470215,
"learning_rate": 1.8696179183135704e-05,
"loss": 0.7964,
"step": 3350
},
{
"epoch": 0.4837275475724702,
"grad_norm": 6.0546417236328125,
"learning_rate": 1.8643478260869568e-05,
"loss": 0.7696,
"step": 3400
},
{
"epoch": 0.49084118797794773,
"grad_norm": 5.521228313446045,
"learning_rate": 1.859077733860343e-05,
"loss": 0.7309,
"step": 3450
},
{
"epoch": 0.4979548283834252,
"grad_norm": 9.225383758544922,
"learning_rate": 1.853807641633729e-05,
"loss": 0.7869,
"step": 3500
},
{
"epoch": 0.5050684687889028,
"grad_norm": 4.0471601486206055,
"learning_rate": 1.8485375494071146e-05,
"loss": 0.7323,
"step": 3550
},
{
"epoch": 0.5121821091943802,
"grad_norm": 4.831805229187012,
"learning_rate": 1.8432674571805006e-05,
"loss": 0.7795,
"step": 3600
},
{
"epoch": 0.5192957495998577,
"grad_norm": 5.822267055511475,
"learning_rate": 1.837997364953887e-05,
"loss": 0.712,
"step": 3650
},
{
"epoch": 0.5264093900053353,
"grad_norm": 4.78722620010376,
"learning_rate": 1.832727272727273e-05,
"loss": 0.7408,
"step": 3700
},
{
"epoch": 0.5335230304108127,
"grad_norm": 5.776618480682373,
"learning_rate": 1.8274571805006588e-05,
"loss": 0.6803,
"step": 3750
},
{
"epoch": 0.5406366708162902,
"grad_norm": 11.762743949890137,
"learning_rate": 1.8221870882740448e-05,
"loss": 0.7343,
"step": 3800
},
{
"epoch": 0.5477503112217678,
"grad_norm": 3.884990692138672,
"learning_rate": 1.816916996047431e-05,
"loss": 0.7551,
"step": 3850
},
{
"epoch": 0.5548639516272452,
"grad_norm": 9.219622611999512,
"learning_rate": 1.8116469038208173e-05,
"loss": 0.764,
"step": 3900
},
{
"epoch": 0.5619775920327228,
"grad_norm": 7.808120250701904,
"learning_rate": 1.806376811594203e-05,
"loss": 0.745,
"step": 3950
},
{
"epoch": 0.5690912324382003,
"grad_norm": 7.188344955444336,
"learning_rate": 1.801106719367589e-05,
"loss": 0.7642,
"step": 4000
},
{
"epoch": 0.5762048728436777,
"grad_norm": 4.781955242156982,
"learning_rate": 1.795836627140975e-05,
"loss": 0.7089,
"step": 4050
},
{
"epoch": 0.5833185132491553,
"grad_norm": 4.489506721496582,
"learning_rate": 1.790566534914361e-05,
"loss": 0.7512,
"step": 4100
},
{
"epoch": 0.5904321536546328,
"grad_norm": 5.0979905128479,
"learning_rate": 1.785296442687747e-05,
"loss": 0.7753,
"step": 4150
},
{
"epoch": 0.5975457940601102,
"grad_norm": 5.302441596984863,
"learning_rate": 1.7800263504611332e-05,
"loss": 0.7412,
"step": 4200
},
{
"epoch": 0.6046594344655878,
"grad_norm": 4.942580223083496,
"learning_rate": 1.7747562582345192e-05,
"loss": 0.7474,
"step": 4250
},
{
"epoch": 0.6117730748710652,
"grad_norm": 4.685510158538818,
"learning_rate": 1.7694861660079053e-05,
"loss": 0.7173,
"step": 4300
},
{
"epoch": 0.6188867152765428,
"grad_norm": 4.8454999923706055,
"learning_rate": 1.7642160737812913e-05,
"loss": 0.7207,
"step": 4350
},
{
"epoch": 0.6260003556820203,
"grad_norm": 7.190736293792725,
"learning_rate": 1.7589459815546774e-05,
"loss": 0.754,
"step": 4400
},
{
"epoch": 0.6331139960874977,
"grad_norm": 7.0048675537109375,
"learning_rate": 1.7536758893280634e-05,
"loss": 0.7282,
"step": 4450
},
{
"epoch": 0.6402276364929753,
"grad_norm": 5.777256011962891,
"learning_rate": 1.7484057971014495e-05,
"loss": 0.6597,
"step": 4500
},
{
"epoch": 0.6473412768984528,
"grad_norm": 8.8043851852417,
"learning_rate": 1.7431357048748355e-05,
"loss": 0.6672,
"step": 4550
},
{
"epoch": 0.6544549173039302,
"grad_norm": 7.159558296203613,
"learning_rate": 1.7378656126482215e-05,
"loss": 0.7266,
"step": 4600
},
{
"epoch": 0.6615685577094078,
"grad_norm": 5.110541820526123,
"learning_rate": 1.7325955204216076e-05,
"loss": 0.6399,
"step": 4650
},
{
"epoch": 0.6686821981148853,
"grad_norm": 3.4727582931518555,
"learning_rate": 1.7273254281949936e-05,
"loss": 0.68,
"step": 4700
},
{
"epoch": 0.6757958385203628,
"grad_norm": 5.788763999938965,
"learning_rate": 1.7220553359683797e-05,
"loss": 0.7349,
"step": 4750
},
{
"epoch": 0.6829094789258403,
"grad_norm": 4.407217979431152,
"learning_rate": 1.7167852437417657e-05,
"loss": 0.695,
"step": 4800
},
{
"epoch": 0.6900231193313178,
"grad_norm": 4.685910224914551,
"learning_rate": 1.7115151515151514e-05,
"loss": 0.7816,
"step": 4850
},
{
"epoch": 0.6971367597367953,
"grad_norm": 2.9753873348236084,
"learning_rate": 1.7062450592885378e-05,
"loss": 0.6595,
"step": 4900
},
{
"epoch": 0.7042504001422728,
"grad_norm": 3.8326473236083984,
"learning_rate": 1.700974967061924e-05,
"loss": 0.6875,
"step": 4950
},
{
"epoch": 0.7113640405477503,
"grad_norm": 5.974637985229492,
"learning_rate": 1.69570487483531e-05,
"loss": 0.7414,
"step": 5000
},
{
"epoch": 0.7184776809532278,
"grad_norm": 3.4977550506591797,
"learning_rate": 1.6904347826086956e-05,
"loss": 0.6832,
"step": 5050
},
{
"epoch": 0.7255913213587053,
"grad_norm": 5.954402446746826,
"learning_rate": 1.6851646903820817e-05,
"loss": 0.7257,
"step": 5100
},
{
"epoch": 0.7327049617641829,
"grad_norm": 4.272510528564453,
"learning_rate": 1.679894598155468e-05,
"loss": 0.6808,
"step": 5150
},
{
"epoch": 0.7398186021696603,
"grad_norm": 4.155337810516357,
"learning_rate": 1.674624505928854e-05,
"loss": 0.7062,
"step": 5200
},
{
"epoch": 0.7469322425751378,
"grad_norm": 4.898423671722412,
"learning_rate": 1.6693544137022398e-05,
"loss": 0.6549,
"step": 5250
},
{
"epoch": 0.7540458829806154,
"grad_norm": 6.021943092346191,
"learning_rate": 1.664084321475626e-05,
"loss": 0.6886,
"step": 5300
},
{
"epoch": 0.7611595233860928,
"grad_norm": 4.547012805938721,
"learning_rate": 1.658814229249012e-05,
"loss": 0.6653,
"step": 5350
},
{
"epoch": 0.7682731637915703,
"grad_norm": 3.5163490772247314,
"learning_rate": 1.6535441370223983e-05,
"loss": 0.6549,
"step": 5400
},
{
"epoch": 0.7753868041970479,
"grad_norm": 4.255113124847412,
"learning_rate": 1.648274044795784e-05,
"loss": 0.6473,
"step": 5450
},
{
"epoch": 0.7825004446025253,
"grad_norm": 8.696562767028809,
"learning_rate": 1.64300395256917e-05,
"loss": 0.67,
"step": 5500
},
{
"epoch": 0.7896140850080029,
"grad_norm": 3.098172903060913,
"learning_rate": 1.637733860342556e-05,
"loss": 0.6626,
"step": 5550
},
{
"epoch": 0.7967277254134804,
"grad_norm": 5.950908660888672,
"learning_rate": 1.632463768115942e-05,
"loss": 0.7476,
"step": 5600
},
{
"epoch": 0.8038413658189578,
"grad_norm": 5.9870686531066895,
"learning_rate": 1.627193675889328e-05,
"loss": 0.7006,
"step": 5650
},
{
"epoch": 0.8109550062244354,
"grad_norm": 8.210447311401367,
"learning_rate": 1.6219235836627142e-05,
"loss": 0.6104,
"step": 5700
},
{
"epoch": 0.8180686466299129,
"grad_norm": 2.8877291679382324,
"learning_rate": 1.6166534914361002e-05,
"loss": 0.6183,
"step": 5750
},
{
"epoch": 0.8251822870353903,
"grad_norm": 3.5387609004974365,
"learning_rate": 1.6113833992094863e-05,
"loss": 0.6816,
"step": 5800
},
{
"epoch": 0.8322959274408679,
"grad_norm": 3.3335936069488525,
"learning_rate": 1.6061133069828723e-05,
"loss": 0.6948,
"step": 5850
},
{
"epoch": 0.8394095678463454,
"grad_norm": 3.515953302383423,
"learning_rate": 1.6008432147562584e-05,
"loss": 0.6664,
"step": 5900
},
{
"epoch": 0.8465232082518229,
"grad_norm": 2.2991325855255127,
"learning_rate": 1.5955731225296444e-05,
"loss": 0.6363,
"step": 5950
},
{
"epoch": 0.8536368486573004,
"grad_norm": 5.262946128845215,
"learning_rate": 1.5903030303030305e-05,
"loss": 0.7375,
"step": 6000
},
{
"epoch": 0.8607504890627778,
"grad_norm": 4.494121074676514,
"learning_rate": 1.5850329380764165e-05,
"loss": 0.6818,
"step": 6050
},
{
"epoch": 0.8678641294682554,
"grad_norm": 4.698576927185059,
"learning_rate": 1.5797628458498026e-05,
"loss": 0.7246,
"step": 6100
},
{
"epoch": 0.8749777698737329,
"grad_norm": 4.571809768676758,
"learning_rate": 1.5744927536231883e-05,
"loss": 0.6112,
"step": 6150
},
{
"epoch": 0.8820914102792103,
"grad_norm": 7.1781110763549805,
"learning_rate": 1.5692226613965746e-05,
"loss": 0.6593,
"step": 6200
},
{
"epoch": 0.8892050506846879,
"grad_norm": 5.867689609527588,
"learning_rate": 1.5639525691699607e-05,
"loss": 0.6491,
"step": 6250
},
{
"epoch": 0.8963186910901654,
"grad_norm": 5.427711009979248,
"learning_rate": 1.5586824769433467e-05,
"loss": 0.6419,
"step": 6300
},
{
"epoch": 0.903432331495643,
"grad_norm": 5.6072869300842285,
"learning_rate": 1.5534123847167328e-05,
"loss": 0.6443,
"step": 6350
},
{
"epoch": 0.9105459719011204,
"grad_norm": 4.355414867401123,
"learning_rate": 1.5481422924901185e-05,
"loss": 0.7021,
"step": 6400
},
{
"epoch": 0.9176596123065979,
"grad_norm": 4.039976596832275,
"learning_rate": 1.542872200263505e-05,
"loss": 0.7351,
"step": 6450
},
{
"epoch": 0.9247732527120754,
"grad_norm": 3.736238956451416,
"learning_rate": 1.537602108036891e-05,
"loss": 0.6629,
"step": 6500
},
{
"epoch": 0.9318868931175529,
"grad_norm": 3.91310453414917,
"learning_rate": 1.532332015810277e-05,
"loss": 0.6856,
"step": 6550
},
{
"epoch": 0.9390005335230304,
"grad_norm": 5.676443576812744,
"learning_rate": 1.5270619235836627e-05,
"loss": 0.6344,
"step": 6600
},
{
"epoch": 0.9461141739285079,
"grad_norm": 3.6127190589904785,
"learning_rate": 1.5217918313570487e-05,
"loss": 0.6845,
"step": 6650
},
{
"epoch": 0.9532278143339854,
"grad_norm": 4.957319259643555,
"learning_rate": 1.516521739130435e-05,
"loss": 0.7074,
"step": 6700
},
{
"epoch": 0.960341454739463,
"grad_norm": 3.351602554321289,
"learning_rate": 1.511251646903821e-05,
"loss": 0.6722,
"step": 6750
},
{
"epoch": 0.9674550951449404,
"grad_norm": 5.316050052642822,
"learning_rate": 1.505981554677207e-05,
"loss": 0.5993,
"step": 6800
},
{
"epoch": 0.9745687355504179,
"grad_norm": 3.5226705074310303,
"learning_rate": 1.5007114624505929e-05,
"loss": 0.6663,
"step": 6850
},
{
"epoch": 0.9816823759558955,
"grad_norm": 4.123106479644775,
"learning_rate": 1.495441370223979e-05,
"loss": 0.6865,
"step": 6900
},
{
"epoch": 0.9887960163613729,
"grad_norm": 3.942507743835449,
"learning_rate": 1.4901712779973652e-05,
"loss": 0.6535,
"step": 6950
},
{
"epoch": 0.9959096567668504,
"grad_norm": 4.765227317810059,
"learning_rate": 1.4849011857707512e-05,
"loss": 0.5895,
"step": 7000
},
{
"epoch": 1.0,
"eval_loss": 0.6490142345428467,
"eval_runtime": 321.1223,
"eval_samples_per_second": 16.414,
"eval_steps_per_second": 2.052,
"step": 7029
},
{
"epoch": 1.0029877289703006,
"grad_norm": 3.913954496383667,
"learning_rate": 1.479631093544137e-05,
"loss": 0.5676,
"step": 7050
},
{
"epoch": 1.010101369375778,
"grad_norm": 2.70705246925354,
"learning_rate": 1.4743610013175231e-05,
"loss": 0.5268,
"step": 7100
},
{
"epoch": 1.0172150097812556,
"grad_norm": 5.550088882446289,
"learning_rate": 1.4690909090909092e-05,
"loss": 0.507,
"step": 7150
},
{
"epoch": 1.024328650186733,
"grad_norm": 6.876676082611084,
"learning_rate": 1.4638208168642954e-05,
"loss": 0.5285,
"step": 7200
},
{
"epoch": 1.0314422905922105,
"grad_norm": 2.911832809448242,
"learning_rate": 1.4585507246376813e-05,
"loss": 0.5357,
"step": 7250
},
{
"epoch": 1.0385559309976882,
"grad_norm": 5.395831108093262,
"learning_rate": 1.4532806324110673e-05,
"loss": 0.5442,
"step": 7300
},
{
"epoch": 1.0456695714031656,
"grad_norm": 8.814352035522461,
"learning_rate": 1.4480105401844533e-05,
"loss": 0.5394,
"step": 7350
},
{
"epoch": 1.052783211808643,
"grad_norm": 2.947053909301758,
"learning_rate": 1.4427404479578392e-05,
"loss": 0.5549,
"step": 7400
},
{
"epoch": 1.0598968522141206,
"grad_norm": 5.564446926116943,
"learning_rate": 1.4374703557312254e-05,
"loss": 0.5328,
"step": 7450
},
{
"epoch": 1.067010492619598,
"grad_norm": 4.452253341674805,
"learning_rate": 1.4322002635046115e-05,
"loss": 0.5197,
"step": 7500
},
{
"epoch": 1.0741241330250757,
"grad_norm": 3.506099224090576,
"learning_rate": 1.4269301712779975e-05,
"loss": 0.5509,
"step": 7550
},
{
"epoch": 1.0812377734305532,
"grad_norm": 4.477612018585205,
"learning_rate": 1.4216600790513834e-05,
"loss": 0.5005,
"step": 7600
},
{
"epoch": 1.0883514138360306,
"grad_norm": 4.912115097045898,
"learning_rate": 1.4163899868247694e-05,
"loss": 0.5351,
"step": 7650
},
{
"epoch": 1.095465054241508,
"grad_norm": 4.499685764312744,
"learning_rate": 1.4111198945981557e-05,
"loss": 0.5218,
"step": 7700
},
{
"epoch": 1.1025786946469855,
"grad_norm": 4.6450934410095215,
"learning_rate": 1.4058498023715417e-05,
"loss": 0.5195,
"step": 7750
},
{
"epoch": 1.109692335052463,
"grad_norm": 4.883353233337402,
"learning_rate": 1.4005797101449276e-05,
"loss": 0.5464,
"step": 7800
},
{
"epoch": 1.1168059754579407,
"grad_norm": 3.312135934829712,
"learning_rate": 1.3953096179183136e-05,
"loss": 0.4766,
"step": 7850
},
{
"epoch": 1.1239196158634182,
"grad_norm": 2.180053234100342,
"learning_rate": 1.3900395256916997e-05,
"loss": 0.5742,
"step": 7900
},
{
"epoch": 1.1310332562688956,
"grad_norm": 4.708530426025391,
"learning_rate": 1.3847694334650859e-05,
"loss": 0.5499,
"step": 7950
},
{
"epoch": 1.138146896674373,
"grad_norm": 3.369694709777832,
"learning_rate": 1.379499341238472e-05,
"loss": 0.4839,
"step": 8000
},
{
"epoch": 1.1452605370798505,
"grad_norm": 3.1507697105407715,
"learning_rate": 1.3742292490118578e-05,
"loss": 0.5118,
"step": 8050
},
{
"epoch": 1.1523741774853282,
"grad_norm": 3.773084878921509,
"learning_rate": 1.3689591567852438e-05,
"loss": 0.5406,
"step": 8100
},
{
"epoch": 1.1594878178908057,
"grad_norm": 3.446446180343628,
"learning_rate": 1.3636890645586299e-05,
"loss": 0.5175,
"step": 8150
},
{
"epoch": 1.1666014582962831,
"grad_norm": 5.0827155113220215,
"learning_rate": 1.3584189723320161e-05,
"loss": 0.4963,
"step": 8200
},
{
"epoch": 1.1737150987017606,
"grad_norm": 5.854413032531738,
"learning_rate": 1.353148880105402e-05,
"loss": 0.6038,
"step": 8250
},
{
"epoch": 1.180828739107238,
"grad_norm": 4.135270595550537,
"learning_rate": 1.347878787878788e-05,
"loss": 0.5876,
"step": 8300
},
{
"epoch": 1.1879423795127155,
"grad_norm": 4.9584197998046875,
"learning_rate": 1.342608695652174e-05,
"loss": 0.5272,
"step": 8350
},
{
"epoch": 1.1950560199181932,
"grad_norm": 2.152341842651367,
"learning_rate": 1.33733860342556e-05,
"loss": 0.5146,
"step": 8400
},
{
"epoch": 1.2021696603236707,
"grad_norm": 3.346052885055542,
"learning_rate": 1.3320685111989462e-05,
"loss": 0.4661,
"step": 8450
},
{
"epoch": 1.2092833007291481,
"grad_norm": 4.768698692321777,
"learning_rate": 1.3267984189723322e-05,
"loss": 0.5354,
"step": 8500
},
{
"epoch": 1.2163969411346256,
"grad_norm": 5.182507514953613,
"learning_rate": 1.3215283267457183e-05,
"loss": 0.4845,
"step": 8550
},
{
"epoch": 1.223510581540103,
"grad_norm": 4.584784030914307,
"learning_rate": 1.3162582345191041e-05,
"loss": 0.5261,
"step": 8600
},
{
"epoch": 1.2306242219455807,
"grad_norm": 3.520695209503174,
"learning_rate": 1.3110935441370225e-05,
"loss": 0.5735,
"step": 8650
},
{
"epoch": 1.2377378623510582,
"grad_norm": 5.633211612701416,
"learning_rate": 1.3058234519104085e-05,
"loss": 0.516,
"step": 8700
},
{
"epoch": 1.2448515027565357,
"grad_norm": 5.559876441955566,
"learning_rate": 1.3005533596837944e-05,
"loss": 0.5597,
"step": 8750
},
{
"epoch": 1.2519651431620131,
"grad_norm": 4.474053382873535,
"learning_rate": 1.2952832674571806e-05,
"loss": 0.5171,
"step": 8800
},
{
"epoch": 1.2590787835674906,
"grad_norm": 3.866060733795166,
"learning_rate": 1.2900131752305667e-05,
"loss": 0.5057,
"step": 8850
},
{
"epoch": 1.266192423972968,
"grad_norm": 5.632756233215332,
"learning_rate": 1.2847430830039527e-05,
"loss": 0.5514,
"step": 8900
},
{
"epoch": 1.2733060643784457,
"grad_norm": 4.323223114013672,
"learning_rate": 1.2794729907773386e-05,
"loss": 0.5052,
"step": 8950
},
{
"epoch": 1.2804197047839232,
"grad_norm": 3.6960432529449463,
"learning_rate": 1.2742028985507246e-05,
"loss": 0.5295,
"step": 9000
},
{
"epoch": 1.2875333451894007,
"grad_norm": 3.777792453765869,
"learning_rate": 1.2689328063241108e-05,
"loss": 0.49,
"step": 9050
},
{
"epoch": 1.2946469855948781,
"grad_norm": 2.103578805923462,
"learning_rate": 1.2636627140974969e-05,
"loss": 0.5437,
"step": 9100
},
{
"epoch": 1.3017606260003558,
"grad_norm": 5.730409622192383,
"learning_rate": 1.258392621870883e-05,
"loss": 0.532,
"step": 9150
},
{
"epoch": 1.3088742664058333,
"grad_norm": 6.736247539520264,
"learning_rate": 1.2532279314888011e-05,
"loss": 0.554,
"step": 9200
},
{
"epoch": 1.3159879068113107,
"grad_norm": 2.0751116275787354,
"learning_rate": 1.2479578392621872e-05,
"loss": 0.5128,
"step": 9250
},
{
"epoch": 1.3231015472167882,
"grad_norm": 2.6663055419921875,
"learning_rate": 1.2426877470355732e-05,
"loss": 0.5251,
"step": 9300
},
{
"epoch": 1.3302151876222656,
"grad_norm": 4.778317928314209,
"learning_rate": 1.2374176548089594e-05,
"loss": 0.5219,
"step": 9350
},
{
"epoch": 1.337328828027743,
"grad_norm": 5.950105667114258,
"learning_rate": 1.2321475625823453e-05,
"loss": 0.5192,
"step": 9400
},
{
"epoch": 1.3444424684332206,
"grad_norm": 3.5832479000091553,
"learning_rate": 1.2268774703557313e-05,
"loss": 0.5185,
"step": 9450
},
{
"epoch": 1.3515561088386983,
"grad_norm": 2.7390801906585693,
"learning_rate": 1.2216073781291174e-05,
"loss": 0.5375,
"step": 9500
},
{
"epoch": 1.3586697492441757,
"grad_norm": 2.688469648361206,
"learning_rate": 1.2163372859025033e-05,
"loss": 0.5352,
"step": 9550
},
{
"epoch": 1.3657833896496532,
"grad_norm": 4.956066608428955,
"learning_rate": 1.2110671936758893e-05,
"loss": 0.5341,
"step": 9600
},
{
"epoch": 1.3728970300551306,
"grad_norm": 5.483800411224365,
"learning_rate": 1.2057971014492755e-05,
"loss": 0.5644,
"step": 9650
},
{
"epoch": 1.3800106704606083,
"grad_norm": 4.272674083709717,
"learning_rate": 1.2005270092226616e-05,
"loss": 0.5044,
"step": 9700
},
{
"epoch": 1.3871243108660858,
"grad_norm": 2.5622191429138184,
"learning_rate": 1.1952569169960474e-05,
"loss": 0.5171,
"step": 9750
},
{
"epoch": 1.3942379512715632,
"grad_norm": 7.005847454071045,
"learning_rate": 1.1899868247694335e-05,
"loss": 0.5156,
"step": 9800
},
{
"epoch": 1.4013515916770407,
"grad_norm": 4.284529685974121,
"learning_rate": 1.1847167325428195e-05,
"loss": 0.5833,
"step": 9850
},
{
"epoch": 1.4084652320825182,
"grad_norm": 3.930690288543701,
"learning_rate": 1.1794466403162057e-05,
"loss": 0.5219,
"step": 9900
},
{
"epoch": 1.4155788724879956,
"grad_norm": 4.899880886077881,
"learning_rate": 1.1741765480895916e-05,
"loss": 0.5605,
"step": 9950
},
{
"epoch": 1.4226925128934733,
"grad_norm": 5.612455368041992,
"learning_rate": 1.1689064558629777e-05,
"loss": 0.5408,
"step": 10000
},
{
"epoch": 1.4298061532989508,
"grad_norm": 2.502410411834717,
"learning_rate": 1.1636363636363637e-05,
"loss": 0.5052,
"step": 10050
},
{
"epoch": 1.4369197937044282,
"grad_norm": 4.965051174163818,
"learning_rate": 1.1583662714097496e-05,
"loss": 0.5036,
"step": 10100
},
{
"epoch": 1.4440334341099057,
"grad_norm": 4.5783371925354,
"learning_rate": 1.1530961791831358e-05,
"loss": 0.5474,
"step": 10150
},
{
"epoch": 1.4511470745153834,
"grad_norm": 5.443408012390137,
"learning_rate": 1.1478260869565218e-05,
"loss": 0.5018,
"step": 10200
},
{
"epoch": 1.4582607149208608,
"grad_norm": 5.504450798034668,
"learning_rate": 1.1425559947299079e-05,
"loss": 0.4877,
"step": 10250
},
{
"epoch": 1.4653743553263383,
"grad_norm": 3.819101572036743,
"learning_rate": 1.137285902503294e-05,
"loss": 0.526,
"step": 10300
},
{
"epoch": 1.4724879957318158,
"grad_norm": 5.407709121704102,
"learning_rate": 1.1320158102766798e-05,
"loss": 0.4949,
"step": 10350
},
{
"epoch": 1.4796016361372932,
"grad_norm": 4.124641418457031,
"learning_rate": 1.126745718050066e-05,
"loss": 0.5923,
"step": 10400
},
{
"epoch": 1.4867152765427707,
"grad_norm": 4.266039848327637,
"learning_rate": 1.121475625823452e-05,
"loss": 0.5545,
"step": 10450
},
{
"epoch": 1.4938289169482482,
"grad_norm": 4.402031421661377,
"learning_rate": 1.1162055335968381e-05,
"loss": 0.505,
"step": 10500
},
{
"epoch": 1.5009425573537256,
"grad_norm": 6.378383159637451,
"learning_rate": 1.110935441370224e-05,
"loss": 0.5015,
"step": 10550
},
{
"epoch": 1.5080561977592033,
"grad_norm": 4.023849964141846,
"learning_rate": 1.10566534914361e-05,
"loss": 0.5371,
"step": 10600
},
{
"epoch": 1.5151698381646808,
"grad_norm": 4.202026844024658,
"learning_rate": 1.1003952569169962e-05,
"loss": 0.5527,
"step": 10650
},
{
"epoch": 1.5222834785701584,
"grad_norm": 4.9058685302734375,
"learning_rate": 1.0951251646903823e-05,
"loss": 0.5193,
"step": 10700
},
{
"epoch": 1.529397118975636,
"grad_norm": 5.402507781982422,
"learning_rate": 1.0898550724637682e-05,
"loss": 0.542,
"step": 10750
},
{
"epoch": 1.5365107593811134,
"grad_norm": 4.229278087615967,
"learning_rate": 1.0845849802371542e-05,
"loss": 0.5284,
"step": 10800
},
{
"epoch": 1.5436243997865908,
"grad_norm": 3.6707725524902344,
"learning_rate": 1.0793148880105403e-05,
"loss": 0.561,
"step": 10850
},
{
"epoch": 1.5507380401920683,
"grad_norm": 3.6205899715423584,
"learning_rate": 1.0740447957839265e-05,
"loss": 0.4981,
"step": 10900
},
{
"epoch": 1.5578516805975458,
"grad_norm": 4.793003559112549,
"learning_rate": 1.0687747035573123e-05,
"loss": 0.5062,
"step": 10950
},
{
"epoch": 1.5649653210030232,
"grad_norm": 3.2560999393463135,
"learning_rate": 1.0635046113306984e-05,
"loss": 0.5305,
"step": 11000
},
{
"epoch": 1.5720789614085007,
"grad_norm": 4.039142608642578,
"learning_rate": 1.0582345191040844e-05,
"loss": 0.5412,
"step": 11050
},
{
"epoch": 1.5791926018139784,
"grad_norm": 4.209068298339844,
"learning_rate": 1.0529644268774703e-05,
"loss": 0.5014,
"step": 11100
},
{
"epoch": 1.5863062422194558,
"grad_norm": 3.390684127807617,
"learning_rate": 1.0476943346508565e-05,
"loss": 0.5605,
"step": 11150
},
{
"epoch": 1.5934198826249333,
"grad_norm": 3.7350914478302,
"learning_rate": 1.0424242424242426e-05,
"loss": 0.5021,
"step": 11200
},
{
"epoch": 1.600533523030411,
"grad_norm": 3.5504932403564453,
"learning_rate": 1.0371541501976286e-05,
"loss": 0.4967,
"step": 11250
},
{
"epoch": 1.6076471634358884,
"grad_norm": 3.8655166625976562,
"learning_rate": 1.0318840579710145e-05,
"loss": 0.4857,
"step": 11300
},
{
"epoch": 1.6147608038413659,
"grad_norm": 6.647157669067383,
"learning_rate": 1.0266139657444005e-05,
"loss": 0.5033,
"step": 11350
},
{
"epoch": 1.6218744442468433,
"grad_norm": 5.384164810180664,
"learning_rate": 1.0213438735177868e-05,
"loss": 0.5486,
"step": 11400
},
{
"epoch": 1.6289880846523208,
"grad_norm": 4.699065685272217,
"learning_rate": 1.016179183135705e-05,
"loss": 0.5565,
"step": 11450
},
{
"epoch": 1.6361017250577983,
"grad_norm": 4.786224365234375,
"learning_rate": 1.010909090909091e-05,
"loss": 0.5155,
"step": 11500
},
{
"epoch": 1.6432153654632757,
"grad_norm": 4.30881929397583,
"learning_rate": 1.005638998682477e-05,
"loss": 0.5303,
"step": 11550
},
{
"epoch": 1.6503290058687532,
"grad_norm": 4.938950538635254,
"learning_rate": 1.000368906455863e-05,
"loss": 0.4891,
"step": 11600
},
{
"epoch": 1.6574426462742309,
"grad_norm": 4.875250339508057,
"learning_rate": 9.950988142292491e-06,
"loss": 0.531,
"step": 11650
},
{
"epoch": 1.6645562866797083,
"grad_norm": 5.96549654006958,
"learning_rate": 9.898287220026352e-06,
"loss": 0.5092,
"step": 11700
},
{
"epoch": 1.6716699270851858,
"grad_norm": 5.401658535003662,
"learning_rate": 9.845586297760212e-06,
"loss": 0.5018,
"step": 11750
},
{
"epoch": 1.6787835674906635,
"grad_norm": 4.49137020111084,
"learning_rate": 9.792885375494072e-06,
"loss": 0.4877,
"step": 11800
},
{
"epoch": 1.685897207896141,
"grad_norm": 3.0070083141326904,
"learning_rate": 9.740184453227933e-06,
"loss": 0.4884,
"step": 11850
},
{
"epoch": 1.6930108483016184,
"grad_norm": 3.979234218597412,
"learning_rate": 9.687483530961793e-06,
"loss": 0.5013,
"step": 11900
},
{
"epoch": 1.7001244887070959,
"grad_norm": 3.876878499984741,
"learning_rate": 9.634782608695654e-06,
"loss": 0.5322,
"step": 11950
},
{
"epoch": 1.7072381291125733,
"grad_norm": 5.199528694152832,
"learning_rate": 9.582081686429513e-06,
"loss": 0.5052,
"step": 12000
},
{
"epoch": 1.7143517695180508,
"grad_norm": 3.97282075881958,
"learning_rate": 9.529380764163375e-06,
"loss": 0.4887,
"step": 12050
},
{
"epoch": 1.7214654099235283,
"grad_norm": 4.776126384735107,
"learning_rate": 9.476679841897233e-06,
"loss": 0.5384,
"step": 12100
},
{
"epoch": 1.7285790503290057,
"grad_norm": 6.539297580718994,
"learning_rate": 9.423978919631094e-06,
"loss": 0.5283,
"step": 12150
},
{
"epoch": 1.7356926907344834,
"grad_norm": 5.380597114562988,
"learning_rate": 9.371277997364954e-06,
"loss": 0.5481,
"step": 12200
},
{
"epoch": 1.7428063311399609,
"grad_norm": 4.773643493652344,
"learning_rate": 9.318577075098815e-06,
"loss": 0.4815,
"step": 12250
},
{
"epoch": 1.7499199715454383,
"grad_norm": 1.9399245977401733,
"learning_rate": 9.265876152832675e-06,
"loss": 0.515,
"step": 12300
},
{
"epoch": 1.757033611950916,
"grad_norm": 3.586733102798462,
"learning_rate": 9.213175230566536e-06,
"loss": 0.4856,
"step": 12350
},
{
"epoch": 1.7641472523563935,
"grad_norm": 3.9114813804626465,
"learning_rate": 9.160474308300396e-06,
"loss": 0.5114,
"step": 12400
},
{
"epoch": 1.771260892761871,
"grad_norm": 3.8798410892486572,
"learning_rate": 9.107773386034257e-06,
"loss": 0.4939,
"step": 12450
},
{
"epoch": 1.7783745331673484,
"grad_norm": 2.8337159156799316,
"learning_rate": 9.055072463768117e-06,
"loss": 0.4807,
"step": 12500
},
{
"epoch": 1.7854881735728259,
"grad_norm": 4.102101802825928,
"learning_rate": 9.002371541501978e-06,
"loss": 0.5632,
"step": 12550
},
{
"epoch": 1.7926018139783033,
"grad_norm": 2.918848991394043,
"learning_rate": 8.949670619235838e-06,
"loss": 0.5012,
"step": 12600
},
{
"epoch": 1.7997154543837808,
"grad_norm": 3.708796739578247,
"learning_rate": 8.896969696969697e-06,
"loss": 0.5159,
"step": 12650
},
{
"epoch": 1.8068290947892582,
"grad_norm": 4.967949867248535,
"learning_rate": 8.844268774703559e-06,
"loss": 0.5037,
"step": 12700
},
{
"epoch": 1.813942735194736,
"grad_norm": 2.7646732330322266,
"learning_rate": 8.791567852437418e-06,
"loss": 0.4842,
"step": 12750
},
{
"epoch": 1.8210563756002134,
"grad_norm": 3.667774200439453,
"learning_rate": 8.73886693017128e-06,
"loss": 0.4744,
"step": 12800
},
{
"epoch": 1.828170016005691,
"grad_norm": 5.304251670837402,
"learning_rate": 8.686166007905139e-06,
"loss": 0.521,
"step": 12850
},
{
"epoch": 1.8352836564111685,
"grad_norm": 4.663897514343262,
"learning_rate": 8.633465085638999e-06,
"loss": 0.5143,
"step": 12900
},
{
"epoch": 1.842397296816646,
"grad_norm": 6.192160129547119,
"learning_rate": 8.58076416337286e-06,
"loss": 0.5489,
"step": 12950
},
{
"epoch": 1.8495109372221235,
"grad_norm": 3.5136263370513916,
"learning_rate": 8.52806324110672e-06,
"loss": 0.5124,
"step": 13000
},
{
"epoch": 1.856624577627601,
"grad_norm": 3.0261971950531006,
"learning_rate": 8.47536231884058e-06,
"loss": 0.5295,
"step": 13050
},
{
"epoch": 1.8637382180330784,
"grad_norm": 3.9308557510375977,
"learning_rate": 8.42266139657444e-06,
"loss": 0.5209,
"step": 13100
},
{
"epoch": 1.8708518584385558,
"grad_norm": 4.964597225189209,
"learning_rate": 8.369960474308301e-06,
"loss": 0.5187,
"step": 13150
},
{
"epoch": 1.8779654988440333,
"grad_norm": 5.161786079406738,
"learning_rate": 8.317259552042162e-06,
"loss": 0.5207,
"step": 13200
},
{
"epoch": 1.885079139249511,
"grad_norm": 2.5828144550323486,
"learning_rate": 8.264558629776022e-06,
"loss": 0.4598,
"step": 13250
},
{
"epoch": 1.8921927796549884,
"grad_norm": 3.169773817062378,
"learning_rate": 8.211857707509883e-06,
"loss": 0.531,
"step": 13300
},
{
"epoch": 1.899306420060466,
"grad_norm": 3.6591272354125977,
"learning_rate": 8.159156785243743e-06,
"loss": 0.5226,
"step": 13350
},
{
"epoch": 1.9064200604659436,
"grad_norm": 2.735844850540161,
"learning_rate": 8.106455862977602e-06,
"loss": 0.5808,
"step": 13400
},
{
"epoch": 1.913533700871421,
"grad_norm": 5.649113655090332,
"learning_rate": 8.053754940711464e-06,
"loss": 0.5154,
"step": 13450
},
{
"epoch": 1.9206473412768985,
"grad_norm": 2.830268621444702,
"learning_rate": 8.001054018445323e-06,
"loss": 0.5017,
"step": 13500
},
{
"epoch": 1.927760981682376,
"grad_norm": 5.496303558349609,
"learning_rate": 7.948353096179183e-06,
"loss": 0.4824,
"step": 13550
},
{
"epoch": 1.9348746220878534,
"grad_norm": 3.1559128761291504,
"learning_rate": 7.895652173913044e-06,
"loss": 0.5465,
"step": 13600
},
{
"epoch": 1.941988262493331,
"grad_norm": 3.265587091445923,
"learning_rate": 7.842951251646904e-06,
"loss": 0.4673,
"step": 13650
},
{
"epoch": 1.9491019028988084,
"grad_norm": 6.156945705413818,
"learning_rate": 7.790250329380764e-06,
"loss": 0.5558,
"step": 13700
},
{
"epoch": 1.9562155433042858,
"grad_norm": 7.1588029861450195,
"learning_rate": 7.737549407114625e-06,
"loss": 0.5013,
"step": 13750
},
{
"epoch": 1.9633291837097635,
"grad_norm": 3.9235057830810547,
"learning_rate": 7.684848484848485e-06,
"loss": 0.5214,
"step": 13800
},
{
"epoch": 1.970442824115241,
"grad_norm": 3.9342610836029053,
"learning_rate": 7.632147562582346e-06,
"loss": 0.5147,
"step": 13850
},
{
"epoch": 1.9775564645207184,
"grad_norm": 2.9213314056396484,
"learning_rate": 7.579446640316206e-06,
"loss": 0.4756,
"step": 13900
},
{
"epoch": 1.984670104926196,
"grad_norm": 2.211418390274048,
"learning_rate": 7.526745718050067e-06,
"loss": 0.4862,
"step": 13950
},
{
"epoch": 1.9917837453316736,
"grad_norm": 3.4956717491149902,
"learning_rate": 7.474044795783927e-06,
"loss": 0.5235,
"step": 14000
},
{
"epoch": 1.998897385737151,
"grad_norm": 4.755037784576416,
"learning_rate": 7.421343873517787e-06,
"loss": 0.5176,
"step": 14050
},
{
"epoch": 2.0,
"eval_loss": 0.6273267865180969,
"eval_runtime": 321.2086,
"eval_samples_per_second": 16.41,
"eval_steps_per_second": 2.052,
"step": 14058
}
],
"logging_steps": 50,
"max_steps": 21084,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 7.608571407630336e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}