japdongsani / trainer_state.json
SEOje's picture
Upload 14 files
37fb9d9 verified
Invalid JSON: Unexpected token 'I', ..."ad_norm": Infinity, "... is not valid JSON
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.998954547717429,
"eval_steps": 500,
"global_step": 7530,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0019913376810872705,
"grad_norm": 4.0654296875,
"learning_rate": 0.0002999996736294919,
"loss": 3.4521,
"step": 5
},
{
"epoch": 0.003982675362174541,
"grad_norm": 3.3652706146240234,
"learning_rate": 0.00029999894256041275,
"loss": 2.5577,
"step": 10
},
{
"epoch": 0.005974013043261811,
"grad_norm": 6.442378997802734,
"learning_rate": 0.0002999977937399738,
"loss": 2.6864,
"step": 15
},
{
"epoch": 0.007965350724349082,
"grad_norm": 3.433056592941284,
"learning_rate": 0.0002999957702565602,
"loss": 2.6988,
"step": 20
},
{
"epoch": 0.009956688405436351,
"grad_norm": 3.1151390075683594,
"learning_rate": 0.0002999930940505367,
"loss": 2.6097,
"step": 25
},
{
"epoch": 0.011948026086523622,
"grad_norm": 3.0310587882995605,
"learning_rate": 0.00029998976513354893,
"loss": 2.5627,
"step": 30
},
{
"epoch": 0.013939363767610893,
"grad_norm": 2.9382193088531494,
"learning_rate": 0.00029998578352008324,
"loss": 2.6011,
"step": 35
},
{
"epoch": 0.015930701448698164,
"grad_norm": 3.8187520503997803,
"learning_rate": 0.00029998114922746583,
"loss": 2.5422,
"step": 40
},
{
"epoch": 0.017922039129785433,
"grad_norm": 3.3968746662139893,
"learning_rate": 0.00029997586227586346,
"loss": 2.5569,
"step": 45
},
{
"epoch": 0.019913376810872702,
"grad_norm": 5.447213649749756,
"learning_rate": 0.00029996992268828285,
"loss": 2.5303,
"step": 50
},
{
"epoch": 0.021904714491959975,
"grad_norm": 3.1602864265441895,
"learning_rate": 0.0002999633304905707,
"loss": 2.5741,
"step": 55
},
{
"epoch": 0.023896052173047244,
"grad_norm": 3.8679628372192383,
"learning_rate": 0.0002999560857114137,
"loss": 2.5897,
"step": 60
},
{
"epoch": 0.025887389854134513,
"grad_norm": 3.5443027019500732,
"learning_rate": 0.00029994818838233823,
"loss": 2.6875,
"step": 65
},
{
"epoch": 0.027878727535221786,
"grad_norm": 3.584503412246704,
"learning_rate": 0.00029993963853771043,
"loss": 2.6568,
"step": 70
},
{
"epoch": 0.029870065216309055,
"grad_norm": 3.158475399017334,
"learning_rate": 0.00029993043621473584,
"loss": 2.6262,
"step": 75
},
{
"epoch": 0.03186140289739633,
"grad_norm": 4.361789703369141,
"learning_rate": 0.0002999205814534593,
"loss": 2.5888,
"step": 80
},
{
"epoch": 0.033852740578483594,
"grad_norm": 3.4383697509765625,
"learning_rate": 0.00029991007429676494,
"loss": 2.5537,
"step": 85
},
{
"epoch": 0.035844078259570866,
"grad_norm": 3.1702070236206055,
"learning_rate": 0.0002998989147903757,
"loss": 2.4183,
"step": 90
},
{
"epoch": 0.03783541594065814,
"grad_norm": 4.857996940612793,
"learning_rate": 0.00029988710298285335,
"loss": 2.572,
"step": 95
},
{
"epoch": 0.039826753621745405,
"grad_norm": 3.1501991748809814,
"learning_rate": 0.0002998746389255983,
"loss": 2.4835,
"step": 100
},
{
"epoch": 0.04181809130283268,
"grad_norm": 3.975405693054199,
"learning_rate": 0.0002998615226728493,
"loss": 2.579,
"step": 105
},
{
"epoch": 0.04380942898391995,
"grad_norm": 3.191920518875122,
"learning_rate": 0.0002998477542816829,
"loss": 2.5179,
"step": 110
},
{
"epoch": 0.045800766665007216,
"grad_norm": 3.55888295173645,
"learning_rate": 0.0002998333338120139,
"loss": 2.6215,
"step": 115
},
{
"epoch": 0.04779210434609449,
"grad_norm": 3.593543767929077,
"learning_rate": 0.00029981826132659446,
"loss": 2.4763,
"step": 120
},
{
"epoch": 0.04978344202718176,
"grad_norm": 4.385670185089111,
"learning_rate": 0.00029980253689101407,
"loss": 2.538,
"step": 125
},
{
"epoch": 0.05177477970826903,
"grad_norm": 3.4999585151672363,
"learning_rate": 0.0002997861605736994,
"loss": 2.5162,
"step": 130
},
{
"epoch": 0.0537661173893563,
"grad_norm": 3.657758951187134,
"learning_rate": 0.00029976913244591363,
"loss": 2.5734,
"step": 135
},
{
"epoch": 0.05575745507044357,
"grad_norm": 4.323859214782715,
"learning_rate": 0.00029975145258175653,
"loss": 2.6802,
"step": 140
},
{
"epoch": 0.05774879275153084,
"grad_norm": 3.3262791633605957,
"learning_rate": 0.0002997331210581639,
"loss": 2.5822,
"step": 145
},
{
"epoch": 0.05974013043261811,
"grad_norm": 3.809830665588379,
"learning_rate": 0.00029971413795490734,
"loss": 2.5107,
"step": 150
},
{
"epoch": 0.06173146811370538,
"grad_norm": 3.283738613128662,
"learning_rate": 0.0002996945033545939,
"loss": 2.6337,
"step": 155
},
{
"epoch": 0.06372280579479266,
"grad_norm": 3.442885398864746,
"learning_rate": 0.0002996742173426655,
"loss": 2.4263,
"step": 160
},
{
"epoch": 0.06571414347587992,
"grad_norm": 3.4212663173675537,
"learning_rate": 0.000299653280007399,
"loss": 2.5735,
"step": 165
},
{
"epoch": 0.06770548115696719,
"grad_norm": 3.8556458950042725,
"learning_rate": 0.0002996316914399054,
"loss": 2.5495,
"step": 170
},
{
"epoch": 0.06969681883805447,
"grad_norm": 4.062867164611816,
"learning_rate": 0.0002996094517341297,
"loss": 2.4665,
"step": 175
},
{
"epoch": 0.07168815651914173,
"grad_norm": 4.151358127593994,
"learning_rate": 0.00029958656098685036,
"loss": 2.5463,
"step": 180
},
{
"epoch": 0.073679494200229,
"grad_norm": 2.917109966278076,
"learning_rate": 0.00029956301929767883,
"loss": 2.5684,
"step": 185
},
{
"epoch": 0.07567083188131628,
"grad_norm": 4.13242769241333,
"learning_rate": 0.0002995388267690593,
"loss": 2.5404,
"step": 190
},
{
"epoch": 0.07766216956240354,
"grad_norm": 3.765427350997925,
"learning_rate": 0.00029951398350626824,
"loss": 2.4203,
"step": 195
},
{
"epoch": 0.07965350724349081,
"grad_norm": 4.968576908111572,
"learning_rate": 0.0002994884896174137,
"loss": 2.5068,
"step": 200
},
{
"epoch": 0.08164484492457809,
"grad_norm": 3.5136470794677734,
"learning_rate": 0.000299462345213435,
"loss": 2.5755,
"step": 205
},
{
"epoch": 0.08363618260566535,
"grad_norm": 3.77406907081604,
"learning_rate": 0.0002994355504081024,
"loss": 2.6048,
"step": 210
},
{
"epoch": 0.08562752028675262,
"grad_norm": 7.271650314331055,
"learning_rate": 0.00029940810531801633,
"loss": 2.6364,
"step": 215
},
{
"epoch": 0.0876188579678399,
"grad_norm": 4.087569713592529,
"learning_rate": 0.00029938001006260697,
"loss": 2.6253,
"step": 220
},
{
"epoch": 0.08961019564892717,
"grad_norm": 4.92678689956665,
"learning_rate": 0.00029935126476413396,
"loss": 2.4876,
"step": 225
},
{
"epoch": 0.09160153333001443,
"grad_norm": 4.242873668670654,
"learning_rate": 0.00029932186954768534,
"loss": 2.5986,
"step": 230
},
{
"epoch": 0.09359287101110171,
"grad_norm": 3.656292676925659,
"learning_rate": 0.0002992918245411777,
"loss": 2.47,
"step": 235
},
{
"epoch": 0.09558420869218898,
"grad_norm": 5.079676151275635,
"learning_rate": 0.000299261129875355,
"loss": 2.5446,
"step": 240
},
{
"epoch": 0.09757554637327624,
"grad_norm": 4.043215751647949,
"learning_rate": 0.0002992297856837884,
"loss": 2.4366,
"step": 245
},
{
"epoch": 0.09956688405436352,
"grad_norm": 3.710261106491089,
"learning_rate": 0.00029919779210287555,
"loss": 2.627,
"step": 250
},
{
"epoch": 0.10155822173545079,
"grad_norm": 6.257110595703125,
"learning_rate": 0.0002991651492718398,
"loss": 2.5499,
"step": 255
},
{
"epoch": 0.10354955941653805,
"grad_norm": 3.418558120727539,
"learning_rate": 0.00029913185733273,
"loss": 2.5487,
"step": 260
},
{
"epoch": 0.10554089709762533,
"grad_norm": 4.4878740310668945,
"learning_rate": 0.0002990979164304195,
"loss": 2.6916,
"step": 265
},
{
"epoch": 0.1075322347787126,
"grad_norm": 4.143045425415039,
"learning_rate": 0.00029906332671260594,
"loss": 2.5951,
"step": 270
},
{
"epoch": 0.10952357245979986,
"grad_norm": 4.116326332092285,
"learning_rate": 0.00029902808832981,
"loss": 2.5271,
"step": 275
},
{
"epoch": 0.11151491014088714,
"grad_norm": 3.435853958129883,
"learning_rate": 0.00029899220143537526,
"loss": 2.4868,
"step": 280
},
{
"epoch": 0.11350624782197441,
"grad_norm": 4.011627674102783,
"learning_rate": 0.0002989556661854675,
"loss": 2.5136,
"step": 285
},
{
"epoch": 0.11549758550306168,
"grad_norm": 4.947673320770264,
"learning_rate": 0.0002989184827390737,
"loss": 2.4441,
"step": 290
},
{
"epoch": 0.11748892318414896,
"grad_norm": 3.991987466812134,
"learning_rate": 0.00029888065125800153,
"loss": 2.4815,
"step": 295
},
{
"epoch": 0.11948026086523622,
"grad_norm": 5.1937360763549805,
"learning_rate": 0.0002988421719068788,
"loss": 2.6206,
"step": 300
},
{
"epoch": 0.12147159854632349,
"grad_norm": 4.464507102966309,
"learning_rate": 0.00029880304485315254,
"loss": 2.5436,
"step": 305
},
{
"epoch": 0.12346293622741077,
"grad_norm": 22.474430084228516,
"learning_rate": 0.00029876327026708824,
"loss": 2.6138,
"step": 310
},
{
"epoch": 0.12545427390849803,
"grad_norm": 4.142147064208984,
"learning_rate": 0.00029872284832176924,
"loss": 2.5605,
"step": 315
},
{
"epoch": 0.1274456115895853,
"grad_norm": 4.879560947418213,
"learning_rate": 0.00029868177919309603,
"loss": 2.609,
"step": 320
},
{
"epoch": 0.12943694927067256,
"grad_norm": 3.78598952293396,
"learning_rate": 0.00029864006305978523,
"loss": 2.5863,
"step": 325
},
{
"epoch": 0.13142828695175984,
"grad_norm": 5.507261753082275,
"learning_rate": 0.00029859770010336905,
"loss": 2.6525,
"step": 330
},
{
"epoch": 0.13341962463284712,
"grad_norm": 3.913325786590576,
"learning_rate": 0.00029855469050819454,
"loss": 2.5584,
"step": 335
},
{
"epoch": 0.13541096231393437,
"grad_norm": 4.263314723968506,
"learning_rate": 0.00029851981737782636,
"loss": 2.5515,
"step": 340
},
{
"epoch": 0.13740229999502165,
"grad_norm": 5.047379016876221,
"learning_rate": 0.0002984756443064224,
"loss": 2.5811,
"step": 345
},
{
"epoch": 0.13939363767610893,
"grad_norm": 3.8118085861206055,
"learning_rate": 0.00029843082512739883,
"loss": 2.647,
"step": 350
},
{
"epoch": 0.14138497535719619,
"grad_norm": 4.154444217681885,
"learning_rate": 0.0002983853600357912,
"loss": 2.4557,
"step": 355
},
{
"epoch": 0.14337631303828346,
"grad_norm": 3.792567253112793,
"learning_rate": 0.0002983392492294455,
"loss": 2.4481,
"step": 360
},
{
"epoch": 0.14536765071937074,
"grad_norm": 4.538722038269043,
"learning_rate": 0.00029829249290901804,
"loss": 2.4257,
"step": 365
},
{
"epoch": 0.147358988400458,
"grad_norm": 3.793282985687256,
"learning_rate": 0.00029824509127797383,
"loss": 2.5358,
"step": 370
},
{
"epoch": 0.14935032608154528,
"grad_norm": 4.034732818603516,
"learning_rate": 0.0002981970445425862,
"loss": 2.6415,
"step": 375
},
{
"epoch": 0.15134166376263256,
"grad_norm": 4.2402520179748535,
"learning_rate": 0.0002981483529119355,
"loss": 2.5689,
"step": 380
},
{
"epoch": 0.1533330014437198,
"grad_norm": 3.5456957817077637,
"learning_rate": 0.0002980990165979088,
"loss": 2.4177,
"step": 385
},
{
"epoch": 0.1553243391248071,
"grad_norm": 4.852187156677246,
"learning_rate": 0.0002980490358151981,
"loss": 2.5833,
"step": 390
},
{
"epoch": 0.15731567680589437,
"grad_norm": 3.8212156295776367,
"learning_rate": 0.0002979984107813002,
"loss": 2.439,
"step": 395
},
{
"epoch": 0.15930701448698162,
"grad_norm": 5.308228015899658,
"learning_rate": 0.0002979471417165154,
"loss": 2.6805,
"step": 400
},
{
"epoch": 0.1612983521680689,
"grad_norm": 3.9180448055267334,
"learning_rate": 0.00029789522884394646,
"loss": 2.8233,
"step": 405
},
{
"epoch": 0.16328968984915618,
"grad_norm": 3.9016313552856445,
"learning_rate": 0.0002978426723894978,
"loss": 2.5363,
"step": 410
},
{
"epoch": 0.16528102753024343,
"grad_norm": 5.2438225746154785,
"learning_rate": 0.00029778947258187437,
"loss": 2.5845,
"step": 415
},
{
"epoch": 0.1672723652113307,
"grad_norm": 4.8523664474487305,
"learning_rate": 0.0002977356296525809,
"loss": 2.5257,
"step": 420
},
{
"epoch": 0.169263702892418,
"grad_norm": 5.205035209655762,
"learning_rate": 0.00029768114383592066,
"loss": 2.5175,
"step": 425
},
{
"epoch": 0.17125504057350524,
"grad_norm": 4.0942606925964355,
"learning_rate": 0.00029762601536899443,
"loss": 2.5718,
"step": 430
},
{
"epoch": 0.17324637825459252,
"grad_norm": 3.762770652770996,
"learning_rate": 0.0002975702444916997,
"loss": 2.5675,
"step": 435
},
{
"epoch": 0.1752377159356798,
"grad_norm": 4.622000694274902,
"learning_rate": 0.0002975138314467293,
"loss": 2.5359,
"step": 440
},
{
"epoch": 0.17722905361676705,
"grad_norm": 3.6756174564361572,
"learning_rate": 0.0002974567764795707,
"loss": 2.5943,
"step": 445
},
{
"epoch": 0.17922039129785433,
"grad_norm": 3.5715696811676025,
"learning_rate": 0.0002973990798385046,
"loss": 2.6092,
"step": 450
},
{
"epoch": 0.1812117289789416,
"grad_norm": 3.91926646232605,
"learning_rate": 0.0002973407417746042,
"loss": 2.5497,
"step": 455
},
{
"epoch": 0.18320306666002886,
"grad_norm": 3.9486913681030273,
"learning_rate": 0.0002972817625417338,
"loss": 2.5328,
"step": 460
},
{
"epoch": 0.18519440434111614,
"grad_norm": 5.621224880218506,
"learning_rate": 0.00029722214239654766,
"loss": 2.6173,
"step": 465
},
{
"epoch": 0.18718574202220342,
"grad_norm": 3.9391846656799316,
"learning_rate": 0.0002971618815984894,
"loss": 2.6796,
"step": 470
},
{
"epoch": 0.18917707970329067,
"grad_norm": 4.538651943206787,
"learning_rate": 0.00029710098040979026,
"loss": 2.4812,
"step": 475
},
{
"epoch": 0.19116841738437795,
"grad_norm": 4.420340538024902,
"learning_rate": 0.0002970394390954683,
"loss": 2.6853,
"step": 480
},
{
"epoch": 0.19315975506546523,
"grad_norm": 3.4120190143585205,
"learning_rate": 0.000296977257923327,
"loss": 2.5475,
"step": 485
},
{
"epoch": 0.19515109274655248,
"grad_norm": 3.7508790493011475,
"learning_rate": 0.00029691443716395446,
"loss": 2.5562,
"step": 490
},
{
"epoch": 0.19714243042763976,
"grad_norm": 6.769453525543213,
"learning_rate": 0.0002968509770907219,
"loss": 2.5752,
"step": 495
},
{
"epoch": 0.19913376810872704,
"grad_norm": 5.990815162658691,
"learning_rate": 0.0002967868779797827,
"loss": 2.5315,
"step": 500
},
{
"epoch": 0.2011251057898143,
"grad_norm": 4.616420745849609,
"learning_rate": 0.0002967221401100708,
"loss": 2.4976,
"step": 505
},
{
"epoch": 0.20311644347090158,
"grad_norm": 4.068687915802002,
"learning_rate": 0.0002966567637633001,
"loss": 2.4871,
"step": 510
},
{
"epoch": 0.20510778115198885,
"grad_norm": 4.577060699462891,
"learning_rate": 0.00029659074922396266,
"loss": 2.6066,
"step": 515
},
{
"epoch": 0.2070991188330761,
"grad_norm": 6.227926254272461,
"learning_rate": 0.00029652409677932793,
"loss": 2.6226,
"step": 520
},
{
"epoch": 0.2090904565141634,
"grad_norm": 4.359228134155273,
"learning_rate": 0.000296456806719441,
"loss": 2.5863,
"step": 525
},
{
"epoch": 0.21108179419525067,
"grad_norm": 4.112326145172119,
"learning_rate": 0.0002963888793371218,
"loss": 2.5719,
"step": 530
},
{
"epoch": 0.21307313187633792,
"grad_norm": 4.560180187225342,
"learning_rate": 0.0002963203149279636,
"loss": 2.5393,
"step": 535
},
{
"epoch": 0.2150644695574252,
"grad_norm": 6.014193058013916,
"learning_rate": 0.00029625111379033174,
"loss": 2.5966,
"step": 540
},
{
"epoch": 0.21705580723851248,
"grad_norm": 3.90463924407959,
"learning_rate": 0.0002961812762253623,
"loss": 2.671,
"step": 545
},
{
"epoch": 0.21904714491959973,
"grad_norm": 5.432832717895508,
"learning_rate": 0.000296110802536961,
"loss": 2.587,
"step": 550
},
{
"epoch": 0.221038482600687,
"grad_norm": 4.703036785125732,
"learning_rate": 0.0002960396930318016,
"loss": 2.6888,
"step": 555
},
{
"epoch": 0.2230298202817743,
"grad_norm": 3.983517646789551,
"learning_rate": 0.00029596794801932467,
"loss": 2.6974,
"step": 560
},
{
"epoch": 0.22502115796286154,
"grad_norm": 5.251948356628418,
"learning_rate": 0.0002958955678117363,
"loss": 2.581,
"step": 565
},
{
"epoch": 0.22701249564394882,
"grad_norm": 3.736345052719116,
"learning_rate": 0.0002958225527240067,
"loss": 2.6273,
"step": 570
},
{
"epoch": 0.2290038333250361,
"grad_norm": 4.199942588806152,
"learning_rate": 0.0002957489030738688,
"loss": 2.5318,
"step": 575
},
{
"epoch": 0.23099517100612335,
"grad_norm": 4.074978828430176,
"learning_rate": 0.00029567461918181693,
"loss": 2.6444,
"step": 580
},
{
"epoch": 0.23298650868721063,
"grad_norm": 4.244055271148682,
"learning_rate": 0.00029559970137110536,
"loss": 2.5639,
"step": 585
},
{
"epoch": 0.2349778463682979,
"grad_norm": 4.564445972442627,
"learning_rate": 0.000295524149967747,
"loss": 2.6057,
"step": 590
},
{
"epoch": 0.23696918404938516,
"grad_norm": 12.886066436767578,
"learning_rate": 0.00029546325287919543,
"loss": 2.5556,
"step": 595
},
{
"epoch": 0.23896052173047244,
"grad_norm": 4.276486396789551,
"learning_rate": 0.00029538656183942505,
"loss": 2.5598,
"step": 600
},
{
"epoch": 0.24095185941155972,
"grad_norm": 3.5438523292541504,
"learning_rate": 0.0002953092381345073,
"loss": 2.6364,
"step": 605
},
{
"epoch": 0.24294319709264697,
"grad_norm": 5.840348243713379,
"learning_rate": 0.00029523128210092445,
"loss": 2.5398,
"step": 610
},
{
"epoch": 0.24493453477373425,
"grad_norm": 4.8936285972595215,
"learning_rate": 0.0002951526940779106,
"loss": 2.5041,
"step": 615
},
{
"epoch": 0.24692587245482153,
"grad_norm": 5.3416337966918945,
"learning_rate": 0.00029507347440744993,
"loss": 2.5853,
"step": 620
},
{
"epoch": 0.24891721013590878,
"grad_norm": 6.661780834197998,
"learning_rate": 0.0002949936234342752,
"loss": 2.574,
"step": 625
},
{
"epoch": 0.25090854781699606,
"grad_norm": 5.31721305847168,
"learning_rate": 0.0002949131415058665,
"loss": 2.6033,
"step": 630
},
{
"epoch": 0.25289988549808334,
"grad_norm": 4.867547988891602,
"learning_rate": 0.0002948320289724496,
"loss": 2.6626,
"step": 635
},
{
"epoch": 0.2548912231791706,
"grad_norm": 5.965942859649658,
"learning_rate": 0.00029475028618699417,
"loss": 2.5723,
"step": 640
},
{
"epoch": 0.2568825608602579,
"grad_norm": 4.902228355407715,
"learning_rate": 0.00029466791350521286,
"loss": 2.5944,
"step": 645
},
{
"epoch": 0.2588738985413451,
"grad_norm": 5.2122650146484375,
"learning_rate": 0.0002945849112855591,
"loss": 2.5094,
"step": 650
},
{
"epoch": 0.2608652362224324,
"grad_norm": 4.673396110534668,
"learning_rate": 0.0002945012798892259,
"loss": 2.5138,
"step": 655
},
{
"epoch": 0.2628565739035197,
"grad_norm": 8.117375373840332,
"learning_rate": 0.00029441701968014423,
"loss": 2.7075,
"step": 660
},
{
"epoch": 0.26484791158460697,
"grad_norm": 6.844805717468262,
"learning_rate": 0.0002943321310249814,
"loss": 2.6431,
"step": 665
},
{
"epoch": 0.26683924926569424,
"grad_norm": 5.7974982261657715,
"learning_rate": 0.0002942466142931395,
"loss": 2.5232,
"step": 670
},
{
"epoch": 0.2688305869467815,
"grad_norm": 4.563413619995117,
"learning_rate": 0.00029416046985675356,
"loss": 2.492,
"step": 675
},
{
"epoch": 0.27082192462786875,
"grad_norm": 4.943353176116943,
"learning_rate": 0.0002940736980906905,
"loss": 2.4998,
"step": 680
},
{
"epoch": 0.27281326230895603,
"grad_norm": 5.243081569671631,
"learning_rate": 0.00029398629937254676,
"loss": 2.5161,
"step": 685
},
{
"epoch": 0.2748045999900433,
"grad_norm": 3.9631588459014893,
"learning_rate": 0.00029389827408264727,
"loss": 2.5674,
"step": 690
},
{
"epoch": 0.2767959376711306,
"grad_norm": 4.865413665771484,
"learning_rate": 0.0002938096226040435,
"loss": 2.5864,
"step": 695
},
{
"epoch": 0.27878727535221787,
"grad_norm": 4.747464179992676,
"learning_rate": 0.00029372034532251177,
"loss": 2.6115,
"step": 700
},
{
"epoch": 0.28077861303330515,
"grad_norm": 4.920633792877197,
"learning_rate": 0.00029363044262655175,
"loss": 2.596,
"step": 705
},
{
"epoch": 0.28276995071439237,
"grad_norm": 4.5279340744018555,
"learning_rate": 0.00029353991490738455,
"loss": 2.6313,
"step": 710
},
{
"epoch": 0.28476128839547965,
"grad_norm": 4.8143486976623535,
"learning_rate": 0.0002934487625589513,
"loss": 2.6689,
"step": 715
},
{
"epoch": 0.28675262607656693,
"grad_norm": 12.043296813964844,
"learning_rate": 0.0002933569859779111,
"loss": 2.6326,
"step": 720
},
{
"epoch": 0.2887439637576542,
"grad_norm": 4.9652557373046875,
"learning_rate": 0.00029326458556363957,
"loss": 2.5531,
"step": 725
},
{
"epoch": 0.2907353014387415,
"grad_norm": 5.17543888092041,
"learning_rate": 0.000293171561718227,
"loss": 2.5285,
"step": 730
},
{
"epoch": 0.29272663911982877,
"grad_norm": 5.931879997253418,
"learning_rate": 0.0002930779148464765,
"loss": 2.5604,
"step": 735
},
{
"epoch": 0.294717976800916,
"grad_norm": 5.292037487030029,
"learning_rate": 0.0002929836453559026,
"loss": 2.5188,
"step": 740
},
{
"epoch": 0.2967093144820033,
"grad_norm": 5.012665748596191,
"learning_rate": 0.00029288875365672887,
"loss": 2.5135,
"step": 745
},
{
"epoch": 0.29870065216309055,
"grad_norm": 5.738946914672852,
"learning_rate": 0.00029279324016188676,
"loss": 2.5426,
"step": 750
},
{
"epoch": 0.30069198984417783,
"grad_norm": 4.243402481079102,
"learning_rate": 0.00029269710528701345,
"loss": 2.4408,
"step": 755
},
{
"epoch": 0.3026833275252651,
"grad_norm": 5.939416408538818,
"learning_rate": 0.0002926003494504501,
"loss": 2.8477,
"step": 760
},
{
"epoch": 0.3046746652063524,
"grad_norm": 4.015962600708008,
"learning_rate": 0.0002925029730732401,
"loss": 2.4087,
"step": 765
},
{
"epoch": 0.3066660028874396,
"grad_norm": 6.449498653411865,
"learning_rate": 0.000292404976579127,
"loss": 2.5785,
"step": 770
},
{
"epoch": 0.3086573405685269,
"grad_norm": 7.409605503082275,
"learning_rate": 0.0002923063603945532,
"loss": 2.5958,
"step": 775
},
{
"epoch": 0.3106486782496142,
"grad_norm": 4.776349067687988,
"learning_rate": 0.0002922071249486575,
"loss": 2.6296,
"step": 780
},
{
"epoch": 0.31264001593070145,
"grad_norm": 4.3293352127075195,
"learning_rate": 0.0002921072706732734,
"loss": 2.5353,
"step": 785
},
{
"epoch": 0.31463135361178873,
"grad_norm": 6.083279609680176,
"learning_rate": 0.00029200679800292766,
"loss": 2.6478,
"step": 790
},
{
"epoch": 0.316622691292876,
"grad_norm": 5.940248012542725,
"learning_rate": 0.0002919057073748377,
"loss": 2.7188,
"step": 795
},
{
"epoch": 0.31861402897396324,
"grad_norm": 4.022637367248535,
"learning_rate": 0.00029180399922891026,
"loss": 2.4939,
"step": 800
},
{
"epoch": 0.3206053666550505,
"grad_norm": 5.022520542144775,
"learning_rate": 0.00029170167400773906,
"loss": 2.5447,
"step": 805
},
{
"epoch": 0.3225967043361378,
"grad_norm": 7.304029941558838,
"learning_rate": 0.00029159873215660337,
"loss": 2.5208,
"step": 810
},
{
"epoch": 0.3245880420172251,
"grad_norm": 4.472403526306152,
"learning_rate": 0.00029149517412346555,
"loss": 2.6699,
"step": 815
},
{
"epoch": 0.32657937969831236,
"grad_norm": 5.33425235748291,
"learning_rate": 0.00029139100035896953,
"loss": 2.5921,
"step": 820
},
{
"epoch": 0.32857071737939963,
"grad_norm": 5.940118312835693,
"learning_rate": 0.00029128621131643844,
"loss": 2.6926,
"step": 825
},
{
"epoch": 0.33056205506048686,
"grad_norm": 5.923150539398193,
"learning_rate": 0.0002911808074518731,
"loss": 2.6289,
"step": 830
},
{
"epoch": 0.33255339274157414,
"grad_norm": 7.126611709594727,
"learning_rate": 0.0002910747892239497,
"loss": 2.5597,
"step": 835
},
{
"epoch": 0.3345447304226614,
"grad_norm": 6.361411094665527,
"learning_rate": 0.00029096815709401776,
"loss": 2.6819,
"step": 840
},
{
"epoch": 0.3365360681037487,
"grad_norm": 4.598730564117432,
"learning_rate": 0.00029086091152609846,
"loss": 2.5476,
"step": 845
},
{
"epoch": 0.338527405784836,
"grad_norm": 4.714099884033203,
"learning_rate": 0.00029075305298688234,
"loss": 2.5689,
"step": 850
},
{
"epoch": 0.34051874346592326,
"grad_norm": 6.404987335205078,
"learning_rate": 0.00029064458194572734,
"loss": 2.6044,
"step": 855
},
{
"epoch": 0.3425100811470105,
"grad_norm": 5.2892022132873535,
"learning_rate": 0.0002905354988746568,
"loss": 2.5326,
"step": 860
},
{
"epoch": 0.34450141882809776,
"grad_norm": 5.470654487609863,
"learning_rate": 0.00029042580424835723,
"loss": 2.4232,
"step": 865
},
{
"epoch": 0.34649275650918504,
"grad_norm": 4.468808174133301,
"learning_rate": 0.00029031549854417667,
"loss": 2.6581,
"step": 870
},
{
"epoch": 0.3484840941902723,
"grad_norm": 5.86844539642334,
"learning_rate": 0.00029020458224212204,
"loss": 2.5621,
"step": 875
},
{
"epoch": 0.3504754318713596,
"grad_norm": 4.8259711265563965,
"learning_rate": 0.0002900930558248576,
"loss": 2.6615,
"step": 880
},
{
"epoch": 0.3524667695524469,
"grad_norm": 5.640230655670166,
"learning_rate": 0.00028998091977770236,
"loss": 2.4399,
"step": 885
},
{
"epoch": 0.3544581072335341,
"grad_norm": 5.823399543762207,
"learning_rate": 0.00028986817458862837,
"loss": 2.6114,
"step": 890
},
{
"epoch": 0.3564494449146214,
"grad_norm": 5.678874492645264,
"learning_rate": 0.0002897548207482583,
"loss": 2.6281,
"step": 895
},
{
"epoch": 0.35844078259570866,
"grad_norm": 13.930092811584473,
"learning_rate": 0.00028964085874986356,
"loss": 2.6551,
"step": 900
},
{
"epoch": 0.36043212027679594,
"grad_norm": 4.730804443359375,
"learning_rate": 0.00028952628908936184,
"loss": 2.7113,
"step": 905
},
{
"epoch": 0.3624234579578832,
"grad_norm": 4.784435749053955,
"learning_rate": 0.0002894111122653153,
"loss": 2.7406,
"step": 910
},
{
"epoch": 0.3644147956389705,
"grad_norm": 5.345913887023926,
"learning_rate": 0.0002892953287789283,
"loss": 2.5814,
"step": 915
},
{
"epoch": 0.3664061333200577,
"grad_norm": 5.723335266113281,
"learning_rate": 0.0002891789391340449,
"loss": 2.7098,
"step": 920
},
{
"epoch": 0.368397471001145,
"grad_norm": 4.860372066497803,
"learning_rate": 0.0002890619438371472,
"loss": 2.5403,
"step": 925
},
{
"epoch": 0.3703888086822323,
"grad_norm": 4.663769245147705,
"learning_rate": 0.0002889443433973525,
"loss": 2.5292,
"step": 930
},
{
"epoch": 0.37238014636331956,
"grad_norm": 4.552435874938965,
"learning_rate": 0.0002888261383264119,
"loss": 2.5331,
"step": 935
},
{
"epoch": 0.37437148404440684,
"grad_norm": 6.1328253746032715,
"learning_rate": 0.0002887073291387073,
"loss": 2.6319,
"step": 940
},
{
"epoch": 0.3763628217254941,
"grad_norm": 4.32132625579834,
"learning_rate": 0.0002885879163512496,
"loss": 2.4976,
"step": 945
},
{
"epoch": 0.37835415940658135,
"grad_norm": 5.826934814453125,
"learning_rate": 0.0002884679004836762,
"loss": 2.5219,
"step": 950
},
{
"epoch": 0.3803454970876686,
"grad_norm": 5.766158103942871,
"learning_rate": 0.00028834728205824904,
"loss": 2.5199,
"step": 955
},
{
"epoch": 0.3823368347687559,
"grad_norm": 4.263400077819824,
"learning_rate": 0.0002882260615998521,
"loss": 2.5299,
"step": 960
},
{
"epoch": 0.3843281724498432,
"grad_norm": 5.239638328552246,
"learning_rate": 0.0002881042396359891,
"loss": 2.5803,
"step": 965
},
{
"epoch": 0.38631951013093047,
"grad_norm": 5.97824239730835,
"learning_rate": 0.00028798181669678136,
"loss": 2.5917,
"step": 970
},
{
"epoch": 0.38831084781201775,
"grad_norm": 5.745943069458008,
"learning_rate": 0.0002878587933149653,
"loss": 2.6519,
"step": 975
},
{
"epoch": 0.39030218549310497,
"grad_norm": 4.912780284881592,
"learning_rate": 0.0002877351700258904,
"loss": 2.7083,
"step": 980
},
{
"epoch": 0.39229352317419225,
"grad_norm": 5.343855857849121,
"learning_rate": 0.0002876109473675165,
"loss": 2.4946,
"step": 985
},
{
"epoch": 0.39428486085527953,
"grad_norm": 5.487871170043945,
"learning_rate": 0.0002874861258804118,
"loss": 2.6007,
"step": 990
},
{
"epoch": 0.3962761985363668,
"grad_norm": 5.249770164489746,
"learning_rate": 0.00028736070610775046,
"loss": 2.5687,
"step": 995
},
{
"epoch": 0.3982675362174541,
"grad_norm": 4.565980434417725,
"learning_rate": 0.00028723468859530977,
"loss": 2.5454,
"step": 1000
},
{
"epoch": 0.40025887389854137,
"grad_norm": 6.998654365539551,
"learning_rate": 0.0002871080738914685,
"loss": 2.5901,
"step": 1005
},
{
"epoch": 0.4022502115796286,
"grad_norm": 5.072119235992432,
"learning_rate": 0.0002869808625472042,
"loss": 2.6213,
"step": 1010
},
{
"epoch": 0.40424154926071587,
"grad_norm": 6.3118720054626465,
"learning_rate": 0.0002868530551160904,
"loss": 2.5593,
"step": 1015
},
{
"epoch": 0.40623288694180315,
"grad_norm": 4.6301984786987305,
"learning_rate": 0.00028672465215429484,
"loss": 2.613,
"step": 1020
},
{
"epoch": 0.40822422462289043,
"grad_norm": 5.474564075469971,
"learning_rate": 0.0002865956542205768,
"loss": 2.559,
"step": 1025
},
{
"epoch": 0.4102155623039777,
"grad_norm": 4.7905168533325195,
"learning_rate": 0.00028646606187628446,
"loss": 2.6535,
"step": 1030
},
{
"epoch": 0.412206899985065,
"grad_norm": 5.30242395401001,
"learning_rate": 0.0002863358756853528,
"loss": 2.4873,
"step": 1035
},
{
"epoch": 0.4141982376661522,
"grad_norm": 4.635322093963623,
"learning_rate": 0.000286205096214301,
"loss": 2.6035,
"step": 1040
},
{
"epoch": 0.4161895753472395,
"grad_norm": 5.1605000495910645,
"learning_rate": 0.00028607372403222976,
"loss": 2.6536,
"step": 1045
},
{
"epoch": 0.4181809130283268,
"grad_norm": 5.128244400024414,
"learning_rate": 0.0002859417597108192,
"loss": 2.5994,
"step": 1050
},
{
"epoch": 0.42017225070941405,
"grad_norm": 5.02494478225708,
"learning_rate": 0.0002858092038243262,
"loss": 2.6867,
"step": 1055
},
{
"epoch": 0.42216358839050133,
"grad_norm": 5.030214309692383,
"learning_rate": 0.0002856760569495818,
"loss": 2.5302,
"step": 1060
},
{
"epoch": 0.4241549260715886,
"grad_norm": 5.012157917022705,
"learning_rate": 0.00028554231966598884,
"loss": 2.5605,
"step": 1065
},
{
"epoch": 0.42614626375267584,
"grad_norm": 5.142708778381348,
"learning_rate": 0.0002854079925555195,
"loss": 2.6312,
"step": 1070
},
{
"epoch": 0.4281376014337631,
"grad_norm": 6.910045623779297,
"learning_rate": 0.0002852730762027124,
"loss": 2.5357,
"step": 1075
},
{
"epoch": 0.4301289391148504,
"grad_norm": 5.449594497680664,
"learning_rate": 0.0002851375711946705,
"loss": 2.5713,
"step": 1080
},
{
"epoch": 0.4321202767959377,
"grad_norm": 6.184391975402832,
"learning_rate": 0.0002850014781210583,
"loss": 2.5214,
"step": 1085
},
{
"epoch": 0.43411161447702495,
"grad_norm": 6.970107078552246,
"learning_rate": 0.0002848647975740994,
"loss": 2.63,
"step": 1090
},
{
"epoch": 0.43610295215811223,
"grad_norm": 7.1215972900390625,
"learning_rate": 0.0002847275301485737,
"loss": 2.5809,
"step": 1095
},
{
"epoch": 0.43809428983919946,
"grad_norm": 5.520656108856201,
"learning_rate": 0.0002845896764418152,
"loss": 2.5623,
"step": 1100
},
{
"epoch": 0.44008562752028674,
"grad_norm": 5.367312908172607,
"learning_rate": 0.0002844512370537089,
"loss": 2.6073,
"step": 1105
},
{
"epoch": 0.442076965201374,
"grad_norm": 5.053874492645264,
"learning_rate": 0.00028431221258668865,
"loss": 2.6203,
"step": 1110
},
{
"epoch": 0.4440683028824613,
"grad_norm": 5.396303176879883,
"learning_rate": 0.0002841726036457342,
"loss": 2.3232,
"step": 1115
},
{
"epoch": 0.4460596405635486,
"grad_norm": 4.401814937591553,
"learning_rate": 0.0002840324108383688,
"loss": 2.559,
"step": 1120
},
{
"epoch": 0.44805097824463586,
"grad_norm": 7.336491584777832,
"learning_rate": 0.0002838916347746564,
"loss": 2.6238,
"step": 1125
},
{
"epoch": 0.4500423159257231,
"grad_norm": 4.719950199127197,
"learning_rate": 0.0002837502760671991,
"loss": 2.5617,
"step": 1130
},
{
"epoch": 0.45203365360681036,
"grad_norm": 7.233749866485596,
"learning_rate": 0.0002836083353311344,
"loss": 2.6933,
"step": 1135
},
{
"epoch": 0.45402499128789764,
"grad_norm": 5.0576491355896,
"learning_rate": 0.00028346581318413255,
"loss": 2.4117,
"step": 1140
},
{
"epoch": 0.4560163289689849,
"grad_norm": 4.500210762023926,
"learning_rate": 0.00028332271024639395,
"loss": 2.6071,
"step": 1145
},
{
"epoch": 0.4580076666500722,
"grad_norm": 5.0376105308532715,
"learning_rate": 0.0002831790271406462,
"loss": 2.6693,
"step": 1150
},
{
"epoch": 0.4599990043311595,
"grad_norm": 5.0912981033325195,
"learning_rate": 0.0002830347644921418,
"loss": 2.5838,
"step": 1155
},
{
"epoch": 0.4619903420122467,
"grad_norm": 4.176620006561279,
"learning_rate": 0.00028288992292865503,
"loss": 2.5671,
"step": 1160
},
{
"epoch": 0.463981679693334,
"grad_norm": 5.332305431365967,
"learning_rate": 0.0002827445030804794,
"loss": 2.6354,
"step": 1165
},
{
"epoch": 0.46597301737442126,
"grad_norm": 5.466835975646973,
"learning_rate": 0.00028259850558042486,
"loss": 2.524,
"step": 1170
},
{
"epoch": 0.46796435505550854,
"grad_norm": 5.8639020919799805,
"learning_rate": 0.00028245193106381515,
"loss": 2.5369,
"step": 1175
},
{
"epoch": 0.4699556927365958,
"grad_norm": 5.681973457336426,
"learning_rate": 0.00028230478016848503,
"loss": 2.503,
"step": 1180
},
{
"epoch": 0.4719470304176831,
"grad_norm": 6.664337158203125,
"learning_rate": 0.0002821570535347772,
"loss": 2.6666,
"step": 1185
},
{
"epoch": 0.4739383680987703,
"grad_norm": 13.131908416748047,
"learning_rate": 0.00028200875180553996,
"loss": 2.5864,
"step": 1190
},
{
"epoch": 0.4759297057798576,
"grad_norm": 4.748403549194336,
"learning_rate": 0.00028185987562612404,
"loss": 2.5652,
"step": 1195
},
{
"epoch": 0.4779210434609449,
"grad_norm": 5.171413421630859,
"learning_rate": 0.00028171042564438016,
"loss": 2.4675,
"step": 1200
},
{
"epoch": 0.47991238114203216,
"grad_norm": 5.2752227783203125,
"learning_rate": 0.00028156040251065575,
"loss": 2.738,
"step": 1205
},
{
"epoch": 0.48190371882311944,
"grad_norm": 5.439636707305908,
"learning_rate": 0.00028140980687779254,
"loss": 2.5925,
"step": 1210
},
{
"epoch": 0.4838950565042067,
"grad_norm": 5.162364482879639,
"learning_rate": 0.00028125863940112346,
"loss": 2.5849,
"step": 1215
},
{
"epoch": 0.48588639418529395,
"grad_norm": 4.560568809509277,
"learning_rate": 0.00028110690073847,
"loss": 2.5966,
"step": 1220
},
{
"epoch": 0.4878777318663812,
"grad_norm": 5.049319744110107,
"learning_rate": 0.0002809545915501392,
"loss": 2.5334,
"step": 1225
},
{
"epoch": 0.4898690695474685,
"grad_norm": 6.197220802307129,
"learning_rate": 0.0002808017124989206,
"loss": 2.4928,
"step": 1230
},
{
"epoch": 0.4918604072285558,
"grad_norm": 4.927899360656738,
"learning_rate": 0.0002806482642500838,
"loss": 2.6438,
"step": 1235
},
{
"epoch": 0.49385174490964306,
"grad_norm": 4.190851211547852,
"learning_rate": 0.00028049424747137537,
"loss": 2.6136,
"step": 1240
},
{
"epoch": 0.49584308259073034,
"grad_norm": 4.721156597137451,
"learning_rate": 0.00028033966283301547,
"loss": 2.7049,
"step": 1245
},
{
"epoch": 0.49783442027181757,
"grad_norm": 5.104554653167725,
"learning_rate": 0.0002801845110076959,
"loss": 2.5448,
"step": 1250
},
{
"epoch": 0.49982575795290485,
"grad_norm": 5.173543453216553,
"learning_rate": 0.00028002879267057624,
"loss": 2.4522,
"step": 1255
},
{
"epoch": 0.5018170956339921,
"grad_norm": 6.031247138977051,
"learning_rate": 0.0002798725084992816,
"loss": 2.7182,
"step": 1260
},
{
"epoch": 0.5038084333150794,
"grad_norm": 7.677878379821777,
"learning_rate": 0.00027971565917389915,
"loss": 2.5468,
"step": 1265
},
{
"epoch": 0.5057997709961667,
"grad_norm": 5.632172584533691,
"learning_rate": 0.0002795582453769755,
"loss": 2.618,
"step": 1270
},
{
"epoch": 0.507791108677254,
"grad_norm": 5.612128257751465,
"learning_rate": 0.0002794002677935136,
"loss": 2.6364,
"step": 1275
},
{
"epoch": 0.5097824463583412,
"grad_norm": 6.009293556213379,
"learning_rate": 0.00027924172711096975,
"loss": 2.6681,
"step": 1280
},
{
"epoch": 0.5117737840394285,
"grad_norm": 4.658798694610596,
"learning_rate": 0.0002790826240192507,
"loss": 2.6049,
"step": 1285
},
{
"epoch": 0.5137651217205158,
"grad_norm": 5.140453815460205,
"learning_rate": 0.00027892295921071056,
"loss": 2.5811,
"step": 1290
},
{
"epoch": 0.515756459401603,
"grad_norm": 6.587934494018555,
"learning_rate": 0.00027876273338014773,
"loss": 2.5577,
"step": 1295
},
{
"epoch": 0.5177477970826903,
"grad_norm": 5.756930828094482,
"learning_rate": 0.000278601947224802,
"loss": 2.6275,
"step": 1300
},
{
"epoch": 0.5197391347637775,
"grad_norm": 4.118381500244141,
"learning_rate": 0.00027844060144435154,
"loss": 2.575,
"step": 1305
},
{
"epoch": 0.5217304724448648,
"grad_norm": 6.136612892150879,
"learning_rate": 0.0002782786967409098,
"loss": 2.5923,
"step": 1310
},
{
"epoch": 0.5237218101259521,
"grad_norm": 5.936057090759277,
"learning_rate": 0.0002781162338190222,
"loss": 2.6065,
"step": 1315
},
{
"epoch": 0.5257131478070394,
"grad_norm": 5.510227203369141,
"learning_rate": 0.0002779532133856637,
"loss": 2.5999,
"step": 1320
},
{
"epoch": 0.5277044854881267,
"grad_norm": 5.084118843078613,
"learning_rate": 0.00027778963615023505,
"loss": 2.6796,
"step": 1325
},
{
"epoch": 0.5296958231692139,
"grad_norm": 8.913198471069336,
"learning_rate": 0.00027762550282456,
"loss": 2.6884,
"step": 1330
},
{
"epoch": 0.5316871608503012,
"grad_norm": 4.569077014923096,
"learning_rate": 0.00027746081412288227,
"loss": 2.5403,
"step": 1335
},
{
"epoch": 0.5336784985313885,
"grad_norm": 6.321774005889893,
"learning_rate": 0.00027729557076186235,
"loss": 2.6237,
"step": 1340
},
{
"epoch": 0.5356698362124758,
"grad_norm": 6.090647220611572,
"learning_rate": 0.0002771297734605744,
"loss": 2.6553,
"step": 1345
},
{
"epoch": 0.537661173893563,
"grad_norm": 4.082389831542969,
"learning_rate": 0.00027696342294050296,
"loss": 2.6116,
"step": 1350
},
{
"epoch": 0.5396525115746502,
"grad_norm": 5.483598709106445,
"learning_rate": 0.0002767965199255402,
"loss": 2.5374,
"step": 1355
},
{
"epoch": 0.5416438492557375,
"grad_norm": 6.810405254364014,
"learning_rate": 0.0002766290651419824,
"loss": 2.5081,
"step": 1360
},
{
"epoch": 0.5436351869368248,
"grad_norm": 5.5257039070129395,
"learning_rate": 0.00027646105931852685,
"loss": 2.578,
"step": 1365
},
{
"epoch": 0.5456265246179121,
"grad_norm": 6.28507137298584,
"learning_rate": 0.00027629250318626894,
"loss": 2.7601,
"step": 1370
},
{
"epoch": 0.5476178622989993,
"grad_norm": 4.6391119956970215,
"learning_rate": 0.0002761233974786985,
"loss": 2.6488,
"step": 1375
},
{
"epoch": 0.5496091999800866,
"grad_norm": 4.203032493591309,
"learning_rate": 0.00027595374293169734,
"loss": 2.5824,
"step": 1380
},
{
"epoch": 0.5516005376611739,
"grad_norm": 5.504507541656494,
"learning_rate": 0.0002757835402835351,
"loss": 2.5392,
"step": 1385
},
{
"epoch": 0.5535918753422612,
"grad_norm": 4.769162178039551,
"learning_rate": 0.00027561279027486687,
"loss": 2.6826,
"step": 1390
},
{
"epoch": 0.5555832130233485,
"grad_norm": 5.556728363037109,
"learning_rate": 0.0002754414936487296,
"loss": 2.6405,
"step": 1395
},
{
"epoch": 0.5575745507044357,
"grad_norm": 4.737353324890137,
"learning_rate": 0.0002752696511505387,
"loss": 2.6097,
"step": 1400
},
{
"epoch": 0.559565888385523,
"grad_norm": 4.895107746124268,
"learning_rate": 0.00027509726352808523,
"loss": 2.7201,
"step": 1405
},
{
"epoch": 0.5615572260666103,
"grad_norm": 5.524222373962402,
"learning_rate": 0.0002749243315315324,
"loss": 2.6121,
"step": 1410
},
{
"epoch": 0.5635485637476975,
"grad_norm": 7.01528263092041,
"learning_rate": 0.0002747508559134122,
"loss": 2.4831,
"step": 1415
},
{
"epoch": 0.5655399014287847,
"grad_norm": 5.396809101104736,
"learning_rate": 0.0002745768374286223,
"loss": 2.4708,
"step": 1420
},
{
"epoch": 0.567531239109872,
"grad_norm": 5.1723833084106445,
"learning_rate": 0.00027440227683442266,
"loss": 2.4762,
"step": 1425
},
{
"epoch": 0.5695225767909593,
"grad_norm": 7.093824863433838,
"learning_rate": 0.0002742271748904324,
"loss": 2.6241,
"step": 1430
},
{
"epoch": 0.5715139144720466,
"grad_norm": 5.00604772567749,
"learning_rate": 0.0002740515323586264,
"loss": 2.6665,
"step": 1435
},
{
"epoch": 0.5735052521531339,
"grad_norm": 4.891804218292236,
"learning_rate": 0.0002738753500033318,
"loss": 2.5111,
"step": 1440
},
{
"epoch": 0.5754965898342211,
"grad_norm": 6.145954132080078,
"learning_rate": 0.0002736986285912249,
"loss": 2.5384,
"step": 1445
},
{
"epoch": 0.5774879275153084,
"grad_norm": 6.900878429412842,
"learning_rate": 0.0002735213688913279,
"loss": 2.6777,
"step": 1450
},
{
"epoch": 0.5794792651963957,
"grad_norm": 4.369724750518799,
"learning_rate": 0.00027334357167500524,
"loss": 2.5877,
"step": 1455
},
{
"epoch": 0.581470602877483,
"grad_norm": 5.170623302459717,
"learning_rate": 0.00027316523771596054,
"loss": 2.599,
"step": 1460
},
{
"epoch": 0.5834619405585703,
"grad_norm": 5.459385395050049,
"learning_rate": 0.000272986367790233,
"loss": 2.6303,
"step": 1465
},
{
"epoch": 0.5854532782396575,
"grad_norm": 5.773386001586914,
"learning_rate": 0.0002728069626761943,
"loss": 2.6477,
"step": 1470
},
{
"epoch": 0.5874446159207447,
"grad_norm": 5.394538402557373,
"learning_rate": 0.0002726270231545448,
"loss": 2.5973,
"step": 1475
},
{
"epoch": 0.589435953601832,
"grad_norm": 4.555385589599609,
"learning_rate": 0.0002724465500083106,
"loss": 2.5756,
"step": 1480
},
{
"epoch": 0.5914272912829193,
"grad_norm": 6.8372650146484375,
"learning_rate": 0.00027226554402283996,
"loss": 2.4608,
"step": 1485
},
{
"epoch": 0.5934186289640065,
"grad_norm": 8.024332046508789,
"learning_rate": 0.00027208400598579966,
"loss": 2.534,
"step": 1490
},
{
"epoch": 0.5954099666450938,
"grad_norm": 8.167983055114746,
"learning_rate": 0.0002719019366871719,
"loss": 2.6079,
"step": 1495
},
{
"epoch": 0.5974013043261811,
"grad_norm": 7.087464809417725,
"learning_rate": 0.00027171933691925066,
"loss": 2.5866,
"step": 1500
},
{
"epoch": 0.5993926420072684,
"grad_norm": 7.60006856918335,
"learning_rate": 0.00027153620747663826,
"loss": 2.5826,
"step": 1505
},
{
"epoch": 0.6013839796883557,
"grad_norm": 5.609342575073242,
"learning_rate": 0.0002713525491562421,
"loss": 2.4498,
"step": 1510
},
{
"epoch": 0.6033753173694429,
"grad_norm": 6.407932758331299,
"learning_rate": 0.00027116836275727096,
"loss": 2.7349,
"step": 1515
},
{
"epoch": 0.6053666550505302,
"grad_norm": 6.405856609344482,
"learning_rate": 0.00027098364908123156,
"loss": 2.4881,
"step": 1520
},
{
"epoch": 0.6073579927316175,
"grad_norm": 5.0175676345825195,
"learning_rate": 0.0002707984089319253,
"loss": 2.6345,
"step": 1525
},
{
"epoch": 0.6093493304127048,
"grad_norm": 4.912682056427002,
"learning_rate": 0.00027061264311544435,
"loss": 2.532,
"step": 1530
},
{
"epoch": 0.611340668093792,
"grad_norm": 5.729355812072754,
"learning_rate": 0.0002704263524401686,
"loss": 2.6031,
"step": 1535
},
{
"epoch": 0.6133320057748792,
"grad_norm": 6.606774806976318,
"learning_rate": 0.0002702395377167617,
"loss": 2.6403,
"step": 1540
},
{
"epoch": 0.6153233434559665,
"grad_norm": 7.419695854187012,
"learning_rate": 0.00027005219975816797,
"loss": 2.5945,
"step": 1545
},
{
"epoch": 0.6173146811370538,
"grad_norm": 9.066940307617188,
"learning_rate": 0.0002698643393796085,
"loss": 2.5915,
"step": 1550
},
{
"epoch": 0.6193060188181411,
"grad_norm": 5.113959312438965,
"learning_rate": 0.00026967595739857777,
"loss": 2.6204,
"step": 1555
},
{
"epoch": 0.6212973564992283,
"grad_norm": 5.126779556274414,
"learning_rate": 0.0002694870546348401,
"loss": 2.6282,
"step": 1560
},
{
"epoch": 0.6232886941803156,
"grad_norm": 4.795478820800781,
"learning_rate": 0.00026929763191042605,
"loss": 2.5167,
"step": 1565
},
{
"epoch": 0.6252800318614029,
"grad_norm": 5.510051727294922,
"learning_rate": 0.0002691076900496288,
"loss": 2.4141,
"step": 1570
},
{
"epoch": 0.6272713695424902,
"grad_norm": 5.027546405792236,
"learning_rate": 0.0002689172298790007,
"loss": 2.6739,
"step": 1575
},
{
"epoch": 0.6292627072235775,
"grad_norm": 6.95448637008667,
"learning_rate": 0.0002687262522273494,
"loss": 2.6941,
"step": 1580
},
{
"epoch": 0.6312540449046647,
"grad_norm": 4.615077972412109,
"learning_rate": 0.00026853475792573467,
"loss": 2.5628,
"step": 1585
},
{
"epoch": 0.633245382585752,
"grad_norm": 7.835186958312988,
"learning_rate": 0.00026834274780746436,
"loss": 2.6974,
"step": 1590
},
{
"epoch": 0.6352367202668392,
"grad_norm": 4.616403579711914,
"learning_rate": 0.00026815022270809094,
"loss": 2.7351,
"step": 1595
},
{
"epoch": 0.6372280579479265,
"grad_norm": 5.727982521057129,
"learning_rate": 0.000267957183465408,
"loss": 2.6939,
"step": 1600
},
{
"epoch": 0.6392193956290138,
"grad_norm": 4.361421585083008,
"learning_rate": 0.0002677636309194465,
"loss": 2.4927,
"step": 1605
},
{
"epoch": 0.641210733310101,
"grad_norm": 4.711188793182373,
"learning_rate": 0.00026756956591247086,
"loss": 2.4791,
"step": 1610
},
{
"epoch": 0.6432020709911883,
"grad_norm": 6.163144588470459,
"learning_rate": 0.00026737498928897574,
"loss": 2.663,
"step": 1615
},
{
"epoch": 0.6451934086722756,
"grad_norm": 7.343597888946533,
"learning_rate": 0.0002671799018956821,
"loss": 2.6147,
"step": 1620
},
{
"epoch": 0.6471847463533629,
"grad_norm": 8.960589408874512,
"learning_rate": 0.00026698430458153356,
"loss": 2.5842,
"step": 1625
},
{
"epoch": 0.6491760840344502,
"grad_norm": 7.366634845733643,
"learning_rate": 0.0002667881981976927,
"loss": 2.6003,
"step": 1630
},
{
"epoch": 0.6511674217155374,
"grad_norm": 6.907988548278809,
"learning_rate": 0.00026659158359753737,
"loss": 2.6356,
"step": 1635
},
{
"epoch": 0.6531587593966247,
"grad_norm": 6.1287641525268555,
"learning_rate": 0.00026639446163665705,
"loss": 2.5649,
"step": 1640
},
{
"epoch": 0.655150097077712,
"grad_norm": 5.350227355957031,
"learning_rate": 0.000266196833172849,
"loss": 2.6785,
"step": 1645
},
{
"epoch": 0.6571414347587993,
"grad_norm": 4.664305686950684,
"learning_rate": 0.00026599869906611455,
"loss": 2.5379,
"step": 1650
},
{
"epoch": 0.6591327724398864,
"grad_norm": 8.54836654663086,
"learning_rate": 0.0002658000601786554,
"loss": 2.6151,
"step": 1655
},
{
"epoch": 0.6611241101209737,
"grad_norm": 5.56497859954834,
"learning_rate": 0.00026560091737487,
"loss": 2.6609,
"step": 1660
},
{
"epoch": 0.663115447802061,
"grad_norm": 4.609283447265625,
"learning_rate": 0.0002654012715213493,
"loss": 2.5232,
"step": 1665
},
{
"epoch": 0.6651067854831483,
"grad_norm": 6.875670433044434,
"learning_rate": 0.00026520112348687373,
"loss": 2.5927,
"step": 1670
},
{
"epoch": 0.6670981231642356,
"grad_norm": 6.808399677276611,
"learning_rate": 0.0002650004741424088,
"loss": 2.6449,
"step": 1675
},
{
"epoch": 0.6690894608453228,
"grad_norm": 4.634307384490967,
"learning_rate": 0.0002647993243611016,
"loss": 2.572,
"step": 1680
},
{
"epoch": 0.6710807985264101,
"grad_norm": 9.085533142089844,
"learning_rate": 0.0002645976750182767,
"loss": 2.6565,
"step": 1685
},
{
"epoch": 0.6730721362074974,
"grad_norm": 6.228301048278809,
"learning_rate": 0.000264395526991433,
"loss": 2.5158,
"step": 1690
},
{
"epoch": 0.6750634738885847,
"grad_norm": 4.803519248962402,
"learning_rate": 0.00026419288116023897,
"loss": 2.4979,
"step": 1695
},
{
"epoch": 0.677054811569672,
"grad_norm": 6.3344621658325195,
"learning_rate": 0.0002639897384065298,
"loss": 2.6457,
"step": 1700
},
{
"epoch": 0.6790461492507592,
"grad_norm": 5.770046234130859,
"learning_rate": 0.00026378609961430266,
"loss": 2.5031,
"step": 1705
},
{
"epoch": 0.6810374869318465,
"grad_norm": 5.835142135620117,
"learning_rate": 0.0002635819656697137,
"loss": 2.5488,
"step": 1710
},
{
"epoch": 0.6830288246129338,
"grad_norm": 5.059028625488281,
"learning_rate": 0.0002633773374610735,
"loss": 2.6686,
"step": 1715
},
{
"epoch": 0.685020162294021,
"grad_norm": 6.267749309539795,
"learning_rate": 0.0002631722158788435,
"loss": 2.6308,
"step": 1720
},
{
"epoch": 0.6870114999751082,
"grad_norm": 4.403576374053955,
"learning_rate": 0.0002629666018156322,
"loss": 2.6768,
"step": 1725
},
{
"epoch": 0.6890028376561955,
"grad_norm": 6.0533905029296875,
"learning_rate": 0.0002627604961661912,
"loss": 2.5665,
"step": 1730
},
{
"epoch": 0.6909941753372828,
"grad_norm": 6.283039569854736,
"learning_rate": 0.0002625538998274112,
"loss": 2.6614,
"step": 1735
},
{
"epoch": 0.6929855130183701,
"grad_norm": 5.365102291107178,
"learning_rate": 0.00026234681369831825,
"loss": 2.6992,
"step": 1740
},
{
"epoch": 0.6949768506994574,
"grad_norm": 4.661876201629639,
"learning_rate": 0.00026213923868006973,
"loss": 2.6264,
"step": 1745
},
{
"epoch": 0.6969681883805446,
"grad_norm": 5.980187892913818,
"learning_rate": 0.0002619311756759505,
"loss": 2.5872,
"step": 1750
},
{
"epoch": 0.6989595260616319,
"grad_norm": 6.6604204177856445,
"learning_rate": 0.0002617226255913689,
"loss": 2.5463,
"step": 1755
},
{
"epoch": 0.7009508637427192,
"grad_norm": 5.904111385345459,
"learning_rate": 0.00026151358933385293,
"loss": 2.6829,
"step": 1760
},
{
"epoch": 0.7029422014238065,
"grad_norm": 4.411726951599121,
"learning_rate": 0.00026130406781304627,
"loss": 2.5868,
"step": 1765
},
{
"epoch": 0.7049335391048938,
"grad_norm": 5.401057720184326,
"learning_rate": 0.00026109406194070405,
"loss": 2.4201,
"step": 1770
},
{
"epoch": 0.706924876785981,
"grad_norm": 10.872917175292969,
"learning_rate": 0.0002608835726306893,
"loss": 2.6233,
"step": 1775
},
{
"epoch": 0.7089162144670682,
"grad_norm": 5.127570629119873,
"learning_rate": 0.0002606726007989687,
"loss": 2.4135,
"step": 1780
},
{
"epoch": 0.7109075521481555,
"grad_norm": 4.496290683746338,
"learning_rate": 0.0002604611473636088,
"loss": 2.6121,
"step": 1785
},
{
"epoch": 0.7128988898292428,
"grad_norm": 5.2560715675354,
"learning_rate": 0.00026024921324477166,
"loss": 2.5794,
"step": 1790
},
{
"epoch": 0.71489022751033,
"grad_norm": 4.1616411209106445,
"learning_rate": 0.00026003679936471136,
"loss": 2.4883,
"step": 1795
},
{
"epoch": 0.7168815651914173,
"grad_norm": 4.719475269317627,
"learning_rate": 0.0002598239066477695,
"loss": 2.5903,
"step": 1800
},
{
"epoch": 0.7188729028725046,
"grad_norm": 7.156427383422852,
"learning_rate": 0.00025961053602037144,
"loss": 2.5673,
"step": 1805
},
{
"epoch": 0.7208642405535919,
"grad_norm": 6.929321765899658,
"learning_rate": 0.00025939668841102237,
"loss": 2.521,
"step": 1810
},
{
"epoch": 0.7228555782346792,
"grad_norm": 6.155335903167725,
"learning_rate": 0.0002591823647503029,
"loss": 2.5703,
"step": 1815
},
{
"epoch": 0.7248469159157664,
"grad_norm": 5.760573863983154,
"learning_rate": 0.00025896756597086526,
"loss": 2.6104,
"step": 1820
},
{
"epoch": 0.7268382535968537,
"grad_norm": 6.244937419891357,
"learning_rate": 0.00025875229300742947,
"loss": 2.6097,
"step": 1825
},
{
"epoch": 0.728829591277941,
"grad_norm": 5.067662239074707,
"learning_rate": 0.00025853654679677867,
"loss": 2.556,
"step": 1830
},
{
"epoch": 0.7308209289590283,
"grad_norm": 5.625128269195557,
"learning_rate": 0.0002583203282777555,
"loss": 2.7103,
"step": 1835
},
{
"epoch": 0.7328122666401155,
"grad_norm": 5.45819616317749,
"learning_rate": 0.0002581036383912581,
"loss": 2.6172,
"step": 1840
},
{
"epoch": 0.7348036043212027,
"grad_norm": 4.656463146209717,
"learning_rate": 0.0002578864780802355,
"loss": 2.6565,
"step": 1845
},
{
"epoch": 0.73679494200229,
"grad_norm": 5.9828715324401855,
"learning_rate": 0.000257668848289684,
"loss": 2.5157,
"step": 1850
},
{
"epoch": 0.7387862796833773,
"grad_norm": 5.619359970092773,
"learning_rate": 0.00025745074996664293,
"loss": 2.4723,
"step": 1855
},
{
"epoch": 0.7407776173644646,
"grad_norm": 5.359917163848877,
"learning_rate": 0.0002572321840601904,
"loss": 2.5944,
"step": 1860
},
{
"epoch": 0.7427689550455518,
"grad_norm": 4.9695353507995605,
"learning_rate": 0.0002570131515214393,
"loss": 2.5379,
"step": 1865
},
{
"epoch": 0.7447602927266391,
"grad_norm": 6.0350341796875,
"learning_rate": 0.000256793653303533,
"loss": 2.5925,
"step": 1870
},
{
"epoch": 0.7467516304077264,
"grad_norm": 5.810332298278809,
"learning_rate": 0.0002565736903616416,
"loss": 2.681,
"step": 1875
},
{
"epoch": 0.7487429680888137,
"grad_norm": 4.983578205108643,
"learning_rate": 0.0002563532636529573,
"loss": 2.6438,
"step": 1880
},
{
"epoch": 0.750734305769901,
"grad_norm": 4.9191131591796875,
"learning_rate": 0.0002561323741366904,
"loss": 2.6089,
"step": 1885
},
{
"epoch": 0.7527256434509882,
"grad_norm": 6.884909152984619,
"learning_rate": 0.0002559110227740652,
"loss": 2.6537,
"step": 1890
},
{
"epoch": 0.7547169811320755,
"grad_norm": 5.462011814117432,
"learning_rate": 0.0002556892105283158,
"loss": 2.4471,
"step": 1895
},
{
"epoch": 0.7567083188131627,
"grad_norm": 5.4949822425842285,
"learning_rate": 0.0002554669383646819,
"loss": 2.7252,
"step": 1900
},
{
"epoch": 0.75869965649425,
"grad_norm": 4.802751064300537,
"learning_rate": 0.00025524420725040457,
"loss": 2.4402,
"step": 1905
},
{
"epoch": 0.7606909941753373,
"grad_norm": 7.37558126449585,
"learning_rate": 0.000255021018154722,
"loss": 2.7359,
"step": 1910
},
{
"epoch": 0.7626823318564245,
"grad_norm": 6.215965747833252,
"learning_rate": 0.00025479737204886537,
"loss": 2.6348,
"step": 1915
},
{
"epoch": 0.7646736695375118,
"grad_norm": 5.106846809387207,
"learning_rate": 0.00025457326990605466,
"loss": 2.6013,
"step": 1920
},
{
"epoch": 0.7666650072185991,
"grad_norm": 6.881631374359131,
"learning_rate": 0.0002543487127014942,
"loss": 2.6093,
"step": 1925
},
{
"epoch": 0.7686563448996864,
"grad_norm": 6.443165302276611,
"learning_rate": 0.0002541237014123686,
"loss": 2.6147,
"step": 1930
},
{
"epoch": 0.7706476825807737,
"grad_norm": 6.010848522186279,
"learning_rate": 0.00025389823701783853,
"loss": 2.597,
"step": 1935
},
{
"epoch": 0.7726390202618609,
"grad_norm": 5.879451274871826,
"learning_rate": 0.00025367232049903647,
"loss": 2.7176,
"step": 1940
},
{
"epoch": 0.7746303579429482,
"grad_norm": 3.990070343017578,
"learning_rate": 0.0002534459528390622,
"loss": 2.5272,
"step": 1945
},
{
"epoch": 0.7766216956240355,
"grad_norm": 4.5337629318237305,
"learning_rate": 0.0002532191350229788,
"loss": 2.5446,
"step": 1950
},
{
"epoch": 0.7786130333051228,
"grad_norm": 6.173877716064453,
"learning_rate": 0.0002529918680378082,
"loss": 2.5218,
"step": 1955
},
{
"epoch": 0.7806043709862099,
"grad_norm": 6.064763069152832,
"learning_rate": 0.0002527641528725269,
"loss": 2.6244,
"step": 1960
},
{
"epoch": 0.7825957086672972,
"grad_norm": 5.518800258636475,
"learning_rate": 0.0002525359905180618,
"loss": 2.6561,
"step": 1965
},
{
"epoch": 0.7845870463483845,
"grad_norm": 5.687634468078613,
"learning_rate": 0.0002523073819672858,
"loss": 2.6324,
"step": 1970
},
{
"epoch": 0.7865783840294718,
"grad_norm": 6.286370754241943,
"learning_rate": 0.00025207832821501336,
"loss": 2.6164,
"step": 1975
},
{
"epoch": 0.7885697217105591,
"grad_norm": 7.250809192657471,
"learning_rate": 0.0002518488302579963,
"loss": 2.6098,
"step": 1980
},
{
"epoch": 0.7905610593916463,
"grad_norm": 5.801024436950684,
"learning_rate": 0.0002516188890949196,
"loss": 2.505,
"step": 1985
},
{
"epoch": 0.7925523970727336,
"grad_norm": 7.584057807922363,
"learning_rate": 0.00025138850572639665,
"loss": 2.5602,
"step": 1990
},
{
"epoch": 0.7945437347538209,
"grad_norm": 11.833077430725098,
"learning_rate": 0.0002511576811549654,
"loss": 2.4686,
"step": 1995
},
{
"epoch": 0.7965350724349082,
"grad_norm": 6.000141620635986,
"learning_rate": 0.00025092641638508353,
"loss": 2.4901,
"step": 2000
},
{
"epoch": 0.7985264101159955,
"grad_norm": 6.463593006134033,
"learning_rate": 0.0002506947124231244,
"loss": 2.5631,
"step": 2005
},
{
"epoch": 0.8005177477970827,
"grad_norm": 6.079593658447266,
"learning_rate": 0.0002504625702773725,
"loss": 2.6703,
"step": 2010
},
{
"epoch": 0.80250908547817,
"grad_norm": 6.521668910980225,
"learning_rate": 0.0002502299909580193,
"loss": 2.4696,
"step": 2015
},
{
"epoch": 0.8045004231592572,
"grad_norm": 5.687950611114502,
"learning_rate": 0.0002499969754771584,
"loss": 2.7125,
"step": 2020
},
{
"epoch": 0.8064917608403445,
"grad_norm": 4.566808223724365,
"learning_rate": 0.00024976352484878164,
"loss": 2.7074,
"step": 2025
},
{
"epoch": 0.8084830985214317,
"grad_norm": 5.849015712738037,
"learning_rate": 0.0002495296400887743,
"loss": 2.7419,
"step": 2030
},
{
"epoch": 0.810474436202519,
"grad_norm": 5.121682167053223,
"learning_rate": 0.0002492953222149109,
"loss": 2.5039,
"step": 2035
},
{
"epoch": 0.8124657738836063,
"grad_norm": 5.572865009307861,
"learning_rate": 0.0002490605722468507,
"loss": 2.6298,
"step": 2040
},
{
"epoch": 0.8144571115646936,
"grad_norm": 5.498860836029053,
"learning_rate": 0.00024882539120613323,
"loss": 2.5468,
"step": 2045
},
{
"epoch": 0.8164484492457809,
"grad_norm": 5.025686264038086,
"learning_rate": 0.0002485897801161739,
"loss": 2.5345,
"step": 2050
},
{
"epoch": 0.8184397869268681,
"grad_norm": 4.758937835693359,
"learning_rate": 0.00024835374000225954,
"loss": 2.5416,
"step": 2055
},
{
"epoch": 0.8204311246079554,
"grad_norm": 4.406644821166992,
"learning_rate": 0.00024811727189154396,
"loss": 2.5523,
"step": 2060
},
{
"epoch": 0.8224224622890427,
"grad_norm": 7.341763019561768,
"learning_rate": 0.0002478803768130433,
"loss": 2.4763,
"step": 2065
},
{
"epoch": 0.82441379997013,
"grad_norm": 5.85587739944458,
"learning_rate": 0.00024764305579763184,
"loss": 2.5462,
"step": 2070
},
{
"epoch": 0.8264051376512173,
"grad_norm": 4.7058796882629395,
"learning_rate": 0.00024740530987803726,
"loss": 2.6977,
"step": 2075
},
{
"epoch": 0.8283964753323044,
"grad_norm": 6.2724995613098145,
"learning_rate": 0.0002471671400888364,
"loss": 2.5808,
"step": 2080
},
{
"epoch": 0.8303878130133917,
"grad_norm": 6.3840179443359375,
"learning_rate": 0.0002469285474664505,
"loss": 2.4861,
"step": 2085
},
{
"epoch": 0.832379150694479,
"grad_norm": 6.859890937805176,
"learning_rate": 0.00024668953304914077,
"loss": 2.6549,
"step": 2090
},
{
"epoch": 0.8343704883755663,
"grad_norm": 5.382211685180664,
"learning_rate": 0.000246450097877004,
"loss": 2.6052,
"step": 2095
},
{
"epoch": 0.8363618260566535,
"grad_norm": 6.212277412414551,
"learning_rate": 0.000246210242991968,
"loss": 2.5116,
"step": 2100
},
{
"epoch": 0.8383531637377408,
"grad_norm": 5.71286678314209,
"learning_rate": 0.0002459699694377868,
"loss": 2.4863,
"step": 2105
},
{
"epoch": 0.8403445014188281,
"grad_norm": 4.8305439949035645,
"learning_rate": 0.0002457292782600364,
"loss": 2.7366,
"step": 2110
},
{
"epoch": 0.8423358390999154,
"grad_norm": 5.540607452392578,
"learning_rate": 0.0002454881705061103,
"loss": 2.4626,
"step": 2115
},
{
"epoch": 0.8443271767810027,
"grad_norm": 4.778063774108887,
"learning_rate": 0.0002452466472252145,
"loss": 2.5505,
"step": 2120
},
{
"epoch": 0.8463185144620899,
"grad_norm": 6.163254261016846,
"learning_rate": 0.00024500470946836345,
"loss": 2.5981,
"step": 2125
},
{
"epoch": 0.8483098521431772,
"grad_norm": 5.865226745605469,
"learning_rate": 0.00024476235828837503,
"loss": 2.6405,
"step": 2130
},
{
"epoch": 0.8503011898242645,
"grad_norm": 5.853811740875244,
"learning_rate": 0.0002445195947398664,
"loss": 2.4575,
"step": 2135
},
{
"epoch": 0.8522925275053517,
"grad_norm": 4.554398059844971,
"learning_rate": 0.000244276419879249,
"loss": 2.6075,
"step": 2140
},
{
"epoch": 0.854283865186439,
"grad_norm": 5.9426493644714355,
"learning_rate": 0.00024403283476472415,
"loss": 2.6052,
"step": 2145
},
{
"epoch": 0.8562752028675262,
"grad_norm": 4.86089563369751,
"learning_rate": 0.00024378884045627854,
"loss": 2.6822,
"step": 2150
},
{
"epoch": 0.8582665405486135,
"grad_norm": 5.2081780433654785,
"learning_rate": 0.00024354443801567942,
"loss": 2.459,
"step": 2155
},
{
"epoch": 0.8602578782297008,
"grad_norm": 4.458292484283447,
"learning_rate": 0.00024329962850647016,
"loss": 2.476,
"step": 2160
},
{
"epoch": 0.8622492159107881,
"grad_norm": 5.3445539474487305,
"learning_rate": 0.0002430544129939654,
"loss": 2.5372,
"step": 2165
},
{
"epoch": 0.8642405535918753,
"grad_norm": 5.6220855712890625,
"learning_rate": 0.0002428087925452467,
"loss": 2.688,
"step": 2170
},
{
"epoch": 0.8662318912729626,
"grad_norm": 5.187811851501465,
"learning_rate": 0.00024256276822915768,
"loss": 2.4423,
"step": 2175
},
{
"epoch": 0.8682232289540499,
"grad_norm": 8.015525817871094,
"learning_rate": 0.00024231634111629934,
"loss": 2.5889,
"step": 2180
},
{
"epoch": 0.8702145666351372,
"grad_norm": 6.496338844299316,
"learning_rate": 0.00024206951227902563,
"loss": 2.5738,
"step": 2185
},
{
"epoch": 0.8722059043162245,
"grad_norm": 4.662912368774414,
"learning_rate": 0.0002418222827914386,
"loss": 2.7525,
"step": 2190
},
{
"epoch": 0.8741972419973117,
"grad_norm": 6.552118301391602,
"learning_rate": 0.0002415746537293837,
"loss": 2.5813,
"step": 2195
},
{
"epoch": 0.8761885796783989,
"grad_norm": 5.152313709259033,
"learning_rate": 0.00024132662617044532,
"loss": 2.5728,
"step": 2200
},
{
"epoch": 0.8781799173594862,
"grad_norm": 4.024521827697754,
"learning_rate": 0.00024107820119394175,
"loss": 2.5748,
"step": 2205
},
{
"epoch": 0.8801712550405735,
"grad_norm": 3.983680248260498,
"learning_rate": 0.00024082937988092096,
"loss": 2.5401,
"step": 2210
},
{
"epoch": 0.8821625927216608,
"grad_norm": 5.457140922546387,
"learning_rate": 0.0002405801633141553,
"loss": 2.5143,
"step": 2215
},
{
"epoch": 0.884153930402748,
"grad_norm": 4.946249008178711,
"learning_rate": 0.00024033055257813732,
"loss": 2.6762,
"step": 2220
},
{
"epoch": 0.8861452680838353,
"grad_norm": 4.660250663757324,
"learning_rate": 0.0002400805487590749,
"loss": 2.5675,
"step": 2225
},
{
"epoch": 0.8881366057649226,
"grad_norm": 7.786635398864746,
"learning_rate": 0.00023983015294488623,
"loss": 2.5946,
"step": 2230
},
{
"epoch": 0.8901279434460099,
"grad_norm": 5.1657395362854,
"learning_rate": 0.00023957936622519542,
"loss": 2.6189,
"step": 2235
},
{
"epoch": 0.8921192811270972,
"grad_norm": 5.05390739440918,
"learning_rate": 0.00023932818969132777,
"loss": 2.5522,
"step": 2240
},
{
"epoch": 0.8941106188081844,
"grad_norm": 7.363577365875244,
"learning_rate": 0.00023907662443630465,
"loss": 2.589,
"step": 2245
},
{
"epoch": 0.8961019564892717,
"grad_norm": 6.585803031921387,
"learning_rate": 0.00023882467155483927,
"loss": 2.5239,
"step": 2250
},
{
"epoch": 0.898093294170359,
"grad_norm": 6.26939582824707,
"learning_rate": 0.0002385723321433314,
"loss": 2.5722,
"step": 2255
},
{
"epoch": 0.9000846318514462,
"grad_norm": 5.3172526359558105,
"learning_rate": 0.00023831960729986292,
"loss": 2.5552,
"step": 2260
},
{
"epoch": 0.9020759695325334,
"grad_norm": 4.980822563171387,
"learning_rate": 0.00023806649812419306,
"loss": 2.6506,
"step": 2265
},
{
"epoch": 0.9040673072136207,
"grad_norm": 5.251991271972656,
"learning_rate": 0.00023781300571775333,
"loss": 2.613,
"step": 2270
},
{
"epoch": 0.906058644894708,
"grad_norm": 4.616904258728027,
"learning_rate": 0.00023755913118364306,
"loss": 2.6393,
"step": 2275
},
{
"epoch": 0.9080499825757953,
"grad_norm": 5.337674617767334,
"learning_rate": 0.00023730487562662438,
"loss": 2.4195,
"step": 2280
},
{
"epoch": 0.9100413202568826,
"grad_norm": 6.516839504241943,
"learning_rate": 0.00023705024015311751,
"loss": 2.5637,
"step": 2285
},
{
"epoch": 0.9120326579379698,
"grad_norm": 5.19506311416626,
"learning_rate": 0.00023679522587119586,
"loss": 2.4586,
"step": 2290
},
{
"epoch": 0.9140239956190571,
"grad_norm": 4.948032855987549,
"learning_rate": 0.00023653983389058128,
"loss": 2.5845,
"step": 2295
},
{
"epoch": 0.9160153333001444,
"grad_norm": 6.443713665008545,
"learning_rate": 0.0002362840653226394,
"loss": 2.664,
"step": 2300
},
{
"epoch": 0.9180066709812317,
"grad_norm": 4.868452548980713,
"learning_rate": 0.00023602792128037426,
"loss": 2.4598,
"step": 2305
},
{
"epoch": 0.919998008662319,
"grad_norm": 5.261866092681885,
"learning_rate": 0.00023577140287842408,
"loss": 2.5356,
"step": 2310
},
{
"epoch": 0.9219893463434062,
"grad_norm": 6.211141586303711,
"learning_rate": 0.00023551451123305614,
"loss": 2.5352,
"step": 2315
},
{
"epoch": 0.9239806840244934,
"grad_norm": 4.445993423461914,
"learning_rate": 0.0002352572474621618,
"loss": 2.5963,
"step": 2320
},
{
"epoch": 0.9259720217055807,
"grad_norm": 4.683290481567383,
"learning_rate": 0.00023499961268525188,
"loss": 2.5556,
"step": 2325
},
{
"epoch": 0.927963359386668,
"grad_norm": 5.033156871795654,
"learning_rate": 0.0002347416080234515,
"loss": 2.6199,
"step": 2330
},
{
"epoch": 0.9299546970677552,
"grad_norm": 5.339117527008057,
"learning_rate": 0.00023448323459949566,
"loss": 2.4189,
"step": 2335
},
{
"epoch": 0.9319460347488425,
"grad_norm": 7.870457649230957,
"learning_rate": 0.0002342244935377238,
"loss": 2.6203,
"step": 2340
},
{
"epoch": 0.9339373724299298,
"grad_norm": 9.246062278747559,
"learning_rate": 0.00023396538596407525,
"loss": 2.5481,
"step": 2345
},
{
"epoch": 0.9359287101110171,
"grad_norm": 5.908375263214111,
"learning_rate": 0.00023370591300608438,
"loss": 2.5498,
"step": 2350
},
{
"epoch": 0.9379200477921044,
"grad_norm": 6.05574893951416,
"learning_rate": 0.0002334460757928754,
"loss": 2.5629,
"step": 2355
},
{
"epoch": 0.9399113854731916,
"grad_norm": 5.820539951324463,
"learning_rate": 0.00023318587545515768,
"loss": 2.525,
"step": 2360
},
{
"epoch": 0.9419027231542789,
"grad_norm": 7.062961101531982,
"learning_rate": 0.00023292531312522083,
"loss": 2.7078,
"step": 2365
},
{
"epoch": 0.9438940608353662,
"grad_norm": 6.001385688781738,
"learning_rate": 0.00023266438993692964,
"loss": 2.6176,
"step": 2370
},
{
"epoch": 0.9458853985164535,
"grad_norm": 5.372692584991455,
"learning_rate": 0.0002324031070257192,
"loss": 2.6501,
"step": 2375
},
{
"epoch": 0.9478767361975406,
"grad_norm": 5.991538047790527,
"learning_rate": 0.00023214146552859,
"loss": 2.436,
"step": 2380
},
{
"epoch": 0.9498680738786279,
"grad_norm": 4.772813320159912,
"learning_rate": 0.0002318794665841029,
"loss": 2.5451,
"step": 2385
},
{
"epoch": 0.9518594115597152,
"grad_norm": 5.644576072692871,
"learning_rate": 0.00023161711133237438,
"loss": 2.5421,
"step": 2390
},
{
"epoch": 0.9538507492408025,
"grad_norm": 6.208878517150879,
"learning_rate": 0.00023135440091507124,
"loss": 2.6161,
"step": 2395
},
{
"epoch": 0.9558420869218898,
"grad_norm": 4.382922649383545,
"learning_rate": 0.00023109133647540602,
"loss": 2.5102,
"step": 2400
},
{
"epoch": 0.957833424602977,
"grad_norm": 5.279038906097412,
"learning_rate": 0.00023082791915813165,
"loss": 2.7259,
"step": 2405
},
{
"epoch": 0.9598247622840643,
"grad_norm": 4.535640716552734,
"learning_rate": 0.00023056415010953667,
"loss": 2.6417,
"step": 2410
},
{
"epoch": 0.9618160999651516,
"grad_norm": 8.062291145324707,
"learning_rate": 0.0002303000304774403,
"loss": 2.439,
"step": 2415
},
{
"epoch": 0.9638074376462389,
"grad_norm": 5.5467305183410645,
"learning_rate": 0.00023003556141118738,
"loss": 2.6322,
"step": 2420
},
{
"epoch": 0.9657987753273262,
"grad_norm": 4.954715251922607,
"learning_rate": 0.00022977074406164318,
"loss": 2.5879,
"step": 2425
},
{
"epoch": 0.9677901130084134,
"grad_norm": 4.517293930053711,
"learning_rate": 0.00022950557958118878,
"loss": 2.4739,
"step": 2430
},
{
"epoch": 0.9697814506895007,
"grad_norm": 7.081440448760986,
"learning_rate": 0.00022924006912371567,
"loss": 2.6461,
"step": 2435
},
{
"epoch": 0.9717727883705879,
"grad_norm": 6.215746879577637,
"learning_rate": 0.00022897421384462092,
"loss": 2.4895,
"step": 2440
},
{
"epoch": 0.9737641260516752,
"grad_norm": 6.380356788635254,
"learning_rate": 0.0002287080149008022,
"loss": 2.6011,
"step": 2445
},
{
"epoch": 0.9757554637327625,
"grad_norm": 5.506686687469482,
"learning_rate": 0.0002284414734506527,
"loss": 2.5588,
"step": 2450
},
{
"epoch": 0.9777468014138497,
"grad_norm": 5.08485746383667,
"learning_rate": 0.00022817459065405593,
"loss": 2.5556,
"step": 2455
},
{
"epoch": 0.979738139094937,
"grad_norm": 4.853709697723389,
"learning_rate": 0.00022790736767238083,
"loss": 2.5654,
"step": 2460
},
{
"epoch": 0.9817294767760243,
"grad_norm": 4.686211109161377,
"learning_rate": 0.0002276398056684769,
"loss": 2.4522,
"step": 2465
},
{
"epoch": 0.9837208144571116,
"grad_norm": 4.961901664733887,
"learning_rate": 0.0002273719058066686,
"loss": 2.5592,
"step": 2470
},
{
"epoch": 0.9857121521381988,
"grad_norm": 5.117202281951904,
"learning_rate": 0.00022710366925275086,
"loss": 2.4735,
"step": 2475
},
{
"epoch": 0.9877034898192861,
"grad_norm": 5.865626811981201,
"learning_rate": 0.00022683509717398367,
"loss": 2.5008,
"step": 2480
},
{
"epoch": 0.9896948275003734,
"grad_norm": 6.076361179351807,
"learning_rate": 0.0002265661907390871,
"loss": 2.5271,
"step": 2485
},
{
"epoch": 0.9916861651814607,
"grad_norm": 6.400584697723389,
"learning_rate": 0.00022629695111823624,
"loss": 2.574,
"step": 2490
},
{
"epoch": 0.993677502862548,
"grad_norm": 5.469146728515625,
"learning_rate": 0.000226027379483056,
"loss": 2.5452,
"step": 2495
},
{
"epoch": 0.9956688405436351,
"grad_norm": 5.671192646026611,
"learning_rate": 0.0002257574770066162,
"loss": 2.5663,
"step": 2500
},
{
"epoch": 0.9976601782247224,
"grad_norm": 5.9548726081848145,
"learning_rate": 0.00022548724486342624,
"loss": 2.4705,
"step": 2505
},
{
"epoch": 0.9996515159058097,
"grad_norm": 5.226431846618652,
"learning_rate": 0.00022521668422943017,
"loss": 2.4196,
"step": 2510
},
{
"epoch": 1.001642853586897,
"grad_norm": 4.7078166007995605,
"learning_rate": 0.00022494579628200146,
"loss": 2.3862,
"step": 2515
},
{
"epoch": 1.0036341912679843,
"grad_norm": 8.271881103515625,
"learning_rate": 0.00022467458219993793,
"loss": 2.2516,
"step": 2520
},
{
"epoch": 1.0056255289490716,
"grad_norm": 6.609498023986816,
"learning_rate": 0.00022440304316345664,
"loss": 2.0925,
"step": 2525
},
{
"epoch": 1.0076168666301588,
"grad_norm": 7.699496269226074,
"learning_rate": 0.00022413118035418867,
"loss": 2.2226,
"step": 2530
},
{
"epoch": 1.009608204311246,
"grad_norm": 6.391448974609375,
"learning_rate": 0.00022385899495517411,
"loss": 2.1925,
"step": 2535
},
{
"epoch": 1.0115995419923334,
"grad_norm": 5.20413064956665,
"learning_rate": 0.00022358648815085674,
"loss": 2.0784,
"step": 2540
},
{
"epoch": 1.0135908796734205,
"grad_norm": 5.514545917510986,
"learning_rate": 0.00022331366112707905,
"loss": 2.1903,
"step": 2545
},
{
"epoch": 1.015582217354508,
"grad_norm": 6.206656455993652,
"learning_rate": 0.00022304051507107691,
"loss": 2.0112,
"step": 2550
},
{
"epoch": 1.017573555035595,
"grad_norm": 5.609284400939941,
"learning_rate": 0.00022276705117147463,
"loss": 2.1828,
"step": 2555
},
{
"epoch": 1.0195648927166825,
"grad_norm": 6.571842193603516,
"learning_rate": 0.0002224932706182795,
"loss": 2.2427,
"step": 2560
},
{
"epoch": 1.0215562303977697,
"grad_norm": 6.576747894287109,
"learning_rate": 0.00022221917460287684,
"loss": 2.0657,
"step": 2565
},
{
"epoch": 1.023547568078857,
"grad_norm": 6.2041754722595215,
"learning_rate": 0.0002219447643180248,
"loss": 2.1984,
"step": 2570
},
{
"epoch": 1.0255389057599442,
"grad_norm": 6.46384334564209,
"learning_rate": 0.00022167004095784889,
"loss": 2.2014,
"step": 2575
},
{
"epoch": 1.0275302434410316,
"grad_norm": 5.901404857635498,
"learning_rate": 0.00022139500571783727,
"loss": 2.1411,
"step": 2580
},
{
"epoch": 1.0295215811221188,
"grad_norm": 6.320353984832764,
"learning_rate": 0.00022111965979483512,
"loss": 2.1498,
"step": 2585
},
{
"epoch": 1.031512918803206,
"grad_norm": 5.76351261138916,
"learning_rate": 0.00022084400438703966,
"loss": 2.1994,
"step": 2590
},
{
"epoch": 1.0335042564842933,
"grad_norm": 6.784847736358643,
"learning_rate": 0.00022056804069399473,
"loss": 2.1888,
"step": 2595
},
{
"epoch": 1.0354955941653805,
"grad_norm": 5.914264678955078,
"learning_rate": 0.00022029176991658587,
"loss": 2.1845,
"step": 2600
},
{
"epoch": 1.037486931846468,
"grad_norm": 5.347561359405518,
"learning_rate": 0.00022001519325703486,
"loss": 2.1797,
"step": 2605
},
{
"epoch": 1.039478269527555,
"grad_norm": 7.8450212478637695,
"learning_rate": 0.0002197383119188946,
"loss": 2.2326,
"step": 2610
},
{
"epoch": 1.0414696072086425,
"grad_norm": 6.4014105796813965,
"learning_rate": 0.00021946112710704375,
"loss": 2.1578,
"step": 2615
},
{
"epoch": 1.0434609448897296,
"grad_norm": 6.188974380493164,
"learning_rate": 0.00021918364002768162,
"loss": 2.163,
"step": 2620
},
{
"epoch": 1.045452282570817,
"grad_norm": 7.119470119476318,
"learning_rate": 0.00021890585188832288,
"loss": 2.131,
"step": 2625
},
{
"epoch": 1.0474436202519042,
"grad_norm": 6.069831848144531,
"learning_rate": 0.00021862776389779226,
"loss": 2.1385,
"step": 2630
},
{
"epoch": 1.0494349579329916,
"grad_norm": 7.496238708496094,
"learning_rate": 0.00021834937726621933,
"loss": 2.1297,
"step": 2635
},
{
"epoch": 1.0514262956140787,
"grad_norm": 5.977787017822266,
"learning_rate": 0.0002180706932050333,
"loss": 2.16,
"step": 2640
},
{
"epoch": 1.0534176332951661,
"grad_norm": 7.515745162963867,
"learning_rate": 0.0002177917129269576,
"loss": 2.1703,
"step": 2645
},
{
"epoch": 1.0554089709762533,
"grad_norm": 6.940252780914307,
"learning_rate": 0.00021751243764600467,
"loss": 2.2871,
"step": 2650
},
{
"epoch": 1.0574003086573405,
"grad_norm": 6.161179065704346,
"learning_rate": 0.0002172328685774707,
"loss": 2.1182,
"step": 2655
},
{
"epoch": 1.0593916463384279,
"grad_norm": 6.096395969390869,
"learning_rate": 0.0002169530069379304,
"loss": 2.1868,
"step": 2660
},
{
"epoch": 1.061382984019515,
"grad_norm": 6.321666717529297,
"learning_rate": 0.00021667285394523153,
"loss": 2.3034,
"step": 2665
},
{
"epoch": 1.0633743217006024,
"grad_norm": 5.823657512664795,
"learning_rate": 0.00021639241081848978,
"loss": 2.1462,
"step": 2670
},
{
"epoch": 1.0653656593816896,
"grad_norm": 6.120224475860596,
"learning_rate": 0.0002161116787780834,
"loss": 2.2264,
"step": 2675
},
{
"epoch": 1.067356997062777,
"grad_norm": 6.145210266113281,
"learning_rate": 0.00021583065904564767,
"loss": 2.1483,
"step": 2680
},
{
"epoch": 1.0693483347438641,
"grad_norm": 5.247588157653809,
"learning_rate": 0.00021554935284407017,
"loss": 2.2671,
"step": 2685
},
{
"epoch": 1.0713396724249515,
"grad_norm": 5.9891252517700195,
"learning_rate": 0.00021526776139748472,
"loss": 2.0288,
"step": 2690
},
{
"epoch": 1.0733310101060387,
"grad_norm": 5.2879204750061035,
"learning_rate": 0.00021498588593126666,
"loss": 2.2455,
"step": 2695
},
{
"epoch": 1.075322347787126,
"grad_norm": 5.61940860748291,
"learning_rate": 0.00021470372767202713,
"loss": 2.0936,
"step": 2700
},
{
"epoch": 1.0773136854682133,
"grad_norm": 7.891155242919922,
"learning_rate": 0.00021442128784760796,
"loss": 2.2633,
"step": 2705
},
{
"epoch": 1.0793050231493004,
"grad_norm": 4.800739765167236,
"learning_rate": 0.00021413856768707617,
"loss": 2.0494,
"step": 2710
},
{
"epoch": 1.0812963608303878,
"grad_norm": 6.422128200531006,
"learning_rate": 0.00021385556842071874,
"loss": 2.2881,
"step": 2715
},
{
"epoch": 1.083287698511475,
"grad_norm": 7.9622297286987305,
"learning_rate": 0.0002135722912800372,
"loss": 2.2561,
"step": 2720
},
{
"epoch": 1.0852790361925624,
"grad_norm": 5.626192092895508,
"learning_rate": 0.00021328873749774225,
"loss": 2.1538,
"step": 2725
},
{
"epoch": 1.0872703738736496,
"grad_norm": 6.245817184448242,
"learning_rate": 0.00021300490830774847,
"loss": 2.2026,
"step": 2730
},
{
"epoch": 1.089261711554737,
"grad_norm": 7.044617652893066,
"learning_rate": 0.00021272080494516888,
"loss": 2.191,
"step": 2735
},
{
"epoch": 1.0912530492358241,
"grad_norm": 5.305728912353516,
"learning_rate": 0.00021243642864630962,
"loss": 2.2054,
"step": 2740
},
{
"epoch": 1.0932443869169115,
"grad_norm": 5.111948490142822,
"learning_rate": 0.00021215178064866444,
"loss": 2.08,
"step": 2745
},
{
"epoch": 1.0952357245979987,
"grad_norm": 5.789201736450195,
"learning_rate": 0.0002118668621909096,
"loss": 2.1121,
"step": 2750
},
{
"epoch": 1.097227062279086,
"grad_norm": 6.3249969482421875,
"learning_rate": 0.00021158167451289814,
"loss": 2.1042,
"step": 2755
},
{
"epoch": 1.0992183999601732,
"grad_norm": 5.103941917419434,
"learning_rate": 0.00021129621885565469,
"loss": 2.1663,
"step": 2760
},
{
"epoch": 1.1012097376412604,
"grad_norm": 7.262547016143799,
"learning_rate": 0.00021101049646137003,
"loss": 2.1616,
"step": 2765
},
{
"epoch": 1.1032010753223478,
"grad_norm": 6.365930080413818,
"learning_rate": 0.00021072450857339563,
"loss": 2.2593,
"step": 2770
},
{
"epoch": 1.105192413003435,
"grad_norm": 5.771877765655518,
"learning_rate": 0.00021043825643623842,
"loss": 1.9991,
"step": 2775
},
{
"epoch": 1.1071837506845223,
"grad_norm": 5.775332927703857,
"learning_rate": 0.00021015174129555502,
"loss": 2.1432,
"step": 2780
},
{
"epoch": 1.1091750883656095,
"grad_norm": 5.588124752044678,
"learning_rate": 0.00020986496439814672,
"loss": 2.1829,
"step": 2785
},
{
"epoch": 1.111166426046697,
"grad_norm": 8.466453552246094,
"learning_rate": 0.00020957792699195378,
"loss": 2.2542,
"step": 2790
},
{
"epoch": 1.113157763727784,
"grad_norm": 5.901578426361084,
"learning_rate": 0.0002092906303260502,
"loss": 2.0991,
"step": 2795
},
{
"epoch": 1.1151491014088715,
"grad_norm": 5.687893867492676,
"learning_rate": 0.000209003075650638,
"loss": 2.1161,
"step": 2800
},
{
"epoch": 1.1171404390899586,
"grad_norm": 7.566319465637207,
"learning_rate": 0.00020871526421704207,
"loss": 2.1824,
"step": 2805
},
{
"epoch": 1.119131776771046,
"grad_norm": 6.398413181304932,
"learning_rate": 0.00020842719727770466,
"loss": 2.0616,
"step": 2810
},
{
"epoch": 1.1211231144521332,
"grad_norm": 5.798287868499756,
"learning_rate": 0.00020813887608617972,
"loss": 2.3593,
"step": 2815
},
{
"epoch": 1.1231144521332206,
"grad_norm": 5.282783031463623,
"learning_rate": 0.00020785030189712783,
"loss": 2.1251,
"step": 2820
},
{
"epoch": 1.1251057898143078,
"grad_norm": 6.402846813201904,
"learning_rate": 0.00020756147596631024,
"loss": 2.1565,
"step": 2825
},
{
"epoch": 1.1270971274953951,
"grad_norm": 6.533949851989746,
"learning_rate": 0.00020727239955058393,
"loss": 2.2081,
"step": 2830
},
{
"epoch": 1.1290884651764823,
"grad_norm": 7.3982954025268555,
"learning_rate": 0.00020698307390789572,
"loss": 2.1653,
"step": 2835
},
{
"epoch": 1.1310798028575695,
"grad_norm": 5.551629066467285,
"learning_rate": 0.00020669350029727708,
"loss": 2.3108,
"step": 2840
},
{
"epoch": 1.1330711405386569,
"grad_norm": 6.518723964691162,
"learning_rate": 0.00020640367997883853,
"loss": 2.1915,
"step": 2845
},
{
"epoch": 1.135062478219744,
"grad_norm": 5.792189121246338,
"learning_rate": 0.00020611361421376403,
"loss": 2.1557,
"step": 2850
},
{
"epoch": 1.1370538159008314,
"grad_norm": 5.768317699432373,
"learning_rate": 0.00020582330426430583,
"loss": 2.1389,
"step": 2855
},
{
"epoch": 1.1390451535819186,
"grad_norm": 7.050993919372559,
"learning_rate": 0.0002055327513937786,
"loss": 2.1938,
"step": 2860
},
{
"epoch": 1.141036491263006,
"grad_norm": 6.800126552581787,
"learning_rate": 0.00020524195686655422,
"loss": 2.1704,
"step": 2865
},
{
"epoch": 1.1430278289440932,
"grad_norm": 6.082620143890381,
"learning_rate": 0.0002049509219480561,
"loss": 2.1409,
"step": 2870
},
{
"epoch": 1.1450191666251806,
"grad_norm": 5.8574018478393555,
"learning_rate": 0.00020465964790475374,
"loss": 2.2212,
"step": 2875
},
{
"epoch": 1.1470105043062677,
"grad_norm": 5.6736931800842285,
"learning_rate": 0.00020436813600415734,
"loss": 2.2452,
"step": 2880
},
{
"epoch": 1.149001841987355,
"grad_norm": 6.067693710327148,
"learning_rate": 0.00020407638751481196,
"loss": 2.1116,
"step": 2885
},
{
"epoch": 1.1509931796684423,
"grad_norm": 6.335114002227783,
"learning_rate": 0.00020378440370629237,
"loss": 2.2473,
"step": 2890
},
{
"epoch": 1.1529845173495294,
"grad_norm": 7.521064281463623,
"learning_rate": 0.00020349218584919722,
"loss": 2.0444,
"step": 2895
},
{
"epoch": 1.1549758550306168,
"grad_norm": 6.531569004058838,
"learning_rate": 0.0002031997352151438,
"loss": 2.2091,
"step": 2900
},
{
"epoch": 1.156967192711704,
"grad_norm": 6.6172027587890625,
"learning_rate": 0.00020290705307676217,
"loss": 2.1326,
"step": 2905
},
{
"epoch": 1.1589585303927914,
"grad_norm": 6.893074989318848,
"learning_rate": 0.00020261414070769,
"loss": 2.1845,
"step": 2910
},
{
"epoch": 1.1609498680738786,
"grad_norm": 5.909544944763184,
"learning_rate": 0.0002023209993825667,
"loss": 2.2273,
"step": 2915
},
{
"epoch": 1.162941205754966,
"grad_norm": 6.50417423248291,
"learning_rate": 0.000202027630377028,
"loss": 2.1177,
"step": 2920
},
{
"epoch": 1.1649325434360531,
"grad_norm": 6.459300518035889,
"learning_rate": 0.00020173403496770057,
"loss": 2.2723,
"step": 2925
},
{
"epoch": 1.1669238811171405,
"grad_norm": 5.6512451171875,
"learning_rate": 0.00020144021443219604,
"loss": 2.2022,
"step": 2930
},
{
"epoch": 1.1689152187982277,
"grad_norm": 6.4242634773254395,
"learning_rate": 0.00020114617004910598,
"loss": 2.1843,
"step": 2935
},
{
"epoch": 1.170906556479315,
"grad_norm": 6.003353118896484,
"learning_rate": 0.0002008519030979958,
"loss": 2.2598,
"step": 2940
},
{
"epoch": 1.1728978941604022,
"grad_norm": 4.6970133781433105,
"learning_rate": 0.00020055741485939963,
"loss": 2.3102,
"step": 2945
},
{
"epoch": 1.1748892318414894,
"grad_norm": 6.059596061706543,
"learning_rate": 0.00020026270661481447,
"loss": 2.2779,
"step": 2950
},
{
"epoch": 1.1768805695225768,
"grad_norm": 5.73154878616333,
"learning_rate": 0.0001999677796466947,
"loss": 2.1848,
"step": 2955
},
{
"epoch": 1.178871907203664,
"grad_norm": 6.712233543395996,
"learning_rate": 0.00019967263523844656,
"loss": 2.1027,
"step": 2960
},
{
"epoch": 1.1808632448847514,
"grad_norm": 7.3542399406433105,
"learning_rate": 0.00019937727467442235,
"loss": 2.254,
"step": 2965
},
{
"epoch": 1.1828545825658385,
"grad_norm": 5.018367767333984,
"learning_rate": 0.00019908169923991522,
"loss": 2.1361,
"step": 2970
},
{
"epoch": 1.184845920246926,
"grad_norm": 5.981570720672607,
"learning_rate": 0.00019878591022115316,
"loss": 2.1516,
"step": 2975
},
{
"epoch": 1.186837257928013,
"grad_norm": 5.398682117462158,
"learning_rate": 0.00019848990890529368,
"loss": 2.2784,
"step": 2980
},
{
"epoch": 1.1888285956091005,
"grad_norm": 4.907093048095703,
"learning_rate": 0.00019819369658041818,
"loss": 2.1853,
"step": 2985
},
{
"epoch": 1.1908199332901876,
"grad_norm": 5.047986030578613,
"learning_rate": 0.00019789727453552606,
"loss": 2.3024,
"step": 2990
},
{
"epoch": 1.192811270971275,
"grad_norm": 5.1892852783203125,
"learning_rate": 0.00019760064406052969,
"loss": 2.1318,
"step": 2995
},
{
"epoch": 1.1948026086523622,
"grad_norm": 5.791012763977051,
"learning_rate": 0.0001973038064462481,
"loss": 2.1913,
"step": 3000
},
{
"epoch": 1.1967939463334494,
"grad_norm": 7.220500946044922,
"learning_rate": 0.00019700676298440195,
"loss": 2.2053,
"step": 3005
},
{
"epoch": 1.1987852840145368,
"grad_norm": 4.852428913116455,
"learning_rate": 0.00019670951496760758,
"loss": 2.1844,
"step": 3010
},
{
"epoch": 1.2007766216956242,
"grad_norm": 6.431066036224365,
"learning_rate": 0.00019641206368937146,
"loss": 2.171,
"step": 3015
},
{
"epoch": 1.2027679593767113,
"grad_norm": 5.527892589569092,
"learning_rate": 0.0001961144104440845,
"loss": 2.284,
"step": 3020
},
{
"epoch": 1.2047592970577985,
"grad_norm": 5.936304092407227,
"learning_rate": 0.00019581655652701671,
"loss": 2.1361,
"step": 3025
},
{
"epoch": 1.2067506347388859,
"grad_norm": 5.172138214111328,
"learning_rate": 0.0001955185032343112,
"loss": 2.158,
"step": 3030
},
{
"epoch": 1.208741972419973,
"grad_norm": 5.692948818206787,
"learning_rate": 0.0001952202518629786,
"loss": 2.1897,
"step": 3035
},
{
"epoch": 1.2107333101010604,
"grad_norm": 5.522965908050537,
"learning_rate": 0.00019492180371089168,
"loss": 2.287,
"step": 3040
},
{
"epoch": 1.2127246477821476,
"grad_norm": 6.14274787902832,
"learning_rate": 0.0001946231600767794,
"loss": 2.1044,
"step": 3045
},
{
"epoch": 1.214715985463235,
"grad_norm": 6.3660101890563965,
"learning_rate": 0.00019432432226022147,
"loss": 2.1893,
"step": 3050
},
{
"epoch": 1.2167073231443222,
"grad_norm": 5.123039722442627,
"learning_rate": 0.0001940252915616425,
"loss": 2.2271,
"step": 3055
},
{
"epoch": 1.2186986608254096,
"grad_norm": 5.454919815063477,
"learning_rate": 0.0001937260692823065,
"loss": 2.1444,
"step": 3060
},
{
"epoch": 1.2206899985064967,
"grad_norm": 5.850592136383057,
"learning_rate": 0.00019342665672431123,
"loss": 2.2168,
"step": 3065
},
{
"epoch": 1.2226813361875841,
"grad_norm": 5.659205913543701,
"learning_rate": 0.00019312705519058238,
"loss": 2.2518,
"step": 3070
},
{
"epoch": 1.2246726738686713,
"grad_norm": 6.5372314453125,
"learning_rate": 0.00019282726598486802,
"loss": 2.2739,
"step": 3075
},
{
"epoch": 1.2266640115497585,
"grad_norm": 5.422939777374268,
"learning_rate": 0.0001925272904117328,
"loss": 2.0801,
"step": 3080
},
{
"epoch": 1.2286553492308459,
"grad_norm": 6.354897975921631,
"learning_rate": 0.00019222712977655262,
"loss": 2.1793,
"step": 3085
},
{
"epoch": 1.230646686911933,
"grad_norm": 6.33408260345459,
"learning_rate": 0.00019192678538550837,
"loss": 2.247,
"step": 3090
},
{
"epoch": 1.2326380245930204,
"grad_norm": 5.345519542694092,
"learning_rate": 0.00019162625854558086,
"loss": 2.1663,
"step": 3095
},
{
"epoch": 1.2346293622741076,
"grad_norm": 5.576323509216309,
"learning_rate": 0.00019132555056454464,
"loss": 2.0648,
"step": 3100
},
{
"epoch": 1.236620699955195,
"grad_norm": 5.9087910652160645,
"learning_rate": 0.00019102466275096262,
"loss": 2.1999,
"step": 3105
},
{
"epoch": 1.2386120376362821,
"grad_norm": 6.147064685821533,
"learning_rate": 0.00019072359641418031,
"loss": 2.4797,
"step": 3110
},
{
"epoch": 1.2406033753173695,
"grad_norm": 5.6060662269592285,
"learning_rate": 0.00019042235286431994,
"loss": 2.3287,
"step": 3115
},
{
"epoch": 1.2425947129984567,
"grad_norm": 5.959619522094727,
"learning_rate": 0.00019012093341227505,
"loss": 2.1932,
"step": 3120
},
{
"epoch": 1.244586050679544,
"grad_norm": 7.9409379959106445,
"learning_rate": 0.00018981933936970446,
"loss": 2.2579,
"step": 3125
},
{
"epoch": 1.2465773883606313,
"grad_norm": 5.92664909362793,
"learning_rate": 0.00018951757204902696,
"loss": 2.3502,
"step": 3130
},
{
"epoch": 1.2485687260417184,
"grad_norm": 5.162164211273193,
"learning_rate": 0.00018921563276341522,
"loss": 2.2192,
"step": 3135
},
{
"epoch": 1.2505600637228058,
"grad_norm": 6.151308059692383,
"learning_rate": 0.00018891352282679027,
"loss": 2.2133,
"step": 3140
},
{
"epoch": 1.252551401403893,
"grad_norm": 5.899823188781738,
"learning_rate": 0.00018861124355381583,
"loss": 2.3228,
"step": 3145
},
{
"epoch": 1.2545427390849804,
"grad_norm": 5.929165840148926,
"learning_rate": 0.0001883087962598923,
"loss": 2.1752,
"step": 3150
},
{
"epoch": 1.2565340767660675,
"grad_norm": 5.03986930847168,
"learning_rate": 0.0001880061822611515,
"loss": 2.196,
"step": 3155
},
{
"epoch": 1.258525414447155,
"grad_norm": 7.291461944580078,
"learning_rate": 0.0001877034028744505,
"loss": 2.1835,
"step": 3160
},
{
"epoch": 1.260516752128242,
"grad_norm": 5.5854620933532715,
"learning_rate": 0.00018740045941736608,
"loss": 2.0839,
"step": 3165
},
{
"epoch": 1.2625080898093295,
"grad_norm": 4.644222259521484,
"learning_rate": 0.00018709735320818918,
"loss": 2.235,
"step": 3170
},
{
"epoch": 1.2644994274904167,
"grad_norm": 4.4264817237854,
"learning_rate": 0.00018679408556591866,
"loss": 2.1371,
"step": 3175
},
{
"epoch": 1.266490765171504,
"grad_norm": 6.052789688110352,
"learning_rate": 0.00018649065781025623,
"loss": 2.286,
"step": 3180
},
{
"epoch": 1.2684821028525912,
"grad_norm": 5.982134819030762,
"learning_rate": 0.00018618707126160004,
"loss": 2.1594,
"step": 3185
},
{
"epoch": 1.2704734405336784,
"grad_norm": 5.037675380706787,
"learning_rate": 0.00018588332724103938,
"loss": 2.2272,
"step": 3190
},
{
"epoch": 1.2724647782147658,
"grad_norm": 4.905061721801758,
"learning_rate": 0.00018557942707034882,
"loss": 2.1714,
"step": 3195
},
{
"epoch": 1.2744561158958532,
"grad_norm": 5.602728843688965,
"learning_rate": 0.0001852753720719824,
"loss": 2.1461,
"step": 3200
},
{
"epoch": 1.2764474535769403,
"grad_norm": 6.72523832321167,
"learning_rate": 0.0001849711635690679,
"loss": 2.3851,
"step": 3205
},
{
"epoch": 1.2784387912580275,
"grad_norm": 7.611420631408691,
"learning_rate": 0.0001846668028854011,
"loss": 2.1879,
"step": 3210
},
{
"epoch": 1.280430128939115,
"grad_norm": 5.812536239624023,
"learning_rate": 0.00018436229134544005,
"loss": 2.2753,
"step": 3215
},
{
"epoch": 1.282421466620202,
"grad_norm": 5.685847282409668,
"learning_rate": 0.00018405763027429917,
"loss": 2.2598,
"step": 3220
},
{
"epoch": 1.2844128043012895,
"grad_norm": 5.823709487915039,
"learning_rate": 0.0001837528209977437,
"loss": 2.1045,
"step": 3225
},
{
"epoch": 1.2864041419823766,
"grad_norm": 5.243285655975342,
"learning_rate": 0.00018344786484218367,
"loss": 2.2614,
"step": 3230
},
{
"epoch": 1.288395479663464,
"grad_norm": 5.421647548675537,
"learning_rate": 0.00018314276313466846,
"loss": 2.1703,
"step": 3235
},
{
"epoch": 1.2903868173445512,
"grad_norm": 6.741246700286865,
"learning_rate": 0.00018283751720288066,
"loss": 2.1891,
"step": 3240
},
{
"epoch": 1.2923781550256384,
"grad_norm": 5.4550676345825195,
"learning_rate": 0.00018253212837513055,
"loss": 2.227,
"step": 3245
},
{
"epoch": 1.2943694927067257,
"grad_norm": 5.21625280380249,
"learning_rate": 0.0001822265979803502,
"loss": 2.246,
"step": 3250
},
{
"epoch": 1.2963608303878131,
"grad_norm": 7.004878044128418,
"learning_rate": 0.00018192092734808772,
"loss": 2.3162,
"step": 3255
},
{
"epoch": 1.2983521680689003,
"grad_norm": 5.326993942260742,
"learning_rate": 0.00018161511780850163,
"loss": 2.1449,
"step": 3260
},
{
"epoch": 1.3003435057499875,
"grad_norm": 6.218927383422852,
"learning_rate": 0.00018130917069235467,
"loss": 2.1936,
"step": 3265
},
{
"epoch": 1.3023348434310749,
"grad_norm": 6.091947555541992,
"learning_rate": 0.00018100308733100847,
"loss": 2.2076,
"step": 3270
},
{
"epoch": 1.304326181112162,
"grad_norm": 7.184474468231201,
"learning_rate": 0.00018069686905641737,
"loss": 2.3472,
"step": 3275
},
{
"epoch": 1.3063175187932494,
"grad_norm": 5.758594036102295,
"learning_rate": 0.00018039051720112297,
"loss": 2.2042,
"step": 3280
},
{
"epoch": 1.3083088564743366,
"grad_norm": 4.715916633605957,
"learning_rate": 0.000180084033098248,
"loss": 2.2247,
"step": 3285
},
{
"epoch": 1.310300194155424,
"grad_norm": 5.105158805847168,
"learning_rate": 0.0001797774180814908,
"loss": 2.1685,
"step": 3290
},
{
"epoch": 1.3122915318365111,
"grad_norm": 7.182875156402588,
"learning_rate": 0.0001794706734851194,
"loss": 2.2263,
"step": 3295
},
{
"epoch": 1.3142828695175983,
"grad_norm": 6.164737224578857,
"learning_rate": 0.00017916380064396557,
"loss": 2.1982,
"step": 3300
},
{
"epoch": 1.3162742071986857,
"grad_norm": 5.850378036499023,
"learning_rate": 0.00017885680089341937,
"loss": 2.1192,
"step": 3305
},
{
"epoch": 1.318265544879773,
"grad_norm": 6.571321964263916,
"learning_rate": 0.00017854967556942285,
"loss": 2.3361,
"step": 3310
},
{
"epoch": 1.3202568825608603,
"grad_norm": 5.506916522979736,
"learning_rate": 0.00017824242600846476,
"loss": 2.194,
"step": 3315
},
{
"epoch": 1.3222482202419474,
"grad_norm": 9.358162879943848,
"learning_rate": 0.00017793505354757428,
"loss": 2.1641,
"step": 3320
},
{
"epoch": 1.3242395579230348,
"grad_norm": 5.8232340812683105,
"learning_rate": 0.00017762755952431558,
"loss": 2.2781,
"step": 3325
},
{
"epoch": 1.326230895604122,
"grad_norm": 6.010040760040283,
"learning_rate": 0.00017731994527678166,
"loss": 2.2773,
"step": 3330
},
{
"epoch": 1.3282222332852094,
"grad_norm": 4.972679138183594,
"learning_rate": 0.0001770122121435888,
"loss": 2.1069,
"step": 3335
},
{
"epoch": 1.3302135709662966,
"grad_norm": 12.89777660369873,
"learning_rate": 0.00017670436146387063,
"loss": 2.044,
"step": 3340
},
{
"epoch": 1.332204908647384,
"grad_norm": 5.31705379486084,
"learning_rate": 0.00017639639457727217,
"loss": 2.153,
"step": 3345
},
{
"epoch": 1.3341962463284711,
"grad_norm": 6.822572708129883,
"learning_rate": 0.00017608831282394426,
"loss": 2.2099,
"step": 3350
},
{
"epoch": 1.3361875840095583,
"grad_norm": 7.124082565307617,
"learning_rate": 0.00017578011754453753,
"loss": 2.2104,
"step": 3355
},
{
"epoch": 1.3381789216906457,
"grad_norm": 5.639463901519775,
"learning_rate": 0.00017547181008019668,
"loss": 2.2408,
"step": 3360
},
{
"epoch": 1.340170259371733,
"grad_norm": 5.687006950378418,
"learning_rate": 0.0001751633917725545,
"loss": 2.1546,
"step": 3365
},
{
"epoch": 1.3421615970528202,
"grad_norm": 5.070754051208496,
"learning_rate": 0.0001748548639637262,
"loss": 2.2619,
"step": 3370
},
{
"epoch": 1.3441529347339074,
"grad_norm": 5.338910102844238,
"learning_rate": 0.0001745462279963035,
"loss": 2.3145,
"step": 3375
},
{
"epoch": 1.3461442724149948,
"grad_norm": 6.1445794105529785,
"learning_rate": 0.0001742374852133488,
"loss": 2.0273,
"step": 3380
},
{
"epoch": 1.348135610096082,
"grad_norm": 6.033137798309326,
"learning_rate": 0.00017392863695838928,
"loss": 2.2441,
"step": 3385
},
{
"epoch": 1.3501269477771694,
"grad_norm": 8.369287490844727,
"learning_rate": 0.000173619684575411,
"loss": 2.2212,
"step": 3390
},
{
"epoch": 1.3521182854582565,
"grad_norm": 7.496065616607666,
"learning_rate": 0.00017331062940885338,
"loss": 2.2526,
"step": 3395
},
{
"epoch": 1.354109623139344,
"grad_norm": 5.870899677276611,
"learning_rate": 0.00017300147280360285,
"loss": 2.2123,
"step": 3400
},
{
"epoch": 1.356100960820431,
"grad_norm": 6.694897651672363,
"learning_rate": 0.00017269221610498748,
"loss": 2.2284,
"step": 3405
},
{
"epoch": 1.3580922985015185,
"grad_norm": 5.430352210998535,
"learning_rate": 0.0001723828606587708,
"loss": 2.1369,
"step": 3410
},
{
"epoch": 1.3600836361826056,
"grad_norm": 7.428508758544922,
"learning_rate": 0.00017207340781114598,
"loss": 2.2149,
"step": 3415
},
{
"epoch": 1.362074973863693,
"grad_norm": 6.232314586639404,
"learning_rate": 0.0001717638589087303,
"loss": 2.1926,
"step": 3420
},
{
"epoch": 1.3640663115447802,
"grad_norm": 5.414062023162842,
"learning_rate": 0.00017145421529855866,
"loss": 2.1174,
"step": 3425
},
{
"epoch": 1.3660576492258674,
"grad_norm": 5.721498012542725,
"learning_rate": 0.00017114447832807845,
"loss": 2.2366,
"step": 3430
},
{
"epoch": 1.3680489869069548,
"grad_norm": 8.127852439880371,
"learning_rate": 0.0001708346493451431,
"loss": 2.3191,
"step": 3435
},
{
"epoch": 1.3700403245880421,
"grad_norm": 5.363738059997559,
"learning_rate": 0.00017052472969800647,
"loss": 2.2159,
"step": 3440
},
{
"epoch": 1.3720316622691293,
"grad_norm": 4.83927059173584,
"learning_rate": 0.00017021472073531712,
"loss": 2.1906,
"step": 3445
},
{
"epoch": 1.3740229999502165,
"grad_norm": 7.324374675750732,
"learning_rate": 0.000169904623806112,
"loss": 2.1427,
"step": 3450
},
{
"epoch": 1.3760143376313039,
"grad_norm": 5.2989373207092285,
"learning_rate": 0.00016959444025981115,
"loss": 2.1721,
"step": 3455
},
{
"epoch": 1.378005675312391,
"grad_norm": 5.679375648498535,
"learning_rate": 0.00016928417144621127,
"loss": 2.1693,
"step": 3460
},
{
"epoch": 1.3799970129934784,
"grad_norm": 5.638728141784668,
"learning_rate": 0.00016897381871548036,
"loss": 2.2256,
"step": 3465
},
{
"epoch": 1.3819883506745656,
"grad_norm": 6.92189884185791,
"learning_rate": 0.00016866338341815134,
"loss": 2.2384,
"step": 3470
},
{
"epoch": 1.383979688355653,
"grad_norm": 6.763597011566162,
"learning_rate": 0.0001683528669051166,
"loss": 2.0807,
"step": 3475
},
{
"epoch": 1.3859710260367402,
"grad_norm": 13.712860107421875,
"learning_rate": 0.00016804227052762195,
"loss": 2.2717,
"step": 3480
},
{
"epoch": 1.3879623637178273,
"grad_norm": 8.10069465637207,
"learning_rate": 0.00016773159563726058,
"loss": 2.0908,
"step": 3485
},
{
"epoch": 1.3899537013989147,
"grad_norm": 5.849442481994629,
"learning_rate": 0.00016742084358596757,
"loss": 2.2445,
"step": 3490
},
{
"epoch": 1.3919450390800021,
"grad_norm": 5.508017063140869,
"learning_rate": 0.00016711001572601358,
"loss": 2.2985,
"step": 3495
},
{
"epoch": 1.3939363767610893,
"grad_norm": 4.959733486175537,
"learning_rate": 0.00016679911340999925,
"loss": 2.1415,
"step": 3500
},
{
"epoch": 1.3959277144421764,
"grad_norm": 4.843476295471191,
"learning_rate": 0.0001664881379908492,
"loss": 2.3455,
"step": 3505
},
{
"epoch": 1.3979190521232638,
"grad_norm": 5.710421085357666,
"learning_rate": 0.00016617709082180615,
"loss": 2.0707,
"step": 3510
},
{
"epoch": 1.399910389804351,
"grad_norm": 5.431413173675537,
"learning_rate": 0.00016586597325642514,
"loss": 2.0146,
"step": 3515
},
{
"epoch": 1.4019017274854384,
"grad_norm": 5.725831031799316,
"learning_rate": 0.00016555478664856738,
"loss": 2.1838,
"step": 3520
},
{
"epoch": 1.4038930651665256,
"grad_norm": 6.110559940338135,
"learning_rate": 0.00016524353235239473,
"loss": 2.2518,
"step": 3525
},
{
"epoch": 1.405884402847613,
"grad_norm": 5.45647668838501,
"learning_rate": 0.00016493221172236337,
"loss": 2.1785,
"step": 3530
},
{
"epoch": 1.4078757405287001,
"grad_norm": 7.214836120605469,
"learning_rate": 0.00016462082611321834,
"loss": 2.2149,
"step": 3535
},
{
"epoch": 1.4098670782097873,
"grad_norm": 5.839008331298828,
"learning_rate": 0.00016430937687998736,
"loss": 2.1552,
"step": 3540
},
{
"epoch": 1.4118584158908747,
"grad_norm": 6.743014812469482,
"learning_rate": 0.000163997865377975,
"loss": 2.195,
"step": 3545
},
{
"epoch": 1.413849753571962,
"grad_norm": 6.176872730255127,
"learning_rate": 0.00016368629296275683,
"loss": 2.1918,
"step": 3550
},
{
"epoch": 1.4158410912530492,
"grad_norm": 5.614704132080078,
"learning_rate": 0.0001633746609901735,
"loss": 2.1733,
"step": 3555
},
{
"epoch": 1.4178324289341364,
"grad_norm": 6.306468963623047,
"learning_rate": 0.00016306297081632482,
"loss": 2.1643,
"step": 3560
},
{
"epoch": 1.4198237666152238,
"grad_norm": 5.466887950897217,
"learning_rate": 0.0001627512237975638,
"loss": 2.1423,
"step": 3565
},
{
"epoch": 1.421815104296311,
"grad_norm": 6.443585395812988,
"learning_rate": 0.00016243942129049097,
"loss": 2.3586,
"step": 3570
},
{
"epoch": 1.4238064419773984,
"grad_norm": 6.56254243850708,
"learning_rate": 0.00016212756465194814,
"loss": 2.1143,
"step": 3575
},
{
"epoch": 1.4257977796584855,
"grad_norm": 6.025684833526611,
"learning_rate": 0.00016181565523901284,
"loss": 2.2403,
"step": 3580
},
{
"epoch": 1.427789117339573,
"grad_norm": 5.992644786834717,
"learning_rate": 0.0001615036944089921,
"loss": 2.1688,
"step": 3585
},
{
"epoch": 1.42978045502066,
"grad_norm": 5.697604179382324,
"learning_rate": 0.00016119168351941688,
"loss": 2.1304,
"step": 3590
},
{
"epoch": 1.4317717927017475,
"grad_norm": 7.877066612243652,
"learning_rate": 0.0001608796239280359,
"loss": 2.2145,
"step": 3595
},
{
"epoch": 1.4337631303828346,
"grad_norm": 7.044979095458984,
"learning_rate": 0.00016056751699280964,
"loss": 2.2152,
"step": 3600
},
{
"epoch": 1.435754468063922,
"grad_norm": 6.05941915512085,
"learning_rate": 0.0001602553640719049,
"loss": 2.2349,
"step": 3605
},
{
"epoch": 1.4377458057450092,
"grad_norm": 6.004716396331787,
"learning_rate": 0.00015994316652368838,
"loss": 2.1943,
"step": 3610
},
{
"epoch": 1.4397371434260964,
"grad_norm": 7.441656112670898,
"learning_rate": 0.00015963092570672106,
"loss": 2.1507,
"step": 3615
},
{
"epoch": 1.4417284811071838,
"grad_norm": 5.681270122528076,
"learning_rate": 0.00015931864297975219,
"loss": 2.2447,
"step": 3620
},
{
"epoch": 1.4437198187882712,
"grad_norm": 4.99393367767334,
"learning_rate": 0.00015900631970171337,
"loss": 2.1024,
"step": 3625
},
{
"epoch": 1.4457111564693583,
"grad_norm": 5.866675853729248,
"learning_rate": 0.00015869395723171277,
"loss": 2.2081,
"step": 3630
},
{
"epoch": 1.4477024941504455,
"grad_norm": 5.251063823699951,
"learning_rate": 0.00015838155692902896,
"loss": 2.1623,
"step": 3635
},
{
"epoch": 1.4496938318315329,
"grad_norm": 5.525659084320068,
"learning_rate": 0.00015806912015310526,
"loss": 2.3001,
"step": 3640
},
{
"epoch": 1.45168516951262,
"grad_norm": 5.020875930786133,
"learning_rate": 0.00015775664826354362,
"loss": 2.2304,
"step": 3645
},
{
"epoch": 1.4536765071937074,
"grad_norm": 5.0340704917907715,
"learning_rate": 0.00015744414262009886,
"loss": 2.1608,
"step": 3650
},
{
"epoch": 1.4556678448747946,
"grad_norm": 4.942144393920898,
"learning_rate": 0.00015713160458267263,
"loss": 2.101,
"step": 3655
},
{
"epoch": 1.457659182555882,
"grad_norm": 6.115739822387695,
"learning_rate": 0.0001568190355113076,
"loss": 2.1676,
"step": 3660
},
{
"epoch": 1.4596505202369692,
"grad_norm": 8.139789581298828,
"learning_rate": 0.00015650643676618146,
"loss": 2.2546,
"step": 3665
},
{
"epoch": 1.4616418579180563,
"grad_norm": 6.888481616973877,
"learning_rate": 0.00015619380970760096,
"loss": 2.2418,
"step": 3670
},
{
"epoch": 1.4636331955991437,
"grad_norm": 6.429262161254883,
"learning_rate": 0.00015588115569599624,
"loss": 2.1296,
"step": 3675
},
{
"epoch": 1.4656245332802311,
"grad_norm": 5.466495037078857,
"learning_rate": 0.0001555684760919145,
"loss": 2.2744,
"step": 3680
},
{
"epoch": 1.4676158709613183,
"grad_norm": 6.42126989364624,
"learning_rate": 0.0001552557722560145,
"loss": 2.1773,
"step": 3685
},
{
"epoch": 1.4696072086424055,
"grad_norm": 5.6322174072265625,
"learning_rate": 0.00015494304554906034,
"loss": 2.2818,
"step": 3690
},
{
"epoch": 1.4715985463234929,
"grad_norm": 6.184334754943848,
"learning_rate": 0.0001546302973319157,
"loss": 2.1761,
"step": 3695
},
{
"epoch": 1.47358988400458,
"grad_norm": 6.0626540184021,
"learning_rate": 0.0001543175289655378,
"loss": 2.2124,
"step": 3700
},
{
"epoch": 1.4755812216856674,
"grad_norm": 5.791230201721191,
"learning_rate": 0.00015400474181097163,
"loss": 2.113,
"step": 3705
},
{
"epoch": 1.4775725593667546,
"grad_norm": 5.994204998016357,
"learning_rate": 0.00015369193722934385,
"loss": 2.1965,
"step": 3710
},
{
"epoch": 1.479563897047842,
"grad_norm": 5.7597856521606445,
"learning_rate": 0.000153379116581857,
"loss": 2.2088,
"step": 3715
},
{
"epoch": 1.4815552347289291,
"grad_norm": 5.297857284545898,
"learning_rate": 0.0001530662812297836,
"loss": 2.1675,
"step": 3720
},
{
"epoch": 1.4835465724100163,
"grad_norm": 6.562660217285156,
"learning_rate": 0.00015275343253445993,
"loss": 2.1832,
"step": 3725
},
{
"epoch": 1.4855379100911037,
"grad_norm": 6.066761493682861,
"learning_rate": 0.00015244057185728066,
"loss": 2.2723,
"step": 3730
},
{
"epoch": 1.487529247772191,
"grad_norm": 5.877458095550537,
"learning_rate": 0.00015212770055969232,
"loss": 2.1688,
"step": 3735
},
{
"epoch": 1.4895205854532783,
"grad_norm": 6.463152885437012,
"learning_rate": 0.00015181482000318786,
"loss": 2.1291,
"step": 3740
},
{
"epoch": 1.4915119231343654,
"grad_norm": 5.8509931564331055,
"learning_rate": 0.00015150193154930033,
"loss": 2.1488,
"step": 3745
},
{
"epoch": 1.4935032608154528,
"grad_norm": 6.462055206298828,
"learning_rate": 0.0001511890365595973,
"loss": 2.1914,
"step": 3750
},
{
"epoch": 1.49549459849654,
"grad_norm": 7.797671318054199,
"learning_rate": 0.0001508761363956748,
"loss": 2.1354,
"step": 3755
},
{
"epoch": 1.4974859361776274,
"grad_norm": 5.754122734069824,
"learning_rate": 0.00015056323241915113,
"loss": 2.1788,
"step": 3760
},
{
"epoch": 1.4994772738587145,
"grad_norm": 6.21970796585083,
"learning_rate": 0.00015025032599166153,
"loss": 2.2506,
"step": 3765
},
{
"epoch": 1.501468611539802,
"grad_norm": 5.987273216247559,
"learning_rate": 0.00014993741847485158,
"loss": 2.3639,
"step": 3770
},
{
"epoch": 1.503459949220889,
"grad_norm": 5.687912940979004,
"learning_rate": 0.00014962451123037187,
"loss": 2.1029,
"step": 3775
},
{
"epoch": 1.5054512869019763,
"grad_norm": 5.01307487487793,
"learning_rate": 0.0001493116056198716,
"loss": 2.2263,
"step": 3780
},
{
"epoch": 1.5074426245830637,
"grad_norm": 5.397319793701172,
"learning_rate": 0.00014899870300499296,
"loss": 2.3244,
"step": 3785
},
{
"epoch": 1.509433962264151,
"grad_norm": 6.498507499694824,
"learning_rate": 0.00014868580474736517,
"loss": 2.2593,
"step": 3790
},
{
"epoch": 1.5114252999452382,
"grad_norm": 5.791152477264404,
"learning_rate": 0.00014837291220859829,
"loss": 2.3214,
"step": 3795
},
{
"epoch": 1.5134166376263254,
"grad_norm": 6.324441432952881,
"learning_rate": 0.00014806002675027768,
"loss": 2.1823,
"step": 3800
},
{
"epoch": 1.5154079753074128,
"grad_norm": 5.276822566986084,
"learning_rate": 0.0001477471497339578,
"loss": 2.0003,
"step": 3805
},
{
"epoch": 1.5173993129885002,
"grad_norm": 6.361807346343994,
"learning_rate": 0.00014743428252115642,
"loss": 2.1975,
"step": 3810
},
{
"epoch": 1.5193906506695873,
"grad_norm": 5.382236003875732,
"learning_rate": 0.0001471214264733486,
"loss": 2.1137,
"step": 3815
},
{
"epoch": 1.5213819883506745,
"grad_norm": 6.161167621612549,
"learning_rate": 0.0001468085829519608,
"loss": 2.206,
"step": 3820
},
{
"epoch": 1.523373326031762,
"grad_norm": 5.9316086769104,
"learning_rate": 0.0001464957533183651,
"loss": 2.2331,
"step": 3825
},
{
"epoch": 1.525364663712849,
"grad_norm": 5.059473514556885,
"learning_rate": 0.00014618293893387297,
"loss": 2.1523,
"step": 3830
},
{
"epoch": 1.5273560013939362,
"grad_norm": 5.21863317489624,
"learning_rate": 0.00014587014115972967,
"loss": 2.2796,
"step": 3835
},
{
"epoch": 1.5293473390750236,
"grad_norm": 4.951760768890381,
"learning_rate": 0.00014555736135710802,
"loss": 2.2487,
"step": 3840
},
{
"epoch": 1.531338676756111,
"grad_norm": 6.017904758453369,
"learning_rate": 0.00014524460088710282,
"loss": 2.1701,
"step": 3845
},
{
"epoch": 1.5333300144371982,
"grad_norm": 6.514730930328369,
"learning_rate": 0.0001449318611107246,
"loss": 2.3521,
"step": 3850
},
{
"epoch": 1.5353213521182854,
"grad_norm": 5.786331653594971,
"learning_rate": 0.0001446191433888939,
"loss": 2.0861,
"step": 3855
},
{
"epoch": 1.5373126897993727,
"grad_norm": 5.961428642272949,
"learning_rate": 0.0001443064490824353,
"loss": 2.2932,
"step": 3860
},
{
"epoch": 1.5393040274804601,
"grad_norm": 5.961484432220459,
"learning_rate": 0.00014399377955207142,
"loss": 2.1267,
"step": 3865
},
{
"epoch": 1.5412953651615473,
"grad_norm": 5.781057834625244,
"learning_rate": 0.00014368113615841714,
"loss": 2.161,
"step": 3870
},
{
"epoch": 1.5432867028426345,
"grad_norm": 6.685031414031982,
"learning_rate": 0.00014336852026197356,
"loss": 2.1416,
"step": 3875
},
{
"epoch": 1.5452780405237219,
"grad_norm": 6.4983649253845215,
"learning_rate": 0.00014305593322312216,
"loss": 2.1847,
"step": 3880
},
{
"epoch": 1.547269378204809,
"grad_norm": 6.511983394622803,
"learning_rate": 0.00014274337640211873,
"loss": 2.0351,
"step": 3885
},
{
"epoch": 1.5492607158858962,
"grad_norm": 5.873143196105957,
"learning_rate": 0.00014243085115908777,
"loss": 2.2255,
"step": 3890
},
{
"epoch": 1.5512520535669836,
"grad_norm": 5.393064498901367,
"learning_rate": 0.00014211835885401613,
"loss": 2.1551,
"step": 3895
},
{
"epoch": 1.553243391248071,
"grad_norm": 5.8017802238464355,
"learning_rate": 0.0001418059008467475,
"loss": 2.1379,
"step": 3900
},
{
"epoch": 1.5552347289291582,
"grad_norm": 5.897889614105225,
"learning_rate": 0.00014149347849697628,
"loss": 2.2812,
"step": 3905
},
{
"epoch": 1.5572260666102453,
"grad_norm": 6.137565612792969,
"learning_rate": 0.0001411810931642416,
"loss": 2.1536,
"step": 3910
},
{
"epoch": 1.5592174042913327,
"grad_norm": 6.100076198577881,
"learning_rate": 0.00014086874620792173,
"loss": 2.065,
"step": 3915
},
{
"epoch": 1.56120874197242,
"grad_norm": 5.261016368865967,
"learning_rate": 0.0001405564389872277,
"loss": 2.2897,
"step": 3920
},
{
"epoch": 1.5632000796535073,
"grad_norm": 5.7907819747924805,
"learning_rate": 0.00014024417286119772,
"loss": 2.1539,
"step": 3925
},
{
"epoch": 1.5651914173345944,
"grad_norm": 4.588942050933838,
"learning_rate": 0.00013993194918869123,
"loss": 2.1315,
"step": 3930
},
{
"epoch": 1.5671827550156818,
"grad_norm": 5.841504096984863,
"learning_rate": 0.0001396197693283828,
"loss": 2.084,
"step": 3935
},
{
"epoch": 1.5691740926967692,
"grad_norm": 5.055356979370117,
"learning_rate": 0.00013930763463875656,
"loss": 2.0406,
"step": 3940
},
{
"epoch": 1.5711654303778562,
"grad_norm": 5.552372455596924,
"learning_rate": 0.0001389955464780998,
"loss": 2.1606,
"step": 3945
},
{
"epoch": 1.5731567680589436,
"grad_norm": 5.9232378005981445,
"learning_rate": 0.00013868350620449758,
"loss": 2.1084,
"step": 3950
},
{
"epoch": 1.575148105740031,
"grad_norm": 6.918461799621582,
"learning_rate": 0.00013837151517582638,
"loss": 2.2465,
"step": 3955
},
{
"epoch": 1.5771394434211181,
"grad_norm": 5.042429447174072,
"learning_rate": 0.00013805957474974855,
"loss": 2.1984,
"step": 3960
},
{
"epoch": 1.5791307811022053,
"grad_norm": 5.3624444007873535,
"learning_rate": 0.0001377476862837061,
"loss": 2.1909,
"step": 3965
},
{
"epoch": 1.5811221187832927,
"grad_norm": 5.48763370513916,
"learning_rate": 0.00013743585113491503,
"loss": 2.2525,
"step": 3970
},
{
"epoch": 1.58311345646438,
"grad_norm": 5.208502292633057,
"learning_rate": 0.0001371240706603593,
"loss": 2.2275,
"step": 3975
},
{
"epoch": 1.5851047941454672,
"grad_norm": 7.1540069580078125,
"learning_rate": 0.00013681234621678486,
"loss": 2.2409,
"step": 3980
},
{
"epoch": 1.5870961318265544,
"grad_norm": 5.815776824951172,
"learning_rate": 0.000136500679160694,
"loss": 2.1772,
"step": 3985
},
{
"epoch": 1.5890874695076418,
"grad_norm": 5.809691429138184,
"learning_rate": 0.00013618907084833912,
"loss": 2.2952,
"step": 3990
},
{
"epoch": 1.5910788071887292,
"grad_norm": 5.1112847328186035,
"learning_rate": 0.00013587752263571713,
"loss": 2.0964,
"step": 3995
},
{
"epoch": 1.5930701448698164,
"grad_norm": 5.776605606079102,
"learning_rate": 0.0001355660358785633,
"loss": 2.0697,
"step": 4000
},
{
"epoch": 1.5950614825509035,
"grad_norm": 5.093209743499756,
"learning_rate": 0.0001352546119323455,
"loss": 2.136,
"step": 4005
},
{
"epoch": 1.597052820231991,
"grad_norm": 6.1989850997924805,
"learning_rate": 0.00013494325215225833,
"loss": 2.283,
"step": 4010
},
{
"epoch": 1.599044157913078,
"grad_norm": 5.202113151550293,
"learning_rate": 0.00013463195789321708,
"loss": 2.2786,
"step": 4015
},
{
"epoch": 1.6010354955941652,
"grad_norm": 5.918675422668457,
"learning_rate": 0.000134320730509852,
"loss": 2.2188,
"step": 4020
},
{
"epoch": 1.6030268332752526,
"grad_norm": 6.093796730041504,
"learning_rate": 0.00013400957135650221,
"loss": 2.0943,
"step": 4025
},
{
"epoch": 1.60501817095634,
"grad_norm": 5.631002902984619,
"learning_rate": 0.00013369848178721005,
"loss": 2.1842,
"step": 4030
},
{
"epoch": 1.6070095086374272,
"grad_norm": 5.93800163269043,
"learning_rate": 0.00013338746315571495,
"loss": 2.1589,
"step": 4035
},
{
"epoch": 1.6090008463185144,
"grad_norm": 5.358720779418945,
"learning_rate": 0.00013307651681544774,
"loss": 2.1383,
"step": 4040
},
{
"epoch": 1.6109921839996018,
"grad_norm": 6.724377155303955,
"learning_rate": 0.0001327656441195246,
"loss": 2.1412,
"step": 4045
},
{
"epoch": 1.6129835216806891,
"grad_norm": 5.9936747550964355,
"learning_rate": 0.00013245484642074128,
"loss": 2.226,
"step": 4050
},
{
"epoch": 1.6149748593617763,
"grad_norm": 5.944963455200195,
"learning_rate": 0.00013214412507156712,
"loss": 2.1877,
"step": 4055
},
{
"epoch": 1.6169661970428635,
"grad_norm": 4.6952972412109375,
"learning_rate": 0.00013183348142413925,
"loss": 2.1925,
"step": 4060
},
{
"epoch": 1.6189575347239509,
"grad_norm": 4.9247331619262695,
"learning_rate": 0.0001315229168302567,
"loss": 2.0597,
"step": 4065
},
{
"epoch": 1.620948872405038,
"grad_norm": 5.743644714355469,
"learning_rate": 0.0001312124326413744,
"loss": 2.2252,
"step": 4070
},
{
"epoch": 1.6229402100861252,
"grad_norm": 7.026399612426758,
"learning_rate": 0.00013090203020859757,
"loss": 2.2459,
"step": 4075
},
{
"epoch": 1.6249315477672126,
"grad_norm": 5.469521999359131,
"learning_rate": 0.0001305917108826754,
"loss": 2.237,
"step": 4080
},
{
"epoch": 1.6269228854483,
"grad_norm": 5.118337631225586,
"learning_rate": 0.0001302814760139956,
"loss": 2.1671,
"step": 4085
},
{
"epoch": 1.6289142231293872,
"grad_norm": 5.142106056213379,
"learning_rate": 0.00012997132695257842,
"loss": 2.1174,
"step": 4090
},
{
"epoch": 1.6309055608104743,
"grad_norm": 5.856767177581787,
"learning_rate": 0.00012966126504807053,
"loss": 2.1862,
"step": 4095
},
{
"epoch": 1.6328968984915617,
"grad_norm": 5.7608842849731445,
"learning_rate": 0.00012935129164973948,
"loss": 2.1811,
"step": 4100
},
{
"epoch": 1.6348882361726491,
"grad_norm": 6.484520435333252,
"learning_rate": 0.0001290414081064675,
"loss": 2.1627,
"step": 4105
},
{
"epoch": 1.6368795738537363,
"grad_norm": 7.269222259521484,
"learning_rate": 0.00012873161576674604,
"loss": 2.2155,
"step": 4110
},
{
"epoch": 1.6388709115348234,
"grad_norm": 5.344924449920654,
"learning_rate": 0.00012842191597866944,
"loss": 2.2358,
"step": 4115
},
{
"epoch": 1.6408622492159108,
"grad_norm": 5.197498798370361,
"learning_rate": 0.00012811231008992945,
"loss": 2.1274,
"step": 4120
},
{
"epoch": 1.642853586896998,
"grad_norm": 5.114546775817871,
"learning_rate": 0.0001278027994478092,
"loss": 2.3124,
"step": 4125
},
{
"epoch": 1.6448449245780852,
"grad_norm": 5.009857177734375,
"learning_rate": 0.00012749338539917718,
"loss": 2.2095,
"step": 4130
},
{
"epoch": 1.6468362622591726,
"grad_norm": 5.543980121612549,
"learning_rate": 0.00012718406929048177,
"loss": 2.1183,
"step": 4135
},
{
"epoch": 1.64882759994026,
"grad_norm": 6.6033196449279785,
"learning_rate": 0.00012687485246774496,
"loss": 2.0655,
"step": 4140
},
{
"epoch": 1.6508189376213471,
"grad_norm": 4.530311584472656,
"learning_rate": 0.00012656573627655685,
"loss": 2.2523,
"step": 4145
},
{
"epoch": 1.6528102753024343,
"grad_norm": 7.305230617523193,
"learning_rate": 0.00012625672206206947,
"loss": 2.3545,
"step": 4150
},
{
"epoch": 1.6548016129835217,
"grad_norm": 6.4055495262146,
"learning_rate": 0.00012594781116899122,
"loss": 2.2055,
"step": 4155
},
{
"epoch": 1.656792950664609,
"grad_norm": 4.996709823608398,
"learning_rate": 0.00012563900494158084,
"loss": 2.13,
"step": 4160
},
{
"epoch": 1.6587842883456962,
"grad_norm": 5.305093765258789,
"learning_rate": 0.00012533030472364158,
"loss": 1.9513,
"step": 4165
},
{
"epoch": 1.6607756260267834,
"grad_norm": 5.111289024353027,
"learning_rate": 0.00012502171185851542,
"loss": 2.2238,
"step": 4170
},
{
"epoch": 1.6627669637078708,
"grad_norm": 5.95400333404541,
"learning_rate": 0.00012471322768907713,
"loss": 2.2567,
"step": 4175
},
{
"epoch": 1.6647583013889582,
"grad_norm": 5.4170918464660645,
"learning_rate": 0.00012440485355772854,
"loss": 2.1921,
"step": 4180
},
{
"epoch": 1.6667496390700451,
"grad_norm": 5.547909736633301,
"learning_rate": 0.00012409659080639258,
"loss": 2.1026,
"step": 4185
},
{
"epoch": 1.6687409767511325,
"grad_norm": 6.24788761138916,
"learning_rate": 0.00012378844077650755,
"loss": 2.2438,
"step": 4190
},
{
"epoch": 1.67073231443222,
"grad_norm": 5.332764148712158,
"learning_rate": 0.00012348040480902121,
"loss": 2.1973,
"step": 4195
},
{
"epoch": 1.672723652113307,
"grad_norm": 5.244167804718018,
"learning_rate": 0.00012317248424438494,
"loss": 2.1514,
"step": 4200
},
{
"epoch": 1.6747149897943943,
"grad_norm": 4.9936676025390625,
"learning_rate": 0.000122864680422548,
"loss": 2.0853,
"step": 4205
},
{
"epoch": 1.6767063274754817,
"grad_norm": 6.823670387268066,
"learning_rate": 0.00012255699468295155,
"loss": 2.1558,
"step": 4210
},
{
"epoch": 1.678697665156569,
"grad_norm": 5.074921607971191,
"learning_rate": 0.00012224942836452295,
"loss": 2.0616,
"step": 4215
},
{
"epoch": 1.6806890028376562,
"grad_norm": 5.293245792388916,
"learning_rate": 0.00012194198280566983,
"loss": 2.1184,
"step": 4220
},
{
"epoch": 1.6826803405187434,
"grad_norm": 5.603009223937988,
"learning_rate": 0.00012163465934427443,
"loss": 2.219,
"step": 4225
},
{
"epoch": 1.6846716781998308,
"grad_norm": 4.916954517364502,
"learning_rate": 0.00012132745931768762,
"loss": 2.0718,
"step": 4230
},
{
"epoch": 1.6866630158809182,
"grad_norm": 8.043903350830078,
"learning_rate": 0.00012102038406272302,
"loss": 2.1528,
"step": 4235
},
{
"epoch": 1.6886543535620053,
"grad_norm": 4.8785505294799805,
"learning_rate": 0.00012071343491565151,
"loss": 2.1236,
"step": 4240
},
{
"epoch": 1.6906456912430925,
"grad_norm": 5.881628513336182,
"learning_rate": 0.00012040661321219498,
"loss": 2.2627,
"step": 4245
},
{
"epoch": 1.6926370289241799,
"grad_norm": 5.179254531860352,
"learning_rate": 0.0001200999202875209,
"loss": 2.0575,
"step": 4250
},
{
"epoch": 1.694628366605267,
"grad_norm": 4.581838607788086,
"learning_rate": 0.0001197933574762362,
"loss": 2.0092,
"step": 4255
},
{
"epoch": 1.6966197042863542,
"grad_norm": 5.721724510192871,
"learning_rate": 0.0001194869261123818,
"loss": 2.0574,
"step": 4260
},
{
"epoch": 1.6986110419674416,
"grad_norm": 6.110555648803711,
"learning_rate": 0.00011918062752942638,
"loss": 2.1512,
"step": 4265
},
{
"epoch": 1.700602379648529,
"grad_norm": 6.568111896514893,
"learning_rate": 0.00011887446306026098,
"loss": 2.1205,
"step": 4270
},
{
"epoch": 1.7025937173296162,
"grad_norm": 6.348562240600586,
"learning_rate": 0.00011856843403719302,
"loss": 2.1388,
"step": 4275
},
{
"epoch": 1.7045850550107033,
"grad_norm": 4.70415735244751,
"learning_rate": 0.00011826254179194041,
"loss": 2.0661,
"step": 4280
},
{
"epoch": 1.7065763926917907,
"grad_norm": 5.290071487426758,
"learning_rate": 0.00011795678765562598,
"loss": 2.0906,
"step": 4285
},
{
"epoch": 1.7085677303728781,
"grad_norm": 4.59420919418335,
"learning_rate": 0.0001176511729587714,
"loss": 2.0528,
"step": 4290
},
{
"epoch": 1.7105590680539653,
"grad_norm": 5.748522758483887,
"learning_rate": 0.00011734569903129178,
"loss": 2.2097,
"step": 4295
},
{
"epoch": 1.7125504057350525,
"grad_norm": 5.9820876121521,
"learning_rate": 0.00011704036720248941,
"loss": 2.1171,
"step": 4300
},
{
"epoch": 1.7145417434161399,
"grad_norm": 5.448683738708496,
"learning_rate": 0.00011673517880104839,
"loss": 2.1346,
"step": 4305
},
{
"epoch": 1.716533081097227,
"grad_norm": 6.049896240234375,
"learning_rate": 0.0001164301351550287,
"loss": 2.1895,
"step": 4310
},
{
"epoch": 1.7185244187783142,
"grad_norm": 5.706995487213135,
"learning_rate": 0.00011612523759186024,
"loss": 2.1633,
"step": 4315
},
{
"epoch": 1.7205157564594016,
"grad_norm": 5.151224136352539,
"learning_rate": 0.00011582048743833739,
"loss": 2.1786,
"step": 4320
},
{
"epoch": 1.722507094140489,
"grad_norm": 7.260252952575684,
"learning_rate": 0.00011551588602061292,
"loss": 2.1119,
"step": 4325
},
{
"epoch": 1.7244984318215761,
"grad_norm": 6.449362754821777,
"learning_rate": 0.00011521143466419248,
"loss": 2.2628,
"step": 4330
},
{
"epoch": 1.7264897695026633,
"grad_norm": 6.293276309967041,
"learning_rate": 0.00011490713469392858,
"loss": 2.1769,
"step": 4335
},
{
"epoch": 1.7284811071837507,
"grad_norm": 5.702457904815674,
"learning_rate": 0.00011460298743401509,
"loss": 2.2575,
"step": 4340
},
{
"epoch": 1.730472444864838,
"grad_norm": 5.36246919631958,
"learning_rate": 0.0001142989942079813,
"loss": 2.2241,
"step": 4345
},
{
"epoch": 1.7324637825459253,
"grad_norm": 6.228604316711426,
"learning_rate": 0.00011399515633868613,
"loss": 2.2833,
"step": 4350
},
{
"epoch": 1.7344551202270124,
"grad_norm": 6.484302997589111,
"learning_rate": 0.00011369147514831259,
"loss": 2.3199,
"step": 4355
},
{
"epoch": 1.7364464579080998,
"grad_norm": 5.6527886390686035,
"learning_rate": 0.00011338795195836171,
"loss": 2.0978,
"step": 4360
},
{
"epoch": 1.738437795589187,
"grad_norm": 5.345698833465576,
"learning_rate": 0.00011308458808964716,
"loss": 2.0553,
"step": 4365
},
{
"epoch": 1.7404291332702742,
"grad_norm": 5.294981479644775,
"learning_rate": 0.00011278138486228912,
"loss": 2.1411,
"step": 4370
},
{
"epoch": 1.7424204709513615,
"grad_norm": 6.168553829193115,
"learning_rate": 0.00011247834359570883,
"loss": 2.251,
"step": 4375
},
{
"epoch": 1.744411808632449,
"grad_norm": 5.573464393615723,
"learning_rate": 0.00011217546560862279,
"loss": 2.14,
"step": 4380
},
{
"epoch": 1.746403146313536,
"grad_norm": 5.3229851722717285,
"learning_rate": 0.0001118727522190368,
"loss": 2.1176,
"step": 4385
},
{
"epoch": 1.7483944839946233,
"grad_norm": 5.849644184112549,
"learning_rate": 0.00011157020474424055,
"loss": 2.1246,
"step": 4390
},
{
"epoch": 1.7503858216757107,
"grad_norm": 6.106393814086914,
"learning_rate": 0.00011126782450080168,
"loss": 2.2544,
"step": 4395
},
{
"epoch": 1.752377159356798,
"grad_norm": 4.8731255531311035,
"learning_rate": 0.0001109656128045601,
"loss": 2.0802,
"step": 4400
},
{
"epoch": 1.7543684970378852,
"grad_norm": 6.5706257820129395,
"learning_rate": 0.00011066357097062226,
"loss": 2.0044,
"step": 4405
},
{
"epoch": 1.7563598347189724,
"grad_norm": 5.224118709564209,
"learning_rate": 0.00011036170031335544,
"loss": 2.2617,
"step": 4410
},
{
"epoch": 1.7583511724000598,
"grad_norm": 6.793256759643555,
"learning_rate": 0.0001100600021463821,
"loss": 2.2929,
"step": 4415
},
{
"epoch": 1.7603425100811472,
"grad_norm": 5.432855606079102,
"learning_rate": 0.00010975847778257396,
"loss": 2.2589,
"step": 4420
},
{
"epoch": 1.7623338477622341,
"grad_norm": 5.692535400390625,
"learning_rate": 0.00010945712853404654,
"loss": 2.2153,
"step": 4425
},
{
"epoch": 1.7643251854433215,
"grad_norm": 5.197839260101318,
"learning_rate": 0.00010915595571215317,
"loss": 2.2352,
"step": 4430
},
{
"epoch": 1.766316523124409,
"grad_norm": 5.767719268798828,
"learning_rate": 0.00010885496062747969,
"loss": 2.2606,
"step": 4435
},
{
"epoch": 1.768307860805496,
"grad_norm": 6.0898356437683105,
"learning_rate": 0.00010855414458983822,
"loss": 2.0726,
"step": 4440
},
{
"epoch": 1.7702991984865832,
"grad_norm": 5.859544277191162,
"learning_rate": 0.00010825350890826198,
"loss": 2.1762,
"step": 4445
},
{
"epoch": 1.7722905361676706,
"grad_norm": 5.864034175872803,
"learning_rate": 0.00010795305489099912,
"loss": 2.1124,
"step": 4450
},
{
"epoch": 1.774281873848758,
"grad_norm": 6.083292007446289,
"learning_rate": 0.00010765278384550751,
"loss": 1.9115,
"step": 4455
},
{
"epoch": 1.7762732115298452,
"grad_norm": 6.627115249633789,
"learning_rate": 0.00010735269707844854,
"loss": 2.111,
"step": 4460
},
{
"epoch": 1.7782645492109324,
"grad_norm": 6.6406168937683105,
"learning_rate": 0.00010705279589568188,
"loss": 2.225,
"step": 4465
},
{
"epoch": 1.7802558868920197,
"grad_norm": 5.369992256164551,
"learning_rate": 0.00010675308160225952,
"loss": 2.2207,
"step": 4470
},
{
"epoch": 1.7822472245731071,
"grad_norm": 5.870012283325195,
"learning_rate": 0.00010645355550242027,
"loss": 2.2278,
"step": 4475
},
{
"epoch": 1.7842385622541943,
"grad_norm": 5.092059135437012,
"learning_rate": 0.00010615421889958387,
"loss": 2.2037,
"step": 4480
},
{
"epoch": 1.7862298999352815,
"grad_norm": 5.079998970031738,
"learning_rate": 0.00010585507309634553,
"loss": 2.1696,
"step": 4485
},
{
"epoch": 1.7882212376163689,
"grad_norm": 5.64705753326416,
"learning_rate": 0.00010555611939447017,
"loss": 2.1788,
"step": 4490
},
{
"epoch": 1.790212575297456,
"grad_norm": 5.797013282775879,
"learning_rate": 0.00010525735909488672,
"loss": 2.102,
"step": 4495
},
{
"epoch": 1.7922039129785432,
"grad_norm": 6.012421131134033,
"learning_rate": 0.00010495879349768255,
"loss": 2.1638,
"step": 4500
},
{
"epoch": 1.7941952506596306,
"grad_norm": 6.438976287841797,
"learning_rate": 0.0001046604239020977,
"loss": 2.1953,
"step": 4505
},
{
"epoch": 1.796186588340718,
"grad_norm": 4.844203948974609,
"learning_rate": 0.00010436225160651935,
"loss": 2.0658,
"step": 4510
},
{
"epoch": 1.7981779260218052,
"grad_norm": 6.363651752471924,
"learning_rate": 0.00010406427790847608,
"loss": 2.1306,
"step": 4515
},
{
"epoch": 1.8001692637028923,
"grad_norm": 4.920433044433594,
"learning_rate": 0.00010376650410463223,
"loss": 2.1241,
"step": 4520
},
{
"epoch": 1.8021606013839797,
"grad_norm": 5.457321643829346,
"learning_rate": 0.00010346893149078234,
"loss": 2.1246,
"step": 4525
},
{
"epoch": 1.804151939065067,
"grad_norm": 4.993035793304443,
"learning_rate": 0.00010317156136184537,
"loss": 2.1185,
"step": 4530
},
{
"epoch": 1.8061432767461543,
"grad_norm": 5.458240032196045,
"learning_rate": 0.00010287439501185923,
"loss": 2.2005,
"step": 4535
},
{
"epoch": 1.8081346144272414,
"grad_norm": 5.164868354797363,
"learning_rate": 0.000102577433733975,
"loss": 2.2367,
"step": 4540
},
{
"epoch": 1.8101259521083288,
"grad_norm": 5.254199028015137,
"learning_rate": 0.0001022806788204514,
"loss": 2.1091,
"step": 4545
},
{
"epoch": 1.812117289789416,
"grad_norm": 5.778502941131592,
"learning_rate": 0.00010198413156264909,
"loss": 2.2249,
"step": 4550
},
{
"epoch": 1.8141086274705032,
"grad_norm": 5.061821460723877,
"learning_rate": 0.00010168779325102513,
"loss": 2.1306,
"step": 4555
},
{
"epoch": 1.8160999651515906,
"grad_norm": 5.960233688354492,
"learning_rate": 0.00010139166517512739,
"loss": 2.2575,
"step": 4560
},
{
"epoch": 1.818091302832678,
"grad_norm": 6.0032429695129395,
"learning_rate": 0.00010109574862358872,
"loss": 2.2546,
"step": 4565
},
{
"epoch": 1.8200826405137651,
"grad_norm": 6.485595226287842,
"learning_rate": 0.00010080004488412167,
"loss": 2.1074,
"step": 4570
},
{
"epoch": 1.8220739781948523,
"grad_norm": 5.097439765930176,
"learning_rate": 0.00010050455524351256,
"loss": 2.2229,
"step": 4575
},
{
"epoch": 1.8240653158759397,
"grad_norm": 5.461282253265381,
"learning_rate": 0.00010020928098761618,
"loss": 2.2268,
"step": 4580
},
{
"epoch": 1.826056653557027,
"grad_norm": 4.748180866241455,
"learning_rate": 9.991422340134994e-05,
"loss": 2.0874,
"step": 4585
},
{
"epoch": 1.8280479912381142,
"grad_norm": 7.213197231292725,
"learning_rate": 9.961938376868843e-05,
"loss": 2.2942,
"step": 4590
},
{
"epoch": 1.8300393289192014,
"grad_norm": 5.269963264465332,
"learning_rate": 9.932476337265785e-05,
"loss": 2.3143,
"step": 4595
},
{
"epoch": 1.8320306666002888,
"grad_norm": 5.104089260101318,
"learning_rate": 9.903036349533025e-05,
"loss": 2.161,
"step": 4600
},
{
"epoch": 1.834022004281376,
"grad_norm": 5.0273661613464355,
"learning_rate": 9.873618541781814e-05,
"loss": 2.121,
"step": 4605
},
{
"epoch": 1.8360133419624631,
"grad_norm": 5.1278557777404785,
"learning_rate": 9.844223042026885e-05,
"loss": 2.1467,
"step": 4610
},
{
"epoch": 1.8380046796435505,
"grad_norm": 6.490189552307129,
"learning_rate": 9.814849978185895e-05,
"loss": 2.1369,
"step": 4615
},
{
"epoch": 1.839996017324638,
"grad_norm": 5.803715705871582,
"learning_rate": 9.785499478078862e-05,
"loss": 2.2422,
"step": 4620
},
{
"epoch": 1.841987355005725,
"grad_norm": 5.247805118560791,
"learning_rate": 9.756171669427626e-05,
"loss": 2.2922,
"step": 4625
},
{
"epoch": 1.8439786926868122,
"grad_norm": 4.887587547302246,
"learning_rate": 9.726866679855279e-05,
"loss": 2.1578,
"step": 4630
},
{
"epoch": 1.8459700303678996,
"grad_norm": 5.077564239501953,
"learning_rate": 9.697584636885608e-05,
"loss": 2.1355,
"step": 4635
},
{
"epoch": 1.847961368048987,
"grad_norm": 4.351661205291748,
"learning_rate": 9.668325667942554e-05,
"loss": 2.1895,
"step": 4640
},
{
"epoch": 1.8499527057300742,
"grad_norm": 5.322985649108887,
"learning_rate": 9.639089900349638e-05,
"loss": 2.22,
"step": 4645
},
{
"epoch": 1.8519440434111614,
"grad_norm": 5.055166721343994,
"learning_rate": 9.609877461329433e-05,
"loss": 2.1724,
"step": 4650
},
{
"epoch": 1.8539353810922488,
"grad_norm": 5.176064968109131,
"learning_rate": 9.580688478002977e-05,
"loss": 2.1442,
"step": 4655
},
{
"epoch": 1.8559267187733361,
"grad_norm": 5.6811418533325195,
"learning_rate": 9.551523077389253e-05,
"loss": 2.1636,
"step": 4660
},
{
"epoch": 1.857918056454423,
"grad_norm": 6.5232038497924805,
"learning_rate": 9.522381386404618e-05,
"loss": 2.2238,
"step": 4665
},
{
"epoch": 1.8599093941355105,
"grad_norm": 7.1458964347839355,
"learning_rate": 9.493263531862245e-05,
"loss": 2.191,
"step": 4670
},
{
"epoch": 1.8619007318165979,
"grad_norm": 6.995946884155273,
"learning_rate": 9.464169640471596e-05,
"loss": 2.2831,
"step": 4675
},
{
"epoch": 1.863892069497685,
"grad_norm": 4.7528486251831055,
"learning_rate": 9.43509983883784e-05,
"loss": 2.1966,
"step": 4680
},
{
"epoch": 1.8658834071787722,
"grad_norm": 4.584712505340576,
"learning_rate": 9.406054253461323e-05,
"loss": 2.1003,
"step": 4685
},
{
"epoch": 1.8678747448598596,
"grad_norm": 5.8391547203063965,
"learning_rate": 9.377033010737011e-05,
"loss": 2.129,
"step": 4690
},
{
"epoch": 1.869866082540947,
"grad_norm": 5.241213798522949,
"learning_rate": 9.348036236953942e-05,
"loss": 2.1238,
"step": 4695
},
{
"epoch": 1.8718574202220342,
"grad_norm": 5.671873092651367,
"learning_rate": 9.319064058294673e-05,
"loss": 2.0219,
"step": 4700
},
{
"epoch": 1.8738487579031213,
"grad_norm": 5.927294731140137,
"learning_rate": 9.290116600834731e-05,
"loss": 2.2963,
"step": 4705
},
{
"epoch": 1.8758400955842087,
"grad_norm": 5.177212715148926,
"learning_rate": 9.261193990542072e-05,
"loss": 2.0044,
"step": 4710
},
{
"epoch": 1.8778314332652961,
"grad_norm": 5.216682434082031,
"learning_rate": 9.232296353276517e-05,
"loss": 2.072,
"step": 4715
},
{
"epoch": 1.8798227709463833,
"grad_norm": 5.458096981048584,
"learning_rate": 9.203423814789228e-05,
"loss": 2.112,
"step": 4720
},
{
"epoch": 1.8818141086274704,
"grad_norm": 6.221094131469727,
"learning_rate": 9.174576500722132e-05,
"loss": 2.1151,
"step": 4725
},
{
"epoch": 1.8838054463085578,
"grad_norm": 5.665980339050293,
"learning_rate": 9.1457545366074e-05,
"loss": 2.0835,
"step": 4730
},
{
"epoch": 1.885796783989645,
"grad_norm": 6.715048789978027,
"learning_rate": 9.116958047866882e-05,
"loss": 2.1682,
"step": 4735
},
{
"epoch": 1.8877881216707322,
"grad_norm": 5.3913679122924805,
"learning_rate": 9.088187159811576e-05,
"loss": 2.1956,
"step": 4740
},
{
"epoch": 1.8897794593518196,
"grad_norm": 4.760240077972412,
"learning_rate": 9.059441997641079e-05,
"loss": 2.1155,
"step": 4745
},
{
"epoch": 1.891770797032907,
"grad_norm": 4.783170700073242,
"learning_rate": 9.030722686443026e-05,
"loss": 2.0731,
"step": 4750
},
{
"epoch": 1.8937621347139941,
"grad_norm": 6.2002482414245605,
"learning_rate": 9.00202935119257e-05,
"loss": 2.15,
"step": 4755
},
{
"epoch": 1.8957534723950813,
"grad_norm": 5.836517810821533,
"learning_rate": 8.973362116751818e-05,
"loss": 2.0689,
"step": 4760
},
{
"epoch": 1.8977448100761687,
"grad_norm": 6.286279201507568,
"learning_rate": 8.944721107869308e-05,
"loss": 2.1428,
"step": 4765
},
{
"epoch": 1.899736147757256,
"grad_norm": 5.639240264892578,
"learning_rate": 8.916106449179444e-05,
"loss": 2.1202,
"step": 4770
},
{
"epoch": 1.9017274854383432,
"grad_norm": 4.700593948364258,
"learning_rate": 8.887518265201971e-05,
"loss": 2.0323,
"step": 4775
},
{
"epoch": 1.9037188231194304,
"grad_norm": 5.773509979248047,
"learning_rate": 8.858956680341422e-05,
"loss": 2.121,
"step": 4780
},
{
"epoch": 1.9057101608005178,
"grad_norm": 5.129843235015869,
"learning_rate": 8.830421818886586e-05,
"loss": 2.0498,
"step": 4785
},
{
"epoch": 1.907701498481605,
"grad_norm": 5.609828948974609,
"learning_rate": 8.801913805009959e-05,
"loss": 2.1391,
"step": 4790
},
{
"epoch": 1.9096928361626921,
"grad_norm": 4.961773872375488,
"learning_rate": 8.773432762767202e-05,
"loss": 2.1965,
"step": 4795
},
{
"epoch": 1.9116841738437795,
"grad_norm": 5.141656875610352,
"learning_rate": 8.744978816096618e-05,
"loss": 2.0892,
"step": 4800
},
{
"epoch": 1.913675511524867,
"grad_norm": 5.264577388763428,
"learning_rate": 8.71655208881859e-05,
"loss": 2.2029,
"step": 4805
},
{
"epoch": 1.915666849205954,
"grad_norm": 5.03646183013916,
"learning_rate": 8.688152704635062e-05,
"loss": 2.1529,
"step": 4810
},
{
"epoch": 1.9176581868870413,
"grad_norm": 5.15019416809082,
"learning_rate": 8.659780787128976e-05,
"loss": 2.1064,
"step": 4815
},
{
"epoch": 1.9196495245681287,
"grad_norm": 4.951735496520996,
"learning_rate": 8.631436459763774e-05,
"loss": 2.0913,
"step": 4820
},
{
"epoch": 1.921640862249216,
"grad_norm": 7.545402526855469,
"learning_rate": 8.603119845882815e-05,
"loss": 2.1739,
"step": 4825
},
{
"epoch": 1.9236321999303032,
"grad_norm": 7.116678237915039,
"learning_rate": 8.574831068708872e-05,
"loss": 2.1532,
"step": 4830
},
{
"epoch": 1.9256235376113904,
"grad_norm": 6.284858226776123,
"learning_rate": 8.546570251343568e-05,
"loss": 2.0795,
"step": 4835
},
{
"epoch": 1.9276148752924778,
"grad_norm": 5.071526050567627,
"learning_rate": 8.518337516766882e-05,
"loss": 2.1881,
"step": 4840
},
{
"epoch": 1.929606212973565,
"grad_norm": 5.03200626373291,
"learning_rate": 8.49013298783657e-05,
"loss": 2.024,
"step": 4845
},
{
"epoch": 1.931597550654652,
"grad_norm": 5.654317855834961,
"learning_rate": 8.461956787287639e-05,
"loss": 2.1415,
"step": 4850
},
{
"epoch": 1.9335888883357395,
"grad_norm": 5.167952060699463,
"learning_rate": 8.433809037731851e-05,
"loss": 2.0673,
"step": 4855
},
{
"epoch": 1.9355802260168269,
"grad_norm": 5.573550701141357,
"learning_rate": 8.405689861657135e-05,
"loss": 2.2054,
"step": 4860
},
{
"epoch": 1.937571563697914,
"grad_norm": 5.166540622711182,
"learning_rate": 8.37759938142709e-05,
"loss": 2.0969,
"step": 4865
},
{
"epoch": 1.9395629013790012,
"grad_norm": 4.89374303817749,
"learning_rate": 8.349537719280432e-05,
"loss": 2.1023,
"step": 4870
},
{
"epoch": 1.9415542390600886,
"grad_norm": 5.91700553894043,
"learning_rate": 8.321504997330495e-05,
"loss": 2.2145,
"step": 4875
},
{
"epoch": 1.943545576741176,
"grad_norm": 5.197847366333008,
"learning_rate": 8.293501337564653e-05,
"loss": 2.1721,
"step": 4880
},
{
"epoch": 1.9455369144222632,
"grad_norm": 4.873611927032471,
"learning_rate": 8.265526861843816e-05,
"loss": 2.1454,
"step": 4885
},
{
"epoch": 1.9475282521033503,
"grad_norm": 5.418996334075928,
"learning_rate": 8.237581691901915e-05,
"loss": 2.1281,
"step": 4890
},
{
"epoch": 1.9495195897844377,
"grad_norm": 5.226861476898193,
"learning_rate": 8.209665949345333e-05,
"loss": 2.2334,
"step": 4895
},
{
"epoch": 1.9515109274655251,
"grad_norm": 6.427994251251221,
"learning_rate": 8.181779755652409e-05,
"loss": 2.1161,
"step": 4900
},
{
"epoch": 1.953502265146612,
"grad_norm": 5.055788993835449,
"learning_rate": 8.153923232172882e-05,
"loss": 2.0538,
"step": 4905
},
{
"epoch": 1.9554936028276995,
"grad_norm": 5.816956996917725,
"learning_rate": 8.126096500127399e-05,
"loss": 2.1474,
"step": 4910
},
{
"epoch": 1.9574849405087869,
"grad_norm": 4.791458606719971,
"learning_rate": 8.098299680606955e-05,
"loss": 2.0556,
"step": 4915
},
{
"epoch": 1.959476278189874,
"grad_norm": 6.023045063018799,
"learning_rate": 8.07053289457237e-05,
"loss": 2.185,
"step": 4920
},
{
"epoch": 1.9614676158709612,
"grad_norm": 5.472064971923828,
"learning_rate": 8.042796262853786e-05,
"loss": 2.0859,
"step": 4925
},
{
"epoch": 1.9634589535520486,
"grad_norm": 4.683407783508301,
"learning_rate": 8.015089906150117e-05,
"loss": 2.1597,
"step": 4930
},
{
"epoch": 1.965450291233136,
"grad_norm": 5.039796352386475,
"learning_rate": 7.98741394502853e-05,
"loss": 2.0879,
"step": 4935
},
{
"epoch": 1.9674416289142231,
"grad_norm": Infinity,
"learning_rate": 7.965295141887069e-05,
"loss": 2.1786,
"step": 4940
},
{
"epoch": 1.9694329665953103,
"grad_norm": 4.787463665008545,
"learning_rate": 7.937674196219953e-05,
"loss": 2.2384,
"step": 4945
},
{
"epoch": 1.9714243042763977,
"grad_norm": 5.068600654602051,
"learning_rate": 7.910083983017654e-05,
"loss": 2.162,
"step": 4950
},
{
"epoch": 1.973415641957485,
"grad_norm": 4.847311496734619,
"learning_rate": 7.882524622341922e-05,
"loss": 2.0265,
"step": 4955
},
{
"epoch": 1.9754069796385723,
"grad_norm": 5.144827365875244,
"learning_rate": 7.854996234120263e-05,
"loss": 2.2106,
"step": 4960
},
{
"epoch": 1.9773983173196594,
"grad_norm": 5.553465366363525,
"learning_rate": 7.827498938145386e-05,
"loss": 2.0393,
"step": 4965
},
{
"epoch": 1.9793896550007468,
"grad_norm": 4.800995826721191,
"learning_rate": 7.800032854074723e-05,
"loss": 2.0138,
"step": 4970
},
{
"epoch": 1.981380992681834,
"grad_norm": 5.342923164367676,
"learning_rate": 7.772598101429869e-05,
"loss": 2.0343,
"step": 4975
},
{
"epoch": 1.9833723303629212,
"grad_norm": 6.1510443687438965,
"learning_rate": 7.745194799596073e-05,
"loss": 2.2479,
"step": 4980
},
{
"epoch": 1.9853636680440085,
"grad_norm": 5.93689489364624,
"learning_rate": 7.717823067821737e-05,
"loss": 2.2015,
"step": 4985
},
{
"epoch": 1.987355005725096,
"grad_norm": 4.906221389770508,
"learning_rate": 7.690483025217871e-05,
"loss": 2.0462,
"step": 4990
},
{
"epoch": 1.989346343406183,
"grad_norm": 5.339593410491943,
"learning_rate": 7.663174790757592e-05,
"loss": 2.1735,
"step": 4995
},
{
"epoch": 1.9913376810872703,
"grad_norm": 5.099614143371582,
"learning_rate": 7.635898483275588e-05,
"loss": 2.1925,
"step": 5000
},
{
"epoch": 1.9933290187683577,
"grad_norm": 4.778752326965332,
"learning_rate": 7.608654221467637e-05,
"loss": 2.1225,
"step": 5005
},
{
"epoch": 1.995320356449445,
"grad_norm": 4.891331672668457,
"learning_rate": 7.58144212389005e-05,
"loss": 2.1722,
"step": 5010
},
{
"epoch": 1.9973116941305322,
"grad_norm": 6.8483381271362305,
"learning_rate": 7.554262308959163e-05,
"loss": 2.0786,
"step": 5015
},
{
"epoch": 1.9993030318116194,
"grad_norm": 6.551092147827148,
"learning_rate": 7.52711489495086e-05,
"loss": 1.9824,
"step": 5020
},
{
"epoch": 2.001294369492707,
"grad_norm": 3.9980087280273438,
"learning_rate": 7.500000000000002e-05,
"loss": 1.6428,
"step": 5025
},
{
"epoch": 2.003285707173794,
"grad_norm": 4.957551956176758,
"learning_rate": 7.472917742099953e-05,
"loss": 1.6547,
"step": 5030
},
{
"epoch": 2.005277044854881,
"grad_norm": 4.297199249267578,
"learning_rate": 7.445868239102042e-05,
"loss": 1.5545,
"step": 5035
},
{
"epoch": 2.0072683825359685,
"grad_norm": 5.14373779296875,
"learning_rate": 7.418851608715084e-05,
"loss": 1.4708,
"step": 5040
},
{
"epoch": 2.009259720217056,
"grad_norm": 6.700247764587402,
"learning_rate": 7.39186796850482e-05,
"loss": 1.4207,
"step": 5045
},
{
"epoch": 2.0112510578981433,
"grad_norm": 8.440861701965332,
"learning_rate": 7.364917435893446e-05,
"loss": 1.5883,
"step": 5050
},
{
"epoch": 2.0132423955792302,
"grad_norm": 6.795040130615234,
"learning_rate": 7.338000128159074e-05,
"loss": 1.3543,
"step": 5055
},
{
"epoch": 2.0152337332603176,
"grad_norm": 9.68851375579834,
"learning_rate": 7.311116162435254e-05,
"loss": 1.5251,
"step": 5060
},
{
"epoch": 2.017225070941405,
"grad_norm": 6.082462787628174,
"learning_rate": 7.284265655710429e-05,
"loss": 1.4305,
"step": 5065
},
{
"epoch": 2.019216408622492,
"grad_norm": 4.950602054595947,
"learning_rate": 7.257448724827442e-05,
"loss": 1.4375,
"step": 5070
},
{
"epoch": 2.0212077463035794,
"grad_norm": 5.117416858673096,
"learning_rate": 7.23066548648304e-05,
"loss": 1.5044,
"step": 5075
},
{
"epoch": 2.0231990839846667,
"grad_norm": 6.734591960906982,
"learning_rate": 7.203916057227338e-05,
"loss": 1.4885,
"step": 5080
},
{
"epoch": 2.025190421665754,
"grad_norm": 5.435779571533203,
"learning_rate": 7.177200553463333e-05,
"loss": 1.4702,
"step": 5085
},
{
"epoch": 2.027181759346841,
"grad_norm": 5.962007999420166,
"learning_rate": 7.15051909144639e-05,
"loss": 1.4807,
"step": 5090
},
{
"epoch": 2.0291730970279285,
"grad_norm": 7.344203948974609,
"learning_rate": 7.123871787283748e-05,
"loss": 1.3754,
"step": 5095
},
{
"epoch": 2.031164434709016,
"grad_norm": 5.459081649780273,
"learning_rate": 7.097258756933995e-05,
"loss": 1.353,
"step": 5100
},
{
"epoch": 2.0331557723901033,
"grad_norm": 8.167816162109375,
"learning_rate": 7.070680116206563e-05,
"loss": 1.4694,
"step": 5105
},
{
"epoch": 2.03514711007119,
"grad_norm": 5.838662147521973,
"learning_rate": 7.044135980761259e-05,
"loss": 1.3246,
"step": 5110
},
{
"epoch": 2.0371384477522776,
"grad_norm": 6.302445888519287,
"learning_rate": 7.017626466107718e-05,
"loss": 1.5211,
"step": 5115
},
{
"epoch": 2.039129785433365,
"grad_norm": 5.905932426452637,
"learning_rate": 6.991151687604922e-05,
"loss": 1.4369,
"step": 5120
},
{
"epoch": 2.041121123114452,
"grad_norm": 5.84182596206665,
"learning_rate": 6.96471176046069e-05,
"loss": 1.3727,
"step": 5125
},
{
"epoch": 2.0431124607955393,
"grad_norm": 5.710243225097656,
"learning_rate": 6.938306799731201e-05,
"loss": 1.5538,
"step": 5130
},
{
"epoch": 2.0451037984766267,
"grad_norm": 5.345917701721191,
"learning_rate": 6.911936920320456e-05,
"loss": 1.4582,
"step": 5135
},
{
"epoch": 2.047095136157714,
"grad_norm": 5.621410369873047,
"learning_rate": 6.885602236979794e-05,
"loss": 1.4463,
"step": 5140
},
{
"epoch": 2.049086473838801,
"grad_norm": 6.868717670440674,
"learning_rate": 6.859302864307412e-05,
"loss": 1.4782,
"step": 5145
},
{
"epoch": 2.0510778115198884,
"grad_norm": 5.948458671569824,
"learning_rate": 6.833038916747834e-05,
"loss": 1.4206,
"step": 5150
},
{
"epoch": 2.053069149200976,
"grad_norm": 6.426418304443359,
"learning_rate": 6.806810508591432e-05,
"loss": 1.4499,
"step": 5155
},
{
"epoch": 2.055060486882063,
"grad_norm": 6.86967658996582,
"learning_rate": 6.780617753973917e-05,
"loss": 1.5276,
"step": 5160
},
{
"epoch": 2.05705182456315,
"grad_norm": 6.395193576812744,
"learning_rate": 6.75446076687587e-05,
"loss": 1.4228,
"step": 5165
},
{
"epoch": 2.0590431622442376,
"grad_norm": 6.809487342834473,
"learning_rate": 6.728339661122211e-05,
"loss": 1.4222,
"step": 5170
},
{
"epoch": 2.061034499925325,
"grad_norm": 5.599830627441406,
"learning_rate": 6.702254550381711e-05,
"loss": 1.4701,
"step": 5175
},
{
"epoch": 2.063025837606412,
"grad_norm": 5.606078147888184,
"learning_rate": 6.676205548166525e-05,
"loss": 1.5687,
"step": 5180
},
{
"epoch": 2.0650171752874993,
"grad_norm": 6.464372634887695,
"learning_rate": 6.650192767831659e-05,
"loss": 1.4237,
"step": 5185
},
{
"epoch": 2.0670085129685867,
"grad_norm": 5.501936435699463,
"learning_rate": 6.624216322574514e-05,
"loss": 1.451,
"step": 5190
},
{
"epoch": 2.068999850649674,
"grad_norm": 6.523349285125732,
"learning_rate": 6.598276325434356e-05,
"loss": 1.3937,
"step": 5195
},
{
"epoch": 2.070991188330761,
"grad_norm": 5.1206278800964355,
"learning_rate": 6.57237288929186e-05,
"loss": 1.4379,
"step": 5200
},
{
"epoch": 2.0729825260118484,
"grad_norm": 8.085687637329102,
"learning_rate": 6.546506126868592e-05,
"loss": 1.6361,
"step": 5205
},
{
"epoch": 2.074973863692936,
"grad_norm": 6.567965507507324,
"learning_rate": 6.52067615072653e-05,
"loss": 1.5065,
"step": 5210
},
{
"epoch": 2.076965201374023,
"grad_norm": 5.279557228088379,
"learning_rate": 6.494883073267566e-05,
"loss": 1.5345,
"step": 5215
},
{
"epoch": 2.07895653905511,
"grad_norm": 5.705742835998535,
"learning_rate": 6.46912700673304e-05,
"loss": 1.3541,
"step": 5220
},
{
"epoch": 2.0809478767361975,
"grad_norm": 5.617422103881836,
"learning_rate": 6.443408063203225e-05,
"loss": 1.4141,
"step": 5225
},
{
"epoch": 2.082939214417285,
"grad_norm": 7.766902923583984,
"learning_rate": 6.417726354596842e-05,
"loss": 1.3367,
"step": 5230
},
{
"epoch": 2.0849305520983723,
"grad_norm": 7.059458255767822,
"learning_rate": 6.392081992670598e-05,
"loss": 1.473,
"step": 5235
},
{
"epoch": 2.0869218897794592,
"grad_norm": 6.633787631988525,
"learning_rate": 6.366475089018668e-05,
"loss": 1.4581,
"step": 5240
},
{
"epoch": 2.0889132274605466,
"grad_norm": 5.493014335632324,
"learning_rate": 6.340905755072227e-05,
"loss": 1.4396,
"step": 5245
},
{
"epoch": 2.090904565141634,
"grad_norm": 6.356673717498779,
"learning_rate": 6.315374102098956e-05,
"loss": 1.4986,
"step": 5250
},
{
"epoch": 2.092895902822721,
"grad_norm": 6.202879905700684,
"learning_rate": 6.28988024120258e-05,
"loss": 1.3253,
"step": 5255
},
{
"epoch": 2.0948872405038084,
"grad_norm": 6.061187744140625,
"learning_rate": 6.264424283322353e-05,
"loss": 1.4252,
"step": 5260
},
{
"epoch": 2.0968785781848958,
"grad_norm": 6.786153316497803,
"learning_rate": 6.239006339232592e-05,
"loss": 1.3983,
"step": 5265
},
{
"epoch": 2.098869915865983,
"grad_norm": 6.876367092132568,
"learning_rate": 6.213626519542189e-05,
"loss": 1.4465,
"step": 5270
},
{
"epoch": 2.10086125354707,
"grad_norm": 5.9195170402526855,
"learning_rate": 6.188284934694152e-05,
"loss": 1.4788,
"step": 5275
},
{
"epoch": 2.1028525912281575,
"grad_norm": 6.0256242752075195,
"learning_rate": 6.162981694965086e-05,
"loss": 1.5226,
"step": 5280
},
{
"epoch": 2.104843928909245,
"grad_norm": 8.436933517456055,
"learning_rate": 6.137716910464735e-05,
"loss": 1.4933,
"step": 5285
},
{
"epoch": 2.1068352665903323,
"grad_norm": 6.069702625274658,
"learning_rate": 6.112490691135518e-05,
"loss": 1.4294,
"step": 5290
},
{
"epoch": 2.108826604271419,
"grad_norm": 6.045875549316406,
"learning_rate": 6.087303146752018e-05,
"loss": 1.4371,
"step": 5295
},
{
"epoch": 2.1108179419525066,
"grad_norm": 7.140870571136475,
"learning_rate": 6.062154386920521e-05,
"loss": 1.4965,
"step": 5300
},
{
"epoch": 2.112809279633594,
"grad_norm": 7.140031814575195,
"learning_rate": 6.037044521078536e-05,
"loss": 1.5039,
"step": 5305
},
{
"epoch": 2.114800617314681,
"grad_norm": 6.5458879470825195,
"learning_rate": 6.011973658494337e-05,
"loss": 1.478,
"step": 5310
},
{
"epoch": 2.1167919549957683,
"grad_norm": 6.794205665588379,
"learning_rate": 5.9869419082664525e-05,
"loss": 1.4578,
"step": 5315
},
{
"epoch": 2.1187832926768557,
"grad_norm": 6.228168487548828,
"learning_rate": 5.961949379323214e-05,
"loss": 1.3406,
"step": 5320
},
{
"epoch": 2.120774630357943,
"grad_norm": 6.367410182952881,
"learning_rate": 5.9369961804222875e-05,
"loss": 1.3709,
"step": 5325
},
{
"epoch": 2.12276596803903,
"grad_norm": 6.626134872436523,
"learning_rate": 5.912082420150179e-05,
"loss": 1.3714,
"step": 5330
},
{
"epoch": 2.1247573057201175,
"grad_norm": 5.566625595092773,
"learning_rate": 5.887208206921774e-05,
"loss": 1.3924,
"step": 5335
},
{
"epoch": 2.126748643401205,
"grad_norm": 5.629339218139648,
"learning_rate": 5.862373648979865e-05,
"loss": 1.5259,
"step": 5340
},
{
"epoch": 2.1287399810822922,
"grad_norm": 6.338861465454102,
"learning_rate": 5.837578854394692e-05,
"loss": 1.4742,
"step": 5345
},
{
"epoch": 2.130731318763379,
"grad_norm": 6.18589448928833,
"learning_rate": 5.812823931063445e-05,
"loss": 1.387,
"step": 5350
},
{
"epoch": 2.1327226564444666,
"grad_norm": 5.2369704246521,
"learning_rate": 5.78810898670981e-05,
"loss": 1.4379,
"step": 5355
},
{
"epoch": 2.134713994125554,
"grad_norm": 6.575031280517578,
"learning_rate": 5.7634341288835196e-05,
"loss": 1.4462,
"step": 5360
},
{
"epoch": 2.136705331806641,
"grad_norm": 5.535890102386475,
"learning_rate": 5.738799464959845e-05,
"loss": 1.4794,
"step": 5365
},
{
"epoch": 2.1386966694877283,
"grad_norm": 6.6877827644348145,
"learning_rate": 5.714205102139157e-05,
"loss": 1.4962,
"step": 5370
},
{
"epoch": 2.1406880071688157,
"grad_norm": 6.816041469573975,
"learning_rate": 5.689651147446448e-05,
"loss": 1.3646,
"step": 5375
},
{
"epoch": 2.142679344849903,
"grad_norm": 5.999199390411377,
"learning_rate": 5.66513770773088e-05,
"loss": 1.4707,
"step": 5380
},
{
"epoch": 2.14467068253099,
"grad_norm": 6.749256610870361,
"learning_rate": 5.6406648896653026e-05,
"loss": 1.4804,
"step": 5385
},
{
"epoch": 2.1466620202120774,
"grad_norm": 5.8225603103637695,
"learning_rate": 5.616232799745788e-05,
"loss": 1.4235,
"step": 5390
},
{
"epoch": 2.148653357893165,
"grad_norm": 6.044665336608887,
"learning_rate": 5.5918415442911974e-05,
"loss": 1.4927,
"step": 5395
},
{
"epoch": 2.150644695574252,
"grad_norm": 7.252010822296143,
"learning_rate": 5.567491229442675e-05,
"loss": 1.4286,
"step": 5400
},
{
"epoch": 2.152636033255339,
"grad_norm": 5.604063987731934,
"learning_rate": 5.543181961163219e-05,
"loss": 1.376,
"step": 5405
},
{
"epoch": 2.1546273709364265,
"grad_norm": 6.4235029220581055,
"learning_rate": 5.518913845237197e-05,
"loss": 1.4612,
"step": 5410
},
{
"epoch": 2.156618708617514,
"grad_norm": 7.064362525939941,
"learning_rate": 5.494686987269922e-05,
"loss": 1.4151,
"step": 5415
},
{
"epoch": 2.158610046298601,
"grad_norm": 5.43442964553833,
"learning_rate": 5.470501492687144e-05,
"loss": 1.3365,
"step": 5420
},
{
"epoch": 2.1606013839796883,
"grad_norm": 6.117883205413818,
"learning_rate": 5.4463574667346224e-05,
"loss": 1.3991,
"step": 5425
},
{
"epoch": 2.1625927216607757,
"grad_norm": 7.441193580627441,
"learning_rate": 5.4222550144776716e-05,
"loss": 1.5015,
"step": 5430
},
{
"epoch": 2.164584059341863,
"grad_norm": 5.909252643585205,
"learning_rate": 5.398194240800683e-05,
"loss": 1.507,
"step": 5435
},
{
"epoch": 2.16657539702295,
"grad_norm": 5.986944198608398,
"learning_rate": 5.374175250406682e-05,
"loss": 1.4358,
"step": 5440
},
{
"epoch": 2.1685667347040374,
"grad_norm": 6.824896335601807,
"learning_rate": 5.350198147816866e-05,
"loss": 1.4419,
"step": 5445
},
{
"epoch": 2.1705580723851248,
"grad_norm": 6.392977237701416,
"learning_rate": 5.326263037370164e-05,
"loss": 1.3984,
"step": 5450
},
{
"epoch": 2.172549410066212,
"grad_norm": 5.7649126052856445,
"learning_rate": 5.302370023222764e-05,
"loss": 1.4324,
"step": 5455
},
{
"epoch": 2.174540747747299,
"grad_norm": 6.149197101593018,
"learning_rate": 5.27851920934766e-05,
"loss": 1.3704,
"step": 5460
},
{
"epoch": 2.1765320854283865,
"grad_norm": 6.189772605895996,
"learning_rate": 5.254710699534225e-05,
"loss": 1.4444,
"step": 5465
},
{
"epoch": 2.178523423109474,
"grad_norm": 6.396068572998047,
"learning_rate": 5.2309445973877304e-05,
"loss": 1.4017,
"step": 5470
},
{
"epoch": 2.180514760790561,
"grad_norm": 7.503396987915039,
"learning_rate": 5.2072210063289014e-05,
"loss": 1.4759,
"step": 5475
},
{
"epoch": 2.1825060984716482,
"grad_norm": 5.856228828430176,
"learning_rate": 5.183540029593478e-05,
"loss": 1.4067,
"step": 5480
},
{
"epoch": 2.1844974361527356,
"grad_norm": 6.6336140632629395,
"learning_rate": 5.1599017702317665e-05,
"loss": 1.3564,
"step": 5485
},
{
"epoch": 2.186488773833823,
"grad_norm": 5.758519172668457,
"learning_rate": 5.1363063311081715e-05,
"loss": 1.4527,
"step": 5490
},
{
"epoch": 2.18848011151491,
"grad_norm": 6.567979335784912,
"learning_rate": 5.1127538149007616e-05,
"loss": 1.4816,
"step": 5495
},
{
"epoch": 2.1904714491959973,
"grad_norm": 5.97678804397583,
"learning_rate": 5.089244324100834e-05,
"loss": 1.5056,
"step": 5500
},
{
"epoch": 2.1924627868770847,
"grad_norm": 6.623924732208252,
"learning_rate": 5.065777961012445e-05,
"loss": 1.5091,
"step": 5505
},
{
"epoch": 2.194454124558172,
"grad_norm": 6.162147521972656,
"learning_rate": 5.0423548277519795e-05,
"loss": 1.4491,
"step": 5510
},
{
"epoch": 2.196445462239259,
"grad_norm": 6.854715824127197,
"learning_rate": 5.0189750262476965e-05,
"loss": 1.4854,
"step": 5515
},
{
"epoch": 2.1984367999203465,
"grad_norm": 6.7204270362854,
"learning_rate": 4.995638658239309e-05,
"loss": 1.3953,
"step": 5520
},
{
"epoch": 2.200428137601434,
"grad_norm": 5.354440212249756,
"learning_rate": 4.9723458252775093e-05,
"loss": 1.3548,
"step": 5525
},
{
"epoch": 2.202419475282521,
"grad_norm": 6.147752285003662,
"learning_rate": 4.9490966287235416e-05,
"loss": 1.4881,
"step": 5530
},
{
"epoch": 2.204410812963608,
"grad_norm": 6.762166500091553,
"learning_rate": 4.9258911697487776e-05,
"loss": 1.3972,
"step": 5535
},
{
"epoch": 2.2064021506446956,
"grad_norm": 7.021646976470947,
"learning_rate": 4.902729549334245e-05,
"loss": 1.4337,
"step": 5540
},
{
"epoch": 2.208393488325783,
"grad_norm": 6.543638706207275,
"learning_rate": 4.879611868270207e-05,
"loss": 1.3513,
"step": 5545
},
{
"epoch": 2.21038482600687,
"grad_norm": 6.03616189956665,
"learning_rate": 4.856538227155718e-05,
"loss": 1.5526,
"step": 5550
},
{
"epoch": 2.2123761636879573,
"grad_norm": 6.789417743682861,
"learning_rate": 4.8335087263981994e-05,
"loss": 1.3557,
"step": 5555
},
{
"epoch": 2.2143675013690447,
"grad_norm": 6.264193058013916,
"learning_rate": 4.810523466212979e-05,
"loss": 1.4205,
"step": 5560
},
{
"epoch": 2.216358839050132,
"grad_norm": 6.371268272399902,
"learning_rate": 4.787582546622864e-05,
"loss": 1.4186,
"step": 5565
},
{
"epoch": 2.218350176731219,
"grad_norm": 6.438905239105225,
"learning_rate": 4.764686067457726e-05,
"loss": 1.4787,
"step": 5570
},
{
"epoch": 2.2203415144123064,
"grad_norm": 6.756014347076416,
"learning_rate": 4.7418341283540354e-05,
"loss": 1.4171,
"step": 5575
},
{
"epoch": 2.222332852093394,
"grad_norm": 6.023416042327881,
"learning_rate": 4.719026828754442e-05,
"loss": 1.4746,
"step": 5580
},
{
"epoch": 2.224324189774481,
"grad_norm": 6.320237636566162,
"learning_rate": 4.696264267907346e-05,
"loss": 1.4176,
"step": 5585
},
{
"epoch": 2.226315527455568,
"grad_norm": 6.24311637878418,
"learning_rate": 4.6735465448664656e-05,
"loss": 1.3984,
"step": 5590
},
{
"epoch": 2.2283068651366555,
"grad_norm": 6.375988006591797,
"learning_rate": 4.650873758490394e-05,
"loss": 1.4745,
"step": 5595
},
{
"epoch": 2.230298202817743,
"grad_norm": 5.514700889587402,
"learning_rate": 4.628246007442182e-05,
"loss": 1.3082,
"step": 5600
},
{
"epoch": 2.23228954049883,
"grad_norm": 6.441384315490723,
"learning_rate": 4.605663390188907e-05,
"loss": 1.3744,
"step": 5605
},
{
"epoch": 2.2342808781799173,
"grad_norm": 5.932528495788574,
"learning_rate": 4.583126005001241e-05,
"loss": 1.4119,
"step": 5610
},
{
"epoch": 2.2362722158610047,
"grad_norm": 6.599247455596924,
"learning_rate": 4.560633949953016e-05,
"loss": 1.4283,
"step": 5615
},
{
"epoch": 2.238263553542092,
"grad_norm": 5.809045314788818,
"learning_rate": 4.5381873229208125e-05,
"loss": 1.4476,
"step": 5620
},
{
"epoch": 2.240254891223179,
"grad_norm": 5.800905227661133,
"learning_rate": 4.51578622158353e-05,
"loss": 1.4383,
"step": 5625
},
{
"epoch": 2.2422462289042664,
"grad_norm": 6.697018146514893,
"learning_rate": 4.493430743421949e-05,
"loss": 1.4488,
"step": 5630
},
{
"epoch": 2.244237566585354,
"grad_norm": 6.055057048797607,
"learning_rate": 4.471120985718314e-05,
"loss": 1.4049,
"step": 5635
},
{
"epoch": 2.246228904266441,
"grad_norm": 6.137016296386719,
"learning_rate": 4.44885704555593e-05,
"loss": 1.3719,
"step": 5640
},
{
"epoch": 2.248220241947528,
"grad_norm": 6.365423679351807,
"learning_rate": 4.4266390198187034e-05,
"loss": 1.4431,
"step": 5645
},
{
"epoch": 2.2502115796286155,
"grad_norm": 6.158652305603027,
"learning_rate": 4.404467005190746e-05,
"loss": 1.4635,
"step": 5650
},
{
"epoch": 2.252202917309703,
"grad_norm": 5.414857387542725,
"learning_rate": 4.38234109815594e-05,
"loss": 1.3857,
"step": 5655
},
{
"epoch": 2.2541942549907903,
"grad_norm": 6.830102920532227,
"learning_rate": 4.360261394997545e-05,
"loss": 1.4411,
"step": 5660
},
{
"epoch": 2.2561855926718772,
"grad_norm": 5.979335308074951,
"learning_rate": 4.338227991797741e-05,
"loss": 1.4889,
"step": 5665
},
{
"epoch": 2.2581769303529646,
"grad_norm": 5.826920032501221,
"learning_rate": 4.3162409844372325e-05,
"loss": 1.3643,
"step": 5670
},
{
"epoch": 2.260168268034052,
"grad_norm": 7.233572959899902,
"learning_rate": 4.2943004685948277e-05,
"loss": 1.566,
"step": 5675
},
{
"epoch": 2.262159605715139,
"grad_norm": 5.2802042961120605,
"learning_rate": 4.2724065397470337e-05,
"loss": 1.2933,
"step": 5680
},
{
"epoch": 2.2641509433962264,
"grad_norm": 5.731015205383301,
"learning_rate": 4.2505592931676116e-05,
"loss": 1.4204,
"step": 5685
},
{
"epoch": 2.2661422810773137,
"grad_norm": 5.166196346282959,
"learning_rate": 4.228758823927187e-05,
"loss": 1.3884,
"step": 5690
},
{
"epoch": 2.268133618758401,
"grad_norm": 7.443281173706055,
"learning_rate": 4.2070052268928364e-05,
"loss": 1.412,
"step": 5695
},
{
"epoch": 2.270124956439488,
"grad_norm": 5.445531845092773,
"learning_rate": 4.185298596727655e-05,
"loss": 1.506,
"step": 5700
},
{
"epoch": 2.2721162941205755,
"grad_norm": 6.483303546905518,
"learning_rate": 4.163639027890362e-05,
"loss": 1.429,
"step": 5705
},
{
"epoch": 2.274107631801663,
"grad_norm": 5.9170403480529785,
"learning_rate": 4.142026614634881e-05,
"loss": 1.4705,
"step": 5710
},
{
"epoch": 2.2760989694827503,
"grad_norm": 6.848852634429932,
"learning_rate": 4.1204614510099436e-05,
"loss": 1.4002,
"step": 5715
},
{
"epoch": 2.278090307163837,
"grad_norm": 6.7346577644348145,
"learning_rate": 4.098943630858658e-05,
"loss": 1.3614,
"step": 5720
},
{
"epoch": 2.2800816448449246,
"grad_norm": 6.159039497375488,
"learning_rate": 4.07747324781811e-05,
"loss": 1.4712,
"step": 5725
},
{
"epoch": 2.282072982526012,
"grad_norm": 6.049747943878174,
"learning_rate": 4.056050395318976e-05,
"loss": 1.398,
"step": 5730
},
{
"epoch": 2.284064320207099,
"grad_norm": 6.871042728424072,
"learning_rate": 4.034675166585079e-05,
"loss": 1.4639,
"step": 5735
},
{
"epoch": 2.2860556578881863,
"grad_norm": 6.381641387939453,
"learning_rate": 4.013347654633009e-05,
"loss": 1.4031,
"step": 5740
},
{
"epoch": 2.2880469955692737,
"grad_norm": 8.244357109069824,
"learning_rate": 3.992067952271708e-05,
"loss": 1.4173,
"step": 5745
},
{
"epoch": 2.290038333250361,
"grad_norm": 5.877044200897217,
"learning_rate": 3.970836152102083e-05,
"loss": 1.3303,
"step": 5750
},
{
"epoch": 2.292029670931448,
"grad_norm": 7.401268005371094,
"learning_rate": 3.9496523465165714e-05,
"loss": 1.4079,
"step": 5755
},
{
"epoch": 2.2940210086125354,
"grad_norm": 6.720442771911621,
"learning_rate": 3.928516627698764e-05,
"loss": 1.2561,
"step": 5760
},
{
"epoch": 2.296012346293623,
"grad_norm": 5.907570838928223,
"learning_rate": 3.907429087623007e-05,
"loss": 1.3237,
"step": 5765
},
{
"epoch": 2.29800368397471,
"grad_norm": 6.501235485076904,
"learning_rate": 3.8863898180539774e-05,
"loss": 1.3766,
"step": 5770
},
{
"epoch": 2.299995021655797,
"grad_norm": 5.766967296600342,
"learning_rate": 3.8653989105463004e-05,
"loss": 1.4697,
"step": 5775
},
{
"epoch": 2.3019863593368846,
"grad_norm": 6.114048480987549,
"learning_rate": 3.8444564564441515e-05,
"loss": 1.4737,
"step": 5780
},
{
"epoch": 2.303977697017972,
"grad_norm": 7.505747318267822,
"learning_rate": 3.8235625468808625e-05,
"loss": 1.4341,
"step": 5785
},
{
"epoch": 2.305969034699059,
"grad_norm": 5.183752059936523,
"learning_rate": 3.802717272778506e-05,
"loss": 1.3941,
"step": 5790
},
{
"epoch": 2.3079603723801463,
"grad_norm": 5.889784812927246,
"learning_rate": 3.781920724847517e-05,
"loss": 1.4962,
"step": 5795
},
{
"epoch": 2.3099517100612337,
"grad_norm": 6.595159530639648,
"learning_rate": 3.7611729935863e-05,
"loss": 1.3548,
"step": 5800
},
{
"epoch": 2.311943047742321,
"grad_norm": 5.552062034606934,
"learning_rate": 3.740474169280819e-05,
"loss": 1.3996,
"step": 5805
},
{
"epoch": 2.313934385423408,
"grad_norm": 6.149967193603516,
"learning_rate": 3.7198243420042205e-05,
"loss": 1.2514,
"step": 5810
},
{
"epoch": 2.3159257231044954,
"grad_norm": 7.16715145111084,
"learning_rate": 3.699223601616425e-05,
"loss": 1.5508,
"step": 5815
},
{
"epoch": 2.317917060785583,
"grad_norm": 6.43556547164917,
"learning_rate": 3.678672037763766e-05,
"loss": 1.4132,
"step": 5820
},
{
"epoch": 2.31990839846667,
"grad_norm": 6.692020893096924,
"learning_rate": 3.6581697398785606e-05,
"loss": 1.4769,
"step": 5825
},
{
"epoch": 2.321899736147757,
"grad_norm": 6.448044300079346,
"learning_rate": 3.637716797178745e-05,
"loss": 1.443,
"step": 5830
},
{
"epoch": 2.3238910738288445,
"grad_norm": 5.569827079772949,
"learning_rate": 3.61731329866749e-05,
"loss": 1.3258,
"step": 5835
},
{
"epoch": 2.325882411509932,
"grad_norm": 6.826900482177734,
"learning_rate": 3.5969593331327945e-05,
"loss": 1.4307,
"step": 5840
},
{
"epoch": 2.327873749191019,
"grad_norm": 6.331065654754639,
"learning_rate": 3.576654989147113e-05,
"loss": 1.3226,
"step": 5845
},
{
"epoch": 2.3298650868721063,
"grad_norm": 6.6188530921936035,
"learning_rate": 3.5564003550669626e-05,
"loss": 1.5756,
"step": 5850
},
{
"epoch": 2.3318564245531936,
"grad_norm": 5.677494049072266,
"learning_rate": 3.536195519032556e-05,
"loss": 1.4443,
"step": 5855
},
{
"epoch": 2.333847762234281,
"grad_norm": 6.42593240737915,
"learning_rate": 3.516040568967386e-05,
"loss": 1.3909,
"step": 5860
},
{
"epoch": 2.335839099915368,
"grad_norm": 7.254830360412598,
"learning_rate": 3.4959355925778686e-05,
"loss": 1.4784,
"step": 5865
},
{
"epoch": 2.3378304375964554,
"grad_norm": 5.911623954772949,
"learning_rate": 3.475880677352964e-05,
"loss": 1.4363,
"step": 5870
},
{
"epoch": 2.3398217752775428,
"grad_norm": 5.863831043243408,
"learning_rate": 3.455875910563771e-05,
"loss": 1.4029,
"step": 5875
},
{
"epoch": 2.34181311295863,
"grad_norm": 6.270824909210205,
"learning_rate": 3.4359213792631684e-05,
"loss": 1.3896,
"step": 5880
},
{
"epoch": 2.343804450639717,
"grad_norm": 6.321516513824463,
"learning_rate": 3.4160171702854285e-05,
"loss": 1.4431,
"step": 5885
},
{
"epoch": 2.3457957883208045,
"grad_norm": 6.566595554351807,
"learning_rate": 3.396163370245847e-05,
"loss": 1.4417,
"step": 5890
},
{
"epoch": 2.347787126001892,
"grad_norm": 7.686715602874756,
"learning_rate": 3.376360065540353e-05,
"loss": 1.3301,
"step": 5895
},
{
"epoch": 2.349778463682979,
"grad_norm": 7.394875526428223,
"learning_rate": 3.356607342345136e-05,
"loss": 1.3286,
"step": 5900
},
{
"epoch": 2.351769801364066,
"grad_norm": 6.7270426750183105,
"learning_rate": 3.3369052866162875e-05,
"loss": 1.3933,
"step": 5905
},
{
"epoch": 2.3537611390451536,
"grad_norm": 6.101992607116699,
"learning_rate": 3.3172539840894026e-05,
"loss": 1.4248,
"step": 5910
},
{
"epoch": 2.355752476726241,
"grad_norm": 7.11561918258667,
"learning_rate": 3.297653520279225e-05,
"loss": 1.3682,
"step": 5915
},
{
"epoch": 2.357743814407328,
"grad_norm": 6.916004657745361,
"learning_rate": 3.2781039804792603e-05,
"loss": 1.405,
"step": 5920
},
{
"epoch": 2.3597351520884153,
"grad_norm": 6.024013996124268,
"learning_rate": 3.258605449761425e-05,
"loss": 1.4808,
"step": 5925
},
{
"epoch": 2.3617264897695027,
"grad_norm": 7.276432991027832,
"learning_rate": 3.239158012975655e-05,
"loss": 1.4218,
"step": 5930
},
{
"epoch": 2.36371782745059,
"grad_norm": 6.543519020080566,
"learning_rate": 3.219761754749545e-05,
"loss": 1.5413,
"step": 5935
},
{
"epoch": 2.365709165131677,
"grad_norm": 6.037833213806152,
"learning_rate": 3.200416759487989e-05,
"loss": 1.4296,
"step": 5940
},
{
"epoch": 2.3677005028127645,
"grad_norm": 6.344012260437012,
"learning_rate": 3.181123111372797e-05,
"loss": 1.3827,
"step": 5945
},
{
"epoch": 2.369691840493852,
"grad_norm": 6.848827838897705,
"learning_rate": 3.1618808943623385e-05,
"loss": 1.5233,
"step": 5950
},
{
"epoch": 2.371683178174939,
"grad_norm": 5.740164756774902,
"learning_rate": 3.1426901921911715e-05,
"loss": 1.3816,
"step": 5955
},
{
"epoch": 2.373674515856026,
"grad_norm": 5.448791027069092,
"learning_rate": 3.1235510883696934e-05,
"loss": 1.3521,
"step": 5960
},
{
"epoch": 2.3756658535371136,
"grad_norm": 7.108462333679199,
"learning_rate": 3.104463666183755e-05,
"loss": 1.4244,
"step": 5965
},
{
"epoch": 2.377657191218201,
"grad_norm": 5.922284126281738,
"learning_rate": 3.085428008694306e-05,
"loss": 1.3821,
"step": 5970
},
{
"epoch": 2.379648528899288,
"grad_norm": 7.061075210571289,
"learning_rate": 3.066444198737053e-05,
"loss": 1.3369,
"step": 5975
},
{
"epoch": 2.3816398665803753,
"grad_norm": 7.461008548736572,
"learning_rate": 3.0475123189220664e-05,
"loss": 1.4149,
"step": 5980
},
{
"epoch": 2.3836312042614627,
"grad_norm": 6.245413303375244,
"learning_rate": 3.028632451633442e-05,
"loss": 1.3382,
"step": 5985
},
{
"epoch": 2.38562254194255,
"grad_norm": 7.291752338409424,
"learning_rate": 3.009804679028936e-05,
"loss": 1.4529,
"step": 5990
},
{
"epoch": 2.387613879623637,
"grad_norm": 6.265554904937744,
"learning_rate": 2.9910290830396166e-05,
"loss": 1.3111,
"step": 5995
},
{
"epoch": 2.3896052173047244,
"grad_norm": 5.732409477233887,
"learning_rate": 2.9723057453694942e-05,
"loss": 1.4022,
"step": 6000
},
{
"epoch": 2.391596554985812,
"grad_norm": 6.244343280792236,
"learning_rate": 2.9536347474951644e-05,
"loss": 1.4136,
"step": 6005
},
{
"epoch": 2.3935878926668988,
"grad_norm": 6.473920822143555,
"learning_rate": 2.9350161706654796e-05,
"loss": 1.4527,
"step": 6010
},
{
"epoch": 2.395579230347986,
"grad_norm": 6.109169960021973,
"learning_rate": 2.9164500959011565e-05,
"loss": 1.3211,
"step": 6015
},
{
"epoch": 2.3975705680290735,
"grad_norm": 6.401641368865967,
"learning_rate": 2.8979366039944552e-05,
"loss": 1.4453,
"step": 6020
},
{
"epoch": 2.399561905710161,
"grad_norm": 5.588070392608643,
"learning_rate": 2.8794757755088076e-05,
"loss": 1.307,
"step": 6025
},
{
"epoch": 2.4015532433912483,
"grad_norm": 5.913764953613281,
"learning_rate": 2.861067690778487e-05,
"loss": 1.3796,
"step": 6030
},
{
"epoch": 2.4035445810723353,
"grad_norm": 6.586962699890137,
"learning_rate": 2.842712429908237e-05,
"loss": 1.3723,
"step": 6035
},
{
"epoch": 2.4055359187534227,
"grad_norm": 5.830726146697998,
"learning_rate": 2.82441007277293e-05,
"loss": 1.6131,
"step": 6040
},
{
"epoch": 2.40752725643451,
"grad_norm": 6.545270919799805,
"learning_rate": 2.806160699017237e-05,
"loss": 1.4068,
"step": 6045
},
{
"epoch": 2.409518594115597,
"grad_norm": 6.324460506439209,
"learning_rate": 2.7879643880552512e-05,
"loss": 1.5127,
"step": 6050
},
{
"epoch": 2.4115099317966844,
"grad_norm": 7.322314262390137,
"learning_rate": 2.7698212190701623e-05,
"loss": 1.3749,
"step": 6055
},
{
"epoch": 2.4135012694777718,
"grad_norm": 5.913839817047119,
"learning_rate": 2.7517312710139027e-05,
"loss": 1.4688,
"step": 6060
},
{
"epoch": 2.4154926071588587,
"grad_norm": 6.860445976257324,
"learning_rate": 2.733694622606824e-05,
"loss": 1.4385,
"step": 6065
},
{
"epoch": 2.417483944839946,
"grad_norm": 6.26818323135376,
"learning_rate": 2.71571135233732e-05,
"loss": 1.3846,
"step": 6070
},
{
"epoch": 2.4194752825210335,
"grad_norm": 5.973329544067383,
"learning_rate": 2.6977815384615132e-05,
"loss": 1.4896,
"step": 6075
},
{
"epoch": 2.421466620202121,
"grad_norm": 6.393100738525391,
"learning_rate": 2.6799052590029003e-05,
"loss": 1.3961,
"step": 6080
},
{
"epoch": 2.4234579578832083,
"grad_norm": 6.824984073638916,
"learning_rate": 2.6620825917520257e-05,
"loss": 1.3814,
"step": 6085
},
{
"epoch": 2.4254492955642952,
"grad_norm": 6.342784404754639,
"learning_rate": 2.6443136142661263e-05,
"loss": 1.431,
"step": 6090
},
{
"epoch": 2.4274406332453826,
"grad_norm": 6.881380558013916,
"learning_rate": 2.6265984038688005e-05,
"loss": 1.3264,
"step": 6095
},
{
"epoch": 2.42943197092647,
"grad_norm": 6.984031677246094,
"learning_rate": 2.6089370376496854e-05,
"loss": 1.553,
"step": 6100
},
{
"epoch": 2.431423308607557,
"grad_norm": 6.409067630767822,
"learning_rate": 2.5913295924641004e-05,
"loss": 1.4004,
"step": 6105
},
{
"epoch": 2.4334146462886443,
"grad_norm": 7.744305610656738,
"learning_rate": 2.57377614493272e-05,
"loss": 1.431,
"step": 6110
},
{
"epoch": 2.4354059839697317,
"grad_norm": 6.909455299377441,
"learning_rate": 2.5562767714412418e-05,
"loss": 1.3365,
"step": 6115
},
{
"epoch": 2.437397321650819,
"grad_norm": 6.19822359085083,
"learning_rate": 2.538831548140069e-05,
"loss": 1.3417,
"step": 6120
},
{
"epoch": 2.439388659331906,
"grad_norm": 6.540812015533447,
"learning_rate": 2.5214405509439422e-05,
"loss": 1.3847,
"step": 6125
},
{
"epoch": 2.4413799970129935,
"grad_norm": 6.4344162940979,
"learning_rate": 2.5041038555316518e-05,
"loss": 1.4237,
"step": 6130
},
{
"epoch": 2.443371334694081,
"grad_norm": 7.177534103393555,
"learning_rate": 2.4868215373456712e-05,
"loss": 1.4874,
"step": 6135
},
{
"epoch": 2.4453626723751682,
"grad_norm": 5.777395725250244,
"learning_rate": 2.4695936715918607e-05,
"loss": 1.3981,
"step": 6140
},
{
"epoch": 2.447354010056255,
"grad_norm": 6.4199347496032715,
"learning_rate": 2.4524203332391144e-05,
"loss": 1.6551,
"step": 6145
},
{
"epoch": 2.4493453477373426,
"grad_norm": 7.651169300079346,
"learning_rate": 2.4353015970190453e-05,
"loss": 1.394,
"step": 6150
},
{
"epoch": 2.45133668541843,
"grad_norm": 6.039809703826904,
"learning_rate": 2.418237537425666e-05,
"loss": 1.3926,
"step": 6155
},
{
"epoch": 2.453328023099517,
"grad_norm": 6.839121341705322,
"learning_rate": 2.401228228715052e-05,
"loss": 1.3861,
"step": 6160
},
{
"epoch": 2.4553193607806043,
"grad_norm": 5.479107856750488,
"learning_rate": 2.3842737449050257e-05,
"loss": 1.2794,
"step": 6165
},
{
"epoch": 2.4573106984616917,
"grad_norm": 6.096414089202881,
"learning_rate": 2.367374159774832e-05,
"loss": 1.3764,
"step": 6170
},
{
"epoch": 2.459302036142779,
"grad_norm": 6.262311935424805,
"learning_rate": 2.3505295468648256e-05,
"loss": 1.3863,
"step": 6175
},
{
"epoch": 2.461293373823866,
"grad_norm": 7.217311382293701,
"learning_rate": 2.3337399794761336e-05,
"loss": 1.4574,
"step": 6180
},
{
"epoch": 2.4632847115049534,
"grad_norm": 5.660340309143066,
"learning_rate": 2.3170055306703512e-05,
"loss": 1.4165,
"step": 6185
},
{
"epoch": 2.465276049186041,
"grad_norm": 6.607390880584717,
"learning_rate": 2.3003262732692217e-05,
"loss": 1.4666,
"step": 6190
},
{
"epoch": 2.467267386867128,
"grad_norm": 5.438241481781006,
"learning_rate": 2.2837022798543132e-05,
"loss": 1.3601,
"step": 6195
},
{
"epoch": 2.469258724548215,
"grad_norm": 6.2114338874816895,
"learning_rate": 2.2671336227667097e-05,
"loss": 1.3728,
"step": 6200
},
{
"epoch": 2.4712500622293025,
"grad_norm": 6.168636798858643,
"learning_rate": 2.250620374106686e-05,
"loss": 1.3483,
"step": 6205
},
{
"epoch": 2.47324139991039,
"grad_norm": 6.182287216186523,
"learning_rate": 2.2341626057334133e-05,
"loss": 1.3469,
"step": 6210
},
{
"epoch": 2.475232737591477,
"grad_norm": 8.469073295593262,
"learning_rate": 2.2177603892646272e-05,
"loss": 1.4293,
"step": 6215
},
{
"epoch": 2.4772240752725643,
"grad_norm": 6.551890850067139,
"learning_rate": 2.2014137960763178e-05,
"loss": 1.4135,
"step": 6220
},
{
"epoch": 2.4792154129536517,
"grad_norm": 5.324213981628418,
"learning_rate": 2.1851228973024387e-05,
"loss": 1.3905,
"step": 6225
},
{
"epoch": 2.481206750634739,
"grad_norm": 7.160348892211914,
"learning_rate": 2.1688877638345725e-05,
"loss": 1.4851,
"step": 6230
},
{
"epoch": 2.483198088315826,
"grad_norm": 5.865573406219482,
"learning_rate": 2.152708466321636e-05,
"loss": 1.4202,
"step": 6235
},
{
"epoch": 2.4851894259969134,
"grad_norm": 6.2235918045043945,
"learning_rate": 2.136585075169565e-05,
"loss": 1.3384,
"step": 6240
},
{
"epoch": 2.487180763678001,
"grad_norm": 6.540336608886719,
"learning_rate": 2.1205176605410256e-05,
"loss": 1.3893,
"step": 6245
},
{
"epoch": 2.489172101359088,
"grad_norm": 6.606202602386475,
"learning_rate": 2.104506292355085e-05,
"loss": 1.4976,
"step": 6250
},
{
"epoch": 2.491163439040175,
"grad_norm": 6.5868659019470215,
"learning_rate": 2.0885510402869167e-05,
"loss": 1.3458,
"step": 6255
},
{
"epoch": 2.4931547767212625,
"grad_norm": 7.185669898986816,
"learning_rate": 2.072651973767512e-05,
"loss": 1.4209,
"step": 6260
},
{
"epoch": 2.49514611440235,
"grad_norm": 7.7647576332092285,
"learning_rate": 2.0568091619833523e-05,
"loss": 1.483,
"step": 6265
},
{
"epoch": 2.497137452083437,
"grad_norm": 6.41839075088501,
"learning_rate": 2.041022673876125e-05,
"loss": 1.3673,
"step": 6270
},
{
"epoch": 2.4991287897645242,
"grad_norm": 6.278010368347168,
"learning_rate": 2.025292578142415e-05,
"loss": 1.3931,
"step": 6275
},
{
"epoch": 2.5011201274456116,
"grad_norm": 6.852733135223389,
"learning_rate": 2.009618943233419e-05,
"loss": 1.4295,
"step": 6280
},
{
"epoch": 2.503111465126699,
"grad_norm": 5.997931957244873,
"learning_rate": 1.9940018373546336e-05,
"loss": 1.2439,
"step": 6285
},
{
"epoch": 2.505102802807786,
"grad_norm": 6.848467826843262,
"learning_rate": 1.9784413284655575e-05,
"loss": 1.348,
"step": 6290
},
{
"epoch": 2.5070941404888734,
"grad_norm": 6.354589462280273,
"learning_rate": 1.9629374842794067e-05,
"loss": 1.4389,
"step": 6295
},
{
"epoch": 2.5090854781699607,
"grad_norm": 5.17344856262207,
"learning_rate": 1.9474903722628192e-05,
"loss": 1.3658,
"step": 6300
},
{
"epoch": 2.511076815851048,
"grad_norm": 6.303539752960205,
"learning_rate": 1.932100059635551e-05,
"loss": 1.4752,
"step": 6305
},
{
"epoch": 2.513068153532135,
"grad_norm": 6.674501895904541,
"learning_rate": 1.9167666133701847e-05,
"loss": 1.3493,
"step": 6310
},
{
"epoch": 2.5150594912132225,
"grad_norm": 4.996592044830322,
"learning_rate": 1.901490100191857e-05,
"loss": 1.4374,
"step": 6315
},
{
"epoch": 2.51705082889431,
"grad_norm": 5.650495529174805,
"learning_rate": 1.886270586577944e-05,
"loss": 1.3024,
"step": 6320
},
{
"epoch": 2.519042166575397,
"grad_norm": 6.74934720993042,
"learning_rate": 1.8711081387577837e-05,
"loss": 1.4318,
"step": 6325
},
{
"epoch": 2.521033504256484,
"grad_norm": 6.7489423751831055,
"learning_rate": 1.8560028227123823e-05,
"loss": 1.462,
"step": 6330
},
{
"epoch": 2.5230248419375716,
"grad_norm": 5.957546710968018,
"learning_rate": 1.8409547041741434e-05,
"loss": 1.3598,
"step": 6335
},
{
"epoch": 2.525016179618659,
"grad_norm": 6.806410789489746,
"learning_rate": 1.825963848626557e-05,
"loss": 1.3982,
"step": 6340
},
{
"epoch": 2.5270075172997464,
"grad_norm": 7.132102966308594,
"learning_rate": 1.8110303213039296e-05,
"loss": 1.348,
"step": 6345
},
{
"epoch": 2.5289988549808333,
"grad_norm": 6.3546061515808105,
"learning_rate": 1.796154187191105e-05,
"loss": 1.4032,
"step": 6350
},
{
"epoch": 2.5309901926619207,
"grad_norm": 5.63569450378418,
"learning_rate": 1.7813355110231677e-05,
"loss": 1.3413,
"step": 6355
},
{
"epoch": 2.532981530343008,
"grad_norm": 6.957376480102539,
"learning_rate": 1.766574357285166e-05,
"loss": 1.3587,
"step": 6360
},
{
"epoch": 2.534972868024095,
"grad_norm": 7.172070026397705,
"learning_rate": 1.7518707902118378e-05,
"loss": 1.3324,
"step": 6365
},
{
"epoch": 2.5369642057051824,
"grad_norm": 6.005961894989014,
"learning_rate": 1.7372248737873278e-05,
"loss": 1.3518,
"step": 6370
},
{
"epoch": 2.53895554338627,
"grad_norm": 6.351656913757324,
"learning_rate": 1.7226366717449048e-05,
"loss": 1.3107,
"step": 6375
},
{
"epoch": 2.5409468810673568,
"grad_norm": 5.603379726409912,
"learning_rate": 1.708106247566683e-05,
"loss": 1.3946,
"step": 6380
},
{
"epoch": 2.542938218748444,
"grad_norm": 5.805838108062744,
"learning_rate": 1.6936336644833614e-05,
"loss": 1.3023,
"step": 6385
},
{
"epoch": 2.5449295564295316,
"grad_norm": 6.263773441314697,
"learning_rate": 1.6792189854739258e-05,
"loss": 1.3593,
"step": 6390
},
{
"epoch": 2.546920894110619,
"grad_norm": 6.895465850830078,
"learning_rate": 1.6648622732653943e-05,
"loss": 1.3542,
"step": 6395
},
{
"epoch": 2.5489122317917063,
"grad_norm": 6.908726692199707,
"learning_rate": 1.6505635903325275e-05,
"loss": 1.419,
"step": 6400
},
{
"epoch": 2.5509035694727933,
"grad_norm": 5.818434715270996,
"learning_rate": 1.6363229988975763e-05,
"loss": 1.4539,
"step": 6405
},
{
"epoch": 2.5528949071538807,
"grad_norm": 6.764200210571289,
"learning_rate": 1.6221405609299948e-05,
"loss": 1.4295,
"step": 6410
},
{
"epoch": 2.554886244834968,
"grad_norm": 6.165177345275879,
"learning_rate": 1.6080163381461702e-05,
"loss": 1.3533,
"step": 6415
},
{
"epoch": 2.556877582516055,
"grad_norm": 8.976422309875488,
"learning_rate": 1.5939503920091738e-05,
"loss": 1.3518,
"step": 6420
},
{
"epoch": 2.5588689201971424,
"grad_norm": 7.087680816650391,
"learning_rate": 1.5799427837284683e-05,
"loss": 1.4809,
"step": 6425
},
{
"epoch": 2.56086025787823,
"grad_norm": 5.952059268951416,
"learning_rate": 1.565993574259659e-05,
"loss": 1.3105,
"step": 6430
},
{
"epoch": 2.5628515955593167,
"grad_norm": 5.57841682434082,
"learning_rate": 1.5521028243042134e-05,
"loss": 1.3695,
"step": 6435
},
{
"epoch": 2.564842933240404,
"grad_norm": 5.929076671600342,
"learning_rate": 1.5382705943092218e-05,
"loss": 1.4548,
"step": 6440
},
{
"epoch": 2.5668342709214915,
"grad_norm": 6.7737202644348145,
"learning_rate": 1.5244969444671051e-05,
"loss": 1.3541,
"step": 6445
},
{
"epoch": 2.568825608602579,
"grad_norm": 6.4556074142456055,
"learning_rate": 1.5107819347153704e-05,
"loss": 1.5401,
"step": 6450
},
{
"epoch": 2.5708169462836663,
"grad_norm": 5.711570739746094,
"learning_rate": 1.4971256247363534e-05,
"loss": 1.4449,
"step": 6455
},
{
"epoch": 2.5728082839647533,
"grad_norm": 5.679041862487793,
"learning_rate": 1.4835280739569388e-05,
"loss": 1.2817,
"step": 6460
},
{
"epoch": 2.5747996216458406,
"grad_norm": 6.76185417175293,
"learning_rate": 1.4699893415483232e-05,
"loss": 1.4374,
"step": 6465
},
{
"epoch": 2.576790959326928,
"grad_norm": 6.594644546508789,
"learning_rate": 1.4565094864257426e-05,
"loss": 1.4975,
"step": 6470
},
{
"epoch": 2.578782297008015,
"grad_norm": 5.892383575439453,
"learning_rate": 1.4430885672482328e-05,
"loss": 1.3731,
"step": 6475
},
{
"epoch": 2.5807736346891024,
"grad_norm": 6.247217655181885,
"learning_rate": 1.4297266424183517e-05,
"loss": 1.3796,
"step": 6480
},
{
"epoch": 2.5827649723701898,
"grad_norm": 6.876238822937012,
"learning_rate": 1.416423770081941e-05,
"loss": 1.4786,
"step": 6485
},
{
"epoch": 2.5847563100512767,
"grad_norm": 5.32161283493042,
"learning_rate": 1.4031800081278754e-05,
"loss": 1.3543,
"step": 6490
},
{
"epoch": 2.586747647732364,
"grad_norm": 6.2356648445129395,
"learning_rate": 1.389995414187795e-05,
"loss": 1.3401,
"step": 6495
},
{
"epoch": 2.5887389854134515,
"grad_norm": 6.550376892089844,
"learning_rate": 1.3768700456358683e-05,
"loss": 1.4265,
"step": 6500
},
{
"epoch": 2.590730323094539,
"grad_norm": 6.80794095993042,
"learning_rate": 1.363803959588537e-05,
"loss": 1.439,
"step": 6505
},
{
"epoch": 2.5927216607756263,
"grad_norm": 7.110243797302246,
"learning_rate": 1.3507972129042737e-05,
"loss": 1.4863,
"step": 6510
},
{
"epoch": 2.594712998456713,
"grad_norm": 6.215062141418457,
"learning_rate": 1.3378498621833233e-05,
"loss": 1.4606,
"step": 6515
},
{
"epoch": 2.5967043361378006,
"grad_norm": 6.820135116577148,
"learning_rate": 1.3249619637674619e-05,
"loss": 1.4164,
"step": 6520
},
{
"epoch": 2.598695673818888,
"grad_norm": 5.777781009674072,
"learning_rate": 1.3121335737397609e-05,
"loss": 1.4159,
"step": 6525
},
{
"epoch": 2.600687011499975,
"grad_norm": 5.947566032409668,
"learning_rate": 1.2993647479243258e-05,
"loss": 1.3885,
"step": 6530
},
{
"epoch": 2.6026783491810623,
"grad_norm": 7.960350036621094,
"learning_rate": 1.2866555418860674e-05,
"loss": 1.3622,
"step": 6535
},
{
"epoch": 2.6046696868621497,
"grad_norm": 7.677003383636475,
"learning_rate": 1.2740060109304478e-05,
"loss": 1.3941,
"step": 6540
},
{
"epoch": 2.6066610245432367,
"grad_norm": 6.016810417175293,
"learning_rate": 1.2614162101032599e-05,
"loss": 1.4176,
"step": 6545
},
{
"epoch": 2.608652362224324,
"grad_norm": 5.751842498779297,
"learning_rate": 1.2488861941903593e-05,
"loss": 1.4261,
"step": 6550
},
{
"epoch": 2.6106436999054115,
"grad_norm": 5.662951469421387,
"learning_rate": 1.2364160177174503e-05,
"loss": 1.4162,
"step": 6555
},
{
"epoch": 2.612635037586499,
"grad_norm": 6.150815010070801,
"learning_rate": 1.22400573494984e-05,
"loss": 1.4112,
"step": 6560
},
{
"epoch": 2.6146263752675862,
"grad_norm": 6.814656734466553,
"learning_rate": 1.2116553998921957e-05,
"loss": 1.3148,
"step": 6565
},
{
"epoch": 2.616617712948673,
"grad_norm": 6.5078558921813965,
"learning_rate": 1.1993650662883219e-05,
"loss": 1.4669,
"step": 6570
},
{
"epoch": 2.6186090506297606,
"grad_norm": 7.931562900543213,
"learning_rate": 1.187134787620912e-05,
"loss": 1.4846,
"step": 6575
},
{
"epoch": 2.620600388310848,
"grad_norm": 6.5228590965271,
"learning_rate": 1.1749646171113364e-05,
"loss": 1.3225,
"step": 6580
},
{
"epoch": 2.622591725991935,
"grad_norm": 6.067448616027832,
"learning_rate": 1.1628546077193868e-05,
"loss": 1.3188,
"step": 6585
},
{
"epoch": 2.6245830636730223,
"grad_norm": 6.385129451751709,
"learning_rate": 1.1508048121430625e-05,
"loss": 1.4801,
"step": 6590
},
{
"epoch": 2.6265744013541097,
"grad_norm": 5.796778202056885,
"learning_rate": 1.1388152828183389e-05,
"loss": 1.3978,
"step": 6595
},
{
"epoch": 2.6285657390351966,
"grad_norm": 6.417087554931641,
"learning_rate": 1.1268860719189316e-05,
"loss": 1.3319,
"step": 6600
},
{
"epoch": 2.630557076716284,
"grad_norm": 6.347975254058838,
"learning_rate": 1.115017231356074e-05,
"loss": 1.3908,
"step": 6605
},
{
"epoch": 2.6325484143973714,
"grad_norm": 5.882072448730469,
"learning_rate": 1.1032088127782917e-05,
"loss": 1.3999,
"step": 6610
},
{
"epoch": 2.634539752078459,
"grad_norm": 5.941750526428223,
"learning_rate": 1.0914608675711839e-05,
"loss": 1.3493,
"step": 6615
},
{
"epoch": 2.636531089759546,
"grad_norm": 6.197680950164795,
"learning_rate": 1.0797734468571867e-05,
"loss": 1.3622,
"step": 6620
},
{
"epoch": 2.638522427440633,
"grad_norm": 7.5896992683410645,
"learning_rate": 1.0681466014953538e-05,
"loss": 1.4264,
"step": 6625
},
{
"epoch": 2.6405137651217205,
"grad_norm": 5.610791206359863,
"learning_rate": 1.056580382081153e-05,
"loss": 1.3576,
"step": 6630
},
{
"epoch": 2.642505102802808,
"grad_norm": 5.284849166870117,
"learning_rate": 1.0450748389462183e-05,
"loss": 1.4206,
"step": 6635
},
{
"epoch": 2.644496440483895,
"grad_norm": 6.58115816116333,
"learning_rate": 1.0336300221581511e-05,
"loss": 1.4038,
"step": 6640
},
{
"epoch": 2.6464877781649823,
"grad_norm": 6.350376129150391,
"learning_rate": 1.0222459815202883e-05,
"loss": 1.3237,
"step": 6645
},
{
"epoch": 2.6484791158460697,
"grad_norm": 5.541227340698242,
"learning_rate": 1.0109227665715059e-05,
"loss": 1.349,
"step": 6650
},
{
"epoch": 2.6504704535271566,
"grad_norm": 6.974477291107178,
"learning_rate": 9.99660426585977e-06,
"loss": 1.3641,
"step": 6655
},
{
"epoch": 2.652461791208244,
"grad_norm": 6.048290729522705,
"learning_rate": 9.884590105729784e-06,
"loss": 1.2903,
"step": 6660
},
{
"epoch": 2.6544531288893314,
"grad_norm": 5.928702354431152,
"learning_rate": 9.773185672766703e-06,
"loss": 1.4015,
"step": 6665
},
{
"epoch": 2.6564444665704188,
"grad_norm": 6.905900478363037,
"learning_rate": 9.662391451758788e-06,
"loss": 1.3503,
"step": 6670
},
{
"epoch": 2.658435804251506,
"grad_norm": 8.249216079711914,
"learning_rate": 9.552207924838923e-06,
"loss": 1.3357,
"step": 6675
},
{
"epoch": 2.660427141932593,
"grad_norm": 6.073180675506592,
"learning_rate": 9.442635571482481e-06,
"loss": 1.4308,
"step": 6680
},
{
"epoch": 2.6624184796136805,
"grad_norm": 7.015928745269775,
"learning_rate": 9.333674868505286e-06,
"loss": 1.2818,
"step": 6685
},
{
"epoch": 2.664409817294768,
"grad_norm": 7.025008678436279,
"learning_rate": 9.225326290061453e-06,
"loss": 1.387,
"step": 6690
},
{
"epoch": 2.666401154975855,
"grad_norm": 7.363583564758301,
"learning_rate": 9.11759030764136e-06,
"loss": 1.3665,
"step": 6695
},
{
"epoch": 2.6683924926569422,
"grad_norm": 7.165689945220947,
"learning_rate": 9.010467390069703e-06,
"loss": 1.5225,
"step": 6700
},
{
"epoch": 2.6703838303380296,
"grad_norm": 6.025979518890381,
"learning_rate": 8.903958003503264e-06,
"loss": 1.3151,
"step": 6705
},
{
"epoch": 2.6723751680191166,
"grad_norm": 6.4776411056518555,
"learning_rate": 8.798062611429007e-06,
"loss": 1.3935,
"step": 6710
},
{
"epoch": 2.674366505700204,
"grad_norm": 6.628039360046387,
"learning_rate": 8.692781674662002e-06,
"loss": 1.3811,
"step": 6715
},
{
"epoch": 2.6763578433812913,
"grad_norm": 6.728973388671875,
"learning_rate": 8.588115651343558e-06,
"loss": 1.4715,
"step": 6720
},
{
"epoch": 2.6783491810623787,
"grad_norm": 7.614314556121826,
"learning_rate": 8.484064996939021e-06,
"loss": 1.2541,
"step": 6725
},
{
"epoch": 2.680340518743466,
"grad_norm": 6.7132391929626465,
"learning_rate": 8.380630164235919e-06,
"loss": 1.3637,
"step": 6730
},
{
"epoch": 2.682331856424553,
"grad_norm": 6.705021381378174,
"learning_rate": 8.277811603341933e-06,
"loss": 1.422,
"step": 6735
},
{
"epoch": 2.6843231941056405,
"grad_norm": 6.249575138092041,
"learning_rate": 8.175609761683093e-06,
"loss": 1.4463,
"step": 6740
},
{
"epoch": 2.686314531786728,
"grad_norm": 6.032909870147705,
"learning_rate": 8.074025084001561e-06,
"loss": 1.4117,
"step": 6745
},
{
"epoch": 2.688305869467815,
"grad_norm": 6.830637454986572,
"learning_rate": 7.973058012353911e-06,
"loss": 1.394,
"step": 6750
},
{
"epoch": 2.690297207148902,
"grad_norm": 6.5258097648620605,
"learning_rate": 7.872708986109183e-06,
"loss": 1.4076,
"step": 6755
},
{
"epoch": 2.6922885448299896,
"grad_norm": 6.043037414550781,
"learning_rate": 7.772978441946836e-06,
"loss": 1.368,
"step": 6760
},
{
"epoch": 2.694279882511077,
"grad_norm": 6.692487716674805,
"learning_rate": 7.673866813855e-06,
"loss": 1.4427,
"step": 6765
},
{
"epoch": 2.696271220192164,
"grad_norm": 6.502307415008545,
"learning_rate": 7.575374533128492e-06,
"loss": 1.3994,
"step": 6770
},
{
"epoch": 2.6982625578732513,
"grad_norm": 5.603662014007568,
"learning_rate": 7.477502028366999e-06,
"loss": 1.4891,
"step": 6775
},
{
"epoch": 2.7002538955543387,
"grad_norm": 6.476789474487305,
"learning_rate": 7.380249725473186e-06,
"loss": 1.4082,
"step": 6780
},
{
"epoch": 2.702245233235426,
"grad_norm": 7.55186653137207,
"learning_rate": 7.2836180476508076e-06,
"loss": 1.3687,
"step": 6785
},
{
"epoch": 2.704236570916513,
"grad_norm": 6.1320977210998535,
"learning_rate": 7.187607415402946e-06,
"loss": 1.3346,
"step": 6790
},
{
"epoch": 2.7062279085976004,
"grad_norm": 5.8300395011901855,
"learning_rate": 7.092218246530129e-06,
"loss": 1.4623,
"step": 6795
},
{
"epoch": 2.708219246278688,
"grad_norm": 7.90705680847168,
"learning_rate": 6.997450956128481e-06,
"loss": 1.3667,
"step": 6800
},
{
"epoch": 2.7102105839597748,
"grad_norm": 5.899392604827881,
"learning_rate": 6.90330595658799e-06,
"loss": 1.3206,
"step": 6805
},
{
"epoch": 2.712201921640862,
"grad_norm": 6.035600662231445,
"learning_rate": 6.809783657590695e-06,
"loss": 1.4675,
"step": 6810
},
{
"epoch": 2.7141932593219495,
"grad_norm": 6.182140827178955,
"learning_rate": 6.7168844661088685e-06,
"loss": 1.2444,
"step": 6815
},
{
"epoch": 2.716184597003037,
"grad_norm": 6.4296698570251465,
"learning_rate": 6.6246087864032186e-06,
"loss": 1.3859,
"step": 6820
},
{
"epoch": 2.7181759346841243,
"grad_norm": 5.676371097564697,
"learning_rate": 6.532957020021257e-06,
"loss": 1.3183,
"step": 6825
},
{
"epoch": 2.7201672723652113,
"grad_norm": 5.844963550567627,
"learning_rate": 6.441929565795434e-06,
"loss": 1.3464,
"step": 6830
},
{
"epoch": 2.7221586100462987,
"grad_norm": 6.410009860992432,
"learning_rate": 6.351526819841407e-06,
"loss": 1.4351,
"step": 6835
},
{
"epoch": 2.724149947727386,
"grad_norm": 5.1571855545043945,
"learning_rate": 6.261749175556391e-06,
"loss": 1.3552,
"step": 6840
},
{
"epoch": 2.726141285408473,
"grad_norm": 7.127869129180908,
"learning_rate": 6.172597023617409e-06,
"loss": 1.4057,
"step": 6845
},
{
"epoch": 2.7281326230895604,
"grad_norm": 5.79174280166626,
"learning_rate": 6.084070751979547e-06,
"loss": 1.3631,
"step": 6850
},
{
"epoch": 2.730123960770648,
"grad_norm": 6.417024612426758,
"learning_rate": 5.996170745874335e-06,
"loss": 1.3564,
"step": 6855
},
{
"epoch": 2.7321152984517347,
"grad_norm": 5.934596061706543,
"learning_rate": 5.9088973878080316e-06,
"loss": 1.3883,
"step": 6860
},
{
"epoch": 2.734106636132822,
"grad_norm": 7.157084941864014,
"learning_rate": 5.8222510575599955e-06,
"loss": 1.4344,
"step": 6865
},
{
"epoch": 2.7360979738139095,
"grad_norm": 6.882761001586914,
"learning_rate": 5.736232132180951e-06,
"loss": 1.4617,
"step": 6870
},
{
"epoch": 2.738089311494997,
"grad_norm": 5.87106990814209,
"learning_rate": 5.6508409859914394e-06,
"loss": 1.3243,
"step": 6875
},
{
"epoch": 2.7400806491760843,
"grad_norm": 6.3300065994262695,
"learning_rate": 5.566077990580187e-06,
"loss": 1.3606,
"step": 6880
},
{
"epoch": 2.7420719868571712,
"grad_norm": 5.391266345977783,
"learning_rate": 5.481943514802372e-06,
"loss": 1.4057,
"step": 6885
},
{
"epoch": 2.7440633245382586,
"grad_norm": 6.474628448486328,
"learning_rate": 5.398437924778182e-06,
"loss": 1.2948,
"step": 6890
},
{
"epoch": 2.746054662219346,
"grad_norm": 5.8539276123046875,
"learning_rate": 5.315561583891104e-06,
"loss": 1.4363,
"step": 6895
},
{
"epoch": 2.748045999900433,
"grad_norm": 5.793940544128418,
"learning_rate": 5.233314852786369e-06,
"loss": 1.4127,
"step": 6900
},
{
"epoch": 2.7500373375815204,
"grad_norm": 5.599001407623291,
"learning_rate": 5.151698089369433e-06,
"loss": 1.3238,
"step": 6905
},
{
"epoch": 2.7520286752626077,
"grad_norm": 6.264495372772217,
"learning_rate": 5.07071164880431e-06,
"loss": 1.4469,
"step": 6910
},
{
"epoch": 2.7540200129436947,
"grad_norm": 6.348090648651123,
"learning_rate": 4.990355883512209e-06,
"loss": 1.4263,
"step": 6915
},
{
"epoch": 2.756011350624782,
"grad_norm": 6.050279140472412,
"learning_rate": 4.910631143169785e-06,
"loss": 1.3478,
"step": 6920
},
{
"epoch": 2.7580026883058695,
"grad_norm": 5.605953216552734,
"learning_rate": 4.831537774707739e-06,
"loss": 1.4268,
"step": 6925
},
{
"epoch": 2.759994025986957,
"grad_norm": 5.3409199714660645,
"learning_rate": 4.7530761223093515e-06,
"loss": 1.3866,
"step": 6930
},
{
"epoch": 2.7619853636680443,
"grad_norm": 6.341844081878662,
"learning_rate": 4.675246527408871e-06,
"loss": 1.4328,
"step": 6935
},
{
"epoch": 2.763976701349131,
"grad_norm": 5.9319586753845215,
"learning_rate": 4.5980493286900626e-06,
"loss": 1.4405,
"step": 6940
},
{
"epoch": 2.7659680390302186,
"grad_norm": 6.037888050079346,
"learning_rate": 4.52148486208479e-06,
"loss": 1.4027,
"step": 6945
},
{
"epoch": 2.767959376711306,
"grad_norm": 6.9271440505981445,
"learning_rate": 4.445553460771523e-06,
"loss": 1.4402,
"step": 6950
},
{
"epoch": 2.769950714392393,
"grad_norm": 6.065907001495361,
"learning_rate": 4.3702554551738656e-06,
"loss": 1.3285,
"step": 6955
},
{
"epoch": 2.7719420520734803,
"grad_norm": 6.596900463104248,
"learning_rate": 4.295591172959078e-06,
"loss": 1.3919,
"step": 6960
},
{
"epoch": 2.7739333897545677,
"grad_norm": 6.205164432525635,
"learning_rate": 4.2215609390368275e-06,
"loss": 1.3757,
"step": 6965
},
{
"epoch": 2.7759247274356547,
"grad_norm": 5.704510688781738,
"learning_rate": 4.148165075557536e-06,
"loss": 1.2786,
"step": 6970
},
{
"epoch": 2.777916065116742,
"grad_norm": 6.643089294433594,
"learning_rate": 4.0754039019111504e-06,
"loss": 1.3229,
"step": 6975
},
{
"epoch": 2.7799074027978294,
"grad_norm": 5.736795425415039,
"learning_rate": 4.0032777347256796e-06,
"loss": 1.3014,
"step": 6980
},
{
"epoch": 2.781898740478917,
"grad_norm": 5.650008201599121,
"learning_rate": 3.931786887865856e-06,
"loss": 1.3824,
"step": 6985
},
{
"epoch": 2.7838900781600042,
"grad_norm": 6.722746849060059,
"learning_rate": 3.860931672431727e-06,
"loss": 1.3002,
"step": 6990
},
{
"epoch": 2.785881415841091,
"grad_norm": 7.014549255371094,
"learning_rate": 3.7907123967573174e-06,
"loss": 1.3736,
"step": 6995
},
{
"epoch": 2.7878727535221786,
"grad_norm": 5.417212009429932,
"learning_rate": 3.721129366409331e-06,
"loss": 1.2088,
"step": 7000
},
{
"epoch": 2.789864091203266,
"grad_norm": 6.103097915649414,
"learning_rate": 3.6521828841857227e-06,
"loss": 1.4813,
"step": 7005
},
{
"epoch": 2.791855428884353,
"grad_norm": 6.893677234649658,
"learning_rate": 3.583873250114494e-06,
"loss": 1.4579,
"step": 7010
},
{
"epoch": 2.7938467665654403,
"grad_norm": 5.980708122253418,
"learning_rate": 3.5162007614522813e-06,
"loss": 1.3265,
"step": 7015
},
{
"epoch": 2.7958381042465277,
"grad_norm": 6.312836647033691,
"learning_rate": 3.4491657126831885e-06,
"loss": 1.2436,
"step": 7020
},
{
"epoch": 2.7978294419276146,
"grad_norm": 7.480289936065674,
"learning_rate": 3.3827683955173544e-06,
"loss": 1.4201,
"step": 7025
},
{
"epoch": 2.799820779608702,
"grad_norm": 6.134524345397949,
"learning_rate": 3.317009098889789e-06,
"loss": 1.2915,
"step": 7030
},
{
"epoch": 2.8018121172897894,
"grad_norm": 5.769630432128906,
"learning_rate": 3.251888108959139e-06,
"loss": 1.3436,
"step": 7035
},
{
"epoch": 2.803803454970877,
"grad_norm": 6.060738563537598,
"learning_rate": 3.1874057091063056e-06,
"loss": 1.3628,
"step": 7040
},
{
"epoch": 2.805794792651964,
"grad_norm": 7.591846942901611,
"learning_rate": 3.1235621799333475e-06,
"loss": 1.3498,
"step": 7045
},
{
"epoch": 2.807786130333051,
"grad_norm": 7.156486988067627,
"learning_rate": 3.0603577992621807e-06,
"loss": 1.4384,
"step": 7050
},
{
"epoch": 2.8097774680141385,
"grad_norm": 6.646254062652588,
"learning_rate": 2.9977928421334454e-06,
"loss": 1.38,
"step": 7055
},
{
"epoch": 2.811768805695226,
"grad_norm": 5.995254993438721,
"learning_rate": 2.9358675808052232e-06,
"loss": 1.307,
"step": 7060
},
{
"epoch": 2.813760143376313,
"grad_norm": 7.127299785614014,
"learning_rate": 2.8745822847518752e-06,
"loss": 1.4385,
"step": 7065
},
{
"epoch": 2.8157514810574003,
"grad_norm": 6.606611251831055,
"learning_rate": 2.813937220662954e-06,
"loss": 1.4435,
"step": 7070
},
{
"epoch": 2.8177428187384876,
"grad_norm": 5.3539605140686035,
"learning_rate": 2.7539326524419092e-06,
"loss": 1.3494,
"step": 7075
},
{
"epoch": 2.8197341564195746,
"grad_norm": 6.307882785797119,
"learning_rate": 2.6945688412050357e-06,
"loss": 1.5222,
"step": 7080
},
{
"epoch": 2.821725494100662,
"grad_norm": 6.9230637550354,
"learning_rate": 2.6358460452802767e-06,
"loss": 1.4859,
"step": 7085
},
{
"epoch": 2.8237168317817494,
"grad_norm": 6.22226619720459,
"learning_rate": 2.577764520206155e-06,
"loss": 1.4203,
"step": 7090
},
{
"epoch": 2.8257081694628368,
"grad_norm": 6.648333549499512,
"learning_rate": 2.52032451873066e-06,
"loss": 1.3568,
"step": 7095
},
{
"epoch": 2.827699507143924,
"grad_norm": 6.351929664611816,
"learning_rate": 2.4635262908100473e-06,
"loss": 1.4753,
"step": 7100
},
{
"epoch": 2.829690844825011,
"grad_norm": 6.392971515655518,
"learning_rate": 2.40737008360789e-06,
"loss": 1.378,
"step": 7105
},
{
"epoch": 2.8316821825060985,
"grad_norm": 6.635252952575684,
"learning_rate": 2.35185614149393e-06,
"loss": 1.4273,
"step": 7110
},
{
"epoch": 2.833673520187186,
"grad_norm": 6.16091251373291,
"learning_rate": 2.2969847060429924e-06,
"loss": 1.4064,
"step": 7115
},
{
"epoch": 2.835664857868273,
"grad_norm": 6.811882019042969,
"learning_rate": 2.2427560160340084e-06,
"loss": 1.4521,
"step": 7120
},
{
"epoch": 2.83765619554936,
"grad_norm": 6.058680534362793,
"learning_rate": 2.189170307448879e-06,
"loss": 1.3218,
"step": 7125
},
{
"epoch": 2.8396475332304476,
"grad_norm": 6.335659980773926,
"learning_rate": 2.136227813471575e-06,
"loss": 1.3948,
"step": 7130
},
{
"epoch": 2.8416388709115346,
"grad_norm": 5.820364952087402,
"learning_rate": 2.0839287644870086e-06,
"loss": 1.3663,
"step": 7135
},
{
"epoch": 2.843630208592622,
"grad_norm": 6.760519981384277,
"learning_rate": 2.0322733880800468e-06,
"loss": 1.442,
"step": 7140
},
{
"epoch": 2.8456215462737093,
"grad_norm": 6.848934650421143,
"learning_rate": 1.981261909034598e-06,
"loss": 1.5015,
"step": 7145
},
{
"epoch": 2.8476128839547967,
"grad_norm": 7.0265021324157715,
"learning_rate": 1.930894549332562e-06,
"loss": 1.3389,
"step": 7150
},
{
"epoch": 2.849604221635884,
"grad_norm": 4.9822235107421875,
"learning_rate": 1.881171528152864e-06,
"loss": 1.3325,
"step": 7155
},
{
"epoch": 2.851595559316971,
"grad_norm": 6.151075839996338,
"learning_rate": 1.8320930618705376e-06,
"loss": 1.4162,
"step": 7160
},
{
"epoch": 2.8535868969980585,
"grad_norm": 6.126955986022949,
"learning_rate": 1.7836593640558116e-06,
"loss": 1.4418,
"step": 7165
},
{
"epoch": 2.855578234679146,
"grad_norm": 6.3819122314453125,
"learning_rate": 1.735870645473042e-06,
"loss": 1.4855,
"step": 7170
},
{
"epoch": 2.857569572360233,
"grad_norm": 5.6472859382629395,
"learning_rate": 1.688727114079963e-06,
"loss": 1.4014,
"step": 7175
},
{
"epoch": 2.85956091004132,
"grad_norm": 6.101297378540039,
"learning_rate": 1.6422289750267047e-06,
"loss": 1.4108,
"step": 7180
},
{
"epoch": 2.8615522477224076,
"grad_norm": 6.445528984069824,
"learning_rate": 1.5963764306548443e-06,
"loss": 1.3386,
"step": 7185
},
{
"epoch": 2.863543585403495,
"grad_norm": 6.503068447113037,
"learning_rate": 1.5511696804966556e-06,
"loss": 1.4485,
"step": 7190
},
{
"epoch": 2.865534923084582,
"grad_norm": 6.986062526702881,
"learning_rate": 1.5066089212741271e-06,
"loss": 1.4852,
"step": 7195
},
{
"epoch": 2.8675262607656693,
"grad_norm": 6.3332366943359375,
"learning_rate": 1.4626943468981467e-06,
"loss": 1.3835,
"step": 7200
},
{
"epoch": 2.8695175984467567,
"grad_norm": 6.295677661895752,
"learning_rate": 1.4194261484677006e-06,
"loss": 1.4499,
"step": 7205
},
{
"epoch": 2.871508936127844,
"grad_norm": 6.975468158721924,
"learning_rate": 1.3768045142689588e-06,
"loss": 1.2806,
"step": 7210
},
{
"epoch": 2.873500273808931,
"grad_norm": 6.046121597290039,
"learning_rate": 1.3348296297745254e-06,
"loss": 1.4481,
"step": 7215
},
{
"epoch": 2.8754916114900184,
"grad_norm": 6.595587253570557,
"learning_rate": 1.2935016776425721e-06,
"loss": 1.4859,
"step": 7220
},
{
"epoch": 2.877482949171106,
"grad_norm": 6.250722885131836,
"learning_rate": 1.2528208377161064e-06,
"loss": 1.3335,
"step": 7225
},
{
"epoch": 2.8794742868521928,
"grad_norm": 5.738440990447998,
"learning_rate": 1.2127872870221378e-06,
"loss": 1.39,
"step": 7230
},
{
"epoch": 2.88146562453328,
"grad_norm": 5.639859676361084,
"learning_rate": 1.1734011997709625e-06,
"loss": 1.379,
"step": 7235
},
{
"epoch": 2.8834569622143675,
"grad_norm": 6.074785232543945,
"learning_rate": 1.1346627473553138e-06,
"loss": 1.3816,
"step": 7240
},
{
"epoch": 2.885448299895455,
"grad_norm": 5.629566669464111,
"learning_rate": 1.0965720983497295e-06,
"loss": 1.3705,
"step": 7245
},
{
"epoch": 2.8874396375765423,
"grad_norm": 6.7186970710754395,
"learning_rate": 1.0591294185097187e-06,
"loss": 1.4106,
"step": 7250
},
{
"epoch": 2.8894309752576293,
"grad_norm": 6.334540843963623,
"learning_rate": 1.0223348707710965e-06,
"loss": 1.3795,
"step": 7255
},
{
"epoch": 2.8914223129387167,
"grad_norm": 6.2255988121032715,
"learning_rate": 9.861886152492838e-07,
"loss": 1.4139,
"step": 7260
},
{
"epoch": 2.893413650619804,
"grad_norm": 6.067513942718506,
"learning_rate": 9.506908092385579e-07,
"loss": 1.3067,
"step": 7265
},
{
"epoch": 2.895404988300891,
"grad_norm": 5.387002468109131,
"learning_rate": 9.158416072114205e-07,
"loss": 1.3011,
"step": 7270
},
{
"epoch": 2.8973963259819784,
"grad_norm": 6.608880996704102,
"learning_rate": 8.816411608178975e-07,
"loss": 1.4344,
"step": 7275
},
{
"epoch": 2.8993876636630658,
"grad_norm": 6.290225028991699,
"learning_rate": 8.480896188848729e-07,
"loss": 1.3629,
"step": 7280
},
{
"epoch": 2.9013790013441527,
"grad_norm": 6.4408183097839355,
"learning_rate": 8.151871274154565e-07,
"loss": 1.2844,
"step": 7285
},
{
"epoch": 2.90337033902524,
"grad_norm": 5.775238990783691,
"learning_rate": 7.829338295883836e-07,
"loss": 1.4011,
"step": 7290
},
{
"epoch": 2.9053616767063275,
"grad_norm": 7.453217506408691,
"learning_rate": 7.513298657572831e-07,
"loss": 1.3853,
"step": 7295
},
{
"epoch": 2.907353014387415,
"grad_norm": 7.316783428192139,
"learning_rate": 7.203753734502104e-07,
"loss": 1.316,
"step": 7300
},
{
"epoch": 2.9093443520685023,
"grad_norm": 6.529441833496094,
"learning_rate": 6.900704873689322e-07,
"loss": 1.3602,
"step": 7305
},
{
"epoch": 2.9113356897495892,
"grad_norm": 6.993330001831055,
"learning_rate": 6.604153393884093e-07,
"loss": 1.4213,
"step": 7310
},
{
"epoch": 2.9133270274306766,
"grad_norm": 5.9941182136535645,
"learning_rate": 6.314100585561477e-07,
"loss": 1.5167,
"step": 7315
},
{
"epoch": 2.915318365111764,
"grad_norm": 6.265793800354004,
"learning_rate": 6.030547710917655e-07,
"loss": 1.403,
"step": 7320
},
{
"epoch": 2.917309702792851,
"grad_norm": 5.922816753387451,
"learning_rate": 5.753496003862935e-07,
"loss": 1.4157,
"step": 7325
},
{
"epoch": 2.9193010404739383,
"grad_norm": 6.88986873626709,
"learning_rate": 5.482946670017585e-07,
"loss": 1.3336,
"step": 7330
},
{
"epoch": 2.9212923781550257,
"grad_norm": 7.298521995544434,
"learning_rate": 5.218900886705845e-07,
"loss": 1.3912,
"step": 7335
},
{
"epoch": 2.9232837158361127,
"grad_norm": 6.366304397583008,
"learning_rate": 4.961359802951259e-07,
"loss": 1.4181,
"step": 7340
},
{
"epoch": 2.9252750535172,
"grad_norm": 6.151203155517578,
"learning_rate": 4.7103245394711773e-07,
"loss": 1.4416,
"step": 7345
},
{
"epoch": 2.9272663911982875,
"grad_norm": 7.088301181793213,
"learning_rate": 4.4657961886724327e-07,
"loss": 1.4493,
"step": 7350
},
{
"epoch": 2.929257728879375,
"grad_norm": 7.647716522216797,
"learning_rate": 4.2277758146461726e-07,
"loss": 1.4544,
"step": 7355
},
{
"epoch": 2.9312490665604622,
"grad_norm": 6.20832633972168,
"learning_rate": 3.9962644531637e-07,
"loss": 1.3583,
"step": 7360
},
{
"epoch": 2.933240404241549,
"grad_norm": 5.964649200439453,
"learning_rate": 3.771263111671141e-07,
"loss": 1.3814,
"step": 7365
},
{
"epoch": 2.9352317419226366,
"grad_norm": 6.578097343444824,
"learning_rate": 3.55277276928595e-07,
"loss": 1.4387,
"step": 7370
},
{
"epoch": 2.937223079603724,
"grad_norm": 5.8518452644348145,
"learning_rate": 3.340794376792244e-07,
"loss": 1.3866,
"step": 7375
},
{
"epoch": 2.939214417284811,
"grad_norm": 6.087634086608887,
"learning_rate": 3.135328856636643e-07,
"loss": 1.3422,
"step": 7380
},
{
"epoch": 2.9412057549658983,
"grad_norm": 7.072717189788818,
"learning_rate": 2.93637710292427e-07,
"loss": 1.3686,
"step": 7385
},
{
"epoch": 2.9431970926469857,
"grad_norm": 5.9594502449035645,
"learning_rate": 2.743939981414922e-07,
"loss": 1.4133,
"step": 7390
},
{
"epoch": 2.9451884303280726,
"grad_norm": 8.152170181274414,
"learning_rate": 2.5580183295190735e-07,
"loss": 1.4139,
"step": 7395
},
{
"epoch": 2.94717976800916,
"grad_norm": 6.129400253295898,
"learning_rate": 2.378612956295045e-07,
"loss": 1.4123,
"step": 7400
},
{
"epoch": 2.9491711056902474,
"grad_norm": 6.710536956787109,
"learning_rate": 2.2057246424441733e-07,
"loss": 1.5799,
"step": 7405
},
{
"epoch": 2.951162443371335,
"grad_norm": 6.588728427886963,
"learning_rate": 2.03935414030848e-07,
"loss": 1.4038,
"step": 7410
},
{
"epoch": 2.953153781052422,
"grad_norm": 6.022589683532715,
"learning_rate": 1.879502173867009e-07,
"loss": 1.3085,
"step": 7415
},
{
"epoch": 2.955145118733509,
"grad_norm": 5.406111240386963,
"learning_rate": 1.7261694387328273e-07,
"loss": 1.4544,
"step": 7420
},
{
"epoch": 2.9571364564145965,
"grad_norm": 6.308793067932129,
"learning_rate": 1.579356602149362e-07,
"loss": 1.3811,
"step": 7425
},
{
"epoch": 2.959127794095684,
"grad_norm": 5.521365165710449,
"learning_rate": 1.439064302988402e-07,
"loss": 1.3174,
"step": 7430
},
{
"epoch": 2.961119131776771,
"grad_norm": 6.406843185424805,
"learning_rate": 1.3052931517471e-07,
"loss": 1.4519,
"step": 7435
},
{
"epoch": 2.9631104694578583,
"grad_norm": 7.547394752502441,
"learning_rate": 1.1780437305448087e-07,
"loss": 1.4098,
"step": 7440
},
{
"epoch": 2.9651018071389457,
"grad_norm": 6.150856018066406,
"learning_rate": 1.057316593120916e-07,
"loss": 1.3563,
"step": 7445
},
{
"epoch": 2.9670931448200326,
"grad_norm": 6.139143466949463,
"learning_rate": 9.431122648325129e-08,
"loss": 1.396,
"step": 7450
},
{
"epoch": 2.96908448250112,
"grad_norm": 6.330423355102539,
"learning_rate": 8.354312426518961e-08,
"loss": 1.3253,
"step": 7455
},
{
"epoch": 2.9710758201822074,
"grad_norm": 6.3016228675842285,
"learning_rate": 7.342739951644028e-08,
"loss": 1.4452,
"step": 7460
},
{
"epoch": 2.973067157863295,
"grad_norm": 5.905858516693115,
"learning_rate": 6.396409625667453e-08,
"loss": 1.3387,
"step": 7465
},
{
"epoch": 2.975058495544382,
"grad_norm": 6.0580549240112305,
"learning_rate": 5.5153255666468e-08,
"loss": 1.4156,
"step": 7470
},
{
"epoch": 2.977049833225469,
"grad_norm": 7.04741907119751,
"learning_rate": 4.6994916087134125e-08,
"loss": 1.3523,
"step": 7475
},
{
"epoch": 2.9790411709065565,
"grad_norm": 5.739146709442139,
"learning_rate": 3.9489113020557685e-08,
"loss": 1.3493,
"step": 7480
},
{
"epoch": 2.981032508587644,
"grad_norm": 5.794787883758545,
"learning_rate": 3.263587912902821e-08,
"loss": 1.3492,
"step": 7485
},
{
"epoch": 2.983023846268731,
"grad_norm": 6.884363174438477,
"learning_rate": 2.643524423515675e-08,
"loss": 1.4875,
"step": 7490
},
{
"epoch": 2.9850151839498182,
"grad_norm": 5.869548797607422,
"learning_rate": 2.0887235321626062e-08,
"loss": 1.3215,
"step": 7495
},
{
"epoch": 2.9870065216309056,
"grad_norm": 6.334482192993164,
"learning_rate": 1.599187653122391e-08,
"loss": 1.3576,
"step": 7500
},
{
"epoch": 2.9889978593119926,
"grad_norm": 7.025340557098389,
"learning_rate": 1.174918916660994e-08,
"loss": 1.4502,
"step": 7505
},
{
"epoch": 2.99098919699308,
"grad_norm": 6.232082366943359,
"learning_rate": 8.15919169028234e-09,
"loss": 1.3403,
"step": 7510
},
{
"epoch": 2.9929805346741674,
"grad_norm": 7.146350860595703,
"learning_rate": 5.221899724527912e-09,
"loss": 1.4512,
"step": 7515
},
{
"epoch": 2.9949718723552547,
"grad_norm": 6.049505233764648,
"learning_rate": 2.937326051272171e-09,
"loss": 1.503,
"step": 7520
},
{
"epoch": 2.996963210036342,
"grad_norm": 5.630764007568359,
"learning_rate": 1.3054806120627036e-09,
"loss": 1.2154,
"step": 7525
},
{
"epoch": 2.998954547717429,
"grad_norm": 6.927404880523682,
"learning_rate": 3.263705080691625e-10,
"loss": 1.5118,
"step": 7530
},
{
"epoch": 2.998954547717429,
"step": 7530,
"total_flos": 3.7457880685726925e+17,
"train_loss": 2.0558883373164245,
"train_runtime": 9070.5505,
"train_samples_per_second": 13.287,
"train_steps_per_second": 0.83
}
],
"logging_steps": 5,
"max_steps": 7530,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.7457880685726925e+17,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}