Inputoutput_SFT_Llama3_3B / trainer_state.json
yujianll's picture
Upload folder using huggingface_hub
b94f677 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 7.361963190184049,
"eval_steps": 500,
"global_step": 1200,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.03067484662576687,
"grad_norm": 2.916747570037842,
"learning_rate": 1.6326530612244901e-07,
"loss": 1.2014,
"step": 5
},
{
"epoch": 0.06134969325153374,
"grad_norm": 2.8873817920684814,
"learning_rate": 3.6734693877551025e-07,
"loss": 1.1968,
"step": 10
},
{
"epoch": 0.09202453987730061,
"grad_norm": 2.65090012550354,
"learning_rate": 5.714285714285715e-07,
"loss": 1.1791,
"step": 15
},
{
"epoch": 0.12269938650306748,
"grad_norm": 2.3519155979156494,
"learning_rate": 7.755102040816327e-07,
"loss": 1.1753,
"step": 20
},
{
"epoch": 0.15337423312883436,
"grad_norm": 1.5954474210739136,
"learning_rate": 9.795918367346939e-07,
"loss": 1.1336,
"step": 25
},
{
"epoch": 0.18404907975460122,
"grad_norm": 1.1593321561813354,
"learning_rate": 1.1836734693877552e-06,
"loss": 1.1015,
"step": 30
},
{
"epoch": 0.2147239263803681,
"grad_norm": 1.082676887512207,
"learning_rate": 1.3877551020408165e-06,
"loss": 1.054,
"step": 35
},
{
"epoch": 0.24539877300613497,
"grad_norm": 1.2616770267486572,
"learning_rate": 1.5918367346938775e-06,
"loss": 1.0152,
"step": 40
},
{
"epoch": 0.27607361963190186,
"grad_norm": 0.7849059700965881,
"learning_rate": 1.7959183673469388e-06,
"loss": 0.9693,
"step": 45
},
{
"epoch": 0.3067484662576687,
"grad_norm": 0.5774770379066467,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.9483,
"step": 50
},
{
"epoch": 0.3374233128834356,
"grad_norm": 0.5419191718101501,
"learning_rate": 2.2040816326530616e-06,
"loss": 0.928,
"step": 55
},
{
"epoch": 0.36809815950920244,
"grad_norm": 0.5010412931442261,
"learning_rate": 2.4081632653061225e-06,
"loss": 0.8946,
"step": 60
},
{
"epoch": 0.3987730061349693,
"grad_norm": 0.45867446064949036,
"learning_rate": 2.6122448979591842e-06,
"loss": 0.8818,
"step": 65
},
{
"epoch": 0.4294478527607362,
"grad_norm": 0.41711899638175964,
"learning_rate": 2.816326530612245e-06,
"loss": 0.8657,
"step": 70
},
{
"epoch": 0.4601226993865031,
"grad_norm": 0.3863469362258911,
"learning_rate": 3.0204081632653064e-06,
"loss": 0.8518,
"step": 75
},
{
"epoch": 0.49079754601226994,
"grad_norm": 0.3639077842235565,
"learning_rate": 3.2244897959183672e-06,
"loss": 0.8378,
"step": 80
},
{
"epoch": 0.5214723926380368,
"grad_norm": 0.3680429458618164,
"learning_rate": 3.428571428571429e-06,
"loss": 0.8328,
"step": 85
},
{
"epoch": 0.5521472392638037,
"grad_norm": 0.3493054211139679,
"learning_rate": 3.6326530612244903e-06,
"loss": 0.8278,
"step": 90
},
{
"epoch": 0.5828220858895705,
"grad_norm": 0.35972270369529724,
"learning_rate": 3.836734693877551e-06,
"loss": 0.8177,
"step": 95
},
{
"epoch": 0.6134969325153374,
"grad_norm": 0.3597475588321686,
"learning_rate": 4.040816326530612e-06,
"loss": 0.8045,
"step": 100
},
{
"epoch": 0.6441717791411042,
"grad_norm": 0.3763820230960846,
"learning_rate": 4.244897959183674e-06,
"loss": 0.8003,
"step": 105
},
{
"epoch": 0.6748466257668712,
"grad_norm": 0.4080606997013092,
"learning_rate": 4.448979591836735e-06,
"loss": 0.7961,
"step": 110
},
{
"epoch": 0.7055214723926381,
"grad_norm": 0.3602585196495056,
"learning_rate": 4.653061224489796e-06,
"loss": 0.7918,
"step": 115
},
{
"epoch": 0.7361963190184049,
"grad_norm": 0.3696184456348419,
"learning_rate": 4.857142857142858e-06,
"loss": 0.7875,
"step": 120
},
{
"epoch": 0.7668711656441718,
"grad_norm": 0.37600934505462646,
"learning_rate": 5.061224489795918e-06,
"loss": 0.7875,
"step": 125
},
{
"epoch": 0.7975460122699386,
"grad_norm": 0.38016051054000854,
"learning_rate": 5.26530612244898e-06,
"loss": 0.7748,
"step": 130
},
{
"epoch": 0.8282208588957055,
"grad_norm": 0.3912647068500519,
"learning_rate": 5.4693877551020415e-06,
"loss": 0.772,
"step": 135
},
{
"epoch": 0.8588957055214724,
"grad_norm": 0.3811889886856079,
"learning_rate": 5.673469387755103e-06,
"loss": 0.7678,
"step": 140
},
{
"epoch": 0.8895705521472392,
"grad_norm": 0.38242849707603455,
"learning_rate": 5.877551020408164e-06,
"loss": 0.761,
"step": 145
},
{
"epoch": 0.9202453987730062,
"grad_norm": 0.4102751910686493,
"learning_rate": 6.0816326530612245e-06,
"loss": 0.763,
"step": 150
},
{
"epoch": 0.950920245398773,
"grad_norm": 0.38672712445259094,
"learning_rate": 6.285714285714286e-06,
"loss": 0.7576,
"step": 155
},
{
"epoch": 0.9815950920245399,
"grad_norm": 0.3906507194042206,
"learning_rate": 6.489795918367348e-06,
"loss": 0.751,
"step": 160
},
{
"epoch": 1.0122699386503067,
"grad_norm": 0.4010821580886841,
"learning_rate": 6.693877551020409e-06,
"loss": 0.7451,
"step": 165
},
{
"epoch": 1.0429447852760736,
"grad_norm": 0.454095721244812,
"learning_rate": 6.8979591836734705e-06,
"loss": 0.7373,
"step": 170
},
{
"epoch": 1.0736196319018405,
"grad_norm": 0.44405630230903625,
"learning_rate": 7.102040816326531e-06,
"loss": 0.7361,
"step": 175
},
{
"epoch": 1.1042944785276074,
"grad_norm": 0.42133402824401855,
"learning_rate": 7.306122448979592e-06,
"loss": 0.7309,
"step": 180
},
{
"epoch": 1.1349693251533743,
"grad_norm": 0.4115142226219177,
"learning_rate": 7.5102040816326536e-06,
"loss": 0.7213,
"step": 185
},
{
"epoch": 1.165644171779141,
"grad_norm": 0.48450756072998047,
"learning_rate": 7.714285714285716e-06,
"loss": 0.7311,
"step": 190
},
{
"epoch": 1.196319018404908,
"grad_norm": 0.44604143500328064,
"learning_rate": 7.918367346938776e-06,
"loss": 0.7265,
"step": 195
},
{
"epoch": 1.2269938650306749,
"grad_norm": 0.4400920271873474,
"learning_rate": 8.122448979591837e-06,
"loss": 0.7231,
"step": 200
},
{
"epoch": 1.2576687116564418,
"grad_norm": 0.4529883861541748,
"learning_rate": 8.326530612244899e-06,
"loss": 0.7177,
"step": 205
},
{
"epoch": 1.2883435582822087,
"grad_norm": 0.4646182954311371,
"learning_rate": 8.530612244897961e-06,
"loss": 0.7119,
"step": 210
},
{
"epoch": 1.3190184049079754,
"grad_norm": 0.45873209834098816,
"learning_rate": 8.734693877551021e-06,
"loss": 0.7131,
"step": 215
},
{
"epoch": 1.3496932515337423,
"grad_norm": 0.45401182770729065,
"learning_rate": 8.938775510204082e-06,
"loss": 0.7136,
"step": 220
},
{
"epoch": 1.3803680981595092,
"grad_norm": 0.4406537711620331,
"learning_rate": 9.142857142857144e-06,
"loss": 0.7148,
"step": 225
},
{
"epoch": 1.4110429447852761,
"grad_norm": 0.5117276310920715,
"learning_rate": 9.346938775510204e-06,
"loss": 0.7112,
"step": 230
},
{
"epoch": 1.441717791411043,
"grad_norm": 0.4318268895149231,
"learning_rate": 9.551020408163266e-06,
"loss": 0.7065,
"step": 235
},
{
"epoch": 1.4723926380368098,
"grad_norm": 0.4682416617870331,
"learning_rate": 9.755102040816327e-06,
"loss": 0.7019,
"step": 240
},
{
"epoch": 1.5030674846625767,
"grad_norm": 0.48544424772262573,
"learning_rate": 9.959183673469387e-06,
"loss": 0.7084,
"step": 245
},
{
"epoch": 1.5337423312883436,
"grad_norm": 0.45956820249557495,
"learning_rate": 9.999918433243253e-06,
"loss": 0.7045,
"step": 250
},
{
"epoch": 1.5644171779141103,
"grad_norm": 0.4829379618167877,
"learning_rate": 9.999587072854989e-06,
"loss": 0.6992,
"step": 255
},
{
"epoch": 1.5950920245398774,
"grad_norm": 0.4817286431789398,
"learning_rate": 9.99900083779239e-06,
"loss": 0.6949,
"step": 260
},
{
"epoch": 1.6257668711656441,
"grad_norm": 0.41464245319366455,
"learning_rate": 9.998159757941219e-06,
"loss": 0.6943,
"step": 265
},
{
"epoch": 1.656441717791411,
"grad_norm": 0.45459651947021484,
"learning_rate": 9.997063876179007e-06,
"loss": 0.6978,
"step": 270
},
{
"epoch": 1.687116564417178,
"grad_norm": 0.4576292335987091,
"learning_rate": 9.99571324837287e-06,
"loss": 0.6944,
"step": 275
},
{
"epoch": 1.7177914110429446,
"grad_norm": 0.4427788555622101,
"learning_rate": 9.994107943376654e-06,
"loss": 0.6914,
"step": 280
},
{
"epoch": 1.7484662576687118,
"grad_norm": 0.4589153528213501,
"learning_rate": 9.992248043027441e-06,
"loss": 0.6818,
"step": 285
},
{
"epoch": 1.7791411042944785,
"grad_norm": 0.4295940399169922,
"learning_rate": 9.990133642141359e-06,
"loss": 0.6902,
"step": 290
},
{
"epoch": 1.8098159509202454,
"grad_norm": 0.4209536910057068,
"learning_rate": 9.987764848508756e-06,
"loss": 0.6856,
"step": 295
},
{
"epoch": 1.8404907975460123,
"grad_norm": 0.4128643870353699,
"learning_rate": 9.985141782888705e-06,
"loss": 0.683,
"step": 300
},
{
"epoch": 1.871165644171779,
"grad_norm": 0.44537627696990967,
"learning_rate": 9.982264579002853e-06,
"loss": 0.6837,
"step": 305
},
{
"epoch": 1.9018404907975461,
"grad_norm": 0.40354472398757935,
"learning_rate": 9.979133383528591e-06,
"loss": 0.686,
"step": 310
},
{
"epoch": 1.9325153374233128,
"grad_norm": 0.43714097142219543,
"learning_rate": 9.975748356091589e-06,
"loss": 0.6809,
"step": 315
},
{
"epoch": 1.9631901840490797,
"grad_norm": 0.4338328540325165,
"learning_rate": 9.972109669257645e-06,
"loss": 0.6787,
"step": 320
},
{
"epoch": 1.9938650306748467,
"grad_norm": 0.4852101802825928,
"learning_rate": 9.968217508523913e-06,
"loss": 0.6696,
"step": 325
},
{
"epoch": 2.0245398773006134,
"grad_norm": 0.49633219838142395,
"learning_rate": 9.964072072309412e-06,
"loss": 0.652,
"step": 330
},
{
"epoch": 2.0552147239263805,
"grad_norm": 0.48235881328582764,
"learning_rate": 9.959673571944939e-06,
"loss": 0.6525,
"step": 335
},
{
"epoch": 2.085889570552147,
"grad_norm": 0.47604846954345703,
"learning_rate": 9.955022231662282e-06,
"loss": 0.6445,
"step": 340
},
{
"epoch": 2.116564417177914,
"grad_norm": 0.4336640238761902,
"learning_rate": 9.95011828858279e-06,
"loss": 0.6557,
"step": 345
},
{
"epoch": 2.147239263803681,
"grad_norm": 0.4802437722682953,
"learning_rate": 9.944961992705288e-06,
"loss": 0.6467,
"step": 350
},
{
"epoch": 2.1779141104294477,
"grad_norm": 0.476906955242157,
"learning_rate": 9.939553606893334e-06,
"loss": 0.6463,
"step": 355
},
{
"epoch": 2.208588957055215,
"grad_norm": 0.49822670221328735,
"learning_rate": 9.933893406861808e-06,
"loss": 0.6483,
"step": 360
},
{
"epoch": 2.2392638036809815,
"grad_norm": 0.49377647042274475,
"learning_rate": 9.927981681162873e-06,
"loss": 0.6413,
"step": 365
},
{
"epoch": 2.2699386503067487,
"grad_norm": 0.4360464811325073,
"learning_rate": 9.921818731171249e-06,
"loss": 0.6478,
"step": 370
},
{
"epoch": 2.3006134969325154,
"grad_norm": 0.5180469155311584,
"learning_rate": 9.915404871068855e-06,
"loss": 0.6458,
"step": 375
},
{
"epoch": 2.331288343558282,
"grad_norm": 0.5449960231781006,
"learning_rate": 9.9087404278288e-06,
"loss": 0.6445,
"step": 380
},
{
"epoch": 2.361963190184049,
"grad_norm": 0.4497530460357666,
"learning_rate": 9.901825741198697e-06,
"loss": 0.6403,
"step": 385
},
{
"epoch": 2.392638036809816,
"grad_norm": 0.4317333698272705,
"learning_rate": 9.894661163683361e-06,
"loss": 0.6434,
"step": 390
},
{
"epoch": 2.4233128834355826,
"grad_norm": 0.47986018657684326,
"learning_rate": 9.887247060526827e-06,
"loss": 0.6422,
"step": 395
},
{
"epoch": 2.4539877300613497,
"grad_norm": 0.4332720637321472,
"learning_rate": 9.879583809693737e-06,
"loss": 0.6438,
"step": 400
},
{
"epoch": 2.4846625766871164,
"grad_norm": 0.4776013493537903,
"learning_rate": 9.871671801850065e-06,
"loss": 0.6393,
"step": 405
},
{
"epoch": 2.5153374233128836,
"grad_norm": 0.48002681136131287,
"learning_rate": 9.863511440343206e-06,
"loss": 0.6385,
"step": 410
},
{
"epoch": 2.5460122699386503,
"grad_norm": 0.4508747458457947,
"learning_rate": 9.855103141181412e-06,
"loss": 0.6362,
"step": 415
},
{
"epoch": 2.5766871165644174,
"grad_norm": 0.5055172443389893,
"learning_rate": 9.846447333012587e-06,
"loss": 0.6388,
"step": 420
},
{
"epoch": 2.607361963190184,
"grad_norm": 0.4450007975101471,
"learning_rate": 9.837544457102428e-06,
"loss": 0.6377,
"step": 425
},
{
"epoch": 2.638036809815951,
"grad_norm": 0.4479919970035553,
"learning_rate": 9.82839496731194e-06,
"loss": 0.6372,
"step": 430
},
{
"epoch": 2.668711656441718,
"grad_norm": 0.45310866832733154,
"learning_rate": 9.818999330074288e-06,
"loss": 0.6336,
"step": 435
},
{
"epoch": 2.6993865030674846,
"grad_norm": 0.4643765091896057,
"learning_rate": 9.809358024371025e-06,
"loss": 0.6363,
"step": 440
},
{
"epoch": 2.7300613496932513,
"grad_norm": 0.48741263151168823,
"learning_rate": 9.799471541707672e-06,
"loss": 0.6382,
"step": 445
},
{
"epoch": 2.7607361963190185,
"grad_norm": 0.45080405473709106,
"learning_rate": 9.789340386088663e-06,
"loss": 0.6357,
"step": 450
},
{
"epoch": 2.791411042944785,
"grad_norm": 0.5053460597991943,
"learning_rate": 9.778965073991652e-06,
"loss": 0.6315,
"step": 455
},
{
"epoch": 2.8220858895705523,
"grad_norm": 0.523303747177124,
"learning_rate": 9.768346134341174e-06,
"loss": 0.6336,
"step": 460
},
{
"epoch": 2.852760736196319,
"grad_norm": 0.4880037307739258,
"learning_rate": 9.757484108481695e-06,
"loss": 0.63,
"step": 465
},
{
"epoch": 2.883435582822086,
"grad_norm": 0.4551119804382324,
"learning_rate": 9.74637955015001e-06,
"loss": 0.6315,
"step": 470
},
{
"epoch": 2.914110429447853,
"grad_norm": 0.529259204864502,
"learning_rate": 9.735033025447e-06,
"loss": 0.6346,
"step": 475
},
{
"epoch": 2.9447852760736195,
"grad_norm": 0.44139307737350464,
"learning_rate": 9.723445112808802e-06,
"loss": 0.6284,
"step": 480
},
{
"epoch": 2.9754601226993866,
"grad_norm": 0.44775474071502686,
"learning_rate": 9.71161640297729e-06,
"loss": 0.6316,
"step": 485
},
{
"epoch": 3.0061349693251533,
"grad_norm": 0.5552355051040649,
"learning_rate": 9.699547498969978e-06,
"loss": 0.6217,
"step": 490
},
{
"epoch": 3.03680981595092,
"grad_norm": 0.5540191531181335,
"learning_rate": 9.687239016049275e-06,
"loss": 0.5991,
"step": 495
},
{
"epoch": 3.067484662576687,
"grad_norm": 0.5161706209182739,
"learning_rate": 9.674691581691114e-06,
"loss": 0.5983,
"step": 500
},
{
"epoch": 3.098159509202454,
"grad_norm": 0.564895510673523,
"learning_rate": 9.661905835552974e-06,
"loss": 0.599,
"step": 505
},
{
"epoch": 3.128834355828221,
"grad_norm": 0.5886032581329346,
"learning_rate": 9.648882429441258e-06,
"loss": 0.5958,
"step": 510
},
{
"epoch": 3.1595092024539877,
"grad_norm": 0.5232917666435242,
"learning_rate": 9.635622027278076e-06,
"loss": 0.5981,
"step": 515
},
{
"epoch": 3.190184049079755,
"grad_norm": 0.5367550849914551,
"learning_rate": 9.622125305067394e-06,
"loss": 0.6009,
"step": 520
},
{
"epoch": 3.2208588957055215,
"grad_norm": 0.5145596861839294,
"learning_rate": 9.608392950860568e-06,
"loss": 0.5964,
"step": 525
},
{
"epoch": 3.2515337423312882,
"grad_norm": 0.48504817485809326,
"learning_rate": 9.594425664721275e-06,
"loss": 0.5954,
"step": 530
},
{
"epoch": 3.2822085889570554,
"grad_norm": 0.44789251685142517,
"learning_rate": 9.580224158689821e-06,
"loss": 0.5963,
"step": 535
},
{
"epoch": 3.312883435582822,
"grad_norm": 0.44385582208633423,
"learning_rate": 9.565789156746843e-06,
"loss": 0.5953,
"step": 540
},
{
"epoch": 3.3435582822085887,
"grad_norm": 0.46638399362564087,
"learning_rate": 9.551121394776395e-06,
"loss": 0.5953,
"step": 545
},
{
"epoch": 3.374233128834356,
"grad_norm": 0.47886648774147034,
"learning_rate": 9.536221620528442e-06,
"loss": 0.5984,
"step": 550
},
{
"epoch": 3.4049079754601226,
"grad_norm": 0.43750545382499695,
"learning_rate": 9.521090593580737e-06,
"loss": 0.5966,
"step": 555
},
{
"epoch": 3.4355828220858897,
"grad_norm": 0.4883286654949188,
"learning_rate": 9.505729085300098e-06,
"loss": 0.5976,
"step": 560
},
{
"epoch": 3.4662576687116564,
"grad_norm": 0.5246576070785522,
"learning_rate": 9.490137878803078e-06,
"loss": 0.596,
"step": 565
},
{
"epoch": 3.4969325153374236,
"grad_norm": 0.47524797916412354,
"learning_rate": 9.47431776891606e-06,
"loss": 0.5991,
"step": 570
},
{
"epoch": 3.5276073619631902,
"grad_norm": 0.4653734266757965,
"learning_rate": 9.458269562134717e-06,
"loss": 0.594,
"step": 575
},
{
"epoch": 3.558282208588957,
"grad_norm": 0.5341756343841553,
"learning_rate": 9.441994076582907e-06,
"loss": 0.5985,
"step": 580
},
{
"epoch": 3.588957055214724,
"grad_norm": 0.5514233708381653,
"learning_rate": 9.425492141970973e-06,
"loss": 0.5992,
"step": 585
},
{
"epoch": 3.6196319018404908,
"grad_norm": 0.4911518096923828,
"learning_rate": 9.408764599553429e-06,
"loss": 0.5935,
"step": 590
},
{
"epoch": 3.6503067484662575,
"grad_norm": 0.4620342552661896,
"learning_rate": 9.391812302086088e-06,
"loss": 0.5977,
"step": 595
},
{
"epoch": 3.6809815950920246,
"grad_norm": 0.44591575860977173,
"learning_rate": 9.374636113782576e-06,
"loss": 0.5928,
"step": 600
},
{
"epoch": 3.7116564417177913,
"grad_norm": 0.40037840604782104,
"learning_rate": 9.357236910270292e-06,
"loss": 0.5883,
"step": 605
},
{
"epoch": 3.7423312883435584,
"grad_norm": 0.4316963255405426,
"learning_rate": 9.339615578545753e-06,
"loss": 0.5916,
"step": 610
},
{
"epoch": 3.773006134969325,
"grad_norm": 0.47416016459465027,
"learning_rate": 9.321773016929382e-06,
"loss": 0.5935,
"step": 615
},
{
"epoch": 3.8036809815950923,
"grad_norm": 0.486749529838562,
"learning_rate": 9.30371013501972e-06,
"loss": 0.5976,
"step": 620
},
{
"epoch": 3.834355828220859,
"grad_norm": 0.46578896045684814,
"learning_rate": 9.285427853647038e-06,
"loss": 0.5964,
"step": 625
},
{
"epoch": 3.8650306748466257,
"grad_norm": 0.4229313135147095,
"learning_rate": 9.26692710482641e-06,
"loss": 0.5941,
"step": 630
},
{
"epoch": 3.895705521472393,
"grad_norm": 0.45090341567993164,
"learning_rate": 9.248208831710195e-06,
"loss": 0.5922,
"step": 635
},
{
"epoch": 3.9263803680981595,
"grad_norm": 0.4940469264984131,
"learning_rate": 9.229273988539951e-06,
"loss": 0.5957,
"step": 640
},
{
"epoch": 3.957055214723926,
"grad_norm": 0.45583370327949524,
"learning_rate": 9.210123540597792e-06,
"loss": 0.592,
"step": 645
},
{
"epoch": 3.9877300613496933,
"grad_norm": 0.43879297375679016,
"learning_rate": 9.190758464157184e-06,
"loss": 0.5936,
"step": 650
},
{
"epoch": 4.0184049079754605,
"grad_norm": 0.5458897948265076,
"learning_rate": 9.171179746433164e-06,
"loss": 0.5735,
"step": 655
},
{
"epoch": 4.049079754601227,
"grad_norm": 0.5021305084228516,
"learning_rate": 9.151388385532022e-06,
"loss": 0.5571,
"step": 660
},
{
"epoch": 4.079754601226994,
"grad_norm": 0.47301429510116577,
"learning_rate": 9.131385390400417e-06,
"loss": 0.5594,
"step": 665
},
{
"epoch": 4.110429447852761,
"grad_norm": 0.46195101737976074,
"learning_rate": 9.111171780773938e-06,
"loss": 0.556,
"step": 670
},
{
"epoch": 4.141104294478527,
"grad_norm": 0.4663168489933014,
"learning_rate": 9.090748587125118e-06,
"loss": 0.5576,
"step": 675
},
{
"epoch": 4.171779141104294,
"grad_norm": 0.5041781067848206,
"learning_rate": 9.070116850610911e-06,
"loss": 0.5546,
"step": 680
},
{
"epoch": 4.2024539877300615,
"grad_norm": 0.571372389793396,
"learning_rate": 9.049277623019603e-06,
"loss": 0.5609,
"step": 685
},
{
"epoch": 4.233128834355828,
"grad_norm": 0.4924178421497345,
"learning_rate": 9.0282319667172e-06,
"loss": 0.5569,
"step": 690
},
{
"epoch": 4.263803680981595,
"grad_norm": 0.4700438678264618,
"learning_rate": 9.006980954593262e-06,
"loss": 0.5622,
"step": 695
},
{
"epoch": 4.294478527607362,
"grad_norm": 0.46645787358283997,
"learning_rate": 8.985525670006225e-06,
"loss": 0.5561,
"step": 700
},
{
"epoch": 4.325153374233129,
"grad_norm": 0.47866857051849365,
"learning_rate": 8.963867206728147e-06,
"loss": 0.5594,
"step": 705
},
{
"epoch": 4.355828220858895,
"grad_norm": 0.4613681733608246,
"learning_rate": 8.942006668888972e-06,
"loss": 0.5556,
"step": 710
},
{
"epoch": 4.386503067484663,
"grad_norm": 0.451261043548584,
"learning_rate": 8.919945170920224e-06,
"loss": 0.558,
"step": 715
},
{
"epoch": 4.41717791411043,
"grad_norm": 0.5014335513114929,
"learning_rate": 8.89768383749821e-06,
"loss": 0.5622,
"step": 720
},
{
"epoch": 4.447852760736196,
"grad_norm": 0.5099634528160095,
"learning_rate": 8.875223803486674e-06,
"loss": 0.566,
"step": 725
},
{
"epoch": 4.478527607361963,
"grad_norm": 0.5409094095230103,
"learning_rate": 8.852566213878947e-06,
"loss": 0.5562,
"step": 730
},
{
"epoch": 4.50920245398773,
"grad_norm": 0.5168077945709229,
"learning_rate": 8.829712223739574e-06,
"loss": 0.5589,
"step": 735
},
{
"epoch": 4.539877300613497,
"grad_norm": 0.5725398063659668,
"learning_rate": 8.80666299814543e-06,
"loss": 0.5567,
"step": 740
},
{
"epoch": 4.570552147239264,
"grad_norm": 0.45501619577407837,
"learning_rate": 8.783419712126335e-06,
"loss": 0.5615,
"step": 745
},
{
"epoch": 4.601226993865031,
"grad_norm": 0.5540989637374878,
"learning_rate": 8.759983550605132e-06,
"loss": 0.5578,
"step": 750
},
{
"epoch": 4.631901840490798,
"grad_norm": 0.5188400745391846,
"learning_rate": 8.736355708337298e-06,
"loss": 0.5605,
"step": 755
},
{
"epoch": 4.662576687116564,
"grad_norm": 0.48551928997039795,
"learning_rate": 8.71253738985003e-06,
"loss": 0.5579,
"step": 760
},
{
"epoch": 4.693251533742331,
"grad_norm": 0.44432246685028076,
"learning_rate": 8.688529809380843e-06,
"loss": 0.5595,
"step": 765
},
{
"epoch": 4.723926380368098,
"grad_norm": 0.45983144640922546,
"learning_rate": 8.66433419081566e-06,
"loss": 0.5611,
"step": 770
},
{
"epoch": 4.754601226993865,
"grad_norm": 0.48174920678138733,
"learning_rate": 8.639951767626429e-06,
"loss": 0.5637,
"step": 775
},
{
"epoch": 4.785276073619632,
"grad_norm": 0.4685279130935669,
"learning_rate": 8.615383782808238e-06,
"loss": 0.5608,
"step": 780
},
{
"epoch": 4.815950920245399,
"grad_norm": 0.5617627501487732,
"learning_rate": 8.590631488815945e-06,
"loss": 0.5596,
"step": 785
},
{
"epoch": 4.846625766871165,
"grad_norm": 0.5751484036445618,
"learning_rate": 8.565696147500338e-06,
"loss": 0.5596,
"step": 790
},
{
"epoch": 4.877300613496932,
"grad_norm": 0.6005717515945435,
"learning_rate": 8.540579030043795e-06,
"loss": 0.5614,
"step": 795
},
{
"epoch": 4.9079754601226995,
"grad_norm": 0.5048331022262573,
"learning_rate": 8.515281416895489e-06,
"loss": 0.5585,
"step": 800
},
{
"epoch": 4.938650306748467,
"grad_norm": 0.4952603578567505,
"learning_rate": 8.48980459770611e-06,
"loss": 0.5599,
"step": 805
},
{
"epoch": 4.969325153374233,
"grad_norm": 0.49712324142456055,
"learning_rate": 8.464149871262118e-06,
"loss": 0.5614,
"step": 810
},
{
"epoch": 5.0,
"grad_norm": 0.5496305227279663,
"learning_rate": 8.43831854541953e-06,
"loss": 0.559,
"step": 815
},
{
"epoch": 5.030674846625767,
"grad_norm": 0.565785825252533,
"learning_rate": 8.412311937037255e-06,
"loss": 0.5256,
"step": 820
},
{
"epoch": 5.061349693251533,
"grad_norm": 0.5519700050354004,
"learning_rate": 8.386131371909948e-06,
"loss": 0.5198,
"step": 825
},
{
"epoch": 5.0920245398773005,
"grad_norm": 0.5226123332977295,
"learning_rate": 8.35977818470044e-06,
"loss": 0.5235,
"step": 830
},
{
"epoch": 5.122699386503068,
"grad_norm": 0.5327089428901672,
"learning_rate": 8.33325371887168e-06,
"loss": 0.5267,
"step": 835
},
{
"epoch": 5.153374233128835,
"grad_norm": 0.5217044949531555,
"learning_rate": 8.30655932661826e-06,
"loss": 0.5215,
"step": 840
},
{
"epoch": 5.184049079754601,
"grad_norm": 0.5113396048545837,
"learning_rate": 8.279696368797471e-06,
"loss": 0.5227,
"step": 845
},
{
"epoch": 5.214723926380368,
"grad_norm": 0.5224789977073669,
"learning_rate": 8.252666214859936e-06,
"loss": 0.5211,
"step": 850
},
{
"epoch": 5.245398773006135,
"grad_norm": 0.47648775577545166,
"learning_rate": 8.225470242779791e-06,
"loss": 0.5262,
"step": 855
},
{
"epoch": 5.276073619631902,
"grad_norm": 0.5127813816070557,
"learning_rate": 8.19810983898444e-06,
"loss": 0.5256,
"step": 860
},
{
"epoch": 5.306748466257669,
"grad_norm": 0.5499957799911499,
"learning_rate": 8.170586398283878e-06,
"loss": 0.5267,
"step": 865
},
{
"epoch": 5.337423312883436,
"grad_norm": 0.5156161189079285,
"learning_rate": 8.142901323799578e-06,
"loss": 0.5252,
"step": 870
},
{
"epoch": 5.368098159509202,
"grad_norm": 0.4825647175312042,
"learning_rate": 8.115056026892965e-06,
"loss": 0.5266,
"step": 875
},
{
"epoch": 5.398773006134969,
"grad_norm": 0.5043969750404358,
"learning_rate": 8.08705192709347e-06,
"loss": 0.5241,
"step": 880
},
{
"epoch": 5.429447852760736,
"grad_norm": 0.47043198347091675,
"learning_rate": 8.058890452026155e-06,
"loss": 0.5272,
"step": 885
},
{
"epoch": 5.460122699386503,
"grad_norm": 0.471513956785202,
"learning_rate": 8.030573037338942e-06,
"loss": 0.5321,
"step": 890
},
{
"epoch": 5.49079754601227,
"grad_norm": 0.48654860258102417,
"learning_rate": 8.002101126629422e-06,
"loss": 0.5267,
"step": 895
},
{
"epoch": 5.521472392638037,
"grad_norm": 0.4844699501991272,
"learning_rate": 7.973476171371255e-06,
"loss": 0.5301,
"step": 900
},
{
"epoch": 5.552147239263804,
"grad_norm": 0.4706033170223236,
"learning_rate": 7.94469963084019e-06,
"loss": 0.5276,
"step": 905
},
{
"epoch": 5.58282208588957,
"grad_norm": 0.47882503271102905,
"learning_rate": 7.91577297203966e-06,
"loss": 0.5267,
"step": 910
},
{
"epoch": 5.613496932515337,
"grad_norm": 0.5027369856834412,
"learning_rate": 7.886697669625995e-06,
"loss": 0.5313,
"step": 915
},
{
"epoch": 5.644171779141105,
"grad_norm": 0.4971833825111389,
"learning_rate": 7.857475205833255e-06,
"loss": 0.5219,
"step": 920
},
{
"epoch": 5.674846625766871,
"grad_norm": 0.488737016916275,
"learning_rate": 7.828107070397657e-06,
"loss": 0.5233,
"step": 925
},
{
"epoch": 5.705521472392638,
"grad_norm": 0.49035337567329407,
"learning_rate": 7.798594760481639e-06,
"loss": 0.5254,
"step": 930
},
{
"epoch": 5.736196319018405,
"grad_norm": 0.5278029441833496,
"learning_rate": 7.768939780597523e-06,
"loss": 0.5246,
"step": 935
},
{
"epoch": 5.766871165644172,
"grad_norm": 0.4724002480506897,
"learning_rate": 7.739143642530833e-06,
"loss": 0.5218,
"step": 940
},
{
"epoch": 5.7975460122699385,
"grad_norm": 0.5203721523284912,
"learning_rate": 7.70920786526321e-06,
"loss": 0.5256,
"step": 945
},
{
"epoch": 5.828220858895706,
"grad_norm": 0.5066553950309753,
"learning_rate": 7.679133974894984e-06,
"loss": 0.5272,
"step": 950
},
{
"epoch": 5.858895705521473,
"grad_norm": 0.5213795304298401,
"learning_rate": 7.648923504567374e-06,
"loss": 0.5293,
"step": 955
},
{
"epoch": 5.889570552147239,
"grad_norm": 0.5189357399940491,
"learning_rate": 7.618577994384324e-06,
"loss": 0.525,
"step": 960
},
{
"epoch": 5.920245398773006,
"grad_norm": 0.45614001154899597,
"learning_rate": 7.588098991334001e-06,
"loss": 0.5254,
"step": 965
},
{
"epoch": 5.950920245398773,
"grad_norm": 0.5392826199531555,
"learning_rate": 7.557488049209921e-06,
"loss": 0.5215,
"step": 970
},
{
"epoch": 5.9815950920245395,
"grad_norm": 0.5028013586997986,
"learning_rate": 7.52674672853174e-06,
"loss": 0.5249,
"step": 975
},
{
"epoch": 6.012269938650307,
"grad_norm": 0.5520371198654175,
"learning_rate": 7.495876596465703e-06,
"loss": 0.5087,
"step": 980
},
{
"epoch": 6.042944785276074,
"grad_norm": 0.6230819225311279,
"learning_rate": 7.464879226744748e-06,
"loss": 0.4881,
"step": 985
},
{
"epoch": 6.07361963190184,
"grad_norm": 0.581471860408783,
"learning_rate": 7.433756199588282e-06,
"loss": 0.4891,
"step": 990
},
{
"epoch": 6.104294478527607,
"grad_norm": 0.6113958358764648,
"learning_rate": 7.402509101621618e-06,
"loss": 0.4848,
"step": 995
},
{
"epoch": 6.134969325153374,
"grad_norm": 0.5332604050636292,
"learning_rate": 7.371139525795094e-06,
"loss": 0.488,
"step": 1000
},
{
"epoch": 6.1656441717791415,
"grad_norm": 0.538058340549469,
"learning_rate": 7.3396490713028674e-06,
"loss": 0.4904,
"step": 1005
},
{
"epoch": 6.196319018404908,
"grad_norm": 0.5761954188346863,
"learning_rate": 7.308039343501381e-06,
"loss": 0.4907,
"step": 1010
},
{
"epoch": 6.226993865030675,
"grad_norm": 0.5636390447616577,
"learning_rate": 7.276311953827533e-06,
"loss": 0.4872,
"step": 1015
},
{
"epoch": 6.257668711656442,
"grad_norm": 0.5386704802513123,
"learning_rate": 7.244468519716521e-06,
"loss": 0.4946,
"step": 1020
},
{
"epoch": 6.288343558282208,
"grad_norm": 0.585161566734314,
"learning_rate": 7.212510664519391e-06,
"loss": 0.4934,
"step": 1025
},
{
"epoch": 6.319018404907975,
"grad_norm": 0.5781230330467224,
"learning_rate": 7.180440017420277e-06,
"loss": 0.494,
"step": 1030
},
{
"epoch": 6.3496932515337425,
"grad_norm": 0.5272154211997986,
"learning_rate": 7.148258213353347e-06,
"loss": 0.4944,
"step": 1035
},
{
"epoch": 6.38036809815951,
"grad_norm": 0.5416070222854614,
"learning_rate": 7.115966892919459e-06,
"loss": 0.4936,
"step": 1040
},
{
"epoch": 6.411042944785276,
"grad_norm": 0.5524059534072876,
"learning_rate": 7.083567702302517e-06,
"loss": 0.4901,
"step": 1045
},
{
"epoch": 6.441717791411043,
"grad_norm": 0.5226433277130127,
"learning_rate": 7.05106229318556e-06,
"loss": 0.4931,
"step": 1050
},
{
"epoch": 6.47239263803681,
"grad_norm": 0.5324561595916748,
"learning_rate": 7.018452322666549e-06,
"loss": 0.4906,
"step": 1055
},
{
"epoch": 6.5030674846625764,
"grad_norm": 0.5099997520446777,
"learning_rate": 6.985739453173903e-06,
"loss": 0.4909,
"step": 1060
},
{
"epoch": 6.533742331288344,
"grad_norm": 0.5172037482261658,
"learning_rate": 6.9529253523817396e-06,
"loss": 0.4936,
"step": 1065
},
{
"epoch": 6.564417177914111,
"grad_norm": 0.5281527638435364,
"learning_rate": 6.9200116931248575e-06,
"loss": 0.4949,
"step": 1070
},
{
"epoch": 6.595092024539877,
"grad_norm": 0.5232874751091003,
"learning_rate": 6.887000153313468e-06,
"loss": 0.493,
"step": 1075
},
{
"epoch": 6.625766871165644,
"grad_norm": 0.5422880053520203,
"learning_rate": 6.853892415847645e-06,
"loss": 0.4944,
"step": 1080
},
{
"epoch": 6.656441717791411,
"grad_norm": 0.5587918758392334,
"learning_rate": 6.8206901685315366e-06,
"loss": 0.4974,
"step": 1085
},
{
"epoch": 6.6871165644171775,
"grad_norm": 0.5716516971588135,
"learning_rate": 6.787395103987323e-06,
"loss": 0.4959,
"step": 1090
},
{
"epoch": 6.717791411042945,
"grad_norm": 0.5290241837501526,
"learning_rate": 6.754008919568927e-06,
"loss": 0.49,
"step": 1095
},
{
"epoch": 6.748466257668712,
"grad_norm": 0.545229971408844,
"learning_rate": 6.72053331727549e-06,
"loss": 0.4913,
"step": 1100
},
{
"epoch": 6.779141104294479,
"grad_norm": 0.5339862108230591,
"learning_rate": 6.686970003664588e-06,
"loss": 0.4933,
"step": 1105
},
{
"epoch": 6.809815950920245,
"grad_norm": 0.5517235398292542,
"learning_rate": 6.653320689765257e-06,
"loss": 0.4957,
"step": 1110
},
{
"epoch": 6.840490797546012,
"grad_norm": 0.5221419334411621,
"learning_rate": 6.619587090990748e-06,
"loss": 0.4938,
"step": 1115
},
{
"epoch": 6.871165644171779,
"grad_norm": 0.5578407049179077,
"learning_rate": 6.585770927051085e-06,
"loss": 0.4992,
"step": 1120
},
{
"epoch": 6.901840490797546,
"grad_norm": 0.5061231255531311,
"learning_rate": 6.551873921865393e-06,
"loss": 0.4968,
"step": 1125
},
{
"epoch": 6.932515337423313,
"grad_norm": 0.5198781490325928,
"learning_rate": 6.517897803474011e-06,
"loss": 0.4982,
"step": 1130
},
{
"epoch": 6.96319018404908,
"grad_norm": 0.5066606998443604,
"learning_rate": 6.483844303950411e-06,
"loss": 0.4975,
"step": 1135
},
{
"epoch": 6.993865030674847,
"grad_norm": 0.5105178952217102,
"learning_rate": 6.4497151593128795e-06,
"loss": 0.5015,
"step": 1140
},
{
"epoch": 7.024539877300613,
"grad_norm": 0.6838037967681885,
"learning_rate": 6.415512109436031e-06,
"loss": 0.4634,
"step": 1145
},
{
"epoch": 7.0552147239263805,
"grad_norm": 0.6098262667655945,
"learning_rate": 6.381236897962102e-06,
"loss": 0.4616,
"step": 1150
},
{
"epoch": 7.085889570552148,
"grad_norm": 0.5857096314430237,
"learning_rate": 6.3468912722120715e-06,
"loss": 0.4557,
"step": 1155
},
{
"epoch": 7.116564417177914,
"grad_norm": 0.5579211115837097,
"learning_rate": 6.312476983096573e-06,
"loss": 0.454,
"step": 1160
},
{
"epoch": 7.147239263803681,
"grad_norm": 0.5579200387001038,
"learning_rate": 6.277995785026642e-06,
"loss": 0.4606,
"step": 1165
},
{
"epoch": 7.177914110429448,
"grad_norm": 0.5602801442146301,
"learning_rate": 6.243449435824276e-06,
"loss": 0.4557,
"step": 1170
},
{
"epoch": 7.208588957055214,
"grad_norm": 0.5627221465110779,
"learning_rate": 6.2088396966328155e-06,
"loss": 0.4564,
"step": 1175
},
{
"epoch": 7.2392638036809815,
"grad_norm": 0.5709859728813171,
"learning_rate": 6.174168331827179e-06,
"loss": 0.4596,
"step": 1180
},
{
"epoch": 7.269938650306749,
"grad_norm": 0.554961621761322,
"learning_rate": 6.139437108923898e-06,
"loss": 0.4583,
"step": 1185
},
{
"epoch": 7.300613496932515,
"grad_norm": 0.5725740790367126,
"learning_rate": 6.1046477984910215e-06,
"loss": 0.458,
"step": 1190
},
{
"epoch": 7.331288343558282,
"grad_norm": 0.5590230226516724,
"learning_rate": 6.069802174057849e-06,
"loss": 0.4604,
"step": 1195
},
{
"epoch": 7.361963190184049,
"grad_norm": 0.5823590159416199,
"learning_rate": 6.034902012024521e-06,
"loss": 0.4575,
"step": 1200
}
],
"logging_steps": 5,
"max_steps": 2445,
"num_input_tokens_seen": 0,
"num_train_epochs": 15,
"save_steps": 400,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.6846057011425575e+19,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}