mt.sft.v2 / trainer_state.json
double7's picture
Upload model
96a926c verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 295,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.003395585738539898,
"grad_norm": 6.320343624704885,
"learning_rate": 0.0,
"loss": 0.9899,
"step": 1
},
{
"epoch": 0.006791171477079796,
"grad_norm": 16.545756923443665,
"learning_rate": 3.3333333333333335e-07,
"loss": 0.9683,
"step": 2
},
{
"epoch": 0.010186757215619695,
"grad_norm": 6.124626290117518,
"learning_rate": 6.666666666666667e-07,
"loss": 0.9722,
"step": 3
},
{
"epoch": 0.013582342954159592,
"grad_norm": 6.907471369837716,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.9667,
"step": 4
},
{
"epoch": 0.01697792869269949,
"grad_norm": 6.275431523928528,
"learning_rate": 1.3333333333333334e-06,
"loss": 0.9814,
"step": 5
},
{
"epoch": 0.02037351443123939,
"grad_norm": 6.074623009583982,
"learning_rate": 1.6666666666666667e-06,
"loss": 0.9502,
"step": 6
},
{
"epoch": 0.023769100169779286,
"grad_norm": 89.99670872123082,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.9411,
"step": 7
},
{
"epoch": 0.027164685908319185,
"grad_norm": 6.224698090768632,
"learning_rate": 2.3333333333333336e-06,
"loss": 0.9075,
"step": 8
},
{
"epoch": 0.030560271646859084,
"grad_norm": 5.913600661145363,
"learning_rate": 2.666666666666667e-06,
"loss": 0.8979,
"step": 9
},
{
"epoch": 0.03395585738539898,
"grad_norm": 7.267004601688997,
"learning_rate": 3e-06,
"loss": 0.8589,
"step": 10
},
{
"epoch": 0.03735144312393888,
"grad_norm": 4.920293108168373,
"learning_rate": 3.3333333333333333e-06,
"loss": 0.8797,
"step": 11
},
{
"epoch": 0.04074702886247878,
"grad_norm": 7.0251120254603325,
"learning_rate": 3.6666666666666666e-06,
"loss": 0.785,
"step": 12
},
{
"epoch": 0.044142614601018676,
"grad_norm": 4.620531235688851,
"learning_rate": 4.000000000000001e-06,
"loss": 0.7932,
"step": 13
},
{
"epoch": 0.04753820033955857,
"grad_norm": 8.017042756220283,
"learning_rate": 4.333333333333334e-06,
"loss": 0.7878,
"step": 14
},
{
"epoch": 0.050933786078098474,
"grad_norm": 7.168335735527789,
"learning_rate": 4.666666666666667e-06,
"loss": 0.7854,
"step": 15
},
{
"epoch": 0.05432937181663837,
"grad_norm": 2.6102783506270137,
"learning_rate": 5e-06,
"loss": 0.7194,
"step": 16
},
{
"epoch": 0.057724957555178265,
"grad_norm": 2.8281275726026074,
"learning_rate": 5.333333333333334e-06,
"loss": 0.7059,
"step": 17
},
{
"epoch": 0.06112054329371817,
"grad_norm": 2.1330350735122505,
"learning_rate": 5.666666666666667e-06,
"loss": 0.7236,
"step": 18
},
{
"epoch": 0.06451612903225806,
"grad_norm": 3.3828007198200694,
"learning_rate": 6e-06,
"loss": 0.7133,
"step": 19
},
{
"epoch": 0.06791171477079797,
"grad_norm": 1.733407506222324,
"learning_rate": 6.333333333333333e-06,
"loss": 0.6807,
"step": 20
},
{
"epoch": 0.07130730050933787,
"grad_norm": 1.5732888077219966,
"learning_rate": 6.666666666666667e-06,
"loss": 0.6717,
"step": 21
},
{
"epoch": 0.07470288624787776,
"grad_norm": 2.869294784618442,
"learning_rate": 7e-06,
"loss": 0.6836,
"step": 22
},
{
"epoch": 0.07809847198641766,
"grad_norm": 1.539403446338188,
"learning_rate": 7.333333333333333e-06,
"loss": 0.656,
"step": 23
},
{
"epoch": 0.08149405772495756,
"grad_norm": 1.390711875015812,
"learning_rate": 7.666666666666667e-06,
"loss": 0.652,
"step": 24
},
{
"epoch": 0.08488964346349745,
"grad_norm": 1.4133171062619616,
"learning_rate": 8.000000000000001e-06,
"loss": 0.6237,
"step": 25
},
{
"epoch": 0.08828522920203735,
"grad_norm": 1.2675100082185478,
"learning_rate": 8.333333333333334e-06,
"loss": 0.5999,
"step": 26
},
{
"epoch": 0.09168081494057725,
"grad_norm": 1.4010718655700412,
"learning_rate": 8.666666666666668e-06,
"loss": 0.6229,
"step": 27
},
{
"epoch": 0.09507640067911714,
"grad_norm": 1.4302934299507057,
"learning_rate": 9e-06,
"loss": 0.5909,
"step": 28
},
{
"epoch": 0.09847198641765705,
"grad_norm": 1.3417528783040296,
"learning_rate": 9.333333333333334e-06,
"loss": 0.6236,
"step": 29
},
{
"epoch": 0.10186757215619695,
"grad_norm": 1.2550326498175997,
"learning_rate": 9.666666666666667e-06,
"loss": 0.5978,
"step": 30
},
{
"epoch": 0.10526315789473684,
"grad_norm": 1.2949203936642646,
"learning_rate": 1e-05,
"loss": 0.6206,
"step": 31
},
{
"epoch": 0.10865874363327674,
"grad_norm": 1.3404316606873072,
"learning_rate": 9.999648647603774e-06,
"loss": 0.6103,
"step": 32
},
{
"epoch": 0.11205432937181664,
"grad_norm": 1.2464543142979883,
"learning_rate": 9.998594639794502e-06,
"loss": 0.5823,
"step": 33
},
{
"epoch": 0.11544991511035653,
"grad_norm": 1.28881560865773,
"learning_rate": 9.996838124703448e-06,
"loss": 0.5724,
"step": 34
},
{
"epoch": 0.11884550084889643,
"grad_norm": 1.236715524722473,
"learning_rate": 9.994379349192927e-06,
"loss": 0.6006,
"step": 35
},
{
"epoch": 0.12224108658743633,
"grad_norm": 1.1748122882399419,
"learning_rate": 9.991218658821609e-06,
"loss": 0.5834,
"step": 36
},
{
"epoch": 0.12563667232597622,
"grad_norm": 1.3335055567165142,
"learning_rate": 9.987356497795944e-06,
"loss": 0.5781,
"step": 37
},
{
"epoch": 0.12903225806451613,
"grad_norm": 1.2573177388462913,
"learning_rate": 9.982793408907747e-06,
"loss": 0.5613,
"step": 38
},
{
"epoch": 0.13242784380305603,
"grad_norm": 614.0974853168025,
"learning_rate": 9.977530033457906e-06,
"loss": 0.5518,
"step": 39
},
{
"epoch": 0.13582342954159593,
"grad_norm": 1.8130801394796245,
"learning_rate": 9.971567111166246e-06,
"loss": 0.5875,
"step": 40
},
{
"epoch": 0.13921901528013583,
"grad_norm": 3.021540974319598,
"learning_rate": 9.964905480067585e-06,
"loss": 0.5666,
"step": 41
},
{
"epoch": 0.14261460101867574,
"grad_norm": 1.4307732415515269,
"learning_rate": 9.957546076393944e-06,
"loss": 0.5829,
"step": 42
},
{
"epoch": 0.1460101867572156,
"grad_norm": 1.2948953962351142,
"learning_rate": 9.949489934442966e-06,
"loss": 0.592,
"step": 43
},
{
"epoch": 0.1494057724957555,
"grad_norm": 1.28013431192408,
"learning_rate": 9.940738186432565e-06,
"loss": 0.554,
"step": 44
},
{
"epoch": 0.15280135823429541,
"grad_norm": 1.364283730780204,
"learning_rate": 9.931292062341793e-06,
"loss": 0.5563,
"step": 45
},
{
"epoch": 0.15619694397283532,
"grad_norm": 1.1789812327801241,
"learning_rate": 9.921152889737985e-06,
"loss": 0.5349,
"step": 46
},
{
"epoch": 0.15959252971137522,
"grad_norm": 1.272995264216437,
"learning_rate": 9.910322093590177e-06,
"loss": 0.5885,
"step": 47
},
{
"epoch": 0.16298811544991512,
"grad_norm": 1.2071797315958241,
"learning_rate": 9.898801196068839e-06,
"loss": 0.555,
"step": 48
},
{
"epoch": 0.166383701188455,
"grad_norm": 1.2466300266575705,
"learning_rate": 9.886591816331953e-06,
"loss": 0.5412,
"step": 49
},
{
"epoch": 0.1697792869269949,
"grad_norm": 1.1951025261533827,
"learning_rate": 9.87369567029745e-06,
"loss": 0.5266,
"step": 50
},
{
"epoch": 0.1731748726655348,
"grad_norm": 1.1409578758956818,
"learning_rate": 9.860114570402055e-06,
"loss": 0.5364,
"step": 51
},
{
"epoch": 0.1765704584040747,
"grad_norm": 1.194968225105999,
"learning_rate": 9.845850425346563e-06,
"loss": 0.544,
"step": 52
},
{
"epoch": 0.1799660441426146,
"grad_norm": 1.105032206844106,
"learning_rate": 9.830905239827592e-06,
"loss": 0.526,
"step": 53
},
{
"epoch": 0.1833616298811545,
"grad_norm": 1.0951143452700236,
"learning_rate": 9.815281114255841e-06,
"loss": 0.5384,
"step": 54
},
{
"epoch": 0.1867572156196944,
"grad_norm": 1.237628011550819,
"learning_rate": 9.798980244460892e-06,
"loss": 0.5216,
"step": 55
},
{
"epoch": 0.19015280135823429,
"grad_norm": 1.2049941993061364,
"learning_rate": 9.782004921382612e-06,
"loss": 0.5602,
"step": 56
},
{
"epoch": 0.1935483870967742,
"grad_norm": 1.0963401467603866,
"learning_rate": 9.764357530749178e-06,
"loss": 0.5211,
"step": 57
},
{
"epoch": 0.1969439728353141,
"grad_norm": 1.1153693044651702,
"learning_rate": 9.74604055274178e-06,
"loss": 0.5408,
"step": 58
},
{
"epoch": 0.200339558573854,
"grad_norm": 1.1446372149192978,
"learning_rate": 9.727056561646067e-06,
"loss": 0.5297,
"step": 59
},
{
"epoch": 0.2037351443123939,
"grad_norm": 1.1379654178567067,
"learning_rate": 9.707408225490343e-06,
"loss": 0.5241,
"step": 60
},
{
"epoch": 0.2071307300509338,
"grad_norm": 1.1133344090884694,
"learning_rate": 9.687098305670606e-06,
"loss": 0.5373,
"step": 61
},
{
"epoch": 0.21052631578947367,
"grad_norm": 1.14642700050245,
"learning_rate": 9.66612965656245e-06,
"loss": 0.5397,
"step": 62
},
{
"epoch": 0.21392190152801357,
"grad_norm": 1.0251599768158084,
"learning_rate": 9.644505225119922e-06,
"loss": 0.5444,
"step": 63
},
{
"epoch": 0.21731748726655348,
"grad_norm": 0.9890854380699226,
"learning_rate": 9.622228050461345e-06,
"loss": 0.4984,
"step": 64
},
{
"epoch": 0.22071307300509338,
"grad_norm": 1.1117068387014335,
"learning_rate": 9.599301263442194e-06,
"loss": 0.5214,
"step": 65
},
{
"epoch": 0.22410865874363328,
"grad_norm": 1.1785090789686485,
"learning_rate": 9.575728086215093e-06,
"loss": 0.5265,
"step": 66
},
{
"epoch": 0.22750424448217318,
"grad_norm": 1.0565597242070186,
"learning_rate": 9.551511831776966e-06,
"loss": 0.5155,
"step": 67
},
{
"epoch": 0.23089983022071306,
"grad_norm": 1.2278748412937952,
"learning_rate": 9.526655903503423e-06,
"loss": 0.5277,
"step": 68
},
{
"epoch": 0.23429541595925296,
"grad_norm": 1.1213279864335683,
"learning_rate": 9.501163794670445e-06,
"loss": 0.5307,
"step": 69
},
{
"epoch": 0.23769100169779286,
"grad_norm": 1.1237007845490015,
"learning_rate": 9.475039087963443e-06,
"loss": 0.51,
"step": 70
},
{
"epoch": 0.24108658743633277,
"grad_norm": 1.0615774210564253,
"learning_rate": 9.448285454973739e-06,
"loss": 0.517,
"step": 71
},
{
"epoch": 0.24448217317487267,
"grad_norm": 1.2104405507373044,
"learning_rate": 9.420906655682553e-06,
"loss": 0.5223,
"step": 72
},
{
"epoch": 0.24787775891341257,
"grad_norm": 1.0985265075768098,
"learning_rate": 9.392906537932582e-06,
"loss": 0.5257,
"step": 73
},
{
"epoch": 0.25127334465195245,
"grad_norm": 1.0626836434651132,
"learning_rate": 9.364289036887214e-06,
"loss": 0.5187,
"step": 74
},
{
"epoch": 0.2546689303904924,
"grad_norm": 1.0234412989238388,
"learning_rate": 9.335058174477472e-06,
"loss": 0.5127,
"step": 75
},
{
"epoch": 0.25806451612903225,
"grad_norm": 1.0702433549745685,
"learning_rate": 9.305218058836778e-06,
"loss": 0.4966,
"step": 76
},
{
"epoch": 0.2614601018675722,
"grad_norm": 1.1228773632449267,
"learning_rate": 9.274772883723587e-06,
"loss": 0.5265,
"step": 77
},
{
"epoch": 0.26485568760611206,
"grad_norm": 1.0076068781933833,
"learning_rate": 9.24372692793199e-06,
"loss": 0.5003,
"step": 78
},
{
"epoch": 0.26825127334465193,
"grad_norm": 1.0987680518642855,
"learning_rate": 9.21208455469037e-06,
"loss": 0.4971,
"step": 79
},
{
"epoch": 0.27164685908319186,
"grad_norm": 1.1051172285171345,
"learning_rate": 9.179850211048193e-06,
"loss": 0.491,
"step": 80
},
{
"epoch": 0.27504244482173174,
"grad_norm": 1.0886761114445709,
"learning_rate": 9.14702842725101e-06,
"loss": 0.5033,
"step": 81
},
{
"epoch": 0.27843803056027167,
"grad_norm": 1.101989710769037,
"learning_rate": 9.113623816103775e-06,
"loss": 0.5157,
"step": 82
},
{
"epoch": 0.28183361629881154,
"grad_norm": 1.0846044794179348,
"learning_rate": 9.079641072322555e-06,
"loss": 0.5103,
"step": 83
},
{
"epoch": 0.28522920203735147,
"grad_norm": 1.0994243701550208,
"learning_rate": 9.045084971874738e-06,
"loss": 0.5118,
"step": 84
},
{
"epoch": 0.28862478777589134,
"grad_norm": 0.982679989364588,
"learning_rate": 9.009960371307798e-06,
"loss": 0.4753,
"step": 85
},
{
"epoch": 0.2920203735144312,
"grad_norm": 1.0974557383394132,
"learning_rate": 8.974272207066767e-06,
"loss": 0.5065,
"step": 86
},
{
"epoch": 0.29541595925297115,
"grad_norm": 1.191767026119928,
"learning_rate": 8.938025494800454e-06,
"loss": 0.496,
"step": 87
},
{
"epoch": 0.298811544991511,
"grad_norm": 1.0863558052080193,
"learning_rate": 8.901225328656543e-06,
"loss": 0.4768,
"step": 88
},
{
"epoch": 0.30220713073005095,
"grad_norm": 1.0292339490206497,
"learning_rate": 8.863876880565656e-06,
"loss": 0.4982,
"step": 89
},
{
"epoch": 0.30560271646859083,
"grad_norm": 1.0263488228624327,
"learning_rate": 8.825985399514488e-06,
"loss": 0.49,
"step": 90
},
{
"epoch": 0.3089983022071307,
"grad_norm": 1.1083934275101859,
"learning_rate": 8.787556210808101e-06,
"loss": 0.508,
"step": 91
},
{
"epoch": 0.31239388794567063,
"grad_norm": 1.0759718773481644,
"learning_rate": 8.748594715321512e-06,
"loss": 0.493,
"step": 92
},
{
"epoch": 0.3157894736842105,
"grad_norm": 1.1275176872041444,
"learning_rate": 8.70910638874064e-06,
"loss": 0.5063,
"step": 93
},
{
"epoch": 0.31918505942275044,
"grad_norm": 1.0268180450930244,
"learning_rate": 8.669096780792754e-06,
"loss": 0.4953,
"step": 94
},
{
"epoch": 0.3225806451612903,
"grad_norm": 1.0957503547012442,
"learning_rate": 8.628571514466502e-06,
"loss": 0.5062,
"step": 95
},
{
"epoch": 0.32597623089983024,
"grad_norm": 1.0476160480506713,
"learning_rate": 8.587536285221656e-06,
"loss": 0.5254,
"step": 96
},
{
"epoch": 0.3293718166383701,
"grad_norm": 0.973467408933069,
"learning_rate": 8.545996860188668e-06,
"loss": 0.5107,
"step": 97
},
{
"epoch": 0.33276740237691,
"grad_norm": 0.9690640110532758,
"learning_rate": 8.503959077358143e-06,
"loss": 0.4767,
"step": 98
},
{
"epoch": 0.3361629881154499,
"grad_norm": 1.1272513638936046,
"learning_rate": 8.46142884476038e-06,
"loss": 0.4991,
"step": 99
},
{
"epoch": 0.3395585738539898,
"grad_norm": 1.089949180068211,
"learning_rate": 8.418412139635026e-06,
"loss": 0.5202,
"step": 100
},
{
"epoch": 0.34295415959252973,
"grad_norm": 1.0904976650398113,
"learning_rate": 8.374915007591053e-06,
"loss": 0.5019,
"step": 101
},
{
"epoch": 0.3463497453310696,
"grad_norm": 1.0132973869482662,
"learning_rate": 8.330943561757092e-06,
"loss": 0.4981,
"step": 102
},
{
"epoch": 0.34974533106960953,
"grad_norm": 1.0144345934272716,
"learning_rate": 8.286503981922284e-06,
"loss": 0.4828,
"step": 103
},
{
"epoch": 0.3531409168081494,
"grad_norm": 1.0708364693456955,
"learning_rate": 8.241602513667775e-06,
"loss": 0.4677,
"step": 104
},
{
"epoch": 0.3565365025466893,
"grad_norm": 1.0734641865863697,
"learning_rate": 8.19624546748895e-06,
"loss": 0.4991,
"step": 105
},
{
"epoch": 0.3599320882852292,
"grad_norm": 1.206730960165882,
"learning_rate": 8.150439217908557e-06,
"loss": 0.4874,
"step": 106
},
{
"epoch": 0.3633276740237691,
"grad_norm": 0.990112721589732,
"learning_rate": 8.104190202580811e-06,
"loss": 0.4613,
"step": 107
},
{
"epoch": 0.366723259762309,
"grad_norm": 1.135017537331381,
"learning_rate": 8.057504921386661e-06,
"loss": 0.4981,
"step": 108
},
{
"epoch": 0.3701188455008489,
"grad_norm": 1.1168906124405507,
"learning_rate": 8.010389935520269e-06,
"loss": 0.4812,
"step": 109
},
{
"epoch": 0.3735144312393888,
"grad_norm": 1.109365518761395,
"learning_rate": 7.962851866566912e-06,
"loss": 0.4901,
"step": 110
},
{
"epoch": 0.3769100169779287,
"grad_norm": 1.1121750297406146,
"learning_rate": 7.914897395572362e-06,
"loss": 0.4841,
"step": 111
},
{
"epoch": 0.38030560271646857,
"grad_norm": 1.10407244769018,
"learning_rate": 7.866533262103937e-06,
"loss": 0.4813,
"step": 112
},
{
"epoch": 0.3837011884550085,
"grad_norm": 1.109502100982312,
"learning_rate": 7.817766263303312e-06,
"loss": 0.4731,
"step": 113
},
{
"epoch": 0.3870967741935484,
"grad_norm": 1.0709239172717544,
"learning_rate": 7.768603252931243e-06,
"loss": 0.4988,
"step": 114
},
{
"epoch": 0.3904923599320883,
"grad_norm": 1.0915174076138725,
"learning_rate": 7.719051140404327e-06,
"loss": 0.4825,
"step": 115
},
{
"epoch": 0.3938879456706282,
"grad_norm": 1.0965896575304275,
"learning_rate": 7.669116889823955e-06,
"loss": 0.4742,
"step": 116
},
{
"epoch": 0.39728353140916806,
"grad_norm": 1.0241307032424805,
"learning_rate": 7.6188075189975644e-06,
"loss": 0.4743,
"step": 117
},
{
"epoch": 0.400679117147708,
"grad_norm": 0.9435780818732129,
"learning_rate": 7.568130098452352e-06,
"loss": 0.4739,
"step": 118
},
{
"epoch": 0.40407470288624786,
"grad_norm": 1.0248298007190784,
"learning_rate": 7.517091750441576e-06,
"loss": 0.4814,
"step": 119
},
{
"epoch": 0.4074702886247878,
"grad_norm": 1.0315243401737155,
"learning_rate": 7.465699647943586e-06,
"loss": 0.4805,
"step": 120
},
{
"epoch": 0.41086587436332767,
"grad_norm": 1.091443614674752,
"learning_rate": 7.413961013653725e-06,
"loss": 0.4953,
"step": 121
},
{
"epoch": 0.4142614601018676,
"grad_norm": 1.0680561484041602,
"learning_rate": 7.361883118969248e-06,
"loss": 0.4588,
"step": 122
},
{
"epoch": 0.41765704584040747,
"grad_norm": 0.9761759048656751,
"learning_rate": 7.309473282967387e-06,
"loss": 0.469,
"step": 123
},
{
"epoch": 0.42105263157894735,
"grad_norm": 1.0119186737160737,
"learning_rate": 7.256738871376733e-06,
"loss": 0.4812,
"step": 124
},
{
"epoch": 0.4244482173174873,
"grad_norm": 0.9670302415957511,
"learning_rate": 7.203687295542032e-06,
"loss": 0.456,
"step": 125
},
{
"epoch": 0.42784380305602715,
"grad_norm": 1.001107021910696,
"learning_rate": 7.1503260113826035e-06,
"loss": 0.4597,
"step": 126
},
{
"epoch": 0.4312393887945671,
"grad_norm": 1.024857257795839,
"learning_rate": 7.09666251834447e-06,
"loss": 0.4884,
"step": 127
},
{
"epoch": 0.43463497453310695,
"grad_norm": 1.0331435407774852,
"learning_rate": 7.042704358346375e-06,
"loss": 0.4581,
"step": 128
},
{
"epoch": 0.4380305602716469,
"grad_norm": 1.00793619623988,
"learning_rate": 6.988459114719849e-06,
"loss": 0.4672,
"step": 129
},
{
"epoch": 0.44142614601018676,
"grad_norm": 0.9145035341259203,
"learning_rate": 6.933934411143419e-06,
"loss": 0.4524,
"step": 130
},
{
"epoch": 0.44482173174872663,
"grad_norm": 1.1118101624400085,
"learning_rate": 6.879137910571191e-06,
"loss": 0.4738,
"step": 131
},
{
"epoch": 0.44821731748726656,
"grad_norm": 1.0044078413498474,
"learning_rate": 6.824077314155877e-06,
"loss": 0.4679,
"step": 132
},
{
"epoch": 0.45161290322580644,
"grad_norm": 1.0568591065810937,
"learning_rate": 6.768760360166471e-06,
"loss": 0.4909,
"step": 133
},
{
"epoch": 0.45500848896434637,
"grad_norm": 0.9937372079053678,
"learning_rate": 6.713194822900707e-06,
"loss": 0.4734,
"step": 134
},
{
"epoch": 0.45840407470288624,
"grad_norm": 1.0380172124065203,
"learning_rate": 6.657388511592453e-06,
"loss": 0.4979,
"step": 135
},
{
"epoch": 0.4617996604414261,
"grad_norm": 0.9710945454163669,
"learning_rate": 6.601349269314188e-06,
"loss": 0.4826,
"step": 136
},
{
"epoch": 0.46519524617996605,
"grad_norm": 1.1195229655523649,
"learning_rate": 6.545084971874738e-06,
"loss": 0.4685,
"step": 137
},
{
"epoch": 0.4685908319185059,
"grad_norm": 1.079871197310449,
"learning_rate": 6.488603526712391e-06,
"loss": 0.483,
"step": 138
},
{
"epoch": 0.47198641765704585,
"grad_norm": 1.0089317127122743,
"learning_rate": 6.431912871783587e-06,
"loss": 0.5028,
"step": 139
},
{
"epoch": 0.47538200339558573,
"grad_norm": 0.9671431137430636,
"learning_rate": 6.3750209744473105e-06,
"loss": 0.4669,
"step": 140
},
{
"epoch": 0.47877758913412566,
"grad_norm": 1.0520557175009426,
"learning_rate": 6.3179358303453386e-06,
"loss": 0.4735,
"step": 141
},
{
"epoch": 0.48217317487266553,
"grad_norm": 1.0538462837406155,
"learning_rate": 6.260665462278544e-06,
"loss": 0.5051,
"step": 142
},
{
"epoch": 0.4855687606112054,
"grad_norm": 1.037965055623141,
"learning_rate": 6.203217919079343e-06,
"loss": 0.4762,
"step": 143
},
{
"epoch": 0.48896434634974534,
"grad_norm": 1.0176275927771237,
"learning_rate": 6.145601274480521e-06,
"loss": 0.4869,
"step": 144
},
{
"epoch": 0.4923599320882852,
"grad_norm": 1.0431340329887582,
"learning_rate": 6.08782362598054e-06,
"loss": 0.4611,
"step": 145
},
{
"epoch": 0.49575551782682514,
"grad_norm": 1.0308571103490511,
"learning_rate": 6.029893093705492e-06,
"loss": 0.4763,
"step": 146
},
{
"epoch": 0.499151103565365,
"grad_norm": 1.0355702334516321,
"learning_rate": 5.971817819267914e-06,
"loss": 0.4654,
"step": 147
},
{
"epoch": 0.5025466893039049,
"grad_norm": 0.9893529623332218,
"learning_rate": 5.9136059646225375e-06,
"loss": 0.5044,
"step": 148
},
{
"epoch": 0.5059422750424448,
"grad_norm": 0.960195438000855,
"learning_rate": 5.855265710919211e-06,
"loss": 0.4435,
"step": 149
},
{
"epoch": 0.5093378607809848,
"grad_norm": 0.9646163806867624,
"learning_rate": 5.796805257353109e-06,
"loss": 0.4834,
"step": 150
},
{
"epoch": 0.5127334465195246,
"grad_norm": 0.9931874900764169,
"learning_rate": 5.738232820012407e-06,
"loss": 0.461,
"step": 151
},
{
"epoch": 0.5161290322580645,
"grad_norm": 1.0149629817842267,
"learning_rate": 5.679556630723592e-06,
"loss": 0.4611,
"step": 152
},
{
"epoch": 0.5195246179966044,
"grad_norm": 1.0022805808479243,
"learning_rate": 5.620784935894548e-06,
"loss": 0.4712,
"step": 153
},
{
"epoch": 0.5229202037351444,
"grad_norm": 0.9091219788824919,
"learning_rate": 5.561925995355595e-06,
"loss": 0.4906,
"step": 154
},
{
"epoch": 0.5263157894736842,
"grad_norm": 0.9508691466403413,
"learning_rate": 5.5029880811986546e-06,
"loss": 0.4836,
"step": 155
},
{
"epoch": 0.5297113752122241,
"grad_norm": 0.9816091460679524,
"learning_rate": 5.443979476614674e-06,
"loss": 0.4568,
"step": 156
},
{
"epoch": 0.533106960950764,
"grad_norm": 0.9890714751673821,
"learning_rate": 5.384908474729501e-06,
"loss": 0.4669,
"step": 157
},
{
"epoch": 0.5365025466893039,
"grad_norm": 0.9852345937629808,
"learning_rate": 5.325783377438357e-06,
"loss": 0.4631,
"step": 158
},
{
"epoch": 0.5398981324278438,
"grad_norm": 0.9992258942689997,
"learning_rate": 5.266612494239088e-06,
"loss": 0.4824,
"step": 159
},
{
"epoch": 0.5432937181663837,
"grad_norm": 1.0010167382867965,
"learning_rate": 5.207404141064334e-06,
"loss": 0.4619,
"step": 160
},
{
"epoch": 0.5466893039049237,
"grad_norm": 1.117194827109753,
"learning_rate": 5.148166639112799e-06,
"loss": 0.518,
"step": 161
},
{
"epoch": 0.5500848896434635,
"grad_norm": 0.9423156223035444,
"learning_rate": 5.088908313679788e-06,
"loss": 0.4567,
"step": 162
},
{
"epoch": 0.5534804753820034,
"grad_norm": 0.919109798014551,
"learning_rate": 5.029637492987153e-06,
"loss": 0.4649,
"step": 163
},
{
"epoch": 0.5568760611205433,
"grad_norm": 1.1232411363329386,
"learning_rate": 4.970362507012848e-06,
"loss": 0.4682,
"step": 164
},
{
"epoch": 0.5602716468590832,
"grad_norm": 0.9560703825281516,
"learning_rate": 4.911091686320213e-06,
"loss": 0.4594,
"step": 165
},
{
"epoch": 0.5636672325976231,
"grad_norm": 0.9927617945956294,
"learning_rate": 4.8518333608872015e-06,
"loss": 0.4266,
"step": 166
},
{
"epoch": 0.567062818336163,
"grad_norm": 1.0053279255521879,
"learning_rate": 4.792595858935668e-06,
"loss": 0.4745,
"step": 167
},
{
"epoch": 0.5704584040747029,
"grad_norm": 1.1247627665383018,
"learning_rate": 4.733387505760913e-06,
"loss": 0.4863,
"step": 168
},
{
"epoch": 0.5738539898132428,
"grad_norm": 0.9506838521386066,
"learning_rate": 4.674216622561645e-06,
"loss": 0.4459,
"step": 169
},
{
"epoch": 0.5772495755517827,
"grad_norm": 0.8918545360346245,
"learning_rate": 4.6150915252705005e-06,
"loss": 0.4571,
"step": 170
},
{
"epoch": 0.5806451612903226,
"grad_norm": 0.9704497436094051,
"learning_rate": 4.556020523385326e-06,
"loss": 0.4701,
"step": 171
},
{
"epoch": 0.5840407470288624,
"grad_norm": 0.9830084007827714,
"learning_rate": 4.497011918801347e-06,
"loss": 0.477,
"step": 172
},
{
"epoch": 0.5874363327674024,
"grad_norm": 1.0446285595393363,
"learning_rate": 4.438074004644407e-06,
"loss": 0.4963,
"step": 173
},
{
"epoch": 0.5908319185059423,
"grad_norm": 0.9430261659822053,
"learning_rate": 4.379215064105454e-06,
"loss": 0.4502,
"step": 174
},
{
"epoch": 0.5942275042444821,
"grad_norm": 0.9528761706191039,
"learning_rate": 4.32044336927641e-06,
"loss": 0.4667,
"step": 175
},
{
"epoch": 0.597623089983022,
"grad_norm": 0.9712945630496029,
"learning_rate": 4.261767179987595e-06,
"loss": 0.469,
"step": 176
},
{
"epoch": 0.601018675721562,
"grad_norm": 0.9203712027671123,
"learning_rate": 4.203194742646893e-06,
"loss": 0.4549,
"step": 177
},
{
"epoch": 0.6044142614601019,
"grad_norm": 0.9335750991253596,
"learning_rate": 4.1447342890807905e-06,
"loss": 0.4589,
"step": 178
},
{
"epoch": 0.6078098471986417,
"grad_norm": 0.9302840956981335,
"learning_rate": 4.086394035377463e-06,
"loss": 0.4525,
"step": 179
},
{
"epoch": 0.6112054329371817,
"grad_norm": 0.9205763667360544,
"learning_rate": 4.028182180732088e-06,
"loss": 0.4652,
"step": 180
},
{
"epoch": 0.6146010186757216,
"grad_norm": 0.9053159533350853,
"learning_rate": 3.970106906294509e-06,
"loss": 0.4887,
"step": 181
},
{
"epoch": 0.6179966044142614,
"grad_norm": 0.9639072424421392,
"learning_rate": 3.912176374019462e-06,
"loss": 0.4928,
"step": 182
},
{
"epoch": 0.6213921901528013,
"grad_norm": 0.8934882136410526,
"learning_rate": 3.85439872551948e-06,
"loss": 0.4613,
"step": 183
},
{
"epoch": 0.6247877758913413,
"grad_norm": 0.9142153731395757,
"learning_rate": 3.796782080920659e-06,
"loss": 0.4489,
"step": 184
},
{
"epoch": 0.6281833616298812,
"grad_norm": 0.9197655133200423,
"learning_rate": 3.7393345377214584e-06,
"loss": 0.4433,
"step": 185
},
{
"epoch": 0.631578947368421,
"grad_norm": 0.9569197839619695,
"learning_rate": 3.682064169654663e-06,
"loss": 0.4665,
"step": 186
},
{
"epoch": 0.634974533106961,
"grad_norm": 0.9362494462879061,
"learning_rate": 3.6249790255526916e-06,
"loss": 0.4363,
"step": 187
},
{
"epoch": 0.6383701188455009,
"grad_norm": 0.8671267136185908,
"learning_rate": 3.568087128216414e-06,
"loss": 0.4646,
"step": 188
},
{
"epoch": 0.6417657045840407,
"grad_norm": 0.9203019838984255,
"learning_rate": 3.511396473287611e-06,
"loss": 0.4571,
"step": 189
},
{
"epoch": 0.6451612903225806,
"grad_norm": 0.9768723974984408,
"learning_rate": 3.4549150281252635e-06,
"loss": 0.4484,
"step": 190
},
{
"epoch": 0.6485568760611206,
"grad_norm": 0.9396504959111149,
"learning_rate": 3.398650730685813e-06,
"loss": 0.457,
"step": 191
},
{
"epoch": 0.6519524617996605,
"grad_norm": 0.9336335579059931,
"learning_rate": 3.3426114884075488e-06,
"loss": 0.4467,
"step": 192
},
{
"epoch": 0.6553480475382003,
"grad_norm": 0.8976643217426382,
"learning_rate": 3.2868051770992935e-06,
"loss": 0.4572,
"step": 193
},
{
"epoch": 0.6587436332767402,
"grad_norm": 0.9373033603378723,
"learning_rate": 3.2312396398335312e-06,
"loss": 0.4687,
"step": 194
},
{
"epoch": 0.6621392190152802,
"grad_norm": 0.9707294952319975,
"learning_rate": 3.175922685844125e-06,
"loss": 0.4336,
"step": 195
},
{
"epoch": 0.66553480475382,
"grad_norm": 0.9583493306758707,
"learning_rate": 3.1208620894288105e-06,
"loss": 0.4486,
"step": 196
},
{
"epoch": 0.6689303904923599,
"grad_norm": 0.9820021819293268,
"learning_rate": 3.0660655888565827e-06,
"loss": 0.4588,
"step": 197
},
{
"epoch": 0.6723259762308998,
"grad_norm": 0.9400174461623624,
"learning_rate": 3.0115408852801535e-06,
"loss": 0.4423,
"step": 198
},
{
"epoch": 0.6757215619694398,
"grad_norm": 0.9603934801371697,
"learning_rate": 2.9572956416536267e-06,
"loss": 0.444,
"step": 199
},
{
"epoch": 0.6791171477079796,
"grad_norm": 0.9537870436181752,
"learning_rate": 2.9033374816555338e-06,
"loss": 0.4556,
"step": 200
},
{
"epoch": 0.6825127334465195,
"grad_norm": 0.9149135094841083,
"learning_rate": 2.8496739886173994e-06,
"loss": 0.4727,
"step": 201
},
{
"epoch": 0.6859083191850595,
"grad_norm": 0.9486316397589806,
"learning_rate": 2.7963127044579697e-06,
"loss": 0.432,
"step": 202
},
{
"epoch": 0.6893039049235993,
"grad_norm": 0.9240748880967119,
"learning_rate": 2.743261128623269e-06,
"loss": 0.4596,
"step": 203
},
{
"epoch": 0.6926994906621392,
"grad_norm": 0.9573847569611018,
"learning_rate": 2.6905267170326143e-06,
"loss": 0.4679,
"step": 204
},
{
"epoch": 0.6960950764006791,
"grad_norm": 0.8845992056166951,
"learning_rate": 2.6381168810307536e-06,
"loss": 0.4631,
"step": 205
},
{
"epoch": 0.6994906621392191,
"grad_norm": 1.004797015683845,
"learning_rate": 2.5860389863462765e-06,
"loss": 0.4475,
"step": 206
},
{
"epoch": 0.7028862478777589,
"grad_norm": 0.9408081397458115,
"learning_rate": 2.534300352056416e-06,
"loss": 0.4529,
"step": 207
},
{
"epoch": 0.7062818336162988,
"grad_norm": 0.8830371296945094,
"learning_rate": 2.4829082495584244e-06,
"loss": 0.4431,
"step": 208
},
{
"epoch": 0.7096774193548387,
"grad_norm": 0.918591969840702,
"learning_rate": 2.4318699015476495e-06,
"loss": 0.4295,
"step": 209
},
{
"epoch": 0.7130730050933786,
"grad_norm": 0.9126233293580099,
"learning_rate": 2.3811924810024385e-06,
"loss": 0.4462,
"step": 210
},
{
"epoch": 0.7164685908319185,
"grad_norm": 0.944049132172904,
"learning_rate": 2.330883110176049e-06,
"loss": 0.4764,
"step": 211
},
{
"epoch": 0.7198641765704584,
"grad_norm": 0.8594916774059891,
"learning_rate": 2.2809488595956746e-06,
"loss": 0.4245,
"step": 212
},
{
"epoch": 0.7232597623089984,
"grad_norm": 0.9004237430035651,
"learning_rate": 2.2313967470687593e-06,
"loss": 0.4618,
"step": 213
},
{
"epoch": 0.7266553480475382,
"grad_norm": 0.9087973453481795,
"learning_rate": 2.18223373669669e-06,
"loss": 0.4614,
"step": 214
},
{
"epoch": 0.7300509337860781,
"grad_norm": 0.8919978483526617,
"learning_rate": 2.1334667378960642e-06,
"loss": 0.462,
"step": 215
},
{
"epoch": 0.733446519524618,
"grad_norm": 0.9099211098465206,
"learning_rate": 2.0851026044276405e-06,
"loss": 0.4634,
"step": 216
},
{
"epoch": 0.7368421052631579,
"grad_norm": 0.8896204484265512,
"learning_rate": 2.0371481334330913e-06,
"loss": 0.463,
"step": 217
},
{
"epoch": 0.7402376910016978,
"grad_norm": 0.9519257493559129,
"learning_rate": 1.9896100644797316e-06,
"loss": 0.4788,
"step": 218
},
{
"epoch": 0.7436332767402377,
"grad_norm": 0.9116718849117885,
"learning_rate": 1.9424950786133414e-06,
"loss": 0.4634,
"step": 219
},
{
"epoch": 0.7470288624787776,
"grad_norm": 0.9263410885293661,
"learning_rate": 1.8958097974191909e-06,
"loss": 0.484,
"step": 220
},
{
"epoch": 0.7504244482173175,
"grad_norm": 0.9023645875611066,
"learning_rate": 1.8495607820914451e-06,
"loss": 0.4442,
"step": 221
},
{
"epoch": 0.7538200339558574,
"grad_norm": 0.8875288134755263,
"learning_rate": 1.8037545325110506e-06,
"loss": 0.4241,
"step": 222
},
{
"epoch": 0.7572156196943973,
"grad_norm": 0.9432820849930534,
"learning_rate": 1.7583974863322272e-06,
"loss": 0.4837,
"step": 223
},
{
"epoch": 0.7606112054329371,
"grad_norm": 0.9658199086993665,
"learning_rate": 1.7134960180777171e-06,
"loss": 0.4728,
"step": 224
},
{
"epoch": 0.7640067911714771,
"grad_norm": 0.8457866400541412,
"learning_rate": 1.6690564382429104e-06,
"loss": 0.4301,
"step": 225
},
{
"epoch": 0.767402376910017,
"grad_norm": 0.8989697214841547,
"learning_rate": 1.6250849924089485e-06,
"loss": 0.4527,
"step": 226
},
{
"epoch": 0.7707979626485568,
"grad_norm": 0.9022700419018949,
"learning_rate": 1.581587860364977e-06,
"loss": 0.4543,
"step": 227
},
{
"epoch": 0.7741935483870968,
"grad_norm": 0.9173241233268217,
"learning_rate": 1.5385711552396227e-06,
"loss": 0.4466,
"step": 228
},
{
"epoch": 0.7775891341256367,
"grad_norm": 0.8617337826425147,
"learning_rate": 1.4960409226418576e-06,
"loss": 0.4386,
"step": 229
},
{
"epoch": 0.7809847198641766,
"grad_norm": 0.8625616538339923,
"learning_rate": 1.4540031398113335e-06,
"loss": 0.4533,
"step": 230
},
{
"epoch": 0.7843803056027164,
"grad_norm": 0.894048701617508,
"learning_rate": 1.4124637147783431e-06,
"loss": 0.4447,
"step": 231
},
{
"epoch": 0.7877758913412564,
"grad_norm": 0.9208844935282854,
"learning_rate": 1.371428485533498e-06,
"loss": 0.4826,
"step": 232
},
{
"epoch": 0.7911714770797963,
"grad_norm": 0.8883039726822113,
"learning_rate": 1.3309032192072463e-06,
"loss": 0.4486,
"step": 233
},
{
"epoch": 0.7945670628183361,
"grad_norm": 0.899703690304073,
"learning_rate": 1.29089361125936e-06,
"loss": 0.4396,
"step": 234
},
{
"epoch": 0.797962648556876,
"grad_norm": 0.942126628969688,
"learning_rate": 1.251405284678488e-06,
"loss": 0.4666,
"step": 235
},
{
"epoch": 0.801358234295416,
"grad_norm": 0.876574882692073,
"learning_rate": 1.2124437891918995e-06,
"loss": 0.4338,
"step": 236
},
{
"epoch": 0.8047538200339559,
"grad_norm": 0.8738541998828565,
"learning_rate": 1.1740146004855141e-06,
"loss": 0.4609,
"step": 237
},
{
"epoch": 0.8081494057724957,
"grad_norm": 0.8732256896169818,
"learning_rate": 1.1361231194343436e-06,
"loss": 0.4505,
"step": 238
},
{
"epoch": 0.8115449915110357,
"grad_norm": 0.8634374293174691,
"learning_rate": 1.0987746713434578e-06,
"loss": 0.4626,
"step": 239
},
{
"epoch": 0.8149405772495756,
"grad_norm": 0.8709360064531477,
"learning_rate": 1.0619745051995473e-06,
"loss": 0.465,
"step": 240
},
{
"epoch": 0.8183361629881154,
"grad_norm": 0.8612129831209381,
"learning_rate": 1.0257277929332332e-06,
"loss": 0.4543,
"step": 241
},
{
"epoch": 0.8217317487266553,
"grad_norm": 0.9089176413771565,
"learning_rate": 9.900396286922025e-07,
"loss": 0.4676,
"step": 242
},
{
"epoch": 0.8251273344651953,
"grad_norm": 0.8856710278778406,
"learning_rate": 9.549150281252633e-07,
"loss": 0.4366,
"step": 243
},
{
"epoch": 0.8285229202037352,
"grad_norm": 0.8188623584178454,
"learning_rate": 9.203589276774438e-07,
"loss": 0.4404,
"step": 244
},
{
"epoch": 0.831918505942275,
"grad_norm": 0.9252917982729971,
"learning_rate": 8.86376183896226e-07,
"loss": 0.4447,
"step": 245
},
{
"epoch": 0.8353140916808149,
"grad_norm": 0.8784423429363009,
"learning_rate": 8.529715727489912e-07,
"loss": 0.4488,
"step": 246
},
{
"epoch": 0.8387096774193549,
"grad_norm": 0.9207047553546756,
"learning_rate": 8.201497889518073e-07,
"loss": 0.4561,
"step": 247
},
{
"epoch": 0.8421052631578947,
"grad_norm": 0.886908726068121,
"learning_rate": 7.879154453096305e-07,
"loss": 0.4556,
"step": 248
},
{
"epoch": 0.8455008488964346,
"grad_norm": 0.8585705366949494,
"learning_rate": 7.562730720680111e-07,
"loss": 0.4333,
"step": 249
},
{
"epoch": 0.8488964346349746,
"grad_norm": 0.8504815527006262,
"learning_rate": 7.25227116276413e-07,
"loss": 0.4512,
"step": 250
},
{
"epoch": 0.8522920203735145,
"grad_norm": 0.8201572218259242,
"learning_rate": 6.947819411632223e-07,
"loss": 0.4492,
"step": 251
},
{
"epoch": 0.8556876061120543,
"grad_norm": 0.8579475897548395,
"learning_rate": 6.649418255225298e-07,
"loss": 0.4633,
"step": 252
},
{
"epoch": 0.8590831918505942,
"grad_norm": 0.8140416489339838,
"learning_rate": 6.357109631127889e-07,
"loss": 0.4418,
"step": 253
},
{
"epoch": 0.8624787775891342,
"grad_norm": 0.8552881051968678,
"learning_rate": 6.07093462067419e-07,
"loss": 0.4572,
"step": 254
},
{
"epoch": 0.865874363327674,
"grad_norm": 0.8971540169404577,
"learning_rate": 5.79093344317449e-07,
"loss": 0.4775,
"step": 255
},
{
"epoch": 0.8692699490662139,
"grad_norm": 0.8058254068183305,
"learning_rate": 5.517145450262639e-07,
"loss": 0.4313,
"step": 256
},
{
"epoch": 0.8726655348047538,
"grad_norm": 0.8530487589073922,
"learning_rate": 5.249609120365579e-07,
"loss": 0.4397,
"step": 257
},
{
"epoch": 0.8760611205432938,
"grad_norm": 0.9429999718413201,
"learning_rate": 4.988362053295564e-07,
"loss": 0.4496,
"step": 258
},
{
"epoch": 0.8794567062818336,
"grad_norm": 0.8375356078246441,
"learning_rate": 4.733440964965791e-07,
"loss": 0.4303,
"step": 259
},
{
"epoch": 0.8828522920203735,
"grad_norm": 0.8256703834683652,
"learning_rate": 4.484881682230341e-07,
"loss": 0.4268,
"step": 260
},
{
"epoch": 0.8862478777589134,
"grad_norm": 0.8543078561007811,
"learning_rate": 4.242719137849077e-07,
"loss": 0.4514,
"step": 261
},
{
"epoch": 0.8896434634974533,
"grad_norm": 0.905414358204414,
"learning_rate": 4.00698736557808e-07,
"loss": 0.4475,
"step": 262
},
{
"epoch": 0.8930390492359932,
"grad_norm": 0.8479836834072867,
"learning_rate": 3.777719495386567e-07,
"loss": 0.4479,
"step": 263
},
{
"epoch": 0.8964346349745331,
"grad_norm": 0.8646327130622585,
"learning_rate": 3.5549477488007853e-07,
"loss": 0.4563,
"step": 264
},
{
"epoch": 0.8998302207130731,
"grad_norm": 0.8363483102479007,
"learning_rate": 3.3387034343755063e-07,
"loss": 0.4447,
"step": 265
},
{
"epoch": 0.9032258064516129,
"grad_norm": 0.8772923486186023,
"learning_rate": 3.1290169432939556e-07,
"loss": 0.4416,
"step": 266
},
{
"epoch": 0.9066213921901528,
"grad_norm": 0.8320996733730202,
"learning_rate": 2.925917745096568e-07,
"loss": 0.4263,
"step": 267
},
{
"epoch": 0.9100169779286927,
"grad_norm": 0.8735391402072182,
"learning_rate": 2.7294343835393366e-07,
"loss": 0.4635,
"step": 268
},
{
"epoch": 0.9134125636672326,
"grad_norm": 0.8476560390812548,
"learning_rate": 2.539594472582213e-07,
"loss": 0.4387,
"step": 269
},
{
"epoch": 0.9168081494057725,
"grad_norm": 0.8265472830061668,
"learning_rate": 2.3564246925082358e-07,
"loss": 0.4409,
"step": 270
},
{
"epoch": 0.9202037351443124,
"grad_norm": 0.8275085794146296,
"learning_rate": 2.179950786173879e-07,
"loss": 0.422,
"step": 271
},
{
"epoch": 0.9235993208828522,
"grad_norm": 0.8378368477541204,
"learning_rate": 2.01019755539108e-07,
"loss": 0.4533,
"step": 272
},
{
"epoch": 0.9269949066213922,
"grad_norm": 0.8225280576542648,
"learning_rate": 1.8471888574415953e-07,
"loss": 0.4504,
"step": 273
},
{
"epoch": 0.9303904923599321,
"grad_norm": 0.8689138287376853,
"learning_rate": 1.690947601724091e-07,
"loss": 0.4616,
"step": 274
},
{
"epoch": 0.933786078098472,
"grad_norm": 0.8153963652578823,
"learning_rate": 1.5414957465343883e-07,
"loss": 0.4362,
"step": 275
},
{
"epoch": 0.9371816638370118,
"grad_norm": 0.7970051219455544,
"learning_rate": 1.3988542959794627e-07,
"loss": 0.4476,
"step": 276
},
{
"epoch": 0.9405772495755518,
"grad_norm": 0.8286263852975556,
"learning_rate": 1.2630432970255014e-07,
"loss": 0.4317,
"step": 277
},
{
"epoch": 0.9439728353140917,
"grad_norm": 0.819703548732724,
"learning_rate": 1.1340818366804728e-07,
"loss": 0.4374,
"step": 278
},
{
"epoch": 0.9473684210526315,
"grad_norm": 0.8630335285550543,
"learning_rate": 1.0119880393116177e-07,
"loss": 0.4463,
"step": 279
},
{
"epoch": 0.9507640067911715,
"grad_norm": 0.8905416016998572,
"learning_rate": 8.967790640982466e-08,
"loss": 0.4534,
"step": 280
},
{
"epoch": 0.9541595925297114,
"grad_norm": 0.869553726844365,
"learning_rate": 7.884711026201586e-08,
"loss": 0.4474,
"step": 281
},
{
"epoch": 0.9575551782682513,
"grad_norm": 0.8516101851144577,
"learning_rate": 6.870793765820783e-08,
"loss": 0.4735,
"step": 282
},
{
"epoch": 0.9609507640067911,
"grad_norm": 0.834166916764426,
"learning_rate": 5.92618135674361e-08,
"loss": 0.4506,
"step": 283
},
{
"epoch": 0.9643463497453311,
"grad_norm": 0.8500281237762543,
"learning_rate": 5.0510065557034526e-08,
"loss": 0.4677,
"step": 284
},
{
"epoch": 0.967741935483871,
"grad_norm": 0.833367943844974,
"learning_rate": 4.245392360605727e-08,
"loss": 0.4582,
"step": 285
},
{
"epoch": 0.9711375212224108,
"grad_norm": 0.8458107328647104,
"learning_rate": 3.5094519932415417e-08,
"loss": 0.4505,
"step": 286
},
{
"epoch": 0.9745331069609507,
"grad_norm": 0.8443895579435003,
"learning_rate": 2.843288883375539e-08,
"loss": 0.4538,
"step": 287
},
{
"epoch": 0.9779286926994907,
"grad_norm": 0.8631958241337926,
"learning_rate": 2.2469966542096323e-08,
"loss": 0.4432,
"step": 288
},
{
"epoch": 0.9813242784380306,
"grad_norm": 0.886260366562854,
"learning_rate": 1.7206591092253642e-08,
"loss": 0.4567,
"step": 289
},
{
"epoch": 0.9847198641765704,
"grad_norm": 0.7994144470255627,
"learning_rate": 1.264350220405719e-08,
"loss": 0.4317,
"step": 290
},
{
"epoch": 0.9881154499151104,
"grad_norm": 0.8700681537853684,
"learning_rate": 8.781341178393244e-09,
"loss": 0.4555,
"step": 291
},
{
"epoch": 0.9915110356536503,
"grad_norm": 0.8575898812488901,
"learning_rate": 5.620650807073857e-09,
"loss": 0.4606,
"step": 292
},
{
"epoch": 0.9949066213921901,
"grad_norm": 0.8616135380601608,
"learning_rate": 3.1618752965534295e-09,
"loss": 0.4408,
"step": 293
},
{
"epoch": 0.99830220713073,
"grad_norm": 0.7784658740487929,
"learning_rate": 1.4053602054991954e-09,
"loss": 0.4157,
"step": 294
},
{
"epoch": 1.0,
"grad_norm": 1.2063313976605035,
"learning_rate": 3.513523962256349e-10,
"loss": 0.4174,
"step": 295
},
{
"epoch": 1.0,
"step": 295,
"total_flos": 44242544295936.0,
"train_loss": 0.5097209928399425,
"train_runtime": 1256.4644,
"train_samples_per_second": 14.995,
"train_steps_per_second": 0.235
}
],
"logging_steps": 1,
"max_steps": 295,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 2500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 44242544295936.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}