mistral7b_hotpot1 / trainer_state.json
mangopy's picture
Upload trainer_state.json with huggingface_hub
0947451 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.9887005649717513,
"eval_steps": 500,
"global_step": 154,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0129136400322841,
"grad_norm": 26.062711090700205,
"learning_rate": 1.25e-07,
"loss": 0.3676,
"step": 1
},
{
"epoch": 0.0258272800645682,
"grad_norm": 25.492701005475226,
"learning_rate": 2.5e-07,
"loss": 0.3636,
"step": 2
},
{
"epoch": 0.0387409200968523,
"grad_norm": 28.061770353754802,
"learning_rate": 3.75e-07,
"loss": 0.3788,
"step": 3
},
{
"epoch": 0.0516545601291364,
"grad_norm": 28.99654186845467,
"learning_rate": 5e-07,
"loss": 0.395,
"step": 4
},
{
"epoch": 0.0645682001614205,
"grad_norm": 20.935596355684492,
"learning_rate": 6.249999999999999e-07,
"loss": 0.3741,
"step": 5
},
{
"epoch": 0.0774818401937046,
"grad_norm": 26.035750645230728,
"learning_rate": 7.5e-07,
"loss": 0.3701,
"step": 6
},
{
"epoch": 0.0903954802259887,
"grad_norm": 36.159284168135784,
"learning_rate": 8.75e-07,
"loss": 0.3642,
"step": 7
},
{
"epoch": 0.1033091202582728,
"grad_norm": 32.251355945752145,
"learning_rate": 1e-06,
"loss": 0.3473,
"step": 8
},
{
"epoch": 0.1162227602905569,
"grad_norm": 35.98055219357963,
"learning_rate": 1.125e-06,
"loss": 0.3728,
"step": 9
},
{
"epoch": 0.129136400322841,
"grad_norm": 40.267311648036056,
"learning_rate": 1.2499999999999999e-06,
"loss": 0.3472,
"step": 10
},
{
"epoch": 0.1420500403551251,
"grad_norm": 54.431376922416256,
"learning_rate": 1.375e-06,
"loss": 0.3941,
"step": 11
},
{
"epoch": 0.1549636803874092,
"grad_norm": 52.787609103487405,
"learning_rate": 1.5e-06,
"loss": 0.3694,
"step": 12
},
{
"epoch": 0.1678773204196933,
"grad_norm": 51.470919986790086,
"learning_rate": 1.625e-06,
"loss": 0.3587,
"step": 13
},
{
"epoch": 0.1807909604519774,
"grad_norm": 44.49429708878717,
"learning_rate": 1.75e-06,
"loss": 0.3369,
"step": 14
},
{
"epoch": 0.1937046004842615,
"grad_norm": 48.61799720707049,
"learning_rate": 1.8749999999999998e-06,
"loss": 0.3811,
"step": 15
},
{
"epoch": 0.2066182405165456,
"grad_norm": 41.649896181802205,
"learning_rate": 2e-06,
"loss": 0.3911,
"step": 16
},
{
"epoch": 0.2195318805488297,
"grad_norm": 42.44898868701815,
"learning_rate": 1.999740884841349e-06,
"loss": 0.387,
"step": 17
},
{
"epoch": 0.2324455205811138,
"grad_norm": 41.183443431747065,
"learning_rate": 1.9989636736467275e-06,
"loss": 0.3946,
"step": 18
},
{
"epoch": 0.2453591606133979,
"grad_norm": 34.50686543223266,
"learning_rate": 1.9976687691905393e-06,
"loss": 0.4088,
"step": 19
},
{
"epoch": 0.258272800645682,
"grad_norm": 27.279515728605386,
"learning_rate": 1.995856842531531e-06,
"loss": 0.3724,
"step": 20
},
{
"epoch": 0.2711864406779661,
"grad_norm": 30.83455319215996,
"learning_rate": 1.993528832665031e-06,
"loss": 0.3866,
"step": 21
},
{
"epoch": 0.2841000807102502,
"grad_norm": 26.247716082750618,
"learning_rate": 1.9906859460363304e-06,
"loss": 0.3521,
"step": 22
},
{
"epoch": 0.2970137207425343,
"grad_norm": 27.831795874032476,
"learning_rate": 1.9873296559154695e-06,
"loss": 0.4159,
"step": 23
},
{
"epoch": 0.3099273607748184,
"grad_norm": 28.5584628971039,
"learning_rate": 1.983461701633742e-06,
"loss": 0.3767,
"step": 24
},
{
"epoch": 0.3228410008071025,
"grad_norm": 21.378018038217974,
"learning_rate": 1.979084087682323e-06,
"loss": 0.3304,
"step": 25
},
{
"epoch": 0.3357546408393866,
"grad_norm": 23.044867887515178,
"learning_rate": 1.9741990826734792e-06,
"loss": 0.3705,
"step": 26
},
{
"epoch": 0.3486682808716707,
"grad_norm": 26.947778108928976,
"learning_rate": 1.968809218164906e-06,
"loss": 0.404,
"step": 27
},
{
"epoch": 0.3615819209039548,
"grad_norm": 22.039479721325353,
"learning_rate": 1.9629172873477994e-06,
"loss": 0.3365,
"step": 28
},
{
"epoch": 0.3744955609362389,
"grad_norm": 21.274934357850633,
"learning_rate": 1.956526343599335e-06,
"loss": 0.3306,
"step": 29
},
{
"epoch": 0.387409200968523,
"grad_norm": 23.40835330420104,
"learning_rate": 1.949639698900319e-06,
"loss": 0.3246,
"step": 30
},
{
"epoch": 0.4003228410008071,
"grad_norm": 26.03684733839572,
"learning_rate": 1.9422609221188204e-06,
"loss": 0.3121,
"step": 31
},
{
"epoch": 0.4132364810330912,
"grad_norm": 22.817733184810667,
"learning_rate": 1.9343938371606708e-06,
"loss": 0.363,
"step": 32
},
{
"epoch": 0.4261501210653753,
"grad_norm": 27.888538264779257,
"learning_rate": 1.926042520987805e-06,
"loss": 0.3599,
"step": 33
},
{
"epoch": 0.4390637610976594,
"grad_norm": 23.664544295074048,
"learning_rate": 1.9172113015054528e-06,
"loss": 0.3259,
"step": 34
},
{
"epoch": 0.4519774011299435,
"grad_norm": 26.814661596965937,
"learning_rate": 1.907904755319289e-06,
"loss": 0.3622,
"step": 35
},
{
"epoch": 0.4648910411622276,
"grad_norm": 23.047494151516773,
"learning_rate": 1.898127705363696e-06,
"loss": 0.3595,
"step": 36
},
{
"epoch": 0.4778046811945117,
"grad_norm": 23.769391103002327,
"learning_rate": 1.887885218402375e-06,
"loss": 0.3418,
"step": 37
},
{
"epoch": 0.4907183212267958,
"grad_norm": 21.448479890321668,
"learning_rate": 1.8771826024025943e-06,
"loss": 0.3112,
"step": 38
},
{
"epoch": 0.5036319612590799,
"grad_norm": 21.618299294862208,
"learning_rate": 1.8660254037844386e-06,
"loss": 0.2935,
"step": 39
},
{
"epoch": 0.516545601291364,
"grad_norm": 23.83881914235937,
"learning_rate": 1.8544194045464886e-06,
"loss": 0.3487,
"step": 40
},
{
"epoch": 0.5294592413236481,
"grad_norm": 22.34191137757626,
"learning_rate": 1.8423706192694115e-06,
"loss": 0.3337,
"step": 41
},
{
"epoch": 0.5423728813559322,
"grad_norm": 22.680526606663665,
"learning_rate": 1.8298852919990251e-06,
"loss": 0.3326,
"step": 42
},
{
"epoch": 0.5552865213882163,
"grad_norm": 21.047157564860804,
"learning_rate": 1.8169698930104419e-06,
"loss": 0.3262,
"step": 43
},
{
"epoch": 0.5682001614205004,
"grad_norm": 23.210220151582366,
"learning_rate": 1.8036311154549781e-06,
"loss": 0.3531,
"step": 44
},
{
"epoch": 0.5811138014527845,
"grad_norm": 23.059400786558072,
"learning_rate": 1.7898758718915585e-06,
"loss": 0.3042,
"step": 45
},
{
"epoch": 0.5940274414850686,
"grad_norm": 20.985554392897924,
"learning_rate": 1.7757112907044198e-06,
"loss": 0.3211,
"step": 46
},
{
"epoch": 0.6069410815173527,
"grad_norm": 24.381494037046473,
"learning_rate": 1.7611447124089646e-06,
"loss": 0.3501,
"step": 47
},
{
"epoch": 0.6198547215496368,
"grad_norm": 21.163889567259087,
"learning_rate": 1.7461836858476855e-06,
"loss": 0.3407,
"step": 48
},
{
"epoch": 0.632768361581921,
"grad_norm": 17.673857406454214,
"learning_rate": 1.7308359642781241e-06,
"loss": 0.2912,
"step": 49
},
{
"epoch": 0.645682001614205,
"grad_norm": 20.851302876331236,
"learning_rate": 1.7151095013548993e-06,
"loss": 0.311,
"step": 50
},
{
"epoch": 0.6585956416464891,
"grad_norm": 20.07386513124459,
"learning_rate": 1.699012447007882e-06,
"loss": 0.2891,
"step": 51
},
{
"epoch": 0.6715092816787732,
"grad_norm": 16.654008916412547,
"learning_rate": 1.682553143218654e-06,
"loss": 0.2724,
"step": 52
},
{
"epoch": 0.6844229217110573,
"grad_norm": 20.770464397707823,
"learning_rate": 1.6657401196974403e-06,
"loss": 0.332,
"step": 53
},
{
"epoch": 0.6973365617433414,
"grad_norm": 22.094352414249297,
"learning_rate": 1.648582089462756e-06,
"loss": 0.344,
"step": 54
},
{
"epoch": 0.7102502017756255,
"grad_norm": 20.494531642603356,
"learning_rate": 1.6310879443260529e-06,
"loss": 0.3041,
"step": 55
},
{
"epoch": 0.7231638418079096,
"grad_norm": 19.8177079163147,
"learning_rate": 1.6132667502837162e-06,
"loss": 0.3162,
"step": 56
},
{
"epoch": 0.7360774818401937,
"grad_norm": 19.049318629916456,
"learning_rate": 1.5951277428187897e-06,
"loss": 0.3094,
"step": 57
},
{
"epoch": 0.7489911218724778,
"grad_norm": 21.597536610054288,
"learning_rate": 1.5766803221148673e-06,
"loss": 0.3325,
"step": 58
},
{
"epoch": 0.7619047619047619,
"grad_norm": 20.70892358284607,
"learning_rate": 1.5579340481846335e-06,
"loss": 0.3397,
"step": 59
},
{
"epoch": 0.774818401937046,
"grad_norm": 20.308915238405316,
"learning_rate": 1.5388986359155755e-06,
"loss": 0.3265,
"step": 60
},
{
"epoch": 0.7877320419693301,
"grad_norm": 22.627261985685145,
"learning_rate": 1.5195839500354335e-06,
"loss": 0.329,
"step": 61
},
{
"epoch": 0.8006456820016142,
"grad_norm": 19.565397419582293,
"learning_rate": 1.5e-06,
"loss": 0.3167,
"step": 62
},
{
"epoch": 0.8135593220338984,
"grad_norm": 18.907828874624407,
"learning_rate": 1.4801569348059155e-06,
"loss": 0.3146,
"step": 63
},
{
"epoch": 0.8264729620661824,
"grad_norm": 22.644125199961817,
"learning_rate": 1.460065037731152e-06,
"loss": 0.3072,
"step": 64
},
{
"epoch": 0.8393866020984665,
"grad_norm": 20.786645375646636,
"learning_rate": 1.4397347210059057e-06,
"loss": 0.3159,
"step": 65
},
{
"epoch": 0.8523002421307506,
"grad_norm": 18.57141588916772,
"learning_rate": 1.419176520416664e-06,
"loss": 0.3103,
"step": 66
},
{
"epoch": 0.8652138821630347,
"grad_norm": 17.225037473257153,
"learning_rate": 1.3984010898462415e-06,
"loss": 0.3027,
"step": 67
},
{
"epoch": 0.8781275221953188,
"grad_norm": 18.21584058115977,
"learning_rate": 1.3774191957526142e-06,
"loss": 0.2895,
"step": 68
},
{
"epoch": 0.8910411622276029,
"grad_norm": 20.509880420628917,
"learning_rate": 1.3562417115894168e-06,
"loss": 0.3098,
"step": 69
},
{
"epoch": 0.903954802259887,
"grad_norm": 17.29318336230985,
"learning_rate": 1.334879612170986e-06,
"loss": 0.2954,
"step": 70
},
{
"epoch": 0.9168684422921711,
"grad_norm": 18.377551002595997,
"learning_rate": 1.3133439679848822e-06,
"loss": 0.301,
"step": 71
},
{
"epoch": 0.9297820823244553,
"grad_norm": 26.11617966128903,
"learning_rate": 1.2916459394548249e-06,
"loss": 0.3493,
"step": 72
},
{
"epoch": 0.9426957223567393,
"grad_norm": 19.11873065402881,
"learning_rate": 1.2697967711570242e-06,
"loss": 0.3055,
"step": 73
},
{
"epoch": 0.9556093623890234,
"grad_norm": 20.15187293059899,
"learning_rate": 1.2478077859929e-06,
"loss": 0.3267,
"step": 74
},
{
"epoch": 0.9685230024213075,
"grad_norm": 17.356504787996283,
"learning_rate": 1.2256903793212105e-06,
"loss": 0.3,
"step": 75
},
{
"epoch": 0.9814366424535916,
"grad_norm": 20.116822076601576,
"learning_rate": 1.203456013052634e-06,
"loss": 0.3139,
"step": 76
},
{
"epoch": 0.9943502824858758,
"grad_norm": 20.23077529953701,
"learning_rate": 1.1811162097098558e-06,
"loss": 0.2941,
"step": 77
},
{
"epoch": 1.0072639225181599,
"grad_norm": 14.472243001169707,
"learning_rate": 1.1586825464562514e-06,
"loss": 0.217,
"step": 78
},
{
"epoch": 1.020177562550444,
"grad_norm": 11.1760098722801,
"learning_rate": 1.1361666490962467e-06,
"loss": 0.1391,
"step": 79
},
{
"epoch": 1.033091202582728,
"grad_norm": 12.929875937980091,
"learning_rate": 1.1135801860504748e-06,
"loss": 0.144,
"step": 80
},
{
"epoch": 1.0460048426150121,
"grad_norm": 14.401479453367932,
"learning_rate": 1.0909348623088471e-06,
"loss": 0.1525,
"step": 81
},
{
"epoch": 1.0589184826472962,
"grad_norm": 11.466941187008539,
"learning_rate": 1.068242413364671e-06,
"loss": 0.1492,
"step": 82
},
{
"epoch": 1.0718321226795804,
"grad_norm": 13.844595316123245,
"learning_rate": 1.0455145991329637e-06,
"loss": 0.1317,
"step": 83
},
{
"epoch": 1.0847457627118644,
"grad_norm": 14.390882441536627,
"learning_rate": 1.0227631978561055e-06,
"loss": 0.1574,
"step": 84
},
{
"epoch": 1.0976594027441484,
"grad_norm": 13.517598225715695,
"learning_rate": 1e-06,
"loss": 0.1409,
"step": 85
},
{
"epoch": 1.1105730427764327,
"grad_norm": 13.993115366941224,
"learning_rate": 9.772368021438942e-07,
"loss": 0.1464,
"step": 86
},
{
"epoch": 1.1234866828087167,
"grad_norm": 12.2423524228492,
"learning_rate": 9.544854008670366e-07,
"loss": 0.1207,
"step": 87
},
{
"epoch": 1.136400322841001,
"grad_norm": 14.122285998417135,
"learning_rate": 9.317575866353291e-07,
"loss": 0.1251,
"step": 88
},
{
"epoch": 1.149313962873285,
"grad_norm": 15.560999029473845,
"learning_rate": 9.09065137691153e-07,
"loss": 0.1209,
"step": 89
},
{
"epoch": 1.162227602905569,
"grad_norm": 13.507106399175507,
"learning_rate": 8.86419813949525e-07,
"loss": 0.1091,
"step": 90
},
{
"epoch": 1.1751412429378532,
"grad_norm": 13.891730510521498,
"learning_rate": 8.638333509037535e-07,
"loss": 0.1256,
"step": 91
},
{
"epoch": 1.1880548829701372,
"grad_norm": 14.992783015617869,
"learning_rate": 8.413174535437487e-07,
"loss": 0.137,
"step": 92
},
{
"epoch": 1.2009685230024214,
"grad_norm": 15.910575644376518,
"learning_rate": 8.188837902901441e-07,
"loss": 0.1347,
"step": 93
},
{
"epoch": 1.2138821630347054,
"grad_norm": 19.326658299865034,
"learning_rate": 7.965439869473663e-07,
"loss": 0.1473,
"step": 94
},
{
"epoch": 1.2267958030669894,
"grad_norm": 21.139159288452323,
"learning_rate": 7.743096206787893e-07,
"loss": 0.1449,
"step": 95
},
{
"epoch": 1.2397094430992737,
"grad_norm": 19.57794361709339,
"learning_rate": 7.521922140071003e-07,
"loss": 0.1429,
"step": 96
},
{
"epoch": 1.2526230831315577,
"grad_norm": 18.770894293602584,
"learning_rate": 7.302032288429756e-07,
"loss": 0.1533,
"step": 97
},
{
"epoch": 1.2655367231638417,
"grad_norm": 19.148357746011182,
"learning_rate": 7.083540605451749e-07,
"loss": 0.164,
"step": 98
},
{
"epoch": 1.278450363196126,
"grad_norm": 15.077187564406879,
"learning_rate": 6.866560320151178e-07,
"loss": 0.1404,
"step": 99
},
{
"epoch": 1.29136400322841,
"grad_norm": 16.058568106787366,
"learning_rate": 6.651203878290138e-07,
"loss": 0.1327,
"step": 100
},
{
"epoch": 1.3042776432606942,
"grad_norm": 14.029168964991912,
"learning_rate": 6.437582884105834e-07,
"loss": 0.136,
"step": 101
},
{
"epoch": 1.3171912832929782,
"grad_norm": 15.10100507251639,
"learning_rate": 6.225808042473857e-07,
"loss": 0.1149,
"step": 102
},
{
"epoch": 1.3301049233252624,
"grad_norm": 11.718976894577146,
"learning_rate": 6.015989101537586e-07,
"loss": 0.1207,
"step": 103
},
{
"epoch": 1.3430185633575464,
"grad_norm": 12.627764504381856,
"learning_rate": 5.808234795833362e-07,
"loss": 0.1081,
"step": 104
},
{
"epoch": 1.3559322033898304,
"grad_norm": 15.041188760891716,
"learning_rate": 5.602652789940941e-07,
"loss": 0.1262,
"step": 105
},
{
"epoch": 1.3688458434221147,
"grad_norm": 21.09158607094679,
"learning_rate": 5.399349622688478e-07,
"loss": 0.1314,
"step": 106
},
{
"epoch": 1.3817594834543987,
"grad_norm": 20.9326453419062,
"learning_rate": 5.198430651940845e-07,
"loss": 0.1346,
"step": 107
},
{
"epoch": 1.3946731234866827,
"grad_norm": 12.206782639953575,
"learning_rate": 5.000000000000002e-07,
"loss": 0.1157,
"step": 108
},
{
"epoch": 1.407586763518967,
"grad_norm": 13.75893208112109,
"learning_rate": 4.804160499645667e-07,
"loss": 0.1384,
"step": 109
},
{
"epoch": 1.420500403551251,
"grad_norm": 13.597298102282235,
"learning_rate": 4.6110136408442445e-07,
"loss": 0.1137,
"step": 110
},
{
"epoch": 1.4334140435835352,
"grad_norm": 13.538763536743675,
"learning_rate": 4.4206595181536664e-07,
"loss": 0.1203,
"step": 111
},
{
"epoch": 1.4463276836158192,
"grad_norm": 12.774053600703025,
"learning_rate": 4.233196778851329e-07,
"loss": 0.1122,
"step": 112
},
{
"epoch": 1.4592413236481034,
"grad_norm": 12.870680344192039,
"learning_rate": 4.0487225718121045e-07,
"loss": 0.1228,
"step": 113
},
{
"epoch": 1.4721549636803875,
"grad_norm": 15.02538486601625,
"learning_rate": 3.867332497162835e-07,
"loss": 0.1615,
"step": 114
},
{
"epoch": 1.4850686037126715,
"grad_norm": 12.30281015942022,
"learning_rate": 3.6891205567394746e-07,
"loss": 0.099,
"step": 115
},
{
"epoch": 1.4979822437449557,
"grad_norm": 10.419235242267765,
"learning_rate": 3.5141791053724404e-07,
"loss": 0.1096,
"step": 116
},
{
"epoch": 1.5108958837772397,
"grad_norm": 11.16767240028713,
"learning_rate": 3.3425988030255945e-07,
"loss": 0.1199,
"step": 117
},
{
"epoch": 1.5238095238095237,
"grad_norm": 13.158987202921047,
"learning_rate": 3.174468567813461e-07,
"loss": 0.1324,
"step": 118
},
{
"epoch": 1.536723163841808,
"grad_norm": 11.844158983050455,
"learning_rate": 3.009875529921181e-07,
"loss": 0.1106,
"step": 119
},
{
"epoch": 1.549636803874092,
"grad_norm": 14.439077228213756,
"learning_rate": 2.848904986451005e-07,
"loss": 0.117,
"step": 120
},
{
"epoch": 1.562550443906376,
"grad_norm": 12.407600386041583,
"learning_rate": 2.6916403572187587e-07,
"loss": 0.1159,
"step": 121
},
{
"epoch": 1.5754640839386602,
"grad_norm": 13.151590889043437,
"learning_rate": 2.538163141523145e-07,
"loss": 0.1236,
"step": 122
},
{
"epoch": 1.5883777239709445,
"grad_norm": 12.342415661391085,
"learning_rate": 2.3885528759103536e-07,
"loss": 0.1191,
"step": 123
},
{
"epoch": 1.6012913640032282,
"grad_norm": 25.46717666484534,
"learning_rate": 2.2428870929558007e-07,
"loss": 0.1464,
"step": 124
},
{
"epoch": 1.6142050040355125,
"grad_norm": 13.835165114851305,
"learning_rate": 2.101241281084416e-07,
"loss": 0.114,
"step": 125
},
{
"epoch": 1.6271186440677967,
"grad_norm": 12.278805533417628,
"learning_rate": 1.9636888454502177e-07,
"loss": 0.1095,
"step": 126
},
{
"epoch": 1.6400322841000807,
"grad_norm": 9.656229159282553,
"learning_rate": 1.83030106989558e-07,
"loss": 0.1048,
"step": 127
},
{
"epoch": 1.6529459241323647,
"grad_norm": 10.723649324385244,
"learning_rate": 1.7011470800097495e-07,
"loss": 0.1144,
"step": 128
},
{
"epoch": 1.665859564164649,
"grad_norm": 11.18317780807273,
"learning_rate": 1.576293807305885e-07,
"loss": 0.1124,
"step": 129
},
{
"epoch": 1.678773204196933,
"grad_norm": 19.22525610881903,
"learning_rate": 1.4558059545351142e-07,
"loss": 0.1185,
"step": 130
},
{
"epoch": 1.691686844229217,
"grad_norm": 13.840544386244854,
"learning_rate": 1.3397459621556128e-07,
"loss": 0.1172,
"step": 131
},
{
"epoch": 1.7046004842615012,
"grad_norm": 9.863106547796383,
"learning_rate": 1.2281739759740574e-07,
"loss": 0.1123,
"step": 132
},
{
"epoch": 1.7175141242937855,
"grad_norm": 12.249146236722565,
"learning_rate": 1.1211478159762478e-07,
"loss": 0.1026,
"step": 133
},
{
"epoch": 1.7304277643260693,
"grad_norm": 19.53823199805843,
"learning_rate": 1.0187229463630398e-07,
"loss": 0.1245,
"step": 134
},
{
"epoch": 1.7433414043583535,
"grad_norm": 13.813268314685585,
"learning_rate": 9.209524468071095e-08,
"loss": 0.1259,
"step": 135
},
{
"epoch": 1.7562550443906377,
"grad_norm": 13.002603579106614,
"learning_rate": 8.278869849454717e-08,
"loss": 0.126,
"step": 136
},
{
"epoch": 1.7691686844229217,
"grad_norm": 11.790731140563368,
"learning_rate": 7.395747901219473e-08,
"loss": 0.1163,
"step": 137
},
{
"epoch": 1.7820823244552058,
"grad_norm": 15.611451697022517,
"learning_rate": 6.560616283932896e-08,
"loss": 0.1226,
"step": 138
},
{
"epoch": 1.79499596448749,
"grad_norm": 24.410099015402505,
"learning_rate": 5.77390778811796e-08,
"loss": 0.1155,
"step": 139
},
{
"epoch": 1.807909604519774,
"grad_norm": 9.829584476780473,
"learning_rate": 5.036030109968081e-08,
"loss": 0.0972,
"step": 140
},
{
"epoch": 1.820823244552058,
"grad_norm": 14.542859998120877,
"learning_rate": 4.347365640066525e-08,
"loss": 0.1126,
"step": 141
},
{
"epoch": 1.8337368845843423,
"grad_norm": 20.231191724041285,
"learning_rate": 3.708271265220087e-08,
"loss": 0.1175,
"step": 142
},
{
"epoch": 1.8466505246166263,
"grad_norm": 12.407632368224395,
"learning_rate": 3.119078183509372e-08,
"loss": 0.1149,
"step": 143
},
{
"epoch": 1.8595641646489103,
"grad_norm": 14.44661290494769,
"learning_rate": 2.580091732652101e-08,
"loss": 0.1181,
"step": 144
},
{
"epoch": 1.8724778046811945,
"grad_norm": 18.79024412476877,
"learning_rate": 2.0915912317677088e-08,
"loss": 0.1475,
"step": 145
},
{
"epoch": 1.8853914447134787,
"grad_norm": 15.376557680187412,
"learning_rate": 1.6538298366257974e-08,
"loss": 0.1255,
"step": 146
},
{
"epoch": 1.8983050847457628,
"grad_norm": 15.33944835023639,
"learning_rate": 1.2670344084530382e-08,
"loss": 0.1034,
"step": 147
},
{
"epoch": 1.9112187247780468,
"grad_norm": 13.112291504434255,
"learning_rate": 9.314053963669244e-09,
"loss": 0.1086,
"step": 148
},
{
"epoch": 1.924132364810331,
"grad_norm": 12.045688670027003,
"learning_rate": 6.471167334968886e-09,
"loss": 0.1096,
"step": 149
},
{
"epoch": 1.937046004842615,
"grad_norm": 15.127302877916637,
"learning_rate": 4.143157468468716e-09,
"loss": 0.1126,
"step": 150
},
{
"epoch": 1.949959644874899,
"grad_norm": 13.774563816455347,
"learning_rate": 2.331230809460738e-09,
"loss": 0.1245,
"step": 151
},
{
"epoch": 1.9628732849071833,
"grad_norm": 32.82358453731887,
"learning_rate": 1.036326353272443e-09,
"loss": 0.1181,
"step": 152
},
{
"epoch": 1.9757869249394673,
"grad_norm": 14.92150992963698,
"learning_rate": 2.5911515865084663e-10,
"loss": 0.1143,
"step": 153
},
{
"epoch": 1.9887005649717513,
"grad_norm": 13.25981887405618,
"learning_rate": 0.0,
"loss": 0.1276,
"step": 154
},
{
"epoch": 1.9887005649717513,
"step": 154,
"total_flos": 72055202119680.0,
"train_loss": 0.2331618887747263,
"train_runtime": 3106.2852,
"train_samples_per_second": 12.759,
"train_steps_per_second": 0.05
}
],
"logging_steps": 1,
"max_steps": 154,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 72055202119680.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}