openthoughts_113957 / trainer_state.json
ryanmarten's picture
Upload folder using huggingface_hub
7694697 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.9788972089857046,
"eval_steps": 500,
"global_step": 273,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.010891763104152484,
"grad_norm": 5.917168704218648,
"learning_rate": 2.8571428571428573e-06,
"loss": 0.8738,
"step": 1
},
{
"epoch": 0.021783526208304968,
"grad_norm": 5.848189651981173,
"learning_rate": 5.7142857142857145e-06,
"loss": 0.8637,
"step": 2
},
{
"epoch": 0.03267528931245745,
"grad_norm": 5.4410485869996865,
"learning_rate": 8.571428571428571e-06,
"loss": 0.8524,
"step": 3
},
{
"epoch": 0.043567052416609936,
"grad_norm": 2.423684029964203,
"learning_rate": 1.1428571428571429e-05,
"loss": 0.783,
"step": 4
},
{
"epoch": 0.05445881552076242,
"grad_norm": 4.0235503087845155,
"learning_rate": 1.4285714285714287e-05,
"loss": 0.7805,
"step": 5
},
{
"epoch": 0.0653505786249149,
"grad_norm": 4.14511666016107,
"learning_rate": 1.7142857142857142e-05,
"loss": 0.7643,
"step": 6
},
{
"epoch": 0.0762423417290674,
"grad_norm": 4.42039479030743,
"learning_rate": 2e-05,
"loss": 0.7512,
"step": 7
},
{
"epoch": 0.08713410483321987,
"grad_norm": 2.857807259160155,
"learning_rate": 2.2857142857142858e-05,
"loss": 0.7182,
"step": 8
},
{
"epoch": 0.09802586793737236,
"grad_norm": 2.690272592093101,
"learning_rate": 2.5714285714285718e-05,
"loss": 0.6835,
"step": 9
},
{
"epoch": 0.10891763104152484,
"grad_norm": 2.0185553871545165,
"learning_rate": 2.8571428571428574e-05,
"loss": 0.666,
"step": 10
},
{
"epoch": 0.11980939414567733,
"grad_norm": 1.6526659811622688,
"learning_rate": 3.142857142857143e-05,
"loss": 0.6551,
"step": 11
},
{
"epoch": 0.1307011572498298,
"grad_norm": 1.3684051055691528,
"learning_rate": 3.4285714285714284e-05,
"loss": 0.6374,
"step": 12
},
{
"epoch": 0.1415929203539823,
"grad_norm": 1.3936888061646935,
"learning_rate": 3.714285714285715e-05,
"loss": 0.6142,
"step": 13
},
{
"epoch": 0.1524846834581348,
"grad_norm": 1.3347032728939032,
"learning_rate": 4e-05,
"loss": 0.6085,
"step": 14
},
{
"epoch": 0.16337644656228728,
"grad_norm": 1.6041646981350726,
"learning_rate": 4.2857142857142856e-05,
"loss": 0.6168,
"step": 15
},
{
"epoch": 0.17426820966643974,
"grad_norm": 1.250156217488812,
"learning_rate": 4.5714285714285716e-05,
"loss": 0.5904,
"step": 16
},
{
"epoch": 0.18515997277059223,
"grad_norm": 1.363684815879282,
"learning_rate": 4.857142857142857e-05,
"loss": 0.5983,
"step": 17
},
{
"epoch": 0.19605173587474473,
"grad_norm": 1.160866980807094,
"learning_rate": 5.1428571428571436e-05,
"loss": 0.5762,
"step": 18
},
{
"epoch": 0.20694349897889722,
"grad_norm": 1.8320088500371636,
"learning_rate": 5.4285714285714295e-05,
"loss": 0.5794,
"step": 19
},
{
"epoch": 0.21783526208304968,
"grad_norm": 1.310331415981151,
"learning_rate": 5.714285714285715e-05,
"loss": 0.5723,
"step": 20
},
{
"epoch": 0.22872702518720217,
"grad_norm": 2.0403045692700275,
"learning_rate": 6.000000000000001e-05,
"loss": 0.576,
"step": 21
},
{
"epoch": 0.23961878829135466,
"grad_norm": 1.0753217952290004,
"learning_rate": 6.285714285714286e-05,
"loss": 0.5657,
"step": 22
},
{
"epoch": 0.25051055139550715,
"grad_norm": 2.0101370583197076,
"learning_rate": 6.571428571428571e-05,
"loss": 0.5751,
"step": 23
},
{
"epoch": 0.2614023144996596,
"grad_norm": 1.8269660980505025,
"learning_rate": 6.857142857142857e-05,
"loss": 0.5622,
"step": 24
},
{
"epoch": 0.27229407760381213,
"grad_norm": 1.333190530233097,
"learning_rate": 7.142857142857143e-05,
"loss": 0.553,
"step": 25
},
{
"epoch": 0.2831858407079646,
"grad_norm": 1.5278657539191043,
"learning_rate": 7.42857142857143e-05,
"loss": 0.5746,
"step": 26
},
{
"epoch": 0.29407760381211706,
"grad_norm": 1.4753801604023495,
"learning_rate": 7.714285714285715e-05,
"loss": 0.5504,
"step": 27
},
{
"epoch": 0.3049693669162696,
"grad_norm": 1.4736990838732313,
"learning_rate": 8e-05,
"loss": 0.5542,
"step": 28
},
{
"epoch": 0.31586113002042204,
"grad_norm": 2.3486277040401893,
"learning_rate": 7.999671154713278e-05,
"loss": 0.5507,
"step": 29
},
{
"epoch": 0.32675289312457456,
"grad_norm": 1.1844138236301467,
"learning_rate": 7.99868467292272e-05,
"loss": 0.5496,
"step": 30
},
{
"epoch": 0.337644656228727,
"grad_norm": 2.4920955827056748,
"learning_rate": 7.997040716828271e-05,
"loss": 0.5575,
"step": 31
},
{
"epoch": 0.3485364193328795,
"grad_norm": 1.7099423006847476,
"learning_rate": 7.994739556733538e-05,
"loss": 0.5596,
"step": 32
},
{
"epoch": 0.359428182437032,
"grad_norm": 2.1342906902090504,
"learning_rate": 7.991781571001347e-05,
"loss": 0.5337,
"step": 33
},
{
"epoch": 0.37031994554118447,
"grad_norm": 1.7483277996825424,
"learning_rate": 7.988167245991528e-05,
"loss": 0.5492,
"step": 34
},
{
"epoch": 0.381211708645337,
"grad_norm": 1.7968259576328387,
"learning_rate": 7.983897175980957e-05,
"loss": 0.5314,
"step": 35
},
{
"epoch": 0.39210347174948945,
"grad_norm": 1.559300515616932,
"learning_rate": 7.97897206306583e-05,
"loss": 0.5307,
"step": 36
},
{
"epoch": 0.4029952348536419,
"grad_norm": 1.0889578452931736,
"learning_rate": 7.973392717046233e-05,
"loss": 0.5408,
"step": 37
},
{
"epoch": 0.41388699795779443,
"grad_norm": 1.313871076023579,
"learning_rate": 7.967160055292984e-05,
"loss": 0.5147,
"step": 38
},
{
"epoch": 0.4247787610619469,
"grad_norm": 1.583430729115433,
"learning_rate": 7.960275102596809e-05,
"loss": 0.5216,
"step": 39
},
{
"epoch": 0.43567052416609936,
"grad_norm": 1.3933127292778362,
"learning_rate": 7.952738990999824e-05,
"loss": 0.5187,
"step": 40
},
{
"epoch": 0.4465622872702519,
"grad_norm": 1.4991075926399846,
"learning_rate": 7.94455295960942e-05,
"loss": 0.5123,
"step": 41
},
{
"epoch": 0.45745405037440434,
"grad_norm": 1.2790673350427315,
"learning_rate": 7.93571835439452e-05,
"loss": 0.5255,
"step": 42
},
{
"epoch": 0.46834581347855686,
"grad_norm": 1.656174041977428,
"learning_rate": 7.926236627964262e-05,
"loss": 0.5224,
"step": 43
},
{
"epoch": 0.4792375765827093,
"grad_norm": 1.3680863968048569,
"learning_rate": 7.916109339329173e-05,
"loss": 0.5104,
"step": 44
},
{
"epoch": 0.4901293396868618,
"grad_norm": 1.1248700116366424,
"learning_rate": 7.905338153644818e-05,
"loss": 0.5036,
"step": 45
},
{
"epoch": 0.5010211027910143,
"grad_norm": 1.4820737812996578,
"learning_rate": 7.89392484193802e-05,
"loss": 0.5169,
"step": 46
},
{
"epoch": 0.5119128658951668,
"grad_norm": 1.1533791684607533,
"learning_rate": 7.881871280815659e-05,
"loss": 0.5071,
"step": 47
},
{
"epoch": 0.5228046289993192,
"grad_norm": 1.9492363638377506,
"learning_rate": 7.869179452156118e-05,
"loss": 0.5093,
"step": 48
},
{
"epoch": 0.5336963921034718,
"grad_norm": 0.9617512636578286,
"learning_rate": 7.855851442783414e-05,
"loss": 0.501,
"step": 49
},
{
"epoch": 0.5445881552076243,
"grad_norm": 1.0817397757606,
"learning_rate": 7.841889444124078e-05,
"loss": 0.4997,
"step": 50
},
{
"epoch": 0.5554799183117767,
"grad_norm": 2.1099922597533074,
"learning_rate": 7.827295751846836e-05,
"loss": 0.5151,
"step": 51
},
{
"epoch": 0.5663716814159292,
"grad_norm": 1.0752679988230807,
"learning_rate": 7.81207276548515e-05,
"loss": 0.512,
"step": 52
},
{
"epoch": 0.5772634445200817,
"grad_norm": 2.033228507248604,
"learning_rate": 7.796222988042676e-05,
"loss": 0.5247,
"step": 53
},
{
"epoch": 0.5881552076242341,
"grad_norm": 1.4290947325043002,
"learning_rate": 7.779749025581717e-05,
"loss": 0.5161,
"step": 54
},
{
"epoch": 0.5990469707283866,
"grad_norm": 1.6107474254316703,
"learning_rate": 7.762653586794731e-05,
"loss": 0.5113,
"step": 55
},
{
"epoch": 0.6099387338325392,
"grad_norm": 1.7403793303180215,
"learning_rate": 7.74493948255895e-05,
"loss": 0.5104,
"step": 56
},
{
"epoch": 0.6208304969366917,
"grad_norm": 1.3273260664695445,
"learning_rate": 7.726609625474218e-05,
"loss": 0.5149,
"step": 57
},
{
"epoch": 0.6317222600408441,
"grad_norm": 1.2835468728323596,
"learning_rate": 7.707667029384088e-05,
"loss": 0.4985,
"step": 58
},
{
"epoch": 0.6426140231449966,
"grad_norm": 1.3493586560972042,
"learning_rate": 7.688114808880283e-05,
"loss": 0.515,
"step": 59
},
{
"epoch": 0.6535057862491491,
"grad_norm": 0.8317273919287447,
"learning_rate": 7.667956178790582e-05,
"loss": 0.5059,
"step": 60
},
{
"epoch": 0.6643975493533015,
"grad_norm": 1.1983241891992835,
"learning_rate": 7.647194453650228e-05,
"loss": 0.4935,
"step": 61
},
{
"epoch": 0.675289312457454,
"grad_norm": 1.1964694002526848,
"learning_rate": 7.625833047156953e-05,
"loss": 0.5058,
"step": 62
},
{
"epoch": 0.6861810755616066,
"grad_norm": 1.1878899519661879,
"learning_rate": 7.603875471609677e-05,
"loss": 0.4903,
"step": 63
},
{
"epoch": 0.697072838665759,
"grad_norm": 1.2583807783190593,
"learning_rate": 7.581325337331013e-05,
"loss": 0.497,
"step": 64
},
{
"epoch": 0.7079646017699115,
"grad_norm": 0.9383067837973975,
"learning_rate": 7.558186352073648e-05,
"loss": 0.4874,
"step": 65
},
{
"epoch": 0.718856364874064,
"grad_norm": 1.114271813149658,
"learning_rate": 7.534462320410702e-05,
"loss": 0.4957,
"step": 66
},
{
"epoch": 0.7297481279782164,
"grad_norm": 0.9889746448795311,
"learning_rate": 7.510157143110172e-05,
"loss": 0.4976,
"step": 67
},
{
"epoch": 0.7406398910823689,
"grad_norm": 1.5156805885931242,
"learning_rate": 7.485274816493558e-05,
"loss": 0.4959,
"step": 68
},
{
"epoch": 0.7515316541865215,
"grad_norm": 0.9749471928652824,
"learning_rate": 7.459819431778775e-05,
"loss": 0.4796,
"step": 69
},
{
"epoch": 0.762423417290674,
"grad_norm": 0.8404206560752534,
"learning_rate": 7.433795174407465e-05,
"loss": 0.4762,
"step": 70
},
{
"epoch": 0.7733151803948264,
"grad_norm": 1.2857273953478379,
"learning_rate": 7.407206323356818e-05,
"loss": 0.4955,
"step": 71
},
{
"epoch": 0.7842069434989789,
"grad_norm": 1.5254696719564749,
"learning_rate": 7.380057250436006e-05,
"loss": 0.4945,
"step": 72
},
{
"epoch": 0.7950987066031314,
"grad_norm": 1.0225507119621404,
"learning_rate": 7.352352419567362e-05,
"loss": 0.4883,
"step": 73
},
{
"epoch": 0.8059904697072838,
"grad_norm": 1.3824406570348136,
"learning_rate": 7.324096386052416e-05,
"loss": 0.4906,
"step": 74
},
{
"epoch": 0.8168822328114363,
"grad_norm": 0.9568602258759329,
"learning_rate": 7.295293795822887e-05,
"loss": 0.4838,
"step": 75
},
{
"epoch": 0.8277739959155889,
"grad_norm": 1.2765315330970022,
"learning_rate": 7.265949384676795e-05,
"loss": 0.4815,
"step": 76
},
{
"epoch": 0.8386657590197413,
"grad_norm": 1.2279324061842216,
"learning_rate": 7.236067977499791e-05,
"loss": 0.4855,
"step": 77
},
{
"epoch": 0.8495575221238938,
"grad_norm": 0.9629770077059606,
"learning_rate": 7.205654487471826e-05,
"loss": 0.4811,
"step": 78
},
{
"epoch": 0.8604492852280463,
"grad_norm": 1.295116932433786,
"learning_rate": 7.174713915259331e-05,
"loss": 0.4889,
"step": 79
},
{
"epoch": 0.8713410483321987,
"grad_norm": 0.9924719057218118,
"learning_rate": 7.143251348192971e-05,
"loss": 0.488,
"step": 80
},
{
"epoch": 0.8822328114363512,
"grad_norm": 1.3773423039136135,
"learning_rate": 7.111271959431189e-05,
"loss": 0.4859,
"step": 81
},
{
"epoch": 0.8931245745405038,
"grad_norm": 0.9416135398394887,
"learning_rate": 7.078781007109625e-05,
"loss": 0.4803,
"step": 82
},
{
"epoch": 0.9040163376446563,
"grad_norm": 1.0674313281470116,
"learning_rate": 7.045783833476538e-05,
"loss": 0.4733,
"step": 83
},
{
"epoch": 0.9149081007488087,
"grad_norm": 1.000464101793752,
"learning_rate": 7.012285864014445e-05,
"loss": 0.4726,
"step": 84
},
{
"epoch": 0.9257998638529612,
"grad_norm": 1.3944228987734595,
"learning_rate": 6.978292606548029e-05,
"loss": 0.4817,
"step": 85
},
{
"epoch": 0.9366916269571137,
"grad_norm": 0.9210477820477416,
"learning_rate": 6.943809650338541e-05,
"loss": 0.4856,
"step": 86
},
{
"epoch": 0.9475833900612661,
"grad_norm": 1.4702724614133884,
"learning_rate": 6.908842665164789e-05,
"loss": 0.4893,
"step": 87
},
{
"epoch": 0.9584751531654186,
"grad_norm": 0.9392028605617796,
"learning_rate": 6.873397400390911e-05,
"loss": 0.4749,
"step": 88
},
{
"epoch": 0.9693669162695712,
"grad_norm": 1.33752177947439,
"learning_rate": 6.837479684021032e-05,
"loss": 0.4642,
"step": 89
},
{
"epoch": 0.9802586793737236,
"grad_norm": 1.0674472905397991,
"learning_rate": 6.80109542174102e-05,
"loss": 0.4721,
"step": 90
},
{
"epoch": 0.9911504424778761,
"grad_norm": 0.8855006430820443,
"learning_rate": 6.76425059594746e-05,
"loss": 0.471,
"step": 91
},
{
"epoch": 1.0047651463580667,
"grad_norm": 0.7477549645216865,
"learning_rate": 6.726951264763998e-05,
"loss": 0.4723,
"step": 92
},
{
"epoch": 1.0156569094622192,
"grad_norm": 0.8956363744921413,
"learning_rate": 6.689203561045268e-05,
"loss": 0.4562,
"step": 93
},
{
"epoch": 1.0265486725663717,
"grad_norm": 1.060602408653099,
"learning_rate": 6.651013691368492e-05,
"loss": 0.4599,
"step": 94
},
{
"epoch": 1.0374404356705242,
"grad_norm": 1.2755687426821307,
"learning_rate": 6.612387935012995e-05,
"loss": 0.4586,
"step": 95
},
{
"epoch": 1.0483321987746765,
"grad_norm": 0.7866789226297869,
"learning_rate": 6.573332642927737e-05,
"loss": 0.4564,
"step": 96
},
{
"epoch": 1.059223961878829,
"grad_norm": 1.0671580779917749,
"learning_rate": 6.53385423668708e-05,
"loss": 0.4532,
"step": 97
},
{
"epoch": 1.0701157249829816,
"grad_norm": 1.8705582630594135,
"learning_rate": 6.493959207434934e-05,
"loss": 0.4518,
"step": 98
},
{
"epoch": 1.081007488087134,
"grad_norm": 0.8747747648963227,
"learning_rate": 6.453654114817467e-05,
"loss": 0.4602,
"step": 99
},
{
"epoch": 1.0918992511912866,
"grad_norm": 2.6708530213907915,
"learning_rate": 6.412945585904545e-05,
"loss": 0.4706,
"step": 100
},
{
"epoch": 1.1027910142954391,
"grad_norm": 2.059630403847963,
"learning_rate": 6.371840314100104e-05,
"loss": 0.4746,
"step": 101
},
{
"epoch": 1.1136827773995917,
"grad_norm": 2.100904334906369,
"learning_rate": 6.330345058041585e-05,
"loss": 0.4645,
"step": 102
},
{
"epoch": 1.124574540503744,
"grad_norm": 1.9589727338791063,
"learning_rate": 6.288466640488679e-05,
"loss": 0.4574,
"step": 103
},
{
"epoch": 1.1354663036078965,
"grad_norm": 1.250427733328813,
"learning_rate": 6.2462119472015e-05,
"loss": 0.453,
"step": 104
},
{
"epoch": 1.146358066712049,
"grad_norm": 1.244933711184395,
"learning_rate": 6.20358792580841e-05,
"loss": 0.4562,
"step": 105
},
{
"epoch": 1.1572498298162015,
"grad_norm": 0.978598635089721,
"learning_rate": 6.160601584663681e-05,
"loss": 0.4544,
"step": 106
},
{
"epoch": 1.168141592920354,
"grad_norm": 1.0133074820438104,
"learning_rate": 6.11725999169515e-05,
"loss": 0.4567,
"step": 107
},
{
"epoch": 1.1790333560245065,
"grad_norm": 0.9079803477398684,
"learning_rate": 6.0735702732421015e-05,
"loss": 0.4537,
"step": 108
},
{
"epoch": 1.189925119128659,
"grad_norm": 0.47916750157430577,
"learning_rate": 6.029539612883529e-05,
"loss": 0.4514,
"step": 109
},
{
"epoch": 1.2008168822328114,
"grad_norm": 0.737478359754882,
"learning_rate": 5.9851752502570015e-05,
"loss": 0.4462,
"step": 110
},
{
"epoch": 1.2117086453369639,
"grad_norm": 0.5899481616784151,
"learning_rate": 5.940484479868288e-05,
"loss": 0.4456,
"step": 111
},
{
"epoch": 1.2226004084411164,
"grad_norm": 0.766606804098348,
"learning_rate": 5.895474649891995e-05,
"loss": 0.4441,
"step": 112
},
{
"epoch": 1.233492171545269,
"grad_norm": 0.6645207203467242,
"learning_rate": 5.8501531609633424e-05,
"loss": 0.4463,
"step": 113
},
{
"epoch": 1.2443839346494214,
"grad_norm": 0.681340548554224,
"learning_rate": 5.8045274649613386e-05,
"loss": 0.4424,
"step": 114
},
{
"epoch": 1.2552756977535737,
"grad_norm": 0.7119900691253739,
"learning_rate": 5.7586050637835295e-05,
"loss": 0.4459,
"step": 115
},
{
"epoch": 1.2661674608577265,
"grad_norm": 0.49479084006465096,
"learning_rate": 5.7123935081125034e-05,
"loss": 0.4408,
"step": 116
},
{
"epoch": 1.2770592239618788,
"grad_norm": 0.7116517379970301,
"learning_rate": 5.6659003961743965e-05,
"loss": 0.4455,
"step": 117
},
{
"epoch": 1.2879509870660313,
"grad_norm": 0.6527860650845239,
"learning_rate": 5.619133372489575e-05,
"loss": 0.4362,
"step": 118
},
{
"epoch": 1.2988427501701838,
"grad_norm": 0.5221316831765591,
"learning_rate": 5.572100126615695e-05,
"loss": 0.4483,
"step": 119
},
{
"epoch": 1.3097345132743363,
"grad_norm": 0.36665207601249333,
"learning_rate": 5.524808391883367e-05,
"loss": 0.4488,
"step": 120
},
{
"epoch": 1.3206262763784888,
"grad_norm": 0.41335991497113145,
"learning_rate": 5.477265944124626e-05,
"loss": 0.4466,
"step": 121
},
{
"epoch": 1.3315180394826411,
"grad_norm": 0.4288901779920896,
"learning_rate": 5.429480600394405e-05,
"loss": 0.4455,
"step": 122
},
{
"epoch": 1.3424098025867937,
"grad_norm": 0.4476404528186165,
"learning_rate": 5.381460217685231e-05,
"loss": 0.4409,
"step": 123
},
{
"epoch": 1.3533015656909462,
"grad_norm": 0.3975466345144197,
"learning_rate": 5.333212691635368e-05,
"loss": 0.4368,
"step": 124
},
{
"epoch": 1.3641933287950987,
"grad_norm": 0.3833210592542072,
"learning_rate": 5.2847459552305834e-05,
"loss": 0.4414,
"step": 125
},
{
"epoch": 1.3750850918992512,
"grad_norm": 0.38028400797503575,
"learning_rate": 5.23606797749979e-05,
"loss": 0.4342,
"step": 126
},
{
"epoch": 1.3859768550034037,
"grad_norm": 0.29336709544411704,
"learning_rate": 5.1871867622047624e-05,
"loss": 0.4349,
"step": 127
},
{
"epoch": 1.3968686181075562,
"grad_norm": 0.4021055378411946,
"learning_rate": 5.13811034652413e-05,
"loss": 0.4359,
"step": 128
},
{
"epoch": 1.4077603812117085,
"grad_norm": 0.36564910628203584,
"learning_rate": 5.088846799731885e-05,
"loss": 0.4449,
"step": 129
},
{
"epoch": 1.418652144315861,
"grad_norm": 0.3687189406492797,
"learning_rate": 5.039404221870612e-05,
"loss": 0.4384,
"step": 130
},
{
"epoch": 1.4295439074200136,
"grad_norm": 0.29150702350417407,
"learning_rate": 4.989790742419658e-05,
"loss": 0.4391,
"step": 131
},
{
"epoch": 1.440435670524166,
"grad_norm": 0.28941702462294044,
"learning_rate": 4.940014518958461e-05,
"loss": 0.435,
"step": 132
},
{
"epoch": 1.4513274336283186,
"grad_norm": 0.3046809035055879,
"learning_rate": 4.890083735825258e-05,
"loss": 0.4423,
"step": 133
},
{
"epoch": 1.4622191967324711,
"grad_norm": 0.3901636771063769,
"learning_rate": 4.8400066027713974e-05,
"loss": 0.4373,
"step": 134
},
{
"epoch": 1.4731109598366237,
"grad_norm": 0.2975253048975319,
"learning_rate": 4.789791353611469e-05,
"loss": 0.441,
"step": 135
},
{
"epoch": 1.484002722940776,
"grad_norm": 0.2861058200191958,
"learning_rate": 4.7394462448694756e-05,
"loss": 0.4392,
"step": 136
},
{
"epoch": 1.4948944860449285,
"grad_norm": 0.30291271659415053,
"learning_rate": 4.688979554421276e-05,
"loss": 0.4437,
"step": 137
},
{
"epoch": 1.505786249149081,
"grad_norm": 0.27992103529188517,
"learning_rate": 4.6383995801335176e-05,
"loss": 0.4448,
"step": 138
},
{
"epoch": 1.5166780122532335,
"grad_norm": 0.25747738933655456,
"learning_rate": 4.5877146384992725e-05,
"loss": 0.4431,
"step": 139
},
{
"epoch": 1.527569775357386,
"grad_norm": 0.341071325514537,
"learning_rate": 4.5369330632706223e-05,
"loss": 0.4331,
"step": 140
},
{
"epoch": 1.5384615384615383,
"grad_norm": 0.2602954860252923,
"learning_rate": 4.486063204088402e-05,
"loss": 0.4433,
"step": 141
},
{
"epoch": 1.549353301565691,
"grad_norm": 0.23767705409963544,
"learning_rate": 4.435113425109324e-05,
"loss": 0.4337,
"step": 142
},
{
"epoch": 1.5602450646698434,
"grad_norm": 0.29322114556856593,
"learning_rate": 4.3840921036307274e-05,
"loss": 0.4387,
"step": 143
},
{
"epoch": 1.5711368277739959,
"grad_norm": 0.2171415398561475,
"learning_rate": 4.333007628713158e-05,
"loss": 0.4333,
"step": 144
},
{
"epoch": 1.5820285908781484,
"grad_norm": 0.26753092128368694,
"learning_rate": 4.281868399801016e-05,
"loss": 0.4375,
"step": 145
},
{
"epoch": 1.592920353982301,
"grad_norm": 0.2577848543389177,
"learning_rate": 4.230682825341498e-05,
"loss": 0.4331,
"step": 146
},
{
"epoch": 1.6038121170864534,
"grad_norm": 0.23037795102518452,
"learning_rate": 4.17945932140206e-05,
"loss": 0.4432,
"step": 147
},
{
"epoch": 1.6147038801906057,
"grad_norm": 0.22432493774064247,
"learning_rate": 4.128206310286622e-05,
"loss": 0.433,
"step": 148
},
{
"epoch": 1.6255956432947585,
"grad_norm": 0.27660950965609943,
"learning_rate": 4.0769322191507485e-05,
"loss": 0.4291,
"step": 149
},
{
"epoch": 1.6364874063989108,
"grad_norm": 0.29639803872823944,
"learning_rate": 4.025645478616045e-05,
"loss": 0.4479,
"step": 150
},
{
"epoch": 1.6473791695030633,
"grad_norm": 0.22551842554518517,
"learning_rate": 3.974354521383956e-05,
"loss": 0.4305,
"step": 151
},
{
"epoch": 1.6582709326072158,
"grad_norm": 0.21855552569143427,
"learning_rate": 3.923067780849252e-05,
"loss": 0.4341,
"step": 152
},
{
"epoch": 1.6691626957113683,
"grad_norm": 0.25345471682479176,
"learning_rate": 3.87179368971338e-05,
"loss": 0.4332,
"step": 153
},
{
"epoch": 1.6800544588155208,
"grad_norm": 0.3294043510791402,
"learning_rate": 3.820540678597942e-05,
"loss": 0.4324,
"step": 154
},
{
"epoch": 1.6909462219196731,
"grad_norm": 0.2915052857817396,
"learning_rate": 3.769317174658503e-05,
"loss": 0.4289,
"step": 155
},
{
"epoch": 1.7018379850238259,
"grad_norm": 0.20123186425930717,
"learning_rate": 3.718131600198984e-05,
"loss": 0.4414,
"step": 156
},
{
"epoch": 1.7127297481279782,
"grad_norm": 0.23019840765145663,
"learning_rate": 3.666992371286843e-05,
"loss": 0.4341,
"step": 157
},
{
"epoch": 1.7236215112321307,
"grad_norm": 0.2075296642104422,
"learning_rate": 3.615907896369273e-05,
"loss": 0.4391,
"step": 158
},
{
"epoch": 1.7345132743362832,
"grad_norm": 0.20499145565748494,
"learning_rate": 3.564886574890677e-05,
"loss": 0.4368,
"step": 159
},
{
"epoch": 1.7454050374404355,
"grad_norm": 0.2270936672913207,
"learning_rate": 3.5139367959115986e-05,
"loss": 0.4362,
"step": 160
},
{
"epoch": 1.7562968005445883,
"grad_norm": 0.20695924414455472,
"learning_rate": 3.4630669367293797e-05,
"loss": 0.4321,
"step": 161
},
{
"epoch": 1.7671885636487406,
"grad_norm": 0.20145634393976244,
"learning_rate": 3.412285361500729e-05,
"loss": 0.4327,
"step": 162
},
{
"epoch": 1.778080326752893,
"grad_norm": 0.1954369301373391,
"learning_rate": 3.3616004198664845e-05,
"loss": 0.4324,
"step": 163
},
{
"epoch": 1.7889720898570456,
"grad_norm": 0.20481400956545093,
"learning_rate": 3.311020445578725e-05,
"loss": 0.4377,
"step": 164
},
{
"epoch": 1.799863852961198,
"grad_norm": 0.2162909320824147,
"learning_rate": 3.260553755130525e-05,
"loss": 0.4299,
"step": 165
},
{
"epoch": 1.8107556160653506,
"grad_norm": 0.1811829339048361,
"learning_rate": 3.210208646388532e-05,
"loss": 0.43,
"step": 166
},
{
"epoch": 1.821647379169503,
"grad_norm": 0.2086574503584923,
"learning_rate": 3.1599933972286026e-05,
"loss": 0.4346,
"step": 167
},
{
"epoch": 1.8325391422736557,
"grad_norm": 0.22526815431719788,
"learning_rate": 3.109916264174743e-05,
"loss": 0.4288,
"step": 168
},
{
"epoch": 1.843430905377808,
"grad_norm": 0.18015782113002662,
"learning_rate": 3.0599854810415393e-05,
"loss": 0.428,
"step": 169
},
{
"epoch": 1.8543226684819605,
"grad_norm": 0.20990114701229817,
"learning_rate": 3.0102092575803435e-05,
"loss": 0.4333,
"step": 170
},
{
"epoch": 1.865214431586113,
"grad_norm": 0.1988483391924283,
"learning_rate": 2.9605957781293893e-05,
"loss": 0.4346,
"step": 171
},
{
"epoch": 1.8761061946902655,
"grad_norm": 0.19078378305907726,
"learning_rate": 2.911153200268116e-05,
"loss": 0.4251,
"step": 172
},
{
"epoch": 1.886997957794418,
"grad_norm": 0.23446180498320415,
"learning_rate": 2.8618896534758707e-05,
"loss": 0.4377,
"step": 173
},
{
"epoch": 1.8978897208985703,
"grad_norm": 0.1761870449513099,
"learning_rate": 2.8128132377952376e-05,
"loss": 0.4227,
"step": 174
},
{
"epoch": 1.908781484002723,
"grad_norm": 0.2332557884998122,
"learning_rate": 2.7639320225002108e-05,
"loss": 0.4325,
"step": 175
},
{
"epoch": 1.9196732471068754,
"grad_norm": 0.2211426717672741,
"learning_rate": 2.715254044769418e-05,
"loss": 0.4308,
"step": 176
},
{
"epoch": 1.9305650102110279,
"grad_norm": 0.18328106510038045,
"learning_rate": 2.666787308364634e-05,
"loss": 0.4322,
"step": 177
},
{
"epoch": 1.9414567733151804,
"grad_norm": 0.2288520554831422,
"learning_rate": 2.6185397823147703e-05,
"loss": 0.4263,
"step": 178
},
{
"epoch": 1.952348536419333,
"grad_norm": 0.17271019835077814,
"learning_rate": 2.5705193996055977e-05,
"loss": 0.4231,
"step": 179
},
{
"epoch": 1.9632402995234854,
"grad_norm": 0.21114296576465388,
"learning_rate": 2.5227340558753755e-05,
"loss": 0.4323,
"step": 180
},
{
"epoch": 1.9741320626276377,
"grad_norm": 0.1708320475165614,
"learning_rate": 2.4751916081166336e-05,
"loss": 0.4318,
"step": 181
},
{
"epoch": 1.9850238257317905,
"grad_norm": 0.20243633225986957,
"learning_rate": 2.427899873384306e-05,
"loss": 0.4284,
"step": 182
},
{
"epoch": 1.9959155888359428,
"grad_norm": 0.17214298347005028,
"learning_rate": 2.3808666275104248e-05,
"loss": 0.4303,
"step": 183
},
{
"epoch": 2.0095302927161334,
"grad_norm": 0.2371948537766053,
"learning_rate": 2.334099603825605e-05,
"loss": 0.4195,
"step": 184
},
{
"epoch": 2.020422055820286,
"grad_norm": 0.20683435017217672,
"learning_rate": 2.2876064918874993e-05,
"loss": 0.4056,
"step": 185
},
{
"epoch": 2.0313138189244384,
"grad_norm": 0.2132114122172093,
"learning_rate": 2.241394936216472e-05,
"loss": 0.4033,
"step": 186
},
{
"epoch": 2.0422055820285907,
"grad_norm": 0.19087121395561507,
"learning_rate": 2.1954725350386614e-05,
"loss": 0.405,
"step": 187
},
{
"epoch": 2.0530973451327434,
"grad_norm": 0.21040243392909405,
"learning_rate": 2.14984683903666e-05,
"loss": 0.4094,
"step": 188
},
{
"epoch": 2.0639891082368957,
"grad_norm": 0.16468816702883052,
"learning_rate": 2.1045253501080058e-05,
"loss": 0.3988,
"step": 189
},
{
"epoch": 2.0748808713410485,
"grad_norm": 0.17523456140653074,
"learning_rate": 2.0595155201317115e-05,
"loss": 0.4009,
"step": 190
},
{
"epoch": 2.085772634445201,
"grad_norm": 0.19773800473297834,
"learning_rate": 2.0148247497430012e-05,
"loss": 0.4016,
"step": 191
},
{
"epoch": 2.096664397549353,
"grad_norm": 0.1647636735264003,
"learning_rate": 1.970460387116472e-05,
"loss": 0.3948,
"step": 192
},
{
"epoch": 2.107556160653506,
"grad_norm": 0.18353188744441096,
"learning_rate": 1.9264297267579e-05,
"loss": 0.3995,
"step": 193
},
{
"epoch": 2.118447923757658,
"grad_norm": 0.1804931450578559,
"learning_rate": 1.8827400083048503e-05,
"loss": 0.404,
"step": 194
},
{
"epoch": 2.129339686861811,
"grad_norm": 0.16921208996470288,
"learning_rate": 1.8393984153363203e-05,
"loss": 0.3886,
"step": 195
},
{
"epoch": 2.140231449965963,
"grad_norm": 0.15129405097453685,
"learning_rate": 1.7964120741915905e-05,
"loss": 0.4082,
"step": 196
},
{
"epoch": 2.151123213070116,
"grad_norm": 0.17224554916641183,
"learning_rate": 1.753788052798501e-05,
"loss": 0.3994,
"step": 197
},
{
"epoch": 2.162014976174268,
"grad_norm": 0.15770956818834928,
"learning_rate": 1.7115333595113225e-05,
"loss": 0.403,
"step": 198
},
{
"epoch": 2.172906739278421,
"grad_norm": 0.1716039599905281,
"learning_rate": 1.669654941958416e-05,
"loss": 0.4032,
"step": 199
},
{
"epoch": 2.1837985023825732,
"grad_norm": 0.16454710212660165,
"learning_rate": 1.628159685899897e-05,
"loss": 0.405,
"step": 200
},
{
"epoch": 2.1946902654867255,
"grad_norm": 0.13993922901377312,
"learning_rate": 1.5870544140954543e-05,
"loss": 0.4077,
"step": 201
},
{
"epoch": 2.2055820285908783,
"grad_norm": 0.16301323668704157,
"learning_rate": 1.5463458851825345e-05,
"loss": 0.4053,
"step": 202
},
{
"epoch": 2.2164737916950306,
"grad_norm": 0.15989174020128882,
"learning_rate": 1.5060407925650662e-05,
"loss": 0.4036,
"step": 203
},
{
"epoch": 2.2273655547991833,
"grad_norm": 0.14155165998806302,
"learning_rate": 1.466145763312922e-05,
"loss": 0.4021,
"step": 204
},
{
"epoch": 2.2382573179033356,
"grad_norm": 0.13197910914100647,
"learning_rate": 1.426667357072265e-05,
"loss": 0.403,
"step": 205
},
{
"epoch": 2.249149081007488,
"grad_norm": 0.1502501179431283,
"learning_rate": 1.3876120649870051e-05,
"loss": 0.4068,
"step": 206
},
{
"epoch": 2.2600408441116406,
"grad_norm": 0.12632548576221497,
"learning_rate": 1.3489863086315085e-05,
"loss": 0.3954,
"step": 207
},
{
"epoch": 2.270932607215793,
"grad_norm": 0.12310652443499624,
"learning_rate": 1.3107964389547326e-05,
"loss": 0.4056,
"step": 208
},
{
"epoch": 2.2818243703199457,
"grad_norm": 0.13232393309386553,
"learning_rate": 1.2730487352360026e-05,
"loss": 0.4028,
"step": 209
},
{
"epoch": 2.292716133424098,
"grad_norm": 0.12337150415039137,
"learning_rate": 1.2357494040525416e-05,
"loss": 0.4125,
"step": 210
},
{
"epoch": 2.3036078965282503,
"grad_norm": 0.11483009808175632,
"learning_rate": 1.1989045782589815e-05,
"loss": 0.404,
"step": 211
},
{
"epoch": 2.314499659632403,
"grad_norm": 0.1213039797258653,
"learning_rate": 1.1625203159789686e-05,
"loss": 0.398,
"step": 212
},
{
"epoch": 2.3253914227365553,
"grad_norm": 0.1296712824505999,
"learning_rate": 1.1266025996090902e-05,
"loss": 0.4103,
"step": 213
},
{
"epoch": 2.336283185840708,
"grad_norm": 0.12739388021409695,
"learning_rate": 1.0911573348352107e-05,
"loss": 0.4054,
"step": 214
},
{
"epoch": 2.3471749489448603,
"grad_norm": 0.11766621480801555,
"learning_rate": 1.0561903496614603e-05,
"loss": 0.4036,
"step": 215
},
{
"epoch": 2.358066712049013,
"grad_norm": 0.12339813858079209,
"learning_rate": 1.0217073934519726e-05,
"loss": 0.4012,
"step": 216
},
{
"epoch": 2.3689584751531654,
"grad_norm": 0.11647805921713034,
"learning_rate": 9.877141359855567e-06,
"loss": 0.4109,
"step": 217
},
{
"epoch": 2.379850238257318,
"grad_norm": 0.109735463051235,
"learning_rate": 9.542161665234623e-06,
"loss": 0.3991,
"step": 218
},
{
"epoch": 2.3907420013614704,
"grad_norm": 0.1181837766256319,
"learning_rate": 9.212189928903758e-06,
"loss": 0.4008,
"step": 219
},
{
"epoch": 2.4016337644656227,
"grad_norm": 0.12247494016760661,
"learning_rate": 8.887280405688106e-06,
"loss": 0.4044,
"step": 220
},
{
"epoch": 2.4125255275697755,
"grad_norm": 0.11192215755173132,
"learning_rate": 8.567486518070306e-06,
"loss": 0.3992,
"step": 221
},
{
"epoch": 2.4234172906739277,
"grad_norm": 0.109297303198497,
"learning_rate": 8.252860847406712e-06,
"loss": 0.3948,
"step": 222
},
{
"epoch": 2.4343090537780805,
"grad_norm": 0.11284991227332902,
"learning_rate": 7.943455125281741e-06,
"loss": 0.3995,
"step": 223
},
{
"epoch": 2.445200816882233,
"grad_norm": 0.11548250618562617,
"learning_rate": 7.639320225002106e-06,
"loss": 0.407,
"step": 224
},
{
"epoch": 2.456092579986385,
"grad_norm": 0.116001123064691,
"learning_rate": 7.340506153232052e-06,
"loss": 0.4,
"step": 225
},
{
"epoch": 2.466984343090538,
"grad_norm": 0.10992561869419722,
"learning_rate": 7.047062041771133e-06,
"loss": 0.4045,
"step": 226
},
{
"epoch": 2.47787610619469,
"grad_norm": 0.10955112339301888,
"learning_rate": 6.759036139475843e-06,
"loss": 0.3964,
"step": 227
},
{
"epoch": 2.488767869298843,
"grad_norm": 0.11910747258190017,
"learning_rate": 6.476475804326377e-06,
"loss": 0.4025,
"step": 228
},
{
"epoch": 2.499659632402995,
"grad_norm": 0.116651055560925,
"learning_rate": 6.199427495639963e-06,
"loss": 0.3994,
"step": 229
},
{
"epoch": 2.5105513955071475,
"grad_norm": 0.1138861345434715,
"learning_rate": 5.927936766431836e-06,
"loss": 0.4022,
"step": 230
},
{
"epoch": 2.5214431586113,
"grad_norm": 0.09848824385777599,
"learning_rate": 5.662048255925357e-06,
"loss": 0.3988,
"step": 231
},
{
"epoch": 2.532334921715453,
"grad_norm": 0.10150492531984864,
"learning_rate": 5.40180568221226e-06,
"loss": 0.4047,
"step": 232
},
{
"epoch": 2.5432266848196052,
"grad_norm": 0.11056977593577616,
"learning_rate": 5.147251835064424e-06,
"loss": 0.4057,
"step": 233
},
{
"epoch": 2.5541184479237575,
"grad_norm": 0.10655441819972683,
"learning_rate": 4.898428568898288e-06,
"loss": 0.3971,
"step": 234
},
{
"epoch": 2.5650102110279103,
"grad_norm": 0.10580636847333873,
"learning_rate": 4.65537679589299e-06,
"loss": 0.4045,
"step": 235
},
{
"epoch": 2.5759019741320626,
"grad_norm": 0.09908356652521799,
"learning_rate": 4.418136479263533e-06,
"loss": 0.4004,
"step": 236
},
{
"epoch": 2.5867937372362153,
"grad_norm": 0.09287282151185479,
"learning_rate": 4.186746626689879e-06,
"loss": 0.4006,
"step": 237
},
{
"epoch": 2.5976855003403676,
"grad_norm": 0.11618401418689138,
"learning_rate": 3.961245283903239e-06,
"loss": 0.4025,
"step": 238
},
{
"epoch": 2.60857726344452,
"grad_norm": 0.09813772046749265,
"learning_rate": 3.7416695284304737e-06,
"loss": 0.4051,
"step": 239
},
{
"epoch": 2.6194690265486726,
"grad_norm": 0.08852232603950476,
"learning_rate": 3.5280554634977217e-06,
"loss": 0.4013,
"step": 240
},
{
"epoch": 2.630360789652825,
"grad_norm": 0.09394838490800748,
"learning_rate": 3.320438212094197e-06,
"loss": 0.4059,
"step": 241
},
{
"epoch": 2.6412525527569777,
"grad_norm": 0.09059231002024923,
"learning_rate": 3.1188519111971804e-06,
"loss": 0.4039,
"step": 242
},
{
"epoch": 2.65214431586113,
"grad_norm": 0.09411366384619929,
"learning_rate": 2.9233297061591346e-06,
"loss": 0.3973,
"step": 243
},
{
"epoch": 2.6630360789652823,
"grad_norm": 0.09939434350976686,
"learning_rate": 2.733903745257838e-06,
"loss": 0.4049,
"step": 244
},
{
"epoch": 2.673927842069435,
"grad_norm": 0.09359826061681908,
"learning_rate": 2.550605174410512e-06,
"loss": 0.4017,
"step": 245
},
{
"epoch": 2.6848196051735873,
"grad_norm": 0.09354919223068338,
"learning_rate": 2.373464132052701e-06,
"loss": 0.4115,
"step": 246
},
{
"epoch": 2.69571136827774,
"grad_norm": 0.09746026612098238,
"learning_rate": 2.202509744182835e-06,
"loss": 0.3982,
"step": 247
},
{
"epoch": 2.7066031313818923,
"grad_norm": 0.09168148902821703,
"learning_rate": 2.0377701195732545e-06,
"loss": 0.4071,
"step": 248
},
{
"epoch": 2.717494894486045,
"grad_norm": 0.09074247743593482,
"learning_rate": 1.879272345148513e-06,
"loss": 0.4101,
"step": 249
},
{
"epoch": 2.7283866575901974,
"grad_norm": 0.09117996023348725,
"learning_rate": 1.727042481531651e-06,
"loss": 0.4035,
"step": 250
},
{
"epoch": 2.73927842069435,
"grad_norm": 0.09180074130441847,
"learning_rate": 1.5811055587592283e-06,
"loss": 0.4009,
"step": 251
},
{
"epoch": 2.7501701837985024,
"grad_norm": 0.08657516729277243,
"learning_rate": 1.4414855721658705e-06,
"loss": 0.4041,
"step": 252
},
{
"epoch": 2.7610619469026547,
"grad_norm": 0.09146051585946402,
"learning_rate": 1.3082054784388221e-06,
"loss": 0.4055,
"step": 253
},
{
"epoch": 2.7719537100068075,
"grad_norm": 0.09305195530663764,
"learning_rate": 1.1812871918434143e-06,
"loss": 0.4027,
"step": 254
},
{
"epoch": 2.7828454731109598,
"grad_norm": 0.09703421946821408,
"learning_rate": 1.0607515806198142e-06,
"loss": 0.4079,
"step": 255
},
{
"epoch": 2.7937372362151125,
"grad_norm": 0.08906174726324725,
"learning_rate": 9.466184635518361e-07,
"loss": 0.4017,
"step": 256
},
{
"epoch": 2.804628999319265,
"grad_norm": 0.08574571986172967,
"learning_rate": 8.389066067082852e-07,
"loss": 0.3967,
"step": 257
},
{
"epoch": 2.815520762423417,
"grad_norm": 0.0880676756279348,
"learning_rate": 7.376337203573824e-07,
"loss": 0.406,
"step": 258
},
{
"epoch": 2.82641252552757,
"grad_norm": 0.08789941779545143,
"learning_rate": 6.428164560548134e-07,
"loss": 0.4076,
"step": 259
},
{
"epoch": 2.837304288631722,
"grad_norm": 0.08553769033856681,
"learning_rate": 5.544704039058025e-07,
"loss": 0.3983,
"step": 260
},
{
"epoch": 2.848196051735875,
"grad_norm": 0.08010149417951533,
"learning_rate": 4.7261009000177274e-07,
"loss": 0.3918,
"step": 261
},
{
"epoch": 2.859087814840027,
"grad_norm": 0.07962434672916259,
"learning_rate": 3.972489740319274e-07,
"loss": 0.3924,
"step": 262
},
{
"epoch": 2.8699795779441795,
"grad_norm": 0.08482433839475353,
"learning_rate": 3.283994470701579e-07,
"loss": 0.4057,
"step": 263
},
{
"epoch": 2.880871341048332,
"grad_norm": 0.08350491927147591,
"learning_rate": 2.66072829537678e-07,
"loss": 0.4099,
"step": 264
},
{
"epoch": 2.891763104152485,
"grad_norm": 0.0880658902471283,
"learning_rate": 2.102793693417038e-07,
"loss": 0.4073,
"step": 265
},
{
"epoch": 2.9026548672566372,
"grad_norm": 0.08328293731348216,
"learning_rate": 1.6102824019043728e-07,
"loss": 0.4002,
"step": 266
},
{
"epoch": 2.9135466303607895,
"grad_norm": 0.07841175895615185,
"learning_rate": 1.1832754008472614e-07,
"loss": 0.389,
"step": 267
},
{
"epoch": 2.9244383934649423,
"grad_norm": 0.08287793053425582,
"learning_rate": 8.21842899865466e-08,
"loss": 0.3955,
"step": 268
},
{
"epoch": 2.9353301565690946,
"grad_norm": 0.08049459789709469,
"learning_rate": 5.260443266462467e-08,
"loss": 0.3949,
"step": 269
},
{
"epoch": 2.9462219196732473,
"grad_norm": 0.086710676105483,
"learning_rate": 2.9592831717293326e-08,
"loss": 0.4072,
"step": 270
},
{
"epoch": 2.9571136827773996,
"grad_norm": 0.08154316295610879,
"learning_rate": 1.3153270772807702e-08,
"loss": 0.3967,
"step": 271
},
{
"epoch": 2.968005445881552,
"grad_norm": 0.08705352570078856,
"learning_rate": 3.2884528672294523e-09,
"loss": 0.395,
"step": 272
},
{
"epoch": 2.9788972089857046,
"grad_norm": 0.08297445165644558,
"learning_rate": 0.0,
"loss": 0.4097,
"step": 273
},
{
"epoch": 2.9788972089857046,
"step": 273,
"total_flos": 7.02510214541004e+18,
"train_loss": 0.464617241979082,
"train_runtime": 57723.5744,
"train_samples_per_second": 2.443,
"train_steps_per_second": 0.005
}
],
"logging_steps": 1,
"max_steps": 273,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 7.02510214541004e+18,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}