global_batchsize_96_lr4e5 / trainer_state.json
sedrickkeh's picture
End of training
d6d2f24 verified
raw
history blame
91.4 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.994263862332696,
"eval_steps": 500,
"global_step": 522,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0057361376673040155,
"grad_norm": 5.963601353196484,
"learning_rate": 7.547169811320755e-07,
"loss": 0.8872,
"step": 1
},
{
"epoch": 0.011472275334608031,
"grad_norm": 6.333467701978284,
"learning_rate": 1.509433962264151e-06,
"loss": 0.9165,
"step": 2
},
{
"epoch": 0.017208413001912046,
"grad_norm": 6.231014140658969,
"learning_rate": 2.2641509433962266e-06,
"loss": 0.9408,
"step": 3
},
{
"epoch": 0.022944550669216062,
"grad_norm": 5.646301524394669,
"learning_rate": 3.018867924528302e-06,
"loss": 0.8925,
"step": 4
},
{
"epoch": 0.028680688336520075,
"grad_norm": 4.6749029765975045,
"learning_rate": 3.7735849056603777e-06,
"loss": 0.8569,
"step": 5
},
{
"epoch": 0.03441682600382409,
"grad_norm": 4.477470504435412,
"learning_rate": 4.528301886792453e-06,
"loss": 0.8331,
"step": 6
},
{
"epoch": 0.040152963671128104,
"grad_norm": 2.469128428269406,
"learning_rate": 5.283018867924529e-06,
"loss": 0.8192,
"step": 7
},
{
"epoch": 0.045889101338432124,
"grad_norm": 3.1842562805095596,
"learning_rate": 6.037735849056604e-06,
"loss": 0.7936,
"step": 8
},
{
"epoch": 0.05162523900573614,
"grad_norm": 4.1014744450012195,
"learning_rate": 6.792452830188679e-06,
"loss": 0.7749,
"step": 9
},
{
"epoch": 0.05736137667304015,
"grad_norm": 4.176176317229083,
"learning_rate": 7.5471698113207555e-06,
"loss": 0.8084,
"step": 10
},
{
"epoch": 0.06309751434034416,
"grad_norm": 3.5718306849844206,
"learning_rate": 8.301886792452832e-06,
"loss": 0.7579,
"step": 11
},
{
"epoch": 0.06883365200764818,
"grad_norm": 2.9545908685425166,
"learning_rate": 9.056603773584907e-06,
"loss": 0.6838,
"step": 12
},
{
"epoch": 0.0745697896749522,
"grad_norm": 2.484202012476629,
"learning_rate": 9.811320754716981e-06,
"loss": 0.7421,
"step": 13
},
{
"epoch": 0.08030592734225621,
"grad_norm": 1.6256320736281968,
"learning_rate": 1.0566037735849058e-05,
"loss": 0.6922,
"step": 14
},
{
"epoch": 0.08604206500956023,
"grad_norm": 1.8910779905338204,
"learning_rate": 1.1320754716981132e-05,
"loss": 0.695,
"step": 15
},
{
"epoch": 0.09177820267686425,
"grad_norm": 2.1618666538191085,
"learning_rate": 1.2075471698113209e-05,
"loss": 0.6969,
"step": 16
},
{
"epoch": 0.09751434034416825,
"grad_norm": 1.6536038409758815,
"learning_rate": 1.2830188679245283e-05,
"loss": 0.6608,
"step": 17
},
{
"epoch": 0.10325047801147227,
"grad_norm": 1.1899674167485872,
"learning_rate": 1.3584905660377358e-05,
"loss": 0.6611,
"step": 18
},
{
"epoch": 0.1089866156787763,
"grad_norm": 1.3839358683063427,
"learning_rate": 1.4339622641509435e-05,
"loss": 0.5937,
"step": 19
},
{
"epoch": 0.1147227533460803,
"grad_norm": 1.2164557799105977,
"learning_rate": 1.5094339622641511e-05,
"loss": 0.6311,
"step": 20
},
{
"epoch": 0.12045889101338432,
"grad_norm": 0.9617228211418869,
"learning_rate": 1.5849056603773586e-05,
"loss": 0.613,
"step": 21
},
{
"epoch": 0.12619502868068833,
"grad_norm": 0.9112538434649583,
"learning_rate": 1.6603773584905664e-05,
"loss": 0.6066,
"step": 22
},
{
"epoch": 0.13193116634799235,
"grad_norm": 0.9079396604897542,
"learning_rate": 1.735849056603774e-05,
"loss": 0.5748,
"step": 23
},
{
"epoch": 0.13766730401529637,
"grad_norm": 1.0350466631308726,
"learning_rate": 1.8113207547169813e-05,
"loss": 0.6075,
"step": 24
},
{
"epoch": 0.14340344168260039,
"grad_norm": 0.7428776620391337,
"learning_rate": 1.8867924528301888e-05,
"loss": 0.5943,
"step": 25
},
{
"epoch": 0.1491395793499044,
"grad_norm": 0.8363548105185957,
"learning_rate": 1.9622641509433963e-05,
"loss": 0.5737,
"step": 26
},
{
"epoch": 0.15487571701720843,
"grad_norm": 1.110084813808486,
"learning_rate": 2.037735849056604e-05,
"loss": 0.6161,
"step": 27
},
{
"epoch": 0.16061185468451242,
"grad_norm": 0.6819468789186882,
"learning_rate": 2.1132075471698115e-05,
"loss": 0.594,
"step": 28
},
{
"epoch": 0.16634799235181644,
"grad_norm": 1.0895642393194147,
"learning_rate": 2.188679245283019e-05,
"loss": 0.5809,
"step": 29
},
{
"epoch": 0.17208413001912046,
"grad_norm": 0.9651676521976803,
"learning_rate": 2.2641509433962265e-05,
"loss": 0.5813,
"step": 30
},
{
"epoch": 0.17782026768642448,
"grad_norm": 0.8124804726360065,
"learning_rate": 2.339622641509434e-05,
"loss": 0.5523,
"step": 31
},
{
"epoch": 0.1835564053537285,
"grad_norm": 0.851159676858117,
"learning_rate": 2.4150943396226418e-05,
"loss": 0.5376,
"step": 32
},
{
"epoch": 0.18929254302103252,
"grad_norm": 0.8124902743389897,
"learning_rate": 2.4905660377358492e-05,
"loss": 0.5652,
"step": 33
},
{
"epoch": 0.1950286806883365,
"grad_norm": 0.8368258661976667,
"learning_rate": 2.5660377358490567e-05,
"loss": 0.5927,
"step": 34
},
{
"epoch": 0.20076481835564053,
"grad_norm": 0.653616572010454,
"learning_rate": 2.641509433962264e-05,
"loss": 0.5668,
"step": 35
},
{
"epoch": 0.20650095602294455,
"grad_norm": 0.7162845997163322,
"learning_rate": 2.7169811320754716e-05,
"loss": 0.5569,
"step": 36
},
{
"epoch": 0.21223709369024857,
"grad_norm": 0.689294680809673,
"learning_rate": 2.7924528301886794e-05,
"loss": 0.5566,
"step": 37
},
{
"epoch": 0.2179732313575526,
"grad_norm": 0.872345722524905,
"learning_rate": 2.867924528301887e-05,
"loss": 0.5757,
"step": 38
},
{
"epoch": 0.2237093690248566,
"grad_norm": 0.8894547605087567,
"learning_rate": 2.9433962264150944e-05,
"loss": 0.5794,
"step": 39
},
{
"epoch": 0.2294455066921606,
"grad_norm": 0.6582795524486181,
"learning_rate": 3.0188679245283022e-05,
"loss": 0.5319,
"step": 40
},
{
"epoch": 0.23518164435946462,
"grad_norm": 2.7193829449558033,
"learning_rate": 3.09433962264151e-05,
"loss": 0.587,
"step": 41
},
{
"epoch": 0.24091778202676864,
"grad_norm": 0.9717090471397071,
"learning_rate": 3.169811320754717e-05,
"loss": 0.5268,
"step": 42
},
{
"epoch": 0.24665391969407266,
"grad_norm": 0.9224645496458574,
"learning_rate": 3.245283018867925e-05,
"loss": 0.5802,
"step": 43
},
{
"epoch": 0.25239005736137665,
"grad_norm": 0.8938202665523605,
"learning_rate": 3.320754716981133e-05,
"loss": 0.5766,
"step": 44
},
{
"epoch": 0.25812619502868067,
"grad_norm": 0.9583812403407647,
"learning_rate": 3.39622641509434e-05,
"loss": 0.561,
"step": 45
},
{
"epoch": 0.2638623326959847,
"grad_norm": 1.228357544928824,
"learning_rate": 3.471698113207548e-05,
"loss": 0.5193,
"step": 46
},
{
"epoch": 0.2695984703632887,
"grad_norm": 1.1698249381113741,
"learning_rate": 3.547169811320755e-05,
"loss": 0.5639,
"step": 47
},
{
"epoch": 0.27533460803059273,
"grad_norm": 0.8489845786136891,
"learning_rate": 3.6226415094339626e-05,
"loss": 0.549,
"step": 48
},
{
"epoch": 0.28107074569789675,
"grad_norm": 1.57687219740748,
"learning_rate": 3.6981132075471704e-05,
"loss": 0.5742,
"step": 49
},
{
"epoch": 0.28680688336520077,
"grad_norm": 0.8718037736589133,
"learning_rate": 3.7735849056603776e-05,
"loss": 0.5326,
"step": 50
},
{
"epoch": 0.2925430210325048,
"grad_norm": 1.2871827675643652,
"learning_rate": 3.8490566037735854e-05,
"loss": 0.5245,
"step": 51
},
{
"epoch": 0.2982791586998088,
"grad_norm": 1.0013962294494378,
"learning_rate": 3.9245283018867925e-05,
"loss": 0.5621,
"step": 52
},
{
"epoch": 0.30401529636711283,
"grad_norm": 1.1733505550953005,
"learning_rate": 4e-05,
"loss": 0.5527,
"step": 53
},
{
"epoch": 0.30975143403441685,
"grad_norm": 0.8228183926039204,
"learning_rate": 3.999955130375398e-05,
"loss": 0.5207,
"step": 54
},
{
"epoch": 0.3154875717017208,
"grad_norm": 0.9720081493962336,
"learning_rate": 3.999820523514873e-05,
"loss": 0.5388,
"step": 55
},
{
"epoch": 0.32122370936902483,
"grad_norm": 0.8768396168768982,
"learning_rate": 3.9995961854581855e-05,
"loss": 0.5443,
"step": 56
},
{
"epoch": 0.32695984703632885,
"grad_norm": 0.9885103652652149,
"learning_rate": 3.999282126271299e-05,
"loss": 0.5475,
"step": 57
},
{
"epoch": 0.3326959847036329,
"grad_norm": 0.7263822248844192,
"learning_rate": 3.998878360045933e-05,
"loss": 0.5471,
"step": 58
},
{
"epoch": 0.3384321223709369,
"grad_norm": 0.7898997625317963,
"learning_rate": 3.998384904898926e-05,
"loss": 0.532,
"step": 59
},
{
"epoch": 0.3441682600382409,
"grad_norm": 0.7388781217225304,
"learning_rate": 3.997801782971423e-05,
"loss": 0.5196,
"step": 60
},
{
"epoch": 0.34990439770554493,
"grad_norm": 0.7257057362841633,
"learning_rate": 3.997129020427888e-05,
"loss": 0.5485,
"step": 61
},
{
"epoch": 0.35564053537284895,
"grad_norm": 0.6438468755093205,
"learning_rate": 3.996366647454924e-05,
"loss": 0.5466,
"step": 62
},
{
"epoch": 0.361376673040153,
"grad_norm": 0.6663878490068373,
"learning_rate": 3.9955146982599194e-05,
"loss": 0.5468,
"step": 63
},
{
"epoch": 0.367112810707457,
"grad_norm": 0.6745599593845647,
"learning_rate": 3.9945732110695145e-05,
"loss": 0.525,
"step": 64
},
{
"epoch": 0.372848948374761,
"grad_norm": 0.7081809867814687,
"learning_rate": 3.993542228127886e-05,
"loss": 0.5382,
"step": 65
},
{
"epoch": 0.37858508604206503,
"grad_norm": 0.7366611396520816,
"learning_rate": 3.992421795694853e-05,
"loss": 0.5431,
"step": 66
},
{
"epoch": 0.384321223709369,
"grad_norm": 0.7027778770800306,
"learning_rate": 3.9912119640437965e-05,
"loss": 0.5154,
"step": 67
},
{
"epoch": 0.390057361376673,
"grad_norm": 0.6666822720365386,
"learning_rate": 3.989912787459409e-05,
"loss": 0.527,
"step": 68
},
{
"epoch": 0.39579349904397704,
"grad_norm": 0.5660749653496684,
"learning_rate": 3.9885243242352565e-05,
"loss": 0.5351,
"step": 69
},
{
"epoch": 0.40152963671128106,
"grad_norm": 0.5576960477632893,
"learning_rate": 3.987046636671162e-05,
"loss": 0.4824,
"step": 70
},
{
"epoch": 0.4072657743785851,
"grad_norm": 0.658340932840387,
"learning_rate": 3.9854797910704124e-05,
"loss": 0.5364,
"step": 71
},
{
"epoch": 0.4130019120458891,
"grad_norm": 0.6117526542482664,
"learning_rate": 3.9838238577367816e-05,
"loss": 0.5014,
"step": 72
},
{
"epoch": 0.4187380497131931,
"grad_norm": 0.5936168327692725,
"learning_rate": 3.982078910971376e-05,
"loss": 0.5349,
"step": 73
},
{
"epoch": 0.42447418738049714,
"grad_norm": 0.687524043308455,
"learning_rate": 3.980245029069302e-05,
"loss": 0.5334,
"step": 74
},
{
"epoch": 0.43021032504780116,
"grad_norm": 0.6835492486583609,
"learning_rate": 3.978322294316153e-05,
"loss": 0.5509,
"step": 75
},
{
"epoch": 0.4359464627151052,
"grad_norm": 0.5286499638619044,
"learning_rate": 3.976310792984315e-05,
"loss": 0.5323,
"step": 76
},
{
"epoch": 0.4416826003824092,
"grad_norm": 0.6642388755587706,
"learning_rate": 3.974210615329098e-05,
"loss": 0.5451,
"step": 77
},
{
"epoch": 0.4474187380497132,
"grad_norm": 0.4999354398609915,
"learning_rate": 3.9720218555846834e-05,
"loss": 0.5339,
"step": 78
},
{
"epoch": 0.45315487571701724,
"grad_norm": 0.642751208835366,
"learning_rate": 3.9697446119599015e-05,
"loss": 0.544,
"step": 79
},
{
"epoch": 0.4588910133843212,
"grad_norm": 0.719086166881077,
"learning_rate": 3.967378986633818e-05,
"loss": 0.5256,
"step": 80
},
{
"epoch": 0.4646271510516252,
"grad_norm": 0.6270310635224952,
"learning_rate": 3.964925085751152e-05,
"loss": 0.53,
"step": 81
},
{
"epoch": 0.47036328871892924,
"grad_norm": 0.6955843238682398,
"learning_rate": 3.9623830194175175e-05,
"loss": 0.552,
"step": 82
},
{
"epoch": 0.47609942638623326,
"grad_norm": 0.6300967187312061,
"learning_rate": 3.9597529016944746e-05,
"loss": 0.5265,
"step": 83
},
{
"epoch": 0.4818355640535373,
"grad_norm": 0.6927514450309129,
"learning_rate": 3.9570348505944185e-05,
"loss": 0.5389,
"step": 84
},
{
"epoch": 0.4875717017208413,
"grad_norm": 0.6581382765482165,
"learning_rate": 3.9542289880752827e-05,
"loss": 0.5258,
"step": 85
},
{
"epoch": 0.4933078393881453,
"grad_norm": 0.6347863955072047,
"learning_rate": 3.951335440035063e-05,
"loss": 0.5295,
"step": 86
},
{
"epoch": 0.49904397705544934,
"grad_norm": 0.7550386421946482,
"learning_rate": 3.948354336306176e-05,
"loss": 0.5516,
"step": 87
},
{
"epoch": 0.5047801147227533,
"grad_norm": 0.5280267134452366,
"learning_rate": 3.945285810649626e-05,
"loss": 0.5315,
"step": 88
},
{
"epoch": 0.5105162523900574,
"grad_norm": 0.649947445862538,
"learning_rate": 3.942130000749008e-05,
"loss": 0.5415,
"step": 89
},
{
"epoch": 0.5162523900573613,
"grad_norm": 0.5480518529513726,
"learning_rate": 3.938887048204326e-05,
"loss": 0.5465,
"step": 90
},
{
"epoch": 0.5219885277246654,
"grad_norm": 0.5610828231392423,
"learning_rate": 3.935557098525644e-05,
"loss": 0.5005,
"step": 91
},
{
"epoch": 0.5277246653919694,
"grad_norm": 0.5706185940759887,
"learning_rate": 3.9321403011265546e-05,
"loss": 0.5189,
"step": 92
},
{
"epoch": 0.5334608030592735,
"grad_norm": 0.5626627671265411,
"learning_rate": 3.9286368093174745e-05,
"loss": 0.5277,
"step": 93
},
{
"epoch": 0.5391969407265774,
"grad_norm": 0.6290463734820702,
"learning_rate": 3.925046780298764e-05,
"loss": 0.5237,
"step": 94
},
{
"epoch": 0.5449330783938815,
"grad_norm": 0.5701628353252369,
"learning_rate": 3.921370375153681e-05,
"loss": 0.5348,
"step": 95
},
{
"epoch": 0.5506692160611855,
"grad_norm": 0.6186142129858236,
"learning_rate": 3.917607758841141e-05,
"loss": 0.5349,
"step": 96
},
{
"epoch": 0.5564053537284895,
"grad_norm": 0.4973723214529781,
"learning_rate": 3.913759100188327e-05,
"loss": 0.5061,
"step": 97
},
{
"epoch": 0.5621414913957935,
"grad_norm": 0.682743340827704,
"learning_rate": 3.9098245718831076e-05,
"loss": 0.5615,
"step": 98
},
{
"epoch": 0.5678776290630975,
"grad_norm": 0.6357989822661191,
"learning_rate": 3.905804350466291e-05,
"loss": 0.5243,
"step": 99
},
{
"epoch": 0.5736137667304015,
"grad_norm": 0.5449769248133205,
"learning_rate": 3.901698616323703e-05,
"loss": 0.5155,
"step": 100
},
{
"epoch": 0.5793499043977055,
"grad_norm": 0.6266775770075623,
"learning_rate": 3.897507553678093e-05,
"loss": 0.5451,
"step": 101
},
{
"epoch": 0.5850860420650096,
"grad_norm": 0.7475144523577113,
"learning_rate": 3.893231350580869e-05,
"loss": 0.5383,
"step": 102
},
{
"epoch": 0.5908221797323135,
"grad_norm": 0.5434812815586433,
"learning_rate": 3.888870198903658e-05,
"loss": 0.5316,
"step": 103
},
{
"epoch": 0.5965583173996176,
"grad_norm": 0.6008597770179454,
"learning_rate": 3.8844242943297e-05,
"loss": 0.523,
"step": 104
},
{
"epoch": 0.6022944550669216,
"grad_norm": 0.6384325056734049,
"learning_rate": 3.879893836345062e-05,
"loss": 0.5153,
"step": 105
},
{
"epoch": 0.6080305927342257,
"grad_norm": 0.5480218749548357,
"learning_rate": 3.875279028229695e-05,
"loss": 0.5432,
"step": 106
},
{
"epoch": 0.6137667304015296,
"grad_norm": 0.6160281910387104,
"learning_rate": 3.870580077048307e-05,
"loss": 0.5296,
"step": 107
},
{
"epoch": 0.6195028680688337,
"grad_norm": 0.6894563819781858,
"learning_rate": 3.865797193641072e-05,
"loss": 0.5355,
"step": 108
},
{
"epoch": 0.6252390057361377,
"grad_norm": 0.4774944165107399,
"learning_rate": 3.8609305926141735e-05,
"loss": 0.4964,
"step": 109
},
{
"epoch": 0.6309751434034416,
"grad_norm": 0.7099537904032711,
"learning_rate": 3.855980492330173e-05,
"loss": 0.5399,
"step": 110
},
{
"epoch": 0.6367112810707457,
"grad_norm": 0.6001030131052923,
"learning_rate": 3.850947114898212e-05,
"loss": 0.54,
"step": 111
},
{
"epoch": 0.6424474187380497,
"grad_norm": 0.6169400599723738,
"learning_rate": 3.8458306861640465e-05,
"loss": 0.5478,
"step": 112
},
{
"epoch": 0.6481835564053537,
"grad_norm": 0.6331995497335126,
"learning_rate": 3.840631435699912e-05,
"loss": 0.5089,
"step": 113
},
{
"epoch": 0.6539196940726577,
"grad_norm": 0.6758544384585873,
"learning_rate": 3.835349596794226e-05,
"loss": 0.524,
"step": 114
},
{
"epoch": 0.6596558317399618,
"grad_norm": 0.6066427921666631,
"learning_rate": 3.829985406441118e-05,
"loss": 0.5371,
"step": 115
},
{
"epoch": 0.6653919694072657,
"grad_norm": 0.6205375496679291,
"learning_rate": 3.8245391053297936e-05,
"loss": 0.4979,
"step": 116
},
{
"epoch": 0.6711281070745698,
"grad_norm": 0.5806033479161776,
"learning_rate": 3.8190109378337413e-05,
"loss": 0.5239,
"step": 117
},
{
"epoch": 0.6768642447418738,
"grad_norm": 0.6460511560696024,
"learning_rate": 3.813401151999759e-05,
"loss": 0.5475,
"step": 118
},
{
"epoch": 0.6826003824091779,
"grad_norm": 0.5453484149762959,
"learning_rate": 3.807709999536834e-05,
"loss": 0.5395,
"step": 119
},
{
"epoch": 0.6883365200764818,
"grad_norm": 0.5318167624766734,
"learning_rate": 3.801937735804838e-05,
"loss": 0.5087,
"step": 120
},
{
"epoch": 0.6940726577437859,
"grad_norm": 0.8271638416761657,
"learning_rate": 3.7960846198030804e-05,
"loss": 0.5453,
"step": 121
},
{
"epoch": 0.6998087954110899,
"grad_norm": 0.6479532137119753,
"learning_rate": 3.790150914158677e-05,
"loss": 0.5043,
"step": 122
},
{
"epoch": 0.7055449330783938,
"grad_norm": 0.7435964195190743,
"learning_rate": 3.784136885114773e-05,
"loss": 0.5129,
"step": 123
},
{
"epoch": 0.7112810707456979,
"grad_norm": 0.8946003605016197,
"learning_rate": 3.7780428025185954e-05,
"loss": 0.5533,
"step": 124
},
{
"epoch": 0.7170172084130019,
"grad_norm": 0.6464116296407747,
"learning_rate": 3.77186893980934e-05,
"loss": 0.5421,
"step": 125
},
{
"epoch": 0.722753346080306,
"grad_norm": 0.7126674023402929,
"learning_rate": 3.765615574005911e-05,
"loss": 0.5139,
"step": 126
},
{
"epoch": 0.7284894837476099,
"grad_norm": 0.8135984381109922,
"learning_rate": 3.7592829856944835e-05,
"loss": 0.511,
"step": 127
},
{
"epoch": 0.734225621414914,
"grad_norm": 0.7828468225319082,
"learning_rate": 3.752871459015918e-05,
"loss": 0.5286,
"step": 128
},
{
"epoch": 0.739961759082218,
"grad_norm": 0.642287015277123,
"learning_rate": 3.74638128165301e-05,
"loss": 0.5329,
"step": 129
},
{
"epoch": 0.745697896749522,
"grad_norm": 0.5442493722807862,
"learning_rate": 3.739812744817581e-05,
"loss": 0.5362,
"step": 130
},
{
"epoch": 0.751434034416826,
"grad_norm": 0.5568845351714643,
"learning_rate": 3.733166143237413e-05,
"loss": 0.5035,
"step": 131
},
{
"epoch": 0.7571701720841301,
"grad_norm": 0.49191832987393536,
"learning_rate": 3.7264417751430244e-05,
"loss": 0.5085,
"step": 132
},
{
"epoch": 0.762906309751434,
"grad_norm": 0.6805242608433807,
"learning_rate": 3.7196399422542866e-05,
"loss": 0.5065,
"step": 133
},
{
"epoch": 0.768642447418738,
"grad_norm": 0.5897633089426585,
"learning_rate": 3.712760949766888e-05,
"loss": 0.5038,
"step": 134
},
{
"epoch": 0.7743785850860421,
"grad_norm": 0.6094976285430489,
"learning_rate": 3.70580510633864e-05,
"loss": 0.5008,
"step": 135
},
{
"epoch": 0.780114722753346,
"grad_norm": 0.5467744310721907,
"learning_rate": 3.6987727240756246e-05,
"loss": 0.5142,
"step": 136
},
{
"epoch": 0.7858508604206501,
"grad_norm": 0.5880627703365717,
"learning_rate": 3.691664118518195e-05,
"loss": 0.4994,
"step": 137
},
{
"epoch": 0.7915869980879541,
"grad_norm": 0.5006717341236084,
"learning_rate": 3.6844796086268136e-05,
"loss": 0.5164,
"step": 138
},
{
"epoch": 0.7973231357552581,
"grad_norm": 0.5642072424368477,
"learning_rate": 3.677219516767743e-05,
"loss": 0.5349,
"step": 139
},
{
"epoch": 0.8030592734225621,
"grad_norm": 0.5741831253551103,
"learning_rate": 3.669884168698578e-05,
"loss": 0.559,
"step": 140
},
{
"epoch": 0.8087954110898662,
"grad_norm": 0.5577991231255239,
"learning_rate": 3.6624738935536336e-05,
"loss": 0.5082,
"step": 141
},
{
"epoch": 0.8145315487571702,
"grad_norm": 0.6263509656905389,
"learning_rate": 3.654989023829174e-05,
"loss": 0.5401,
"step": 142
},
{
"epoch": 0.8202676864244742,
"grad_norm": 0.562818325762248,
"learning_rate": 3.6474298953684945e-05,
"loss": 0.5071,
"step": 143
},
{
"epoch": 0.8260038240917782,
"grad_norm": 0.5540399435211395,
"learning_rate": 3.639796847346849e-05,
"loss": 0.5343,
"step": 144
},
{
"epoch": 0.8317399617590823,
"grad_norm": 0.5580743271553504,
"learning_rate": 3.6320902222562405e-05,
"loss": 0.515,
"step": 145
},
{
"epoch": 0.8374760994263862,
"grad_norm": 0.5846568206692573,
"learning_rate": 3.62431036589004e-05,
"loss": 0.5235,
"step": 146
},
{
"epoch": 0.8432122370936902,
"grad_norm": 0.6959565954811155,
"learning_rate": 3.616457627327484e-05,
"loss": 0.5075,
"step": 147
},
{
"epoch": 0.8489483747609943,
"grad_norm": 0.565057362201928,
"learning_rate": 3.6085323589180046e-05,
"loss": 0.5092,
"step": 148
},
{
"epoch": 0.8546845124282982,
"grad_norm": 0.6281295561065597,
"learning_rate": 3.600534916265419e-05,
"loss": 0.4943,
"step": 149
},
{
"epoch": 0.8604206500956023,
"grad_norm": 0.6023614171022365,
"learning_rate": 3.592465658211977e-05,
"loss": 0.5133,
"step": 150
},
{
"epoch": 0.8661567877629063,
"grad_norm": 0.6315201878036348,
"learning_rate": 3.5843249468222585e-05,
"loss": 0.5264,
"step": 151
},
{
"epoch": 0.8718929254302104,
"grad_norm": 0.6634687345841984,
"learning_rate": 3.5761131473669285e-05,
"loss": 0.4759,
"step": 152
},
{
"epoch": 0.8776290630975143,
"grad_norm": 0.6850113724699497,
"learning_rate": 3.567830628306344e-05,
"loss": 0.5335,
"step": 153
},
{
"epoch": 0.8833652007648184,
"grad_norm": 0.7083577814147579,
"learning_rate": 3.5594777612740263e-05,
"loss": 0.4993,
"step": 154
},
{
"epoch": 0.8891013384321224,
"grad_norm": 0.7378709536755256,
"learning_rate": 3.551054921059985e-05,
"loss": 0.5176,
"step": 155
},
{
"epoch": 0.8948374760994264,
"grad_norm": 0.7648244850245753,
"learning_rate": 3.542562485593897e-05,
"loss": 0.5238,
"step": 156
},
{
"epoch": 0.9005736137667304,
"grad_norm": 0.6956157652186921,
"learning_rate": 3.5340008359281546e-05,
"loss": 0.501,
"step": 157
},
{
"epoch": 0.9063097514340345,
"grad_norm": 0.7645790590701518,
"learning_rate": 3.525370356220764e-05,
"loss": 0.5345,
"step": 158
},
{
"epoch": 0.9120458891013384,
"grad_norm": 0.6517490678032604,
"learning_rate": 3.5166714337181104e-05,
"loss": 0.5018,
"step": 159
},
{
"epoch": 0.9177820267686424,
"grad_norm": 0.5764447491349046,
"learning_rate": 3.50790445873758e-05,
"loss": 0.5078,
"step": 160
},
{
"epoch": 0.9235181644359465,
"grad_norm": 0.6747308853100928,
"learning_rate": 3.4990698246500484e-05,
"loss": 0.5256,
"step": 161
},
{
"epoch": 0.9292543021032504,
"grad_norm": 0.47711018256334037,
"learning_rate": 3.490167927862233e-05,
"loss": 0.4722,
"step": 162
},
{
"epoch": 0.9349904397705545,
"grad_norm": 0.6778658770774263,
"learning_rate": 3.481199167798899e-05,
"loss": 0.5223,
"step": 163
},
{
"epoch": 0.9407265774378585,
"grad_norm": 0.4558533116863511,
"learning_rate": 3.4721639468849436e-05,
"loss": 0.5213,
"step": 164
},
{
"epoch": 0.9464627151051626,
"grad_norm": 0.6385762668945828,
"learning_rate": 3.463062670527338e-05,
"loss": 0.5273,
"step": 165
},
{
"epoch": 0.9521988527724665,
"grad_norm": 0.4616066439052544,
"learning_rate": 3.4538957470969365e-05,
"loss": 0.4904,
"step": 166
},
{
"epoch": 0.9579349904397706,
"grad_norm": 0.6031035010764613,
"learning_rate": 3.444663587910151e-05,
"loss": 0.4922,
"step": 167
},
{
"epoch": 0.9636711281070746,
"grad_norm": 0.4890495218308744,
"learning_rate": 3.435366607210499e-05,
"loss": 0.4877,
"step": 168
},
{
"epoch": 0.9694072657743786,
"grad_norm": 0.6620804135308059,
"learning_rate": 3.426005222150014e-05,
"loss": 0.5106,
"step": 169
},
{
"epoch": 0.9751434034416826,
"grad_norm": 0.5095511322768239,
"learning_rate": 3.41657985277053e-05,
"loss": 0.5008,
"step": 170
},
{
"epoch": 0.9808795411089866,
"grad_norm": 0.6827188978217906,
"learning_rate": 3.407090921984832e-05,
"loss": 0.5477,
"step": 171
},
{
"epoch": 0.9866156787762906,
"grad_norm": 0.5541219385903691,
"learning_rate": 3.397538855557684e-05,
"loss": 0.5009,
"step": 172
},
{
"epoch": 0.9923518164435946,
"grad_norm": 0.6443008399746032,
"learning_rate": 3.387924082086718e-05,
"loss": 0.5332,
"step": 173
},
{
"epoch": 0.9980879541108987,
"grad_norm": 0.4936165614613112,
"learning_rate": 3.378247032983213e-05,
"loss": 0.522,
"step": 174
},
{
"epoch": 1.0038240917782026,
"grad_norm": 0.9746986187989416,
"learning_rate": 3.368508142452728e-05,
"loss": 0.7212,
"step": 175
},
{
"epoch": 1.0095602294455066,
"grad_norm": 0.6720620316515331,
"learning_rate": 3.358707847475626e-05,
"loss": 0.4198,
"step": 176
},
{
"epoch": 1.0152963671128108,
"grad_norm": 0.6268806895754554,
"learning_rate": 3.348846587787462e-05,
"loss": 0.461,
"step": 177
},
{
"epoch": 1.0210325047801148,
"grad_norm": 0.7216148037661133,
"learning_rate": 3.3389248058592576e-05,
"loss": 0.4066,
"step": 178
},
{
"epoch": 1.0267686424474187,
"grad_norm": 0.6818445339544676,
"learning_rate": 3.328942946877644e-05,
"loss": 0.3977,
"step": 179
},
{
"epoch": 1.0325047801147227,
"grad_norm": 0.6950993166160578,
"learning_rate": 3.318901458724885e-05,
"loss": 0.4345,
"step": 180
},
{
"epoch": 1.0382409177820269,
"grad_norm": 0.6428355368495283,
"learning_rate": 3.308800791958785e-05,
"loss": 0.3577,
"step": 181
},
{
"epoch": 1.0439770554493308,
"grad_norm": 0.5373947494002963,
"learning_rate": 3.2986413997924706e-05,
"loss": 0.4369,
"step": 182
},
{
"epoch": 1.0497131931166348,
"grad_norm": 0.6925796288170799,
"learning_rate": 3.2884237380740545e-05,
"loss": 0.4328,
"step": 183
},
{
"epoch": 1.0554493307839388,
"grad_norm": 0.5233829993400329,
"learning_rate": 3.278148265266182e-05,
"loss": 0.4174,
"step": 184
},
{
"epoch": 1.0611854684512427,
"grad_norm": 0.6185297505581466,
"learning_rate": 3.267815442425459e-05,
"loss": 0.3828,
"step": 185
},
{
"epoch": 1.066921606118547,
"grad_norm": 0.6274152460265636,
"learning_rate": 3.25742573318177e-05,
"loss": 0.4549,
"step": 186
},
{
"epoch": 1.0726577437858509,
"grad_norm": 0.6366347848494356,
"learning_rate": 3.246979603717467e-05,
"loss": 0.4203,
"step": 187
},
{
"epoch": 1.0783938814531548,
"grad_norm": 0.6227490389766099,
"learning_rate": 3.236477522746459e-05,
"loss": 0.4675,
"step": 188
},
{
"epoch": 1.0841300191204588,
"grad_norm": 0.5094295149535436,
"learning_rate": 3.2259199614931745e-05,
"loss": 0.3542,
"step": 189
},
{
"epoch": 1.089866156787763,
"grad_norm": 0.7037793847977521,
"learning_rate": 3.215307393671426e-05,
"loss": 0.4785,
"step": 190
},
{
"epoch": 1.095602294455067,
"grad_norm": 0.5619740614381676,
"learning_rate": 3.204640295463146e-05,
"loss": 0.3886,
"step": 191
},
{
"epoch": 1.101338432122371,
"grad_norm": 0.519062950155675,
"learning_rate": 3.193919145497028e-05,
"loss": 0.3842,
"step": 192
},
{
"epoch": 1.107074569789675,
"grad_norm": 0.5853788545600012,
"learning_rate": 3.1831444248270455e-05,
"loss": 0.423,
"step": 193
},
{
"epoch": 1.1128107074569789,
"grad_norm": 0.4775535757004622,
"learning_rate": 3.17231661691087e-05,
"loss": 0.4239,
"step": 194
},
{
"epoch": 1.118546845124283,
"grad_norm": 0.5172569912420082,
"learning_rate": 3.161436207588178e-05,
"loss": 0.425,
"step": 195
},
{
"epoch": 1.124282982791587,
"grad_norm": 0.5147855471578512,
"learning_rate": 3.1505036850588516e-05,
"loss": 0.3557,
"step": 196
},
{
"epoch": 1.130019120458891,
"grad_norm": 0.5476099573273011,
"learning_rate": 3.139519539861074e-05,
"loss": 0.4791,
"step": 197
},
{
"epoch": 1.135755258126195,
"grad_norm": 0.5619357244873963,
"learning_rate": 3.128484264849314e-05,
"loss": 0.4214,
"step": 198
},
{
"epoch": 1.1414913957934991,
"grad_norm": 0.48517078514017925,
"learning_rate": 3.1173983551722205e-05,
"loss": 0.3698,
"step": 199
},
{
"epoch": 1.147227533460803,
"grad_norm": 0.6343332709061799,
"learning_rate": 3.106262308250399e-05,
"loss": 0.3882,
"step": 200
},
{
"epoch": 1.152963671128107,
"grad_norm": 0.4706809203743212,
"learning_rate": 3.0950766237540946e-05,
"loss": 0.3937,
"step": 201
},
{
"epoch": 1.158699808795411,
"grad_norm": 0.6399176928230911,
"learning_rate": 3.083841803580771e-05,
"loss": 0.4585,
"step": 202
},
{
"epoch": 1.1644359464627152,
"grad_norm": 0.4919902699693636,
"learning_rate": 3.072558351832592e-05,
"loss": 0.3764,
"step": 203
},
{
"epoch": 1.1701720841300192,
"grad_norm": 0.6185212786022399,
"learning_rate": 3.0612267747938015e-05,
"loss": 0.4736,
"step": 204
},
{
"epoch": 1.1759082217973231,
"grad_norm": 0.5805762135464895,
"learning_rate": 3.0498475809080082e-05,
"loss": 0.4208,
"step": 205
},
{
"epoch": 1.181644359464627,
"grad_norm": 0.5097240529240963,
"learning_rate": 3.03842128075537e-05,
"loss": 0.4024,
"step": 206
},
{
"epoch": 1.1873804971319313,
"grad_norm": 0.591913249803808,
"learning_rate": 3.026948387029684e-05,
"loss": 0.4055,
"step": 207
},
{
"epoch": 1.1931166347992352,
"grad_norm": 0.5073228529458789,
"learning_rate": 3.0154294145153865e-05,
"loss": 0.4051,
"step": 208
},
{
"epoch": 1.1988527724665392,
"grad_norm": 0.5492611084308193,
"learning_rate": 3.003864880064449e-05,
"loss": 0.41,
"step": 209
},
{
"epoch": 1.2045889101338432,
"grad_norm": 0.49487668182682537,
"learning_rate": 2.9922553025731907e-05,
"loss": 0.4053,
"step": 210
},
{
"epoch": 1.2103250478011471,
"grad_norm": 0.5668757218011339,
"learning_rate": 2.9806012029589966e-05,
"loss": 0.4272,
"step": 211
},
{
"epoch": 1.2160611854684513,
"grad_norm": 0.4975892481813516,
"learning_rate": 2.9689031041369406e-05,
"loss": 0.4182,
"step": 212
},
{
"epoch": 1.2217973231357553,
"grad_norm": 0.5294244433263008,
"learning_rate": 2.9571615309963255e-05,
"loss": 0.3871,
"step": 213
},
{
"epoch": 1.2275334608030593,
"grad_norm": 0.5243327767554817,
"learning_rate": 2.9453770103771308e-05,
"loss": 0.39,
"step": 214
},
{
"epoch": 1.2332695984703632,
"grad_norm": 0.48558339355845753,
"learning_rate": 2.9335500710463725e-05,
"loss": 0.435,
"step": 215
},
{
"epoch": 1.2390057361376674,
"grad_norm": 0.5519168520623398,
"learning_rate": 2.9216812436743788e-05,
"loss": 0.4313,
"step": 216
},
{
"epoch": 1.2447418738049714,
"grad_norm": 0.4381015639162095,
"learning_rate": 2.9097710608109776e-05,
"loss": 0.3601,
"step": 217
},
{
"epoch": 1.2504780114722753,
"grad_norm": 0.5354350555855661,
"learning_rate": 2.8978200568616036e-05,
"loss": 0.3711,
"step": 218
},
{
"epoch": 1.2562141491395793,
"grad_norm": 0.48392108469684075,
"learning_rate": 2.885828768063317e-05,
"loss": 0.4385,
"step": 219
},
{
"epoch": 1.2619502868068833,
"grad_norm": 0.5147905754022055,
"learning_rate": 2.8737977324607466e-05,
"loss": 0.4318,
"step": 220
},
{
"epoch": 1.2676864244741874,
"grad_norm": 0.45875460028412374,
"learning_rate": 2.861727489881941e-05,
"loss": 0.4165,
"step": 221
},
{
"epoch": 1.2734225621414914,
"grad_norm": 0.46767203325180406,
"learning_rate": 2.8496185819141547e-05,
"loss": 0.4346,
"step": 222
},
{
"epoch": 1.2791586998087954,
"grad_norm": 0.5325780804765026,
"learning_rate": 2.837471551879543e-05,
"loss": 0.4488,
"step": 223
},
{
"epoch": 1.2848948374760996,
"grad_norm": 0.45604591940023564,
"learning_rate": 2.825286944810783e-05,
"loss": 0.4172,
"step": 224
},
{
"epoch": 1.2906309751434035,
"grad_norm": 0.48182478279469737,
"learning_rate": 2.8130653074266188e-05,
"loss": 0.3899,
"step": 225
},
{
"epoch": 1.2963671128107075,
"grad_norm": 0.4285203303941567,
"learning_rate": 2.8008071881073342e-05,
"loss": 0.4119,
"step": 226
},
{
"epoch": 1.3021032504780115,
"grad_norm": 0.5203451219224136,
"learning_rate": 2.78851313687014e-05,
"loss": 0.4167,
"step": 227
},
{
"epoch": 1.3078393881453154,
"grad_norm": 0.44867616077783923,
"learning_rate": 2.7761837053444994e-05,
"loss": 0.4428,
"step": 228
},
{
"epoch": 1.3135755258126194,
"grad_norm": 0.5176774633567216,
"learning_rate": 2.7638194467473776e-05,
"loss": 0.4623,
"step": 229
},
{
"epoch": 1.3193116634799236,
"grad_norm": 0.41453804129004856,
"learning_rate": 2.7514209158584164e-05,
"loss": 0.3855,
"step": 230
},
{
"epoch": 1.3250478011472275,
"grad_norm": 0.44139671451368523,
"learning_rate": 2.7389886689950428e-05,
"loss": 0.3704,
"step": 231
},
{
"epoch": 1.3307839388145315,
"grad_norm": 0.5091495369537983,
"learning_rate": 2.7265232639875043e-05,
"loss": 0.4346,
"step": 232
},
{
"epoch": 1.3365200764818357,
"grad_norm": 0.4627275608117313,
"learning_rate": 2.7140252601538466e-05,
"loss": 0.4339,
"step": 233
},
{
"epoch": 1.3422562141491396,
"grad_norm": 0.48547168450981104,
"learning_rate": 2.7014952182748083e-05,
"loss": 0.4307,
"step": 234
},
{
"epoch": 1.3479923518164436,
"grad_norm": 0.47854324223280925,
"learning_rate": 2.688933700568666e-05,
"loss": 0.4091,
"step": 235
},
{
"epoch": 1.3537284894837476,
"grad_norm": 0.4845854941728673,
"learning_rate": 2.676341270666003e-05,
"loss": 0.4544,
"step": 236
},
{
"epoch": 1.3594646271510515,
"grad_norm": 0.45196894051562736,
"learning_rate": 2.663718493584422e-05,
"loss": 0.3461,
"step": 237
},
{
"epoch": 1.3652007648183555,
"grad_norm": 0.543415301575274,
"learning_rate": 2.6510659357031917e-05,
"loss": 0.4336,
"step": 238
},
{
"epoch": 1.3709369024856597,
"grad_norm": 0.4471607463074475,
"learning_rate": 2.6383841647378356e-05,
"loss": 0.38,
"step": 239
},
{
"epoch": 1.3766730401529637,
"grad_norm": 0.6022022438673463,
"learning_rate": 2.6256737497146544e-05,
"loss": 0.4709,
"step": 240
},
{
"epoch": 1.3824091778202676,
"grad_norm": 0.44800638598268433,
"learning_rate": 2.6129352609452003e-05,
"loss": 0.4105,
"step": 241
},
{
"epoch": 1.3881453154875718,
"grad_norm": 0.4385968231933793,
"learning_rate": 2.600169270000682e-05,
"loss": 0.3782,
"step": 242
},
{
"epoch": 1.3938814531548758,
"grad_norm": 0.48914377606117937,
"learning_rate": 2.5873763496863203e-05,
"loss": 0.3954,
"step": 243
},
{
"epoch": 1.3996175908221797,
"grad_norm": 0.5851945762803631,
"learning_rate": 2.574557074015648e-05,
"loss": 0.458,
"step": 244
},
{
"epoch": 1.4053537284894837,
"grad_norm": 0.46457478071444297,
"learning_rate": 2.561712018184752e-05,
"loss": 0.438,
"step": 245
},
{
"epoch": 1.4110898661567877,
"grad_norm": 0.5204965149805247,
"learning_rate": 2.5488417585464648e-05,
"loss": 0.4171,
"step": 246
},
{
"epoch": 1.4168260038240919,
"grad_norm": 0.47781885084741565,
"learning_rate": 2.535946872584506e-05,
"loss": 0.4265,
"step": 247
},
{
"epoch": 1.4225621414913958,
"grad_norm": 0.5009930687067327,
"learning_rate": 2.523027938887567e-05,
"loss": 0.39,
"step": 248
},
{
"epoch": 1.4282982791586998,
"grad_norm": 0.5407445516705907,
"learning_rate": 2.5100855371233533e-05,
"loss": 0.4255,
"step": 249
},
{
"epoch": 1.4340344168260037,
"grad_norm": 0.4288547555150977,
"learning_rate": 2.4971202480125737e-05,
"loss": 0.3925,
"step": 250
},
{
"epoch": 1.439770554493308,
"grad_norm": 0.4567682163636214,
"learning_rate": 2.4841326533028838e-05,
"loss": 0.4283,
"step": 251
},
{
"epoch": 1.445506692160612,
"grad_norm": 0.47732054782696987,
"learning_rate": 2.4711233357427817e-05,
"loss": 0.429,
"step": 252
},
{
"epoch": 1.4512428298279159,
"grad_norm": 0.3827981714632121,
"learning_rate": 2.458092879055464e-05,
"loss": 0.3684,
"step": 253
},
{
"epoch": 1.4569789674952198,
"grad_norm": 0.4566303964863475,
"learning_rate": 2.445041867912629e-05,
"loss": 0.4141,
"step": 254
},
{
"epoch": 1.4627151051625238,
"grad_norm": 0.41932149526107115,
"learning_rate": 2.431970887908249e-05,
"loss": 0.428,
"step": 255
},
{
"epoch": 1.468451242829828,
"grad_norm": 0.4777606018140004,
"learning_rate": 2.4188805255322895e-05,
"loss": 0.422,
"step": 256
},
{
"epoch": 1.474187380497132,
"grad_norm": 0.4146884687116362,
"learning_rate": 2.405771368144395e-05,
"loss": 0.4407,
"step": 257
},
{
"epoch": 1.479923518164436,
"grad_norm": 0.43435634926442834,
"learning_rate": 2.3926440039475382e-05,
"loss": 0.4004,
"step": 258
},
{
"epoch": 1.48565965583174,
"grad_norm": 0.4184729261005328,
"learning_rate": 2.3794990219616217e-05,
"loss": 0.433,
"step": 259
},
{
"epoch": 1.491395793499044,
"grad_norm": 0.4475976249485021,
"learning_rate": 2.3663370119970527e-05,
"loss": 0.4281,
"step": 260
},
{
"epoch": 1.497131931166348,
"grad_norm": 0.4366125484429576,
"learning_rate": 2.353158564628277e-05,
"loss": 0.3873,
"step": 261
},
{
"epoch": 1.502868068833652,
"grad_norm": 0.4323743190887814,
"learning_rate": 2.339964271167282e-05,
"loss": 0.3906,
"step": 262
},
{
"epoch": 1.508604206500956,
"grad_norm": 0.4349582181643138,
"learning_rate": 2.3267547236370604e-05,
"loss": 0.4153,
"step": 263
},
{
"epoch": 1.51434034416826,
"grad_norm": 0.4300055874488965,
"learning_rate": 2.3135305147450525e-05,
"loss": 0.4243,
"step": 264
},
{
"epoch": 1.520076481835564,
"grad_norm": 0.41191944546871323,
"learning_rate": 2.3002922378565468e-05,
"loss": 0.379,
"step": 265
},
{
"epoch": 1.525812619502868,
"grad_norm": 0.4205849071289299,
"learning_rate": 2.2870404869680573e-05,
"loss": 0.3992,
"step": 266
},
{
"epoch": 1.5315487571701722,
"grad_norm": 0.4167651488072966,
"learning_rate": 2.273775856680672e-05,
"loss": 0.4029,
"step": 267
},
{
"epoch": 1.5372848948374762,
"grad_norm": 0.39362974229347514,
"learning_rate": 2.2604989421733718e-05,
"loss": 0.3964,
"step": 268
},
{
"epoch": 1.5430210325047802,
"grad_norm": 0.4554936101687405,
"learning_rate": 2.2472103391763268e-05,
"loss": 0.4573,
"step": 269
},
{
"epoch": 1.5487571701720841,
"grad_norm": 0.4225697521408909,
"learning_rate": 2.2339106439441656e-05,
"loss": 0.4331,
"step": 270
},
{
"epoch": 1.554493307839388,
"grad_norm": 0.3839293540385542,
"learning_rate": 2.2206004532292192e-05,
"loss": 0.4301,
"step": 271
},
{
"epoch": 1.560229445506692,
"grad_norm": 0.39780099781282624,
"learning_rate": 2.20728036425475e-05,
"loss": 0.4029,
"step": 272
},
{
"epoch": 1.565965583173996,
"grad_norm": 0.39896699881977027,
"learning_rate": 2.1939509746881486e-05,
"loss": 0.4204,
"step": 273
},
{
"epoch": 1.5717017208413002,
"grad_norm": 0.4011711852263823,
"learning_rate": 2.1806128826141223e-05,
"loss": 0.4556,
"step": 274
},
{
"epoch": 1.5774378585086042,
"grad_norm": 0.3780449285964431,
"learning_rate": 2.1672666865078544e-05,
"loss": 0.3854,
"step": 275
},
{
"epoch": 1.5831739961759084,
"grad_norm": 0.3815103294469426,
"learning_rate": 2.153912985208154e-05,
"loss": 0.4304,
"step": 276
},
{
"epoch": 1.5889101338432123,
"grad_norm": 0.39420124826230263,
"learning_rate": 2.140552377890586e-05,
"loss": 0.413,
"step": 277
},
{
"epoch": 1.5946462715105163,
"grad_norm": 0.37547444058482843,
"learning_rate": 2.1271854640405856e-05,
"loss": 0.4033,
"step": 278
},
{
"epoch": 1.6003824091778203,
"grad_norm": 0.3676214604101024,
"learning_rate": 2.1138128434265583e-05,
"loss": 0.3809,
"step": 279
},
{
"epoch": 1.6061185468451242,
"grad_norm": 0.402628978813413,
"learning_rate": 2.1004351160729714e-05,
"loss": 0.4356,
"step": 280
},
{
"epoch": 1.6118546845124282,
"grad_norm": 0.4259466481121817,
"learning_rate": 2.0870528822334296e-05,
"loss": 0.4071,
"step": 281
},
{
"epoch": 1.6175908221797322,
"grad_norm": 0.3980303656074581,
"learning_rate": 2.073666742363742e-05,
"loss": 0.4566,
"step": 282
},
{
"epoch": 1.6233269598470363,
"grad_norm": 0.3793866215066068,
"learning_rate": 2.0602772970949777e-05,
"loss": 0.4084,
"step": 283
},
{
"epoch": 1.6290630975143403,
"grad_norm": 0.4043370635632085,
"learning_rate": 2.0468851472065213e-05,
"loss": 0.4276,
"step": 284
},
{
"epoch": 1.6347992351816445,
"grad_norm": 0.4295543543276556,
"learning_rate": 2.03349089359911e-05,
"loss": 0.3797,
"step": 285
},
{
"epoch": 1.6405353728489485,
"grad_norm": 0.4415259564693047,
"learning_rate": 2.020095137267876e-05,
"loss": 0.4344,
"step": 286
},
{
"epoch": 1.6462715105162524,
"grad_norm": 0.3894742791133544,
"learning_rate": 2.0066984792753757e-05,
"loss": 0.4187,
"step": 287
},
{
"epoch": 1.6520076481835564,
"grad_norm": 0.4128327836910975,
"learning_rate": 1.993301520724625e-05,
"loss": 0.4036,
"step": 288
},
{
"epoch": 1.6577437858508604,
"grad_norm": 0.35013244959729184,
"learning_rate": 1.9799048627321247e-05,
"loss": 0.4141,
"step": 289
},
{
"epoch": 1.6634799235181643,
"grad_norm": 0.3924223109946273,
"learning_rate": 1.966509106400891e-05,
"loss": 0.4098,
"step": 290
},
{
"epoch": 1.6692160611854685,
"grad_norm": 0.375681252496487,
"learning_rate": 1.9531148527934794e-05,
"loss": 0.372,
"step": 291
},
{
"epoch": 1.6749521988527725,
"grad_norm": 0.3583704837247699,
"learning_rate": 1.9397227029050233e-05,
"loss": 0.3977,
"step": 292
},
{
"epoch": 1.6806883365200764,
"grad_norm": 0.4225858356870719,
"learning_rate": 1.9263332576362586e-05,
"loss": 0.4654,
"step": 293
},
{
"epoch": 1.6864244741873806,
"grad_norm": 0.43936703075105055,
"learning_rate": 1.9129471177665714e-05,
"loss": 0.5054,
"step": 294
},
{
"epoch": 1.6921606118546846,
"grad_norm": 0.36215467983870736,
"learning_rate": 1.899564883927029e-05,
"loss": 0.3574,
"step": 295
},
{
"epoch": 1.6978967495219885,
"grad_norm": 0.3732898278812787,
"learning_rate": 1.8861871565734427e-05,
"loss": 0.408,
"step": 296
},
{
"epoch": 1.7036328871892925,
"grad_norm": 0.4033046527093117,
"learning_rate": 1.8728145359594147e-05,
"loss": 0.414,
"step": 297
},
{
"epoch": 1.7093690248565965,
"grad_norm": 0.48233034016482407,
"learning_rate": 1.859447622109415e-05,
"loss": 0.4452,
"step": 298
},
{
"epoch": 1.7151051625239004,
"grad_norm": 0.35022422393865005,
"learning_rate": 1.8460870147918464e-05,
"loss": 0.4098,
"step": 299
},
{
"epoch": 1.7208413001912046,
"grad_norm": 0.3503756135613716,
"learning_rate": 1.832733313492147e-05,
"loss": 0.3665,
"step": 300
},
{
"epoch": 1.7265774378585086,
"grad_norm": 0.453240573589918,
"learning_rate": 1.8193871173858784e-05,
"loss": 0.4017,
"step": 301
},
{
"epoch": 1.7323135755258128,
"grad_norm": 0.4106965211287676,
"learning_rate": 1.8060490253118524e-05,
"loss": 0.457,
"step": 302
},
{
"epoch": 1.7380497131931167,
"grad_norm": 0.3580924920858784,
"learning_rate": 1.7927196357452507e-05,
"loss": 0.3524,
"step": 303
},
{
"epoch": 1.7437858508604207,
"grad_norm": 0.40465378202399943,
"learning_rate": 1.7793995467707808e-05,
"loss": 0.4125,
"step": 304
},
{
"epoch": 1.7495219885277247,
"grad_norm": 0.42671419730243126,
"learning_rate": 1.766089356055835e-05,
"loss": 0.4564,
"step": 305
},
{
"epoch": 1.7552581261950286,
"grad_norm": 0.392872725834381,
"learning_rate": 1.7527896608236735e-05,
"loss": 0.3906,
"step": 306
},
{
"epoch": 1.7609942638623326,
"grad_norm": 0.37117518719191533,
"learning_rate": 1.7395010578266292e-05,
"loss": 0.4225,
"step": 307
},
{
"epoch": 1.7667304015296366,
"grad_norm": 0.36968574125016307,
"learning_rate": 1.7262241433193284e-05,
"loss": 0.3477,
"step": 308
},
{
"epoch": 1.7724665391969407,
"grad_norm": 0.38953378877838585,
"learning_rate": 1.712959513031943e-05,
"loss": 0.3823,
"step": 309
},
{
"epoch": 1.7782026768642447,
"grad_norm": 0.36085297573229624,
"learning_rate": 1.6997077621434535e-05,
"loss": 0.4066,
"step": 310
},
{
"epoch": 1.783938814531549,
"grad_norm": 0.4252176810480222,
"learning_rate": 1.6864694852549478e-05,
"loss": 0.4268,
"step": 311
},
{
"epoch": 1.7896749521988529,
"grad_norm": 0.4412549537519878,
"learning_rate": 1.6732452763629396e-05,
"loss": 0.4326,
"step": 312
},
{
"epoch": 1.7954110898661568,
"grad_norm": 0.3812059474272243,
"learning_rate": 1.6600357288327186e-05,
"loss": 0.379,
"step": 313
},
{
"epoch": 1.8011472275334608,
"grad_norm": 0.47503505255362904,
"learning_rate": 1.646841435371723e-05,
"loss": 0.432,
"step": 314
},
{
"epoch": 1.8068833652007648,
"grad_norm": 0.41481678543056555,
"learning_rate": 1.6336629880029483e-05,
"loss": 0.4023,
"step": 315
},
{
"epoch": 1.8126195028680687,
"grad_norm": 0.4124532353847639,
"learning_rate": 1.6205009780383783e-05,
"loss": 0.4308,
"step": 316
},
{
"epoch": 1.8183556405353727,
"grad_norm": 0.38098821964138535,
"learning_rate": 1.6073559960524624e-05,
"loss": 0.3976,
"step": 317
},
{
"epoch": 1.8240917782026769,
"grad_norm": 0.3811565535669034,
"learning_rate": 1.594228631855605e-05,
"loss": 0.3835,
"step": 318
},
{
"epoch": 1.8298279158699808,
"grad_norm": 0.36589564234813393,
"learning_rate": 1.5811194744677116e-05,
"loss": 0.4129,
"step": 319
},
{
"epoch": 1.835564053537285,
"grad_norm": 0.3653636034932473,
"learning_rate": 1.5680291120917512e-05,
"loss": 0.36,
"step": 320
},
{
"epoch": 1.841300191204589,
"grad_norm": 0.40122714496956696,
"learning_rate": 1.5549581320873715e-05,
"loss": 0.3892,
"step": 321
},
{
"epoch": 1.847036328871893,
"grad_norm": 0.38832451390149825,
"learning_rate": 1.541907120944537e-05,
"loss": 0.3865,
"step": 322
},
{
"epoch": 1.852772466539197,
"grad_norm": 0.437571883751778,
"learning_rate": 1.528876664257219e-05,
"loss": 0.4356,
"step": 323
},
{
"epoch": 1.8585086042065009,
"grad_norm": 0.36968778383694306,
"learning_rate": 1.5158673466971168e-05,
"loss": 0.4257,
"step": 324
},
{
"epoch": 1.8642447418738048,
"grad_norm": 0.35073154051519584,
"learning_rate": 1.502879751987427e-05,
"loss": 0.3947,
"step": 325
},
{
"epoch": 1.869980879541109,
"grad_norm": 0.3805701203026406,
"learning_rate": 1.4899144628766473e-05,
"loss": 0.4197,
"step": 326
},
{
"epoch": 1.875717017208413,
"grad_norm": 0.36051788842397653,
"learning_rate": 1.4769720611124342e-05,
"loss": 0.3862,
"step": 327
},
{
"epoch": 1.8814531548757172,
"grad_norm": 0.390822308734316,
"learning_rate": 1.4640531274154946e-05,
"loss": 0.4547,
"step": 328
},
{
"epoch": 1.8871892925430211,
"grad_norm": 0.35188013033472365,
"learning_rate": 1.4511582414535359e-05,
"loss": 0.372,
"step": 329
},
{
"epoch": 1.892925430210325,
"grad_norm": 0.3492988111143338,
"learning_rate": 1.4382879818152486e-05,
"loss": 0.391,
"step": 330
},
{
"epoch": 1.898661567877629,
"grad_norm": 0.3506998309752745,
"learning_rate": 1.425442925984353e-05,
"loss": 0.3807,
"step": 331
},
{
"epoch": 1.904397705544933,
"grad_norm": 0.4030721263174961,
"learning_rate": 1.41262365031368e-05,
"loss": 0.376,
"step": 332
},
{
"epoch": 1.910133843212237,
"grad_norm": 0.3504037718749043,
"learning_rate": 1.399830729999319e-05,
"loss": 0.3761,
"step": 333
},
{
"epoch": 1.915869980879541,
"grad_norm": 0.3571642792832744,
"learning_rate": 1.3870647390548004e-05,
"loss": 0.3989,
"step": 334
},
{
"epoch": 1.9216061185468452,
"grad_norm": 0.3995469257986992,
"learning_rate": 1.3743262502853458e-05,
"loss": 0.4614,
"step": 335
},
{
"epoch": 1.9273422562141491,
"grad_norm": 0.3374515530857516,
"learning_rate": 1.3616158352621653e-05,
"loss": 0.3545,
"step": 336
},
{
"epoch": 1.9330783938814533,
"grad_norm": 0.3785045797006251,
"learning_rate": 1.3489340642968088e-05,
"loss": 0.3725,
"step": 337
},
{
"epoch": 1.9388145315487573,
"grad_norm": 0.3651730179303691,
"learning_rate": 1.336281506415579e-05,
"loss": 0.4091,
"step": 338
},
{
"epoch": 1.9445506692160612,
"grad_norm": 0.35782792229426036,
"learning_rate": 1.3236587293339974e-05,
"loss": 0.4097,
"step": 339
},
{
"epoch": 1.9502868068833652,
"grad_norm": 0.3715183571626738,
"learning_rate": 1.3110662994313343e-05,
"loss": 0.4238,
"step": 340
},
{
"epoch": 1.9560229445506692,
"grad_norm": 0.3754585864645323,
"learning_rate": 1.2985047817251916e-05,
"loss": 0.4284,
"step": 341
},
{
"epoch": 1.9617590822179731,
"grad_norm": 0.37208739976821165,
"learning_rate": 1.2859747398461542e-05,
"loss": 0.4178,
"step": 342
},
{
"epoch": 1.967495219885277,
"grad_norm": 0.3608587542718569,
"learning_rate": 1.2734767360124955e-05,
"loss": 0.3667,
"step": 343
},
{
"epoch": 1.9732313575525813,
"grad_norm": 0.37784527768886184,
"learning_rate": 1.261011331004958e-05,
"loss": 0.3512,
"step": 344
},
{
"epoch": 1.9789674952198852,
"grad_norm": 0.37238567685296753,
"learning_rate": 1.2485790841415834e-05,
"loss": 0.4461,
"step": 345
},
{
"epoch": 1.9847036328871894,
"grad_norm": 0.3926981132666437,
"learning_rate": 1.2361805532526225e-05,
"loss": 0.4303,
"step": 346
},
{
"epoch": 1.9904397705544934,
"grad_norm": 0.386796605292204,
"learning_rate": 1.2238162946555004e-05,
"loss": 0.4182,
"step": 347
},
{
"epoch": 1.9961759082217974,
"grad_norm": 0.32573202133902224,
"learning_rate": 1.2114868631298608e-05,
"loss": 0.3843,
"step": 348
},
{
"epoch": 2.0019120458891013,
"grad_norm": 0.856534667430809,
"learning_rate": 1.1991928118926661e-05,
"loss": 0.5702,
"step": 349
},
{
"epoch": 2.0076481835564053,
"grad_norm": 0.4595192074096811,
"learning_rate": 1.1869346925733813e-05,
"loss": 0.3478,
"step": 350
},
{
"epoch": 2.0133843212237093,
"grad_norm": 0.8854412030509314,
"learning_rate": 1.1747130551892176e-05,
"loss": 0.3504,
"step": 351
},
{
"epoch": 2.019120458891013,
"grad_norm": 0.5095848588351017,
"learning_rate": 1.1625284481204577e-05,
"loss": 0.3247,
"step": 352
},
{
"epoch": 2.024856596558317,
"grad_norm": 0.45215495379418424,
"learning_rate": 1.1503814180858451e-05,
"loss": 0.3293,
"step": 353
},
{
"epoch": 2.0305927342256216,
"grad_norm": 0.4983547198588867,
"learning_rate": 1.1382725101180593e-05,
"loss": 0.3068,
"step": 354
},
{
"epoch": 2.0363288718929256,
"grad_norm": 0.4664207227186065,
"learning_rate": 1.1262022675392544e-05,
"loss": 0.3031,
"step": 355
},
{
"epoch": 2.0420650095602295,
"grad_norm": 0.4171482692909067,
"learning_rate": 1.1141712319366835e-05,
"loss": 0.3104,
"step": 356
},
{
"epoch": 2.0478011472275335,
"grad_norm": 0.48388082613280514,
"learning_rate": 1.1021799431383969e-05,
"loss": 0.3355,
"step": 357
},
{
"epoch": 2.0535372848948374,
"grad_norm": 0.4405839370097391,
"learning_rate": 1.0902289391890232e-05,
"loss": 0.3181,
"step": 358
},
{
"epoch": 2.0592734225621414,
"grad_norm": 0.38572618966224764,
"learning_rate": 1.0783187563256218e-05,
"loss": 0.255,
"step": 359
},
{
"epoch": 2.0650095602294454,
"grad_norm": 0.4124079647312535,
"learning_rate": 1.0664499289536283e-05,
"loss": 0.3179,
"step": 360
},
{
"epoch": 2.0707456978967493,
"grad_norm": 0.36220901351046625,
"learning_rate": 1.0546229896228692e-05,
"loss": 0.2607,
"step": 361
},
{
"epoch": 2.0764818355640537,
"grad_norm": 0.4049509171035621,
"learning_rate": 1.0428384690036749e-05,
"loss": 0.2878,
"step": 362
},
{
"epoch": 2.0822179732313577,
"grad_norm": 0.48057213951850675,
"learning_rate": 1.0310968958630601e-05,
"loss": 0.3607,
"step": 363
},
{
"epoch": 2.0879541108986617,
"grad_norm": 0.33983057117406296,
"learning_rate": 1.0193987970410046e-05,
"loss": 0.2914,
"step": 364
},
{
"epoch": 2.0936902485659656,
"grad_norm": 0.4031998241867442,
"learning_rate": 1.0077446974268098e-05,
"loss": 0.3372,
"step": 365
},
{
"epoch": 2.0994263862332696,
"grad_norm": 0.3811905438226048,
"learning_rate": 9.961351199355513e-06,
"loss": 0.311,
"step": 366
},
{
"epoch": 2.1051625239005736,
"grad_norm": 0.38408888784962325,
"learning_rate": 9.84570585484614e-06,
"loss": 0.3563,
"step": 367
},
{
"epoch": 2.1108986615678775,
"grad_norm": 0.33981693233156296,
"learning_rate": 9.730516129703158e-06,
"loss": 0.2871,
"step": 368
},
{
"epoch": 2.1166347992351815,
"grad_norm": 0.3518672236516868,
"learning_rate": 9.615787192446304e-06,
"loss": 0.3125,
"step": 369
},
{
"epoch": 2.1223709369024855,
"grad_norm": 0.3802739825741615,
"learning_rate": 9.50152419091992e-06,
"loss": 0.3254,
"step": 370
},
{
"epoch": 2.12810707456979,
"grad_norm": 0.3350283722430752,
"learning_rate": 9.38773225206199e-06,
"loss": 0.2669,
"step": 371
},
{
"epoch": 2.133843212237094,
"grad_norm": 0.37579814126338135,
"learning_rate": 9.274416481674084e-06,
"loss": 0.334,
"step": 372
},
{
"epoch": 2.139579349904398,
"grad_norm": 0.32704016552087445,
"learning_rate": 9.161581964192298e-06,
"loss": 0.2568,
"step": 373
},
{
"epoch": 2.1453154875717018,
"grad_norm": 0.414672276708601,
"learning_rate": 9.049233762459057e-06,
"loss": 0.4001,
"step": 374
},
{
"epoch": 2.1510516252390057,
"grad_norm": 0.3112648437196225,
"learning_rate": 8.937376917496012e-06,
"loss": 0.2447,
"step": 375
},
{
"epoch": 2.1567877629063097,
"grad_norm": 0.3535914113780811,
"learning_rate": 8.826016448277795e-06,
"loss": 0.3157,
"step": 376
},
{
"epoch": 2.1625239005736137,
"grad_norm": 0.3434081422193142,
"learning_rate": 8.715157351506864e-06,
"loss": 0.2992,
"step": 377
},
{
"epoch": 2.1682600382409176,
"grad_norm": 0.3422884659233757,
"learning_rate": 8.604804601389271e-06,
"loss": 0.3132,
"step": 378
},
{
"epoch": 2.173996175908222,
"grad_norm": 0.3430270987205145,
"learning_rate": 8.494963149411489e-06,
"loss": 0.2712,
"step": 379
},
{
"epoch": 2.179732313575526,
"grad_norm": 0.36832674053444375,
"learning_rate": 8.385637924118224e-06,
"loss": 0.2803,
"step": 380
},
{
"epoch": 2.18546845124283,
"grad_norm": 0.338652106601731,
"learning_rate": 8.276833830891312e-06,
"loss": 0.2937,
"step": 381
},
{
"epoch": 2.191204588910134,
"grad_norm": 0.347795356928778,
"learning_rate": 8.168555751729552e-06,
"loss": 0.3159,
"step": 382
},
{
"epoch": 2.196940726577438,
"grad_norm": 0.37649404917042467,
"learning_rate": 8.060808545029727e-06,
"loss": 0.3159,
"step": 383
},
{
"epoch": 2.202676864244742,
"grad_norm": 0.4275027691702314,
"learning_rate": 7.95359704536854e-06,
"loss": 0.3051,
"step": 384
},
{
"epoch": 2.208413001912046,
"grad_norm": 0.36709280229253105,
"learning_rate": 7.846926063285745e-06,
"loss": 0.3082,
"step": 385
},
{
"epoch": 2.21414913957935,
"grad_norm": 0.3197758439418705,
"learning_rate": 7.740800385068256e-06,
"loss": 0.3044,
"step": 386
},
{
"epoch": 2.2198852772466537,
"grad_norm": 0.36467098789844404,
"learning_rate": 7.63522477253542e-06,
"loss": 0.3192,
"step": 387
},
{
"epoch": 2.2256214149139577,
"grad_norm": 0.3792955542610236,
"learning_rate": 7.530203962825331e-06,
"loss": 0.3374,
"step": 388
},
{
"epoch": 2.231357552581262,
"grad_norm": 0.33901384143705177,
"learning_rate": 7.425742668182308e-06,
"loss": 0.3056,
"step": 389
},
{
"epoch": 2.237093690248566,
"grad_norm": 0.33544808925545067,
"learning_rate": 7.3218455757454125e-06,
"loss": 0.2674,
"step": 390
},
{
"epoch": 2.24282982791587,
"grad_norm": 0.32597707317340413,
"learning_rate": 7.218517347338194e-06,
"loss": 0.2843,
"step": 391
},
{
"epoch": 2.248565965583174,
"grad_norm": 0.3555015121926318,
"learning_rate": 7.115762619259459e-06,
"loss": 0.3234,
"step": 392
},
{
"epoch": 2.254302103250478,
"grad_norm": 0.3593977217633907,
"learning_rate": 7.013586002075297e-06,
"loss": 0.3573,
"step": 393
},
{
"epoch": 2.260038240917782,
"grad_norm": 0.326928535932432,
"learning_rate": 6.911992080412153e-06,
"loss": 0.3053,
"step": 394
},
{
"epoch": 2.265774378585086,
"grad_norm": 0.3661810608640239,
"learning_rate": 6.810985412751159e-06,
"loss": 0.3411,
"step": 395
},
{
"epoch": 2.27151051625239,
"grad_norm": 0.33182512098109557,
"learning_rate": 6.710570531223568e-06,
"loss": 0.2988,
"step": 396
},
{
"epoch": 2.2772466539196943,
"grad_norm": 0.37198993314527246,
"learning_rate": 6.610751941407423e-06,
"loss": 0.2987,
"step": 397
},
{
"epoch": 2.2829827915869982,
"grad_norm": 0.3605089217084053,
"learning_rate": 6.511534122125385e-06,
"loss": 0.3126,
"step": 398
},
{
"epoch": 2.288718929254302,
"grad_norm": 0.35417912926194517,
"learning_rate": 6.412921525243747e-06,
"loss": 0.3179,
"step": 399
},
{
"epoch": 2.294455066921606,
"grad_norm": 0.34739676511363254,
"learning_rate": 6.314918575472724e-06,
"loss": 0.3213,
"step": 400
},
{
"epoch": 2.30019120458891,
"grad_norm": 0.34538986538232874,
"learning_rate": 6.2175296701678765e-06,
"loss": 0.3267,
"step": 401
},
{
"epoch": 2.305927342256214,
"grad_norm": 0.33419006023335046,
"learning_rate": 6.120759179132825e-06,
"loss": 0.3022,
"step": 402
},
{
"epoch": 2.311663479923518,
"grad_norm": 0.3490791821453638,
"learning_rate": 6.024611444423167e-06,
"loss": 0.3331,
"step": 403
},
{
"epoch": 2.317399617590822,
"grad_norm": 0.3423057899783048,
"learning_rate": 5.929090780151683e-06,
"loss": 0.3235,
"step": 404
},
{
"epoch": 2.323135755258126,
"grad_norm": 0.3160266995434583,
"learning_rate": 5.8342014722947025e-06,
"loss": 0.2911,
"step": 405
},
{
"epoch": 2.3288718929254304,
"grad_norm": 0.3271039700649619,
"learning_rate": 5.739947778499866e-06,
"loss": 0.2625,
"step": 406
},
{
"epoch": 2.3346080305927344,
"grad_norm": 0.357397125079873,
"learning_rate": 5.6463339278950135e-06,
"loss": 0.2686,
"step": 407
},
{
"epoch": 2.3403441682600383,
"grad_norm": 0.3626417105072833,
"learning_rate": 5.553364120898495e-06,
"loss": 0.3078,
"step": 408
},
{
"epoch": 2.3460803059273423,
"grad_norm": 0.330368230833986,
"learning_rate": 5.461042529030643e-06,
"loss": 0.2669,
"step": 409
},
{
"epoch": 2.3518164435946463,
"grad_norm": 0.35751740719995606,
"learning_rate": 5.369373294726625e-06,
"loss": 0.3035,
"step": 410
},
{
"epoch": 2.35755258126195,
"grad_norm": 0.35220579168854316,
"learning_rate": 5.2783605311505705e-06,
"loss": 0.2959,
"step": 411
},
{
"epoch": 2.363288718929254,
"grad_norm": 0.3719450476826788,
"learning_rate": 5.188008322011022e-06,
"loss": 0.3271,
"step": 412
},
{
"epoch": 2.369024856596558,
"grad_norm": 0.3413354729055759,
"learning_rate": 5.098320721377677e-06,
"loss": 0.2821,
"step": 413
},
{
"epoch": 2.3747609942638626,
"grad_norm": 0.31970016663963907,
"learning_rate": 5.00930175349952e-06,
"loss": 0.2978,
"step": 414
},
{
"epoch": 2.3804971319311665,
"grad_norm": 0.3405014390669701,
"learning_rate": 4.920955412624206e-06,
"loss": 0.3298,
"step": 415
},
{
"epoch": 2.3862332695984705,
"grad_norm": 0.32022737076210306,
"learning_rate": 4.833285662818903e-06,
"loss": 0.3205,
"step": 416
},
{
"epoch": 2.3919694072657744,
"grad_norm": 0.326492045130511,
"learning_rate": 4.746296437792364e-06,
"loss": 0.3073,
"step": 417
},
{
"epoch": 2.3977055449330784,
"grad_norm": 0.32202773276443253,
"learning_rate": 4.65999164071846e-06,
"loss": 0.3132,
"step": 418
},
{
"epoch": 2.4034416826003824,
"grad_norm": 0.30026610991923225,
"learning_rate": 4.5743751440610315e-06,
"loss": 0.2999,
"step": 419
},
{
"epoch": 2.4091778202676863,
"grad_norm": 0.3263265838303172,
"learning_rate": 4.489450789400158e-06,
"loss": 0.3396,
"step": 420
},
{
"epoch": 2.4149139579349903,
"grad_norm": 0.31887864810796523,
"learning_rate": 4.405222387259737e-06,
"loss": 0.2785,
"step": 421
},
{
"epoch": 2.4206500956022943,
"grad_norm": 0.33803434408500127,
"learning_rate": 4.32169371693657e-06,
"loss": 0.3396,
"step": 422
},
{
"epoch": 2.4263862332695982,
"grad_norm": 0.30385763232746726,
"learning_rate": 4.238868526330722e-06,
"loss": 0.2888,
"step": 423
},
{
"epoch": 2.4321223709369026,
"grad_norm": 0.32177073146894075,
"learning_rate": 4.156750531777414e-06,
"loss": 0.2814,
"step": 424
},
{
"epoch": 2.4378585086042066,
"grad_norm": 0.3331670974871312,
"learning_rate": 4.075343417880233e-06,
"loss": 0.3166,
"step": 425
},
{
"epoch": 2.4435946462715106,
"grad_norm": 0.34367786083186397,
"learning_rate": 3.994650837345817e-06,
"loss": 0.2981,
"step": 426
},
{
"epoch": 2.4493307839388145,
"grad_norm": 0.3379368579815228,
"learning_rate": 3.914676410819957e-06,
"loss": 0.3439,
"step": 427
},
{
"epoch": 2.4550669216061185,
"grad_norm": 0.32413796456612975,
"learning_rate": 3.835423726725162e-06,
"loss": 0.3026,
"step": 428
},
{
"epoch": 2.4608030592734225,
"grad_norm": 0.33686529783092345,
"learning_rate": 3.7568963410996028e-06,
"loss": 0.3366,
"step": 429
},
{
"epoch": 2.4665391969407264,
"grad_norm": 0.29819612441201854,
"learning_rate": 3.6790977774376013e-06,
"loss": 0.2838,
"step": 430
},
{
"epoch": 2.472275334608031,
"grad_norm": 0.33348914715995587,
"learning_rate": 3.6020315265315087e-06,
"loss": 0.3133,
"step": 431
},
{
"epoch": 2.478011472275335,
"grad_norm": 0.30518415557095047,
"learning_rate": 3.5257010463150642e-06,
"loss": 0.3132,
"step": 432
},
{
"epoch": 2.4837476099426388,
"grad_norm": 0.33485851224923785,
"learning_rate": 3.4501097617082648e-06,
"loss": 0.3567,
"step": 433
},
{
"epoch": 2.4894837476099427,
"grad_norm": 0.3176169459499719,
"learning_rate": 3.375261064463666e-06,
"loss": 0.3047,
"step": 434
},
{
"epoch": 2.4952198852772467,
"grad_norm": 0.35821675236801026,
"learning_rate": 3.3011583130142277e-06,
"loss": 0.3054,
"step": 435
},
{
"epoch": 2.5009560229445507,
"grad_norm": 0.31447192397671203,
"learning_rate": 3.2278048323225764e-06,
"loss": 0.3002,
"step": 436
},
{
"epoch": 2.5066921606118546,
"grad_norm": 0.29814490505492547,
"learning_rate": 3.155203913731868e-06,
"loss": 0.3145,
"step": 437
},
{
"epoch": 2.5124282982791586,
"grad_norm": 0.31338084805310423,
"learning_rate": 3.0833588148180516e-06,
"loss": 0.316,
"step": 438
},
{
"epoch": 2.5181644359464626,
"grad_norm": 0.33273295840043154,
"learning_rate": 3.0122727592437574e-06,
"loss": 0.3052,
"step": 439
},
{
"epoch": 2.5239005736137665,
"grad_norm": 0.3586562727340161,
"learning_rate": 2.941948936613608e-06,
"loss": 0.3584,
"step": 440
},
{
"epoch": 2.5296367112810705,
"grad_norm": 0.3149636280687063,
"learning_rate": 2.8723905023311237e-06,
"loss": 0.3323,
"step": 441
},
{
"epoch": 2.535372848948375,
"grad_norm": 0.29354487416787345,
"learning_rate": 2.8036005774571373e-06,
"loss": 0.27,
"step": 442
},
{
"epoch": 2.541108986615679,
"grad_norm": 0.32300649834326484,
"learning_rate": 2.7355822485697615e-06,
"loss": 0.2927,
"step": 443
},
{
"epoch": 2.546845124282983,
"grad_norm": 0.32585893999440413,
"learning_rate": 2.668338567625872e-06,
"loss": 0.2917,
"step": 444
},
{
"epoch": 2.552581261950287,
"grad_norm": 0.3206957295285485,
"learning_rate": 2.601872551824196e-06,
"loss": 0.3305,
"step": 445
},
{
"epoch": 2.5583173996175907,
"grad_norm": 0.31706654167791043,
"learning_rate": 2.536187183469905e-06,
"loss": 0.3036,
"step": 446
},
{
"epoch": 2.5640535372848947,
"grad_norm": 0.29637565255206266,
"learning_rate": 2.4712854098408245e-06,
"loss": 0.296,
"step": 447
},
{
"epoch": 2.569789674952199,
"grad_norm": 0.3242672724124806,
"learning_rate": 2.407170143055173e-06,
"loss": 0.3165,
"step": 448
},
{
"epoch": 2.575525812619503,
"grad_norm": 0.31283308105656626,
"learning_rate": 2.3438442599408995e-06,
"loss": 0.2685,
"step": 449
},
{
"epoch": 2.581261950286807,
"grad_norm": 0.31850570032908976,
"learning_rate": 2.2813106019066055e-06,
"loss": 0.2677,
"step": 450
},
{
"epoch": 2.586998087954111,
"grad_norm": 0.33792122182008616,
"learning_rate": 2.219571974814059e-06,
"loss": 0.3382,
"step": 451
},
{
"epoch": 2.592734225621415,
"grad_norm": 0.31634274129393675,
"learning_rate": 2.1586311488522705e-06,
"loss": 0.2952,
"step": 452
},
{
"epoch": 2.598470363288719,
"grad_norm": 0.3044886559468434,
"learning_rate": 2.0984908584132356e-06,
"loss": 0.322,
"step": 453
},
{
"epoch": 2.604206500956023,
"grad_norm": 0.3025610232808951,
"learning_rate": 2.0391538019691983e-06,
"loss": 0.2895,
"step": 454
},
{
"epoch": 2.609942638623327,
"grad_norm": 0.31674447409176193,
"learning_rate": 1.9806226419516195e-06,
"loss": 0.3372,
"step": 455
},
{
"epoch": 2.615678776290631,
"grad_norm": 0.3163602428641736,
"learning_rate": 1.922900004631667e-06,
"loss": 0.2956,
"step": 456
},
{
"epoch": 2.621414913957935,
"grad_norm": 0.29617871455956285,
"learning_rate": 1.8659884800024119e-06,
"loss": 0.281,
"step": 457
},
{
"epoch": 2.6271510516252388,
"grad_norm": 0.3054337634119753,
"learning_rate": 1.8098906216625934e-06,
"loss": 0.2779,
"step": 458
},
{
"epoch": 2.632887189292543,
"grad_norm": 0.3473718579857251,
"learning_rate": 1.7546089467020677e-06,
"loss": 0.3322,
"step": 459
},
{
"epoch": 2.638623326959847,
"grad_norm": 0.3369008816792107,
"learning_rate": 1.700145935588826e-06,
"loss": 0.333,
"step": 460
},
{
"epoch": 2.644359464627151,
"grad_norm": 0.3127691638231103,
"learning_rate": 1.6465040320577408e-06,
"loss": 0.3012,
"step": 461
},
{
"epoch": 2.650095602294455,
"grad_norm": 0.30464242733494856,
"learning_rate": 1.593685643000884e-06,
"loss": 0.2875,
"step": 462
},
{
"epoch": 2.655831739961759,
"grad_norm": 0.33106117803370116,
"learning_rate": 1.5416931383595436e-06,
"loss": 0.3258,
"step": 463
},
{
"epoch": 2.661567877629063,
"grad_norm": 0.3246679490688405,
"learning_rate": 1.490528851017885e-06,
"loss": 0.3027,
"step": 464
},
{
"epoch": 2.667304015296367,
"grad_norm": 0.30458861657265374,
"learning_rate": 1.440195076698272e-06,
"loss": 0.2916,
"step": 465
},
{
"epoch": 2.6730401529636714,
"grad_norm": 0.3186347058123626,
"learning_rate": 1.3906940738582698e-06,
"loss": 0.2863,
"step": 466
},
{
"epoch": 2.6787762906309753,
"grad_norm": 0.31369585810148204,
"learning_rate": 1.3420280635892847e-06,
"loss": 0.3104,
"step": 467
},
{
"epoch": 2.6845124282982793,
"grad_norm": 0.316446279408055,
"learning_rate": 1.2941992295169369e-06,
"loss": 0.3303,
"step": 468
},
{
"epoch": 2.6902485659655833,
"grad_norm": 0.32320749119463515,
"learning_rate": 1.2472097177030485e-06,
"loss": 0.2923,
"step": 469
},
{
"epoch": 2.6959847036328872,
"grad_norm": 0.3397456933911356,
"learning_rate": 1.2010616365493811e-06,
"loss": 0.3006,
"step": 470
},
{
"epoch": 2.701720841300191,
"grad_norm": 0.3135400345614286,
"learning_rate": 1.1557570567030108e-06,
"loss": 0.2936,
"step": 471
},
{
"epoch": 2.707456978967495,
"grad_norm": 0.3021698093355975,
"learning_rate": 1.1112980109634263e-06,
"loss": 0.3341,
"step": 472
},
{
"epoch": 2.713193116634799,
"grad_norm": 0.30592288821677366,
"learning_rate": 1.067686494191318e-06,
"loss": 0.318,
"step": 473
},
{
"epoch": 2.718929254302103,
"grad_norm": 0.3320062239605175,
"learning_rate": 1.0249244632190769e-06,
"loss": 0.3299,
"step": 474
},
{
"epoch": 2.724665391969407,
"grad_norm": 0.30860296916444196,
"learning_rate": 9.83013836762976e-07,
"loss": 0.2836,
"step": 475
},
{
"epoch": 2.730401529636711,
"grad_norm": 0.3361749125512604,
"learning_rate": 9.419564953370952e-07,
"loss": 0.3298,
"step": 476
},
{
"epoch": 2.7361376673040154,
"grad_norm": 0.30689246179518487,
"learning_rate": 9.017542811689272e-07,
"loss": 0.3033,
"step": 477
},
{
"epoch": 2.7418738049713194,
"grad_norm": 0.2838154253297985,
"learning_rate": 8.624089981167349e-07,
"loss": 0.2356,
"step": 478
},
{
"epoch": 2.7476099426386233,
"grad_norm": 0.31969377880658495,
"learning_rate": 8.239224115885957e-07,
"loss": 0.3049,
"step": 479
},
{
"epoch": 2.7533460803059273,
"grad_norm": 0.30060258657145644,
"learning_rate": 7.862962484631986e-07,
"loss": 0.3063,
"step": 480
},
{
"epoch": 2.7590822179732313,
"grad_norm": 0.3022284718933698,
"learning_rate": 7.49532197012357e-07,
"loss": 0.2886,
"step": 481
},
{
"epoch": 2.7648183556405352,
"grad_norm": 0.31537349530205033,
"learning_rate": 7.136319068252629e-07,
"loss": 0.327,
"step": 482
},
{
"epoch": 2.7705544933078396,
"grad_norm": 0.3214439225809495,
"learning_rate": 6.785969887344546e-07,
"loss": 0.3239,
"step": 483
},
{
"epoch": 2.7762906309751436,
"grad_norm": 0.30881041163907574,
"learning_rate": 6.444290147435617e-07,
"loss": 0.2923,
"step": 484
},
{
"epoch": 2.7820267686424476,
"grad_norm": 0.2968321026529436,
"learning_rate": 6.111295179567434e-07,
"loss": 0.2962,
"step": 485
},
{
"epoch": 2.7877629063097515,
"grad_norm": 0.32452460610512424,
"learning_rate": 5.786999925099257e-07,
"loss": 0.3019,
"step": 486
},
{
"epoch": 2.7934990439770555,
"grad_norm": 0.28214977997629037,
"learning_rate": 5.471418935037398e-07,
"loss": 0.2949,
"step": 487
},
{
"epoch": 2.7992351816443595,
"grad_norm": 0.30056333465804613,
"learning_rate": 5.164566369382407e-07,
"loss": 0.3396,
"step": 488
},
{
"epoch": 2.8049713193116634,
"grad_norm": 0.3025564251244908,
"learning_rate": 4.866455996493691e-07,
"loss": 0.2909,
"step": 489
},
{
"epoch": 2.8107074569789674,
"grad_norm": 0.30674628776531615,
"learning_rate": 4.577101192471811e-07,
"loss": 0.3128,
"step": 490
},
{
"epoch": 2.8164435946462714,
"grad_norm": 0.31971943977257217,
"learning_rate": 4.296514940558161e-07,
"loss": 0.3299,
"step": 491
},
{
"epoch": 2.8221797323135753,
"grad_norm": 0.32117510140100713,
"learning_rate": 4.0247098305525645e-07,
"loss": 0.3237,
"step": 492
},
{
"epoch": 2.8279158699808793,
"grad_norm": 0.30441550056900496,
"learning_rate": 3.7616980582482866e-07,
"loss": 0.3431,
"step": 493
},
{
"epoch": 2.8336520076481837,
"grad_norm": 0.2937689706082808,
"learning_rate": 3.507491424884779e-07,
"loss": 0.279,
"step": 494
},
{
"epoch": 2.8393881453154877,
"grad_norm": 0.3243702100672169,
"learning_rate": 3.262101336618262e-07,
"loss": 0.3258,
"step": 495
},
{
"epoch": 2.8451242829827916,
"grad_norm": 0.29048595506865943,
"learning_rate": 3.0255388040098864e-07,
"loss": 0.3202,
"step": 496
},
{
"epoch": 2.8508604206500956,
"grad_norm": 0.32709366282528524,
"learning_rate": 2.7978144415316656e-07,
"loss": 0.3105,
"step": 497
},
{
"epoch": 2.8565965583173996,
"grad_norm": 0.28160883870197856,
"learning_rate": 2.5789384670902753e-07,
"loss": 0.2349,
"step": 498
},
{
"epoch": 2.8623326959847035,
"grad_norm": 0.32964689233644207,
"learning_rate": 2.3689207015685334e-07,
"loss": 0.3767,
"step": 499
},
{
"epoch": 2.8680688336520075,
"grad_norm": 0.30156232678290884,
"learning_rate": 2.1677705683847082e-07,
"loss": 0.2827,
"step": 500
},
{
"epoch": 2.873804971319312,
"grad_norm": 0.29788501422144353,
"learning_rate": 1.9754970930698115e-07,
"loss": 0.2988,
"step": 501
},
{
"epoch": 2.879541108986616,
"grad_norm": 0.3025299588676407,
"learning_rate": 1.792108902862455e-07,
"loss": 0.2925,
"step": 502
},
{
"epoch": 2.88527724665392,
"grad_norm": 0.3248363141852775,
"learning_rate": 1.6176142263219173e-07,
"loss": 0.3148,
"step": 503
},
{
"epoch": 2.891013384321224,
"grad_norm": 0.3189914861271819,
"learning_rate": 1.4520208929587942e-07,
"loss": 0.2563,
"step": 504
},
{
"epoch": 2.8967495219885278,
"grad_norm": 0.2861804455486256,
"learning_rate": 1.2953363328838342e-07,
"loss": 0.2708,
"step": 505
},
{
"epoch": 2.9024856596558317,
"grad_norm": 0.3137865936214283,
"learning_rate": 1.1475675764743843e-07,
"loss": 0.3803,
"step": 506
},
{
"epoch": 2.9082217973231357,
"grad_norm": 0.321251565354294,
"learning_rate": 1.0087212540591307e-07,
"loss": 0.3431,
"step": 507
},
{
"epoch": 2.9139579349904396,
"grad_norm": 0.2951230413126827,
"learning_rate": 8.78803595620381e-08,
"loss": 0.27,
"step": 508
},
{
"epoch": 2.9196940726577436,
"grad_norm": 0.3247014517591854,
"learning_rate": 7.57820430514733e-08,
"loss": 0.2853,
"step": 509
},
{
"epoch": 2.9254302103250476,
"grad_norm": 0.3229719006514364,
"learning_rate": 6.457771872113716e-08,
"loss": 0.3274,
"step": 510
},
{
"epoch": 2.9311663479923515,
"grad_norm": 0.30781258564862424,
"learning_rate": 5.4267889304859824e-08,
"loss": 0.2917,
"step": 511
},
{
"epoch": 2.936902485659656,
"grad_norm": 0.3008043286280689,
"learning_rate": 4.485301740080994e-08,
"loss": 0.3029,
"step": 512
},
{
"epoch": 2.94263862332696,
"grad_norm": 0.2969030376672524,
"learning_rate": 3.633352545076241e-08,
"loss": 0.2627,
"step": 513
},
{
"epoch": 2.948374760994264,
"grad_norm": 0.2974333758480683,
"learning_rate": 2.8709795721117984e-08,
"loss": 0.3265,
"step": 514
},
{
"epoch": 2.954110898661568,
"grad_norm": 0.3173221368962474,
"learning_rate": 2.198217028577254e-08,
"loss": 0.3039,
"step": 515
},
{
"epoch": 2.959847036328872,
"grad_norm": 0.33002144156652646,
"learning_rate": 1.6150951010747152e-08,
"loss": 0.2785,
"step": 516
},
{
"epoch": 2.9655831739961758,
"grad_norm": 0.30529921643631025,
"learning_rate": 1.1216399540669998e-08,
"loss": 0.2788,
"step": 517
},
{
"epoch": 2.97131931166348,
"grad_norm": 0.31427963839680284,
"learning_rate": 7.178737287005799e-09,
"loss": 0.3469,
"step": 518
},
{
"epoch": 2.977055449330784,
"grad_norm": 0.31476300056888123,
"learning_rate": 4.038145418148176e-09,
"loss": 0.2671,
"step": 519
},
{
"epoch": 2.982791586998088,
"grad_norm": 0.3064284484610742,
"learning_rate": 1.7947648512728343e-09,
"loss": 0.32,
"step": 520
},
{
"epoch": 2.988527724665392,
"grad_norm": 0.30798998882484835,
"learning_rate": 4.486962460270583e-10,
"loss": 0.3146,
"step": 521
},
{
"epoch": 2.994263862332696,
"grad_norm": 0.28978511666305123,
"learning_rate": 0.0,
"loss": 0.2861,
"step": 522
},
{
"epoch": 2.994263862332696,
"step": 522,
"total_flos": 5.620362022127534e+17,
"train_loss": 0.4276364711452261,
"train_runtime": 8560.5176,
"train_samples_per_second": 5.856,
"train_steps_per_second": 0.061
}
],
"logging_steps": 1.0,
"max_steps": 522,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.620362022127534e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}