P4Ms-hackathon-vision-task / shadow_lmm /trainer_state.json
IvoHoese's picture
Upload 8 files
4367e37 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9996906897618311,
"eval_steps": 500,
"global_step": 1414,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0007069948301003049,
"grad_norm": 867.8281860351562,
"learning_rate": 1e-06,
"loss": 69.0521,
"step": 1
},
{
"epoch": 0.0014139896602006097,
"grad_norm": 522.1878662109375,
"learning_rate": 1.9244594481721914e-05,
"loss": 75.8844,
"step": 2
},
{
"epoch": 0.0021209844903009147,
"grad_norm": 447.0108337402344,
"learning_rate": 2.991699809439337e-05,
"loss": 74.97,
"step": 3
},
{
"epoch": 0.0028279793204012194,
"grad_norm": 5898.1640625,
"learning_rate": 3.748918896344382e-05,
"loss": 81.0476,
"step": 4
},
{
"epoch": 0.0035349741505015246,
"grad_norm": 226.5022430419922,
"learning_rate": 4.336263650693704e-05,
"loss": 68.3264,
"step": 5
},
{
"epoch": 0.004241968980601829,
"grad_norm": 112.70651245117188,
"learning_rate": 4.8161592576115274e-05,
"loss": 60.3818,
"step": 6
},
{
"epoch": 0.0049489638107021346,
"grad_norm": 93.56769561767578,
"learning_rate": 5.2219052119207004e-05,
"loss": 59.0934,
"step": 7
},
{
"epoch": 0.005655958640802439,
"grad_norm": 241.02914428710938,
"learning_rate": 5.5733783445165726e-05,
"loss": 58.6418,
"step": 8
},
{
"epoch": 0.006362953470902744,
"grad_norm": 431.45977783203125,
"learning_rate": 5.883399618878674e-05,
"loss": 58.1764,
"step": 9
},
{
"epoch": 0.007069948301003049,
"grad_norm": 349.70782470703125,
"learning_rate": 6.160723098865897e-05,
"loss": 56.8595,
"step": 10
},
{
"epoch": 0.007776943131103354,
"grad_norm": 48.183162689208984,
"learning_rate": 6.411592701928434e-05,
"loss": 56.9374,
"step": 11
},
{
"epoch": 0.008483937961203659,
"grad_norm": 28.044233322143555,
"learning_rate": 6.64061870578372e-05,
"loss": 55.5306,
"step": 12
},
{
"epoch": 0.009190932791303964,
"grad_norm": 30.87534523010254,
"learning_rate": 6.851302206154156e-05,
"loss": 55.0557,
"step": 13
},
{
"epoch": 0.009897927621404269,
"grad_norm": 24.83538818359375,
"learning_rate": 7.046364660092892e-05,
"loss": 50.3177,
"step": 14
},
{
"epoch": 0.010604922451504573,
"grad_norm": 28.679935455322266,
"learning_rate": 7.22796346013304e-05,
"loss": 54.1606,
"step": 15
},
{
"epoch": 0.011311917281604878,
"grad_norm": 29.871997833251953,
"learning_rate": 7.397837792688765e-05,
"loss": 52.391,
"step": 16
},
{
"epoch": 0.012018912111705183,
"grad_norm": 25.709745407104492,
"learning_rate": 7.557410199771932e-05,
"loss": 52.2417,
"step": 17
},
{
"epoch": 0.012725906941805488,
"grad_norm": 25.342071533203125,
"learning_rate": 7.707859067050864e-05,
"loss": 55.4544,
"step": 18
},
{
"epoch": 0.013432901771905793,
"grad_norm": 25.05597686767578,
"learning_rate": 7.850171487052752e-05,
"loss": 52.1423,
"step": 19
},
{
"epoch": 0.014139896602006099,
"grad_norm": 23.790668487548828,
"learning_rate": 7.985182547038086e-05,
"loss": 52.9892,
"step": 20
},
{
"epoch": 0.014846891432106402,
"grad_norm": 24.078054428100586,
"learning_rate": 8.113605021360038e-05,
"loss": 52.7495,
"step": 21
},
{
"epoch": 0.015553886262206707,
"grad_norm": 22.381742477416992,
"learning_rate": 8.236052150100626e-05,
"loss": 52.8961,
"step": 22
},
{
"epoch": 0.016260881092307014,
"grad_norm": 23.896352767944336,
"learning_rate": 8.353055350120496e-05,
"loss": 51.6978,
"step": 23
},
{
"epoch": 0.016967875922407318,
"grad_norm": 22.911603927612305,
"learning_rate": 8.46507815395591e-05,
"loss": 51.3604,
"step": 24
},
{
"epoch": 0.01767487075250762,
"grad_norm": 22.95720672607422,
"learning_rate": 8.572527301387408e-05,
"loss": 50.2589,
"step": 25
},
{
"epoch": 0.018381865582607928,
"grad_norm": 24.270532608032227,
"learning_rate": 8.675761654326347e-05,
"loss": 53.2253,
"step": 26
},
{
"epoch": 0.01908886041270823,
"grad_norm": 22.97658920288086,
"learning_rate": 8.77509942831801e-05,
"loss": 51.9815,
"step": 27
},
{
"epoch": 0.019795855242808538,
"grad_norm": 24.958389282226562,
"learning_rate": 8.870824108265084e-05,
"loss": 52.542,
"step": 28
},
{
"epoch": 0.02050285007290884,
"grad_norm": 22.157556533813477,
"learning_rate": 8.963189325601443e-05,
"loss": 49.7963,
"step": 29
},
{
"epoch": 0.021209844903009145,
"grad_norm": 21.693815231323242,
"learning_rate": 9.052422908305234e-05,
"loss": 50.776,
"step": 30
},
{
"epoch": 0.021916839733109452,
"grad_norm": 23.532215118408203,
"learning_rate": 9.138730266585143e-05,
"loss": 53.4715,
"step": 31
},
{
"epoch": 0.022623834563209756,
"grad_norm": 23.66779899597168,
"learning_rate": 9.222297240860955e-05,
"loss": 51.4714,
"step": 32
},
{
"epoch": 0.023330829393310062,
"grad_norm": 21.056167602539062,
"learning_rate": 9.30329251136777e-05,
"loss": 50.4526,
"step": 33
},
{
"epoch": 0.024037824223410366,
"grad_norm": 20.85773468017578,
"learning_rate": 9.381869647944122e-05,
"loss": 49.2931,
"step": 34
},
{
"epoch": 0.024744819053510673,
"grad_norm": 21.533872604370117,
"learning_rate": 9.458168862614404e-05,
"loss": 49.6295,
"step": 35
},
{
"epoch": 0.025451813883610976,
"grad_norm": 23.687660217285156,
"learning_rate": 9.532318515223055e-05,
"loss": 51.6082,
"step": 36
},
{
"epoch": 0.02615880871371128,
"grad_norm": 25.589447021484375,
"learning_rate": 9.604436412734158e-05,
"loss": 49.9036,
"step": 37
},
{
"epoch": 0.026865803543811587,
"grad_norm": 23.40888786315918,
"learning_rate": 9.674630935224942e-05,
"loss": 49.4263,
"step": 38
},
{
"epoch": 0.02757279837391189,
"grad_norm": 24.978803634643555,
"learning_rate": 9.743002015593493e-05,
"loss": 50.1287,
"step": 39
},
{
"epoch": 0.028279793204012197,
"grad_norm": 27.688739776611328,
"learning_rate": 9.809641995210277e-05,
"loss": 49.186,
"step": 40
},
{
"epoch": 0.0289867880341125,
"grad_norm": 28.714271545410156,
"learning_rate": 9.874636373899326e-05,
"loss": 49.1801,
"step": 41
},
{
"epoch": 0.029693782864212804,
"grad_norm": 58.4798583984375,
"learning_rate": 9.938064469532229e-05,
"loss": 48.9665,
"step": 42
},
{
"epoch": 0.03040077769431311,
"grad_norm": 54.95769119262695,
"learning_rate": 9.999999999999999e-05,
"loss": 51.0544,
"step": 43
},
{
"epoch": 0.031107772524413414,
"grad_norm": 38.86529541015625,
"learning_rate": 0.0001,
"loss": 46.6882,
"step": 44
},
{
"epoch": 0.03181476735451372,
"grad_norm": 49.0484619140625,
"learning_rate": 9.99277899343545e-05,
"loss": 45.8064,
"step": 45
},
{
"epoch": 0.03252176218461403,
"grad_norm": 50.259254455566406,
"learning_rate": 9.985557986870897e-05,
"loss": 44.7941,
"step": 46
},
{
"epoch": 0.03322875701471433,
"grad_norm": 44.155155181884766,
"learning_rate": 9.978336980306346e-05,
"loss": 45.6409,
"step": 47
},
{
"epoch": 0.033935751844814635,
"grad_norm": 48.193572998046875,
"learning_rate": 9.971115973741795e-05,
"loss": 42.198,
"step": 48
},
{
"epoch": 0.03464274667491494,
"grad_norm": 58.65378189086914,
"learning_rate": 9.963894967177244e-05,
"loss": 41.381,
"step": 49
},
{
"epoch": 0.03534974150501524,
"grad_norm": 101.705810546875,
"learning_rate": 9.956673960612691e-05,
"loss": 41.7807,
"step": 50
},
{
"epoch": 0.03605673633511555,
"grad_norm": 52.42247009277344,
"learning_rate": 9.94945295404814e-05,
"loss": 39.5665,
"step": 51
},
{
"epoch": 0.036763731165215856,
"grad_norm": 56.90324020385742,
"learning_rate": 9.942231947483589e-05,
"loss": 37.5718,
"step": 52
},
{
"epoch": 0.03747072599531616,
"grad_norm": 63.50370788574219,
"learning_rate": 9.935010940919037e-05,
"loss": 37.1649,
"step": 53
},
{
"epoch": 0.03817772082541646,
"grad_norm": 81.95431518554688,
"learning_rate": 9.927789934354487e-05,
"loss": 36.9419,
"step": 54
},
{
"epoch": 0.038884715655516766,
"grad_norm": 70.1633071899414,
"learning_rate": 9.920568927789935e-05,
"loss": 36.1688,
"step": 55
},
{
"epoch": 0.039591710485617077,
"grad_norm": 55.53812026977539,
"learning_rate": 9.913347921225384e-05,
"loss": 34.972,
"step": 56
},
{
"epoch": 0.04029870531571738,
"grad_norm": 56.5380859375,
"learning_rate": 9.906126914660831e-05,
"loss": 33.7794,
"step": 57
},
{
"epoch": 0.04100570014581768,
"grad_norm": 63.752079010009766,
"learning_rate": 9.89890590809628e-05,
"loss": 32.1362,
"step": 58
},
{
"epoch": 0.04171269497591799,
"grad_norm": 63.47700500488281,
"learning_rate": 9.891684901531729e-05,
"loss": 31.9549,
"step": 59
},
{
"epoch": 0.04241968980601829,
"grad_norm": 64.81684875488281,
"learning_rate": 9.884463894967178e-05,
"loss": 32.808,
"step": 60
},
{
"epoch": 0.0431266846361186,
"grad_norm": 64.31057739257812,
"learning_rate": 9.877242888402627e-05,
"loss": 30.4557,
"step": 61
},
{
"epoch": 0.043833679466218904,
"grad_norm": 106.54881286621094,
"learning_rate": 9.870021881838075e-05,
"loss": 29.8655,
"step": 62
},
{
"epoch": 0.04454067429631921,
"grad_norm": 83.85649108886719,
"learning_rate": 9.862800875273524e-05,
"loss": 30.1123,
"step": 63
},
{
"epoch": 0.04524766912641951,
"grad_norm": 97.56901550292969,
"learning_rate": 9.855579868708971e-05,
"loss": 30.7334,
"step": 64
},
{
"epoch": 0.045954663956519815,
"grad_norm": 47.09314727783203,
"learning_rate": 9.84835886214442e-05,
"loss": 28.8072,
"step": 65
},
{
"epoch": 0.046661658786620125,
"grad_norm": 64.4630355834961,
"learning_rate": 9.841137855579869e-05,
"loss": 28.9196,
"step": 66
},
{
"epoch": 0.04736865361672043,
"grad_norm": 86.87482452392578,
"learning_rate": 9.833916849015318e-05,
"loss": 28.0566,
"step": 67
},
{
"epoch": 0.04807564844682073,
"grad_norm": 55.95505905151367,
"learning_rate": 9.826695842450767e-05,
"loss": 27.4693,
"step": 68
},
{
"epoch": 0.048782643276921035,
"grad_norm": 45.14374923706055,
"learning_rate": 9.819474835886215e-05,
"loss": 27.5724,
"step": 69
},
{
"epoch": 0.049489638107021346,
"grad_norm": 70.64986419677734,
"learning_rate": 9.812253829321663e-05,
"loss": 27.9396,
"step": 70
},
{
"epoch": 0.05019663293712165,
"grad_norm": 45.24726104736328,
"learning_rate": 9.805032822757111e-05,
"loss": 25.0216,
"step": 71
},
{
"epoch": 0.05090362776722195,
"grad_norm": 60.04133605957031,
"learning_rate": 9.797811816192561e-05,
"loss": 25.2692,
"step": 72
},
{
"epoch": 0.051610622597322256,
"grad_norm": 39.81468200683594,
"learning_rate": 9.790590809628009e-05,
"loss": 24.609,
"step": 73
},
{
"epoch": 0.05231761742742256,
"grad_norm": 47.94467544555664,
"learning_rate": 9.783369803063458e-05,
"loss": 26.2778,
"step": 74
},
{
"epoch": 0.05302461225752287,
"grad_norm": 34.34703826904297,
"learning_rate": 9.776148796498907e-05,
"loss": 23.7904,
"step": 75
},
{
"epoch": 0.05373160708762317,
"grad_norm": 53.06298828125,
"learning_rate": 9.768927789934354e-05,
"loss": 24.0161,
"step": 76
},
{
"epoch": 0.05443860191772348,
"grad_norm": 50.771827697753906,
"learning_rate": 9.761706783369803e-05,
"loss": 24.3783,
"step": 77
},
{
"epoch": 0.05514559674782378,
"grad_norm": 39.64093017578125,
"learning_rate": 9.754485776805252e-05,
"loss": 24.1985,
"step": 78
},
{
"epoch": 0.055852591577924084,
"grad_norm": 40.369510650634766,
"learning_rate": 9.747264770240701e-05,
"loss": 24.0333,
"step": 79
},
{
"epoch": 0.056559586408024394,
"grad_norm": 43.88887023925781,
"learning_rate": 9.740043763676149e-05,
"loss": 23.6668,
"step": 80
},
{
"epoch": 0.0572665812381247,
"grad_norm": 42.15245056152344,
"learning_rate": 9.732822757111598e-05,
"loss": 24.157,
"step": 81
},
{
"epoch": 0.057973576068225,
"grad_norm": 45.39596939086914,
"learning_rate": 9.725601750547047e-05,
"loss": 23.0294,
"step": 82
},
{
"epoch": 0.058680570898325304,
"grad_norm": 41.89541244506836,
"learning_rate": 9.718380743982494e-05,
"loss": 22.6679,
"step": 83
},
{
"epoch": 0.05938756572842561,
"grad_norm": 38.90410614013672,
"learning_rate": 9.711159737417943e-05,
"loss": 20.6634,
"step": 84
},
{
"epoch": 0.06009456055852592,
"grad_norm": 44.17354965209961,
"learning_rate": 9.703938730853392e-05,
"loss": 22.3198,
"step": 85
},
{
"epoch": 0.06080155538862622,
"grad_norm": 44.40744400024414,
"learning_rate": 9.696717724288841e-05,
"loss": 22.6737,
"step": 86
},
{
"epoch": 0.061508550218726525,
"grad_norm": 49.779388427734375,
"learning_rate": 9.689496717724289e-05,
"loss": 21.4484,
"step": 87
},
{
"epoch": 0.06221554504882683,
"grad_norm": 38.24262237548828,
"learning_rate": 9.682275711159738e-05,
"loss": 20.6562,
"step": 88
},
{
"epoch": 0.06292253987892714,
"grad_norm": 34.11885070800781,
"learning_rate": 9.675054704595187e-05,
"loss": 21.33,
"step": 89
},
{
"epoch": 0.06362953470902744,
"grad_norm": 37.97296142578125,
"learning_rate": 9.667833698030636e-05,
"loss": 19.8617,
"step": 90
},
{
"epoch": 0.06433652953912775,
"grad_norm": 35.56360626220703,
"learning_rate": 9.660612691466084e-05,
"loss": 21.5979,
"step": 91
},
{
"epoch": 0.06504352436922806,
"grad_norm": 39.539939880371094,
"learning_rate": 9.653391684901532e-05,
"loss": 19.6249,
"step": 92
},
{
"epoch": 0.06575051919932835,
"grad_norm": 32.25932693481445,
"learning_rate": 9.646170678336981e-05,
"loss": 21.2756,
"step": 93
},
{
"epoch": 0.06645751402942866,
"grad_norm": 38.31186294555664,
"learning_rate": 9.638949671772429e-05,
"loss": 20.3971,
"step": 94
},
{
"epoch": 0.06716450885952896,
"grad_norm": 40.35636901855469,
"learning_rate": 9.631728665207878e-05,
"loss": 19.8489,
"step": 95
},
{
"epoch": 0.06787150368962927,
"grad_norm": 31.688522338867188,
"learning_rate": 9.624507658643326e-05,
"loss": 19.9785,
"step": 96
},
{
"epoch": 0.06857849851972958,
"grad_norm": 31.74046516418457,
"learning_rate": 9.617286652078775e-05,
"loss": 19.8483,
"step": 97
},
{
"epoch": 0.06928549334982988,
"grad_norm": 36.49064636230469,
"learning_rate": 9.610065645514224e-05,
"loss": 20.2419,
"step": 98
},
{
"epoch": 0.06999248817993019,
"grad_norm": 31.634632110595703,
"learning_rate": 9.602844638949672e-05,
"loss": 18.9565,
"step": 99
},
{
"epoch": 0.07069948301003048,
"grad_norm": 37.62042236328125,
"learning_rate": 9.595623632385121e-05,
"loss": 19.4741,
"step": 100
},
{
"epoch": 0.0714064778401308,
"grad_norm": 40.450321197509766,
"learning_rate": 9.588402625820568e-05,
"loss": 20.3956,
"step": 101
},
{
"epoch": 0.0721134726702311,
"grad_norm": 36.62712478637695,
"learning_rate": 9.581181619256017e-05,
"loss": 18.8533,
"step": 102
},
{
"epoch": 0.0728204675003314,
"grad_norm": 34.651710510253906,
"learning_rate": 9.573960612691466e-05,
"loss": 18.2028,
"step": 103
},
{
"epoch": 0.07352746233043171,
"grad_norm": 39.63356018066406,
"learning_rate": 9.566739606126915e-05,
"loss": 19.2314,
"step": 104
},
{
"epoch": 0.07423445716053201,
"grad_norm": 33.7768669128418,
"learning_rate": 9.559518599562364e-05,
"loss": 18.8913,
"step": 105
},
{
"epoch": 0.07494145199063232,
"grad_norm": 33.89146041870117,
"learning_rate": 9.552297592997812e-05,
"loss": 19.5537,
"step": 106
},
{
"epoch": 0.07564844682073263,
"grad_norm": 33.79356002807617,
"learning_rate": 9.545076586433261e-05,
"loss": 20.1578,
"step": 107
},
{
"epoch": 0.07635544165083293,
"grad_norm": 28.928699493408203,
"learning_rate": 9.53785557986871e-05,
"loss": 19.4719,
"step": 108
},
{
"epoch": 0.07706243648093324,
"grad_norm": 41.88840866088867,
"learning_rate": 9.530634573304159e-05,
"loss": 18.8494,
"step": 109
},
{
"epoch": 0.07776943131103353,
"grad_norm": 36.432682037353516,
"learning_rate": 9.523413566739606e-05,
"loss": 19.1342,
"step": 110
},
{
"epoch": 0.07847642614113384,
"grad_norm": 35.987552642822266,
"learning_rate": 9.516192560175055e-05,
"loss": 19.1107,
"step": 111
},
{
"epoch": 0.07918342097123415,
"grad_norm": 44.01519012451172,
"learning_rate": 9.508971553610504e-05,
"loss": 18.7917,
"step": 112
},
{
"epoch": 0.07989041580133445,
"grad_norm": 31.156208038330078,
"learning_rate": 9.501750547045952e-05,
"loss": 18.7651,
"step": 113
},
{
"epoch": 0.08059741063143476,
"grad_norm": 27.615875244140625,
"learning_rate": 9.4945295404814e-05,
"loss": 17.4872,
"step": 114
},
{
"epoch": 0.08130440546153506,
"grad_norm": 30.91271209716797,
"learning_rate": 9.48730853391685e-05,
"loss": 19.8748,
"step": 115
},
{
"epoch": 0.08201140029163537,
"grad_norm": 33.25605773925781,
"learning_rate": 9.480087527352299e-05,
"loss": 19.1411,
"step": 116
},
{
"epoch": 0.08271839512173568,
"grad_norm": 29.771507263183594,
"learning_rate": 9.472866520787746e-05,
"loss": 18.2639,
"step": 117
},
{
"epoch": 0.08342538995183597,
"grad_norm": 32.23770523071289,
"learning_rate": 9.465645514223195e-05,
"loss": 18.257,
"step": 118
},
{
"epoch": 0.08413238478193628,
"grad_norm": 36.424015045166016,
"learning_rate": 9.458424507658644e-05,
"loss": 17.0622,
"step": 119
},
{
"epoch": 0.08483937961203658,
"grad_norm": 39.868797302246094,
"learning_rate": 9.451203501094092e-05,
"loss": 19.4193,
"step": 120
},
{
"epoch": 0.08554637444213689,
"grad_norm": 29.13055992126465,
"learning_rate": 9.443982494529542e-05,
"loss": 19.4787,
"step": 121
},
{
"epoch": 0.0862533692722372,
"grad_norm": 32.32482147216797,
"learning_rate": 9.43676148796499e-05,
"loss": 18.933,
"step": 122
},
{
"epoch": 0.0869603641023375,
"grad_norm": 30.419885635375977,
"learning_rate": 9.429540481400438e-05,
"loss": 17.8486,
"step": 123
},
{
"epoch": 0.08766735893243781,
"grad_norm": 28.050142288208008,
"learning_rate": 9.422319474835886e-05,
"loss": 20.1933,
"step": 124
},
{
"epoch": 0.0883743537625381,
"grad_norm": 29.47435188293457,
"learning_rate": 9.415098468271335e-05,
"loss": 17.4255,
"step": 125
},
{
"epoch": 0.08908134859263842,
"grad_norm": 29.57215118408203,
"learning_rate": 9.407877461706784e-05,
"loss": 20.1363,
"step": 126
},
{
"epoch": 0.08978834342273873,
"grad_norm": 38.4536247253418,
"learning_rate": 9.400656455142233e-05,
"loss": 18.2052,
"step": 127
},
{
"epoch": 0.09049533825283902,
"grad_norm": 28.376087188720703,
"learning_rate": 9.393435448577682e-05,
"loss": 19.2259,
"step": 128
},
{
"epoch": 0.09120233308293933,
"grad_norm": 26.830381393432617,
"learning_rate": 9.38621444201313e-05,
"loss": 19.3953,
"step": 129
},
{
"epoch": 0.09190932791303963,
"grad_norm": 27.166439056396484,
"learning_rate": 9.378993435448578e-05,
"loss": 18.9521,
"step": 130
},
{
"epoch": 0.09261632274313994,
"grad_norm": 28.81890869140625,
"learning_rate": 9.371772428884026e-05,
"loss": 17.7015,
"step": 131
},
{
"epoch": 0.09332331757324025,
"grad_norm": 32.495296478271484,
"learning_rate": 9.364551422319475e-05,
"loss": 19.0709,
"step": 132
},
{
"epoch": 0.09403031240334055,
"grad_norm": 27.91164779663086,
"learning_rate": 9.357330415754924e-05,
"loss": 18.7305,
"step": 133
},
{
"epoch": 0.09473730723344086,
"grad_norm": 32.352169036865234,
"learning_rate": 9.350109409190373e-05,
"loss": 17.947,
"step": 134
},
{
"epoch": 0.09544430206354117,
"grad_norm": 41.09231948852539,
"learning_rate": 9.342888402625822e-05,
"loss": 18.0234,
"step": 135
},
{
"epoch": 0.09615129689364146,
"grad_norm": 29.327669143676758,
"learning_rate": 9.335667396061269e-05,
"loss": 18.0342,
"step": 136
},
{
"epoch": 0.09685829172374177,
"grad_norm": 31.431499481201172,
"learning_rate": 9.328446389496718e-05,
"loss": 18.1777,
"step": 137
},
{
"epoch": 0.09756528655384207,
"grad_norm": 28.50454330444336,
"learning_rate": 9.321225382932166e-05,
"loss": 17.4591,
"step": 138
},
{
"epoch": 0.09827228138394238,
"grad_norm": 29.735980987548828,
"learning_rate": 9.314004376367616e-05,
"loss": 17.5018,
"step": 139
},
{
"epoch": 0.09897927621404269,
"grad_norm": 34.070980072021484,
"learning_rate": 9.306783369803064e-05,
"loss": 17.5913,
"step": 140
},
{
"epoch": 0.09968627104414299,
"grad_norm": 30.582719802856445,
"learning_rate": 9.299562363238513e-05,
"loss": 17.0948,
"step": 141
},
{
"epoch": 0.1003932658742433,
"grad_norm": 26.284706115722656,
"learning_rate": 9.292341356673962e-05,
"loss": 18.7132,
"step": 142
},
{
"epoch": 0.1011002607043436,
"grad_norm": 33.66963195800781,
"learning_rate": 9.285120350109409e-05,
"loss": 18.8174,
"step": 143
},
{
"epoch": 0.1018072555344439,
"grad_norm": 33.1568603515625,
"learning_rate": 9.277899343544858e-05,
"loss": 17.7703,
"step": 144
},
{
"epoch": 0.10251425036454422,
"grad_norm": 24.135400772094727,
"learning_rate": 9.270678336980307e-05,
"loss": 17.4647,
"step": 145
},
{
"epoch": 0.10322124519464451,
"grad_norm": 29.076414108276367,
"learning_rate": 9.263457330415756e-05,
"loss": 17.4553,
"step": 146
},
{
"epoch": 0.10392824002474482,
"grad_norm": 35.003414154052734,
"learning_rate": 9.256236323851204e-05,
"loss": 16.8381,
"step": 147
},
{
"epoch": 0.10463523485484512,
"grad_norm": 35.33264923095703,
"learning_rate": 9.249015317286652e-05,
"loss": 17.2061,
"step": 148
},
{
"epoch": 0.10534222968494543,
"grad_norm": 25.581878662109375,
"learning_rate": 9.241794310722101e-05,
"loss": 18.4266,
"step": 149
},
{
"epoch": 0.10604922451504574,
"grad_norm": 36.04518127441406,
"learning_rate": 9.234573304157549e-05,
"loss": 17.6767,
"step": 150
},
{
"epoch": 0.10675621934514604,
"grad_norm": 25.03700828552246,
"learning_rate": 9.227352297592998e-05,
"loss": 18.6159,
"step": 151
},
{
"epoch": 0.10746321417524635,
"grad_norm": 37.83554458618164,
"learning_rate": 9.220131291028447e-05,
"loss": 17.1416,
"step": 152
},
{
"epoch": 0.10817020900534664,
"grad_norm": 32.96514129638672,
"learning_rate": 9.212910284463896e-05,
"loss": 18.1587,
"step": 153
},
{
"epoch": 0.10887720383544695,
"grad_norm": 25.70465850830078,
"learning_rate": 9.205689277899343e-05,
"loss": 17.3172,
"step": 154
},
{
"epoch": 0.10958419866554726,
"grad_norm": 24.10750961303711,
"learning_rate": 9.198468271334792e-05,
"loss": 17.2018,
"step": 155
},
{
"epoch": 0.11029119349564756,
"grad_norm": 24.535715103149414,
"learning_rate": 9.191247264770241e-05,
"loss": 17.2166,
"step": 156
},
{
"epoch": 0.11099818832574787,
"grad_norm": 24.88667869567871,
"learning_rate": 9.184026258205689e-05,
"loss": 20.2791,
"step": 157
},
{
"epoch": 0.11170518315584817,
"grad_norm": 24.478866577148438,
"learning_rate": 9.176805251641139e-05,
"loss": 18.1817,
"step": 158
},
{
"epoch": 0.11241217798594848,
"grad_norm": 30.161104202270508,
"learning_rate": 9.169584245076587e-05,
"loss": 17.7581,
"step": 159
},
{
"epoch": 0.11311917281604879,
"grad_norm": 29.541017532348633,
"learning_rate": 9.162363238512036e-05,
"loss": 17.4958,
"step": 160
},
{
"epoch": 0.11382616764614908,
"grad_norm": 32.381771087646484,
"learning_rate": 9.155142231947483e-05,
"loss": 16.9631,
"step": 161
},
{
"epoch": 0.1145331624762494,
"grad_norm": 31.770675659179688,
"learning_rate": 9.147921225382932e-05,
"loss": 17.7155,
"step": 162
},
{
"epoch": 0.11524015730634969,
"grad_norm": 30.45294952392578,
"learning_rate": 9.140700218818381e-05,
"loss": 19.232,
"step": 163
},
{
"epoch": 0.11594715213645,
"grad_norm": 26.106555938720703,
"learning_rate": 9.13347921225383e-05,
"loss": 17.4079,
"step": 164
},
{
"epoch": 0.11665414696655031,
"grad_norm": 25.87933921813965,
"learning_rate": 9.126258205689279e-05,
"loss": 17.6005,
"step": 165
},
{
"epoch": 0.11736114179665061,
"grad_norm": 27.44808578491211,
"learning_rate": 9.119037199124727e-05,
"loss": 16.6392,
"step": 166
},
{
"epoch": 0.11806813662675092,
"grad_norm": 25.52296257019043,
"learning_rate": 9.111816192560176e-05,
"loss": 17.2663,
"step": 167
},
{
"epoch": 0.11877513145685122,
"grad_norm": 24.483516693115234,
"learning_rate": 9.104595185995623e-05,
"loss": 15.9337,
"step": 168
},
{
"epoch": 0.11948212628695153,
"grad_norm": 23.002857208251953,
"learning_rate": 9.097374179431072e-05,
"loss": 16.215,
"step": 169
},
{
"epoch": 0.12018912111705184,
"grad_norm": 26.76035499572754,
"learning_rate": 9.090153172866521e-05,
"loss": 16.72,
"step": 170
},
{
"epoch": 0.12089611594715213,
"grad_norm": 29.699464797973633,
"learning_rate": 9.08293216630197e-05,
"loss": 17.7319,
"step": 171
},
{
"epoch": 0.12160311077725244,
"grad_norm": 30.24201202392578,
"learning_rate": 9.075711159737419e-05,
"loss": 16.9052,
"step": 172
},
{
"epoch": 0.12231010560735274,
"grad_norm": 48.706581115722656,
"learning_rate": 9.068490153172867e-05,
"loss": 17.2807,
"step": 173
},
{
"epoch": 0.12301710043745305,
"grad_norm": 25.212533950805664,
"learning_rate": 9.061269146608315e-05,
"loss": 16.5931,
"step": 174
},
{
"epoch": 0.12372409526755336,
"grad_norm": 29.620086669921875,
"learning_rate": 9.054048140043763e-05,
"loss": 16.2194,
"step": 175
},
{
"epoch": 0.12443109009765366,
"grad_norm": 26.384183883666992,
"learning_rate": 9.046827133479213e-05,
"loss": 16.1872,
"step": 176
},
{
"epoch": 0.12513808492775397,
"grad_norm": 25.466659545898438,
"learning_rate": 9.039606126914661e-05,
"loss": 16.7155,
"step": 177
},
{
"epoch": 0.12584507975785428,
"grad_norm": 31.032506942749023,
"learning_rate": 9.03238512035011e-05,
"loss": 17.6649,
"step": 178
},
{
"epoch": 0.1265520745879546,
"grad_norm": 24.821929931640625,
"learning_rate": 9.025164113785559e-05,
"loss": 17.3971,
"step": 179
},
{
"epoch": 0.12725906941805487,
"grad_norm": 25.913917541503906,
"learning_rate": 9.017943107221006e-05,
"loss": 16.564,
"step": 180
},
{
"epoch": 0.12796606424815518,
"grad_norm": 27.99454116821289,
"learning_rate": 9.010722100656455e-05,
"loss": 19.4556,
"step": 181
},
{
"epoch": 0.1286730590782555,
"grad_norm": 23.89288902282715,
"learning_rate": 9.003501094091904e-05,
"loss": 17.549,
"step": 182
},
{
"epoch": 0.1293800539083558,
"grad_norm": 26.861974716186523,
"learning_rate": 8.996280087527353e-05,
"loss": 16.8727,
"step": 183
},
{
"epoch": 0.1300870487384561,
"grad_norm": 24.672622680664062,
"learning_rate": 8.989059080962801e-05,
"loss": 15.7491,
"step": 184
},
{
"epoch": 0.1307940435685564,
"grad_norm": 24.098909378051758,
"learning_rate": 8.98183807439825e-05,
"loss": 16.9385,
"step": 185
},
{
"epoch": 0.1315010383986567,
"grad_norm": 26.149381637573242,
"learning_rate": 8.974617067833699e-05,
"loss": 17.5179,
"step": 186
},
{
"epoch": 0.13220803322875702,
"grad_norm": 33.0228271484375,
"learning_rate": 8.967396061269146e-05,
"loss": 18.6741,
"step": 187
},
{
"epoch": 0.13291502805885733,
"grad_norm": 30.52547264099121,
"learning_rate": 8.960175054704595e-05,
"loss": 16.0168,
"step": 188
},
{
"epoch": 0.13362202288895764,
"grad_norm": 24.271406173706055,
"learning_rate": 8.952954048140044e-05,
"loss": 17.562,
"step": 189
},
{
"epoch": 0.13432901771905792,
"grad_norm": 31.642528533935547,
"learning_rate": 8.945733041575493e-05,
"loss": 16.4493,
"step": 190
},
{
"epoch": 0.13503601254915823,
"grad_norm": 24.342866897583008,
"learning_rate": 8.938512035010941e-05,
"loss": 15.8088,
"step": 191
},
{
"epoch": 0.13574300737925854,
"grad_norm": 33.57277297973633,
"learning_rate": 8.93129102844639e-05,
"loss": 18.1225,
"step": 192
},
{
"epoch": 0.13645000220935885,
"grad_norm": 25.293317794799805,
"learning_rate": 8.924070021881839e-05,
"loss": 17.1738,
"step": 193
},
{
"epoch": 0.13715699703945916,
"grad_norm": 25.74216079711914,
"learning_rate": 8.916849015317288e-05,
"loss": 17.5808,
"step": 194
},
{
"epoch": 0.13786399186955944,
"grad_norm": 25.824739456176758,
"learning_rate": 8.909628008752736e-05,
"loss": 16.29,
"step": 195
},
{
"epoch": 0.13857098669965975,
"grad_norm": 21.56652069091797,
"learning_rate": 8.902407002188184e-05,
"loss": 17.3717,
"step": 196
},
{
"epoch": 0.13927798152976006,
"grad_norm": 27.956878662109375,
"learning_rate": 8.895185995623633e-05,
"loss": 16.8826,
"step": 197
},
{
"epoch": 0.13998497635986037,
"grad_norm": 25.947261810302734,
"learning_rate": 8.88796498905908e-05,
"loss": 17.7311,
"step": 198
},
{
"epoch": 0.14069197118996069,
"grad_norm": 26.35049057006836,
"learning_rate": 8.88074398249453e-05,
"loss": 16.7053,
"step": 199
},
{
"epoch": 0.14139896602006097,
"grad_norm": 23.005706787109375,
"learning_rate": 8.873522975929978e-05,
"loss": 17.5012,
"step": 200
},
{
"epoch": 0.14210596085016128,
"grad_norm": 24.649900436401367,
"learning_rate": 8.866301969365427e-05,
"loss": 16.9559,
"step": 201
},
{
"epoch": 0.1428129556802616,
"grad_norm": 23.98031997680664,
"learning_rate": 8.859080962800876e-05,
"loss": 15.1581,
"step": 202
},
{
"epoch": 0.1435199505103619,
"grad_norm": 27.078948974609375,
"learning_rate": 8.851859956236324e-05,
"loss": 16.7947,
"step": 203
},
{
"epoch": 0.1442269453404622,
"grad_norm": 26.9580020904541,
"learning_rate": 8.844638949671773e-05,
"loss": 16.5367,
"step": 204
},
{
"epoch": 0.1449339401705625,
"grad_norm": 22.63777732849121,
"learning_rate": 8.83741794310722e-05,
"loss": 16.4217,
"step": 205
},
{
"epoch": 0.1456409350006628,
"grad_norm": 26.905527114868164,
"learning_rate": 8.83019693654267e-05,
"loss": 15.5522,
"step": 206
},
{
"epoch": 0.1463479298307631,
"grad_norm": 23.968271255493164,
"learning_rate": 8.822975929978118e-05,
"loss": 14.9917,
"step": 207
},
{
"epoch": 0.14705492466086342,
"grad_norm": 26.18745994567871,
"learning_rate": 8.815754923413567e-05,
"loss": 16.8354,
"step": 208
},
{
"epoch": 0.14776191949096373,
"grad_norm": 24.03843879699707,
"learning_rate": 8.808533916849016e-05,
"loss": 16.5202,
"step": 209
},
{
"epoch": 0.14846891432106402,
"grad_norm": 22.780261993408203,
"learning_rate": 8.801312910284464e-05,
"loss": 15.2645,
"step": 210
},
{
"epoch": 0.14917590915116433,
"grad_norm": 24.574827194213867,
"learning_rate": 8.794091903719913e-05,
"loss": 16.0331,
"step": 211
},
{
"epoch": 0.14988290398126464,
"grad_norm": 26.111984252929688,
"learning_rate": 8.786870897155362e-05,
"loss": 16.5817,
"step": 212
},
{
"epoch": 0.15058989881136495,
"grad_norm": 20.612037658691406,
"learning_rate": 8.77964989059081e-05,
"loss": 17.2985,
"step": 213
},
{
"epoch": 0.15129689364146526,
"grad_norm": 20.90341567993164,
"learning_rate": 8.772428884026258e-05,
"loss": 16.5036,
"step": 214
},
{
"epoch": 0.15200388847156554,
"grad_norm": 20.831926345825195,
"learning_rate": 8.765207877461707e-05,
"loss": 15.8133,
"step": 215
},
{
"epoch": 0.15271088330166585,
"grad_norm": 26.699604034423828,
"learning_rate": 8.757986870897156e-05,
"loss": 15.9572,
"step": 216
},
{
"epoch": 0.15341787813176616,
"grad_norm": 27.643829345703125,
"learning_rate": 8.750765864332604e-05,
"loss": 16.6718,
"step": 217
},
{
"epoch": 0.15412487296186647,
"grad_norm": 27.599285125732422,
"learning_rate": 8.743544857768053e-05,
"loss": 16.9441,
"step": 218
},
{
"epoch": 0.15483186779196678,
"grad_norm": 24.13196563720703,
"learning_rate": 8.736323851203502e-05,
"loss": 17.9485,
"step": 219
},
{
"epoch": 0.15553886262206706,
"grad_norm": 23.68052864074707,
"learning_rate": 8.72910284463895e-05,
"loss": 15.4747,
"step": 220
},
{
"epoch": 0.15624585745216737,
"grad_norm": 25.603376388549805,
"learning_rate": 8.721881838074398e-05,
"loss": 16.7561,
"step": 221
},
{
"epoch": 0.15695285228226769,
"grad_norm": 24.765710830688477,
"learning_rate": 8.714660831509847e-05,
"loss": 15.5788,
"step": 222
},
{
"epoch": 0.157659847112368,
"grad_norm": 21.479066848754883,
"learning_rate": 8.707439824945296e-05,
"loss": 17.7476,
"step": 223
},
{
"epoch": 0.1583668419424683,
"grad_norm": 20.06366539001465,
"learning_rate": 8.700218818380744e-05,
"loss": 15.1917,
"step": 224
},
{
"epoch": 0.1590738367725686,
"grad_norm": 22.646343231201172,
"learning_rate": 8.692997811816194e-05,
"loss": 16.2958,
"step": 225
},
{
"epoch": 0.1597808316026689,
"grad_norm": 22.053787231445312,
"learning_rate": 8.685776805251641e-05,
"loss": 16.4618,
"step": 226
},
{
"epoch": 0.1604878264327692,
"grad_norm": 22.88997459411621,
"learning_rate": 8.67855579868709e-05,
"loss": 16.0649,
"step": 227
},
{
"epoch": 0.16119482126286952,
"grad_norm": 22.48101806640625,
"learning_rate": 8.671334792122538e-05,
"loss": 15.7039,
"step": 228
},
{
"epoch": 0.16190181609296983,
"grad_norm": 25.803606033325195,
"learning_rate": 8.664113785557987e-05,
"loss": 16.5133,
"step": 229
},
{
"epoch": 0.1626088109230701,
"grad_norm": 22.281761169433594,
"learning_rate": 8.656892778993436e-05,
"loss": 16.2889,
"step": 230
},
{
"epoch": 0.16331580575317042,
"grad_norm": 26.732994079589844,
"learning_rate": 8.649671772428885e-05,
"loss": 16.9495,
"step": 231
},
{
"epoch": 0.16402280058327073,
"grad_norm": 22.665224075317383,
"learning_rate": 8.642450765864334e-05,
"loss": 17.289,
"step": 232
},
{
"epoch": 0.16472979541337104,
"grad_norm": 21.641260147094727,
"learning_rate": 8.635229759299781e-05,
"loss": 16.0455,
"step": 233
},
{
"epoch": 0.16543679024347135,
"grad_norm": 25.28057098388672,
"learning_rate": 8.62800875273523e-05,
"loss": 17.1432,
"step": 234
},
{
"epoch": 0.16614378507357164,
"grad_norm": 25.27671241760254,
"learning_rate": 8.620787746170678e-05,
"loss": 17.2155,
"step": 235
},
{
"epoch": 0.16685077990367195,
"grad_norm": 27.388463973999023,
"learning_rate": 8.613566739606127e-05,
"loss": 16.2077,
"step": 236
},
{
"epoch": 0.16755777473377226,
"grad_norm": 21.950359344482422,
"learning_rate": 8.606345733041576e-05,
"loss": 17.8045,
"step": 237
},
{
"epoch": 0.16826476956387257,
"grad_norm": 23.083728790283203,
"learning_rate": 8.599124726477025e-05,
"loss": 17.1346,
"step": 238
},
{
"epoch": 0.16897176439397288,
"grad_norm": 24.17853355407715,
"learning_rate": 8.591903719912474e-05,
"loss": 15.6759,
"step": 239
},
{
"epoch": 0.16967875922407316,
"grad_norm": 21.20966148376465,
"learning_rate": 8.584682713347921e-05,
"loss": 17.0444,
"step": 240
},
{
"epoch": 0.17038575405417347,
"grad_norm": 22.764076232910156,
"learning_rate": 8.57746170678337e-05,
"loss": 15.188,
"step": 241
},
{
"epoch": 0.17109274888427378,
"grad_norm": 21.44521713256836,
"learning_rate": 8.570240700218818e-05,
"loss": 16.7055,
"step": 242
},
{
"epoch": 0.1717997437143741,
"grad_norm": 23.313913345336914,
"learning_rate": 8.563019693654267e-05,
"loss": 16.3735,
"step": 243
},
{
"epoch": 0.1725067385444744,
"grad_norm": 21.593088150024414,
"learning_rate": 8.555798687089716e-05,
"loss": 14.0391,
"step": 244
},
{
"epoch": 0.17321373337457469,
"grad_norm": 20.44053840637207,
"learning_rate": 8.548577680525165e-05,
"loss": 16.2655,
"step": 245
},
{
"epoch": 0.173920728204675,
"grad_norm": 21.79986572265625,
"learning_rate": 8.541356673960614e-05,
"loss": 16.0105,
"step": 246
},
{
"epoch": 0.1746277230347753,
"grad_norm": 31.781938552856445,
"learning_rate": 8.534135667396061e-05,
"loss": 16.5409,
"step": 247
},
{
"epoch": 0.17533471786487562,
"grad_norm": 33.61064910888672,
"learning_rate": 8.52691466083151e-05,
"loss": 17.2873,
"step": 248
},
{
"epoch": 0.17604171269497593,
"grad_norm": 22.703304290771484,
"learning_rate": 8.519693654266959e-05,
"loss": 15.2044,
"step": 249
},
{
"epoch": 0.1767487075250762,
"grad_norm": 20.650047302246094,
"learning_rate": 8.512472647702408e-05,
"loss": 16.8391,
"step": 250
},
{
"epoch": 0.17745570235517652,
"grad_norm": 25.48831558227539,
"learning_rate": 8.505251641137856e-05,
"loss": 16.8433,
"step": 251
},
{
"epoch": 0.17816269718527683,
"grad_norm": 21.704578399658203,
"learning_rate": 8.498030634573304e-05,
"loss": 16.3913,
"step": 252
},
{
"epoch": 0.17886969201537714,
"grad_norm": 23.87643814086914,
"learning_rate": 8.490809628008753e-05,
"loss": 15.8773,
"step": 253
},
{
"epoch": 0.17957668684547745,
"grad_norm": 23.842498779296875,
"learning_rate": 8.483588621444201e-05,
"loss": 17.0596,
"step": 254
},
{
"epoch": 0.18028368167557773,
"grad_norm": 24.21112823486328,
"learning_rate": 8.47636761487965e-05,
"loss": 16.4233,
"step": 255
},
{
"epoch": 0.18099067650567804,
"grad_norm": 21.40100860595703,
"learning_rate": 8.469146608315099e-05,
"loss": 17.5836,
"step": 256
},
{
"epoch": 0.18169767133577835,
"grad_norm": 22.97970199584961,
"learning_rate": 8.461925601750548e-05,
"loss": 17.3076,
"step": 257
},
{
"epoch": 0.18240466616587866,
"grad_norm": 23.0300235748291,
"learning_rate": 8.454704595185995e-05,
"loss": 16.2172,
"step": 258
},
{
"epoch": 0.18311166099597898,
"grad_norm": 21.156230926513672,
"learning_rate": 8.447483588621444e-05,
"loss": 14.648,
"step": 259
},
{
"epoch": 0.18381865582607926,
"grad_norm": 22.934730529785156,
"learning_rate": 8.440262582056893e-05,
"loss": 17.373,
"step": 260
},
{
"epoch": 0.18452565065617957,
"grad_norm": 23.799009323120117,
"learning_rate": 8.433041575492341e-05,
"loss": 15.3648,
"step": 261
},
{
"epoch": 0.18523264548627988,
"grad_norm": 22.114885330200195,
"learning_rate": 8.425820568927791e-05,
"loss": 16.3895,
"step": 262
},
{
"epoch": 0.1859396403163802,
"grad_norm": 26.017728805541992,
"learning_rate": 8.418599562363239e-05,
"loss": 15.9444,
"step": 263
},
{
"epoch": 0.1866466351464805,
"grad_norm": 25.28594970703125,
"learning_rate": 8.411378555798688e-05,
"loss": 16.4829,
"step": 264
},
{
"epoch": 0.1873536299765808,
"grad_norm": 21.194791793823242,
"learning_rate": 8.404157549234135e-05,
"loss": 15.6376,
"step": 265
},
{
"epoch": 0.1880606248066811,
"grad_norm": 25.057748794555664,
"learning_rate": 8.396936542669584e-05,
"loss": 16.9039,
"step": 266
},
{
"epoch": 0.1887676196367814,
"grad_norm": 22.067426681518555,
"learning_rate": 8.389715536105033e-05,
"loss": 16.105,
"step": 267
},
{
"epoch": 0.1894746144668817,
"grad_norm": 23.50616455078125,
"learning_rate": 8.382494529540482e-05,
"loss": 15.3005,
"step": 268
},
{
"epoch": 0.19018160929698202,
"grad_norm": 22.99486541748047,
"learning_rate": 8.375273522975931e-05,
"loss": 16.2793,
"step": 269
},
{
"epoch": 0.19088860412708233,
"grad_norm": 28.1767578125,
"learning_rate": 8.368052516411379e-05,
"loss": 14.8503,
"step": 270
},
{
"epoch": 0.19159559895718262,
"grad_norm": 21.05082893371582,
"learning_rate": 8.360831509846828e-05,
"loss": 14.3509,
"step": 271
},
{
"epoch": 0.19230259378728293,
"grad_norm": 27.14815902709961,
"learning_rate": 8.353610503282275e-05,
"loss": 16.2584,
"step": 272
},
{
"epoch": 0.19300958861738324,
"grad_norm": 24.146907806396484,
"learning_rate": 8.346389496717724e-05,
"loss": 16.2285,
"step": 273
},
{
"epoch": 0.19371658344748355,
"grad_norm": 21.399860382080078,
"learning_rate": 8.339168490153173e-05,
"loss": 16.4681,
"step": 274
},
{
"epoch": 0.19442357827758386,
"grad_norm": 23.532634735107422,
"learning_rate": 8.331947483588622e-05,
"loss": 17.9858,
"step": 275
},
{
"epoch": 0.19513057310768414,
"grad_norm": 22.816146850585938,
"learning_rate": 8.324726477024071e-05,
"loss": 14.9456,
"step": 276
},
{
"epoch": 0.19583756793778445,
"grad_norm": 19.56248664855957,
"learning_rate": 8.317505470459519e-05,
"loss": 15.6442,
"step": 277
},
{
"epoch": 0.19654456276788476,
"grad_norm": 24.974672317504883,
"learning_rate": 8.310284463894968e-05,
"loss": 18.1326,
"step": 278
},
{
"epoch": 0.19725155759798507,
"grad_norm": 24.85466766357422,
"learning_rate": 8.303063457330415e-05,
"loss": 14.3476,
"step": 279
},
{
"epoch": 0.19795855242808538,
"grad_norm": 18.94011878967285,
"learning_rate": 8.295842450765865e-05,
"loss": 15.3077,
"step": 280
},
{
"epoch": 0.19866554725818567,
"grad_norm": 18.339811325073242,
"learning_rate": 8.288621444201313e-05,
"loss": 15.5157,
"step": 281
},
{
"epoch": 0.19937254208828598,
"grad_norm": 36.65610122680664,
"learning_rate": 8.281400437636762e-05,
"loss": 16.4754,
"step": 282
},
{
"epoch": 0.20007953691838629,
"grad_norm": 20.609729766845703,
"learning_rate": 8.274179431072211e-05,
"loss": 14.7549,
"step": 283
},
{
"epoch": 0.2007865317484866,
"grad_norm": 18.10039710998535,
"learning_rate": 8.266958424507658e-05,
"loss": 16.8433,
"step": 284
},
{
"epoch": 0.2014935265785869,
"grad_norm": 22.75161361694336,
"learning_rate": 8.259737417943107e-05,
"loss": 14.9722,
"step": 285
},
{
"epoch": 0.2022005214086872,
"grad_norm": 23.14825439453125,
"learning_rate": 8.252516411378556e-05,
"loss": 15.1921,
"step": 286
},
{
"epoch": 0.2029075162387875,
"grad_norm": 32.537208557128906,
"learning_rate": 8.245295404814005e-05,
"loss": 17.0106,
"step": 287
},
{
"epoch": 0.2036145110688878,
"grad_norm": 23.347766876220703,
"learning_rate": 8.238074398249453e-05,
"loss": 17.4193,
"step": 288
},
{
"epoch": 0.20432150589898812,
"grad_norm": 20.956504821777344,
"learning_rate": 8.230853391684902e-05,
"loss": 16.9236,
"step": 289
},
{
"epoch": 0.20502850072908843,
"grad_norm": 19.826589584350586,
"learning_rate": 8.223632385120351e-05,
"loss": 15.4246,
"step": 290
},
{
"epoch": 0.2057354955591887,
"grad_norm": 20.190919876098633,
"learning_rate": 8.216411378555798e-05,
"loss": 15.6196,
"step": 291
},
{
"epoch": 0.20644249038928902,
"grad_norm": 21.368432998657227,
"learning_rate": 8.209190371991247e-05,
"loss": 15.6062,
"step": 292
},
{
"epoch": 0.20714948521938933,
"grad_norm": 19.60832405090332,
"learning_rate": 8.201969365426696e-05,
"loss": 14.8065,
"step": 293
},
{
"epoch": 0.20785648004948964,
"grad_norm": 20.27286720275879,
"learning_rate": 8.194748358862145e-05,
"loss": 16.176,
"step": 294
},
{
"epoch": 0.20856347487958996,
"grad_norm": 20.443204879760742,
"learning_rate": 8.187527352297593e-05,
"loss": 15.325,
"step": 295
},
{
"epoch": 0.20927046970969024,
"grad_norm": 25.870630264282227,
"learning_rate": 8.180306345733042e-05,
"loss": 16.4023,
"step": 296
},
{
"epoch": 0.20997746453979055,
"grad_norm": 20.376365661621094,
"learning_rate": 8.17308533916849e-05,
"loss": 15.4194,
"step": 297
},
{
"epoch": 0.21068445936989086,
"grad_norm": 18.827817916870117,
"learning_rate": 8.16586433260394e-05,
"loss": 15.4671,
"step": 298
},
{
"epoch": 0.21139145419999117,
"grad_norm": 24.46773910522461,
"learning_rate": 8.158643326039389e-05,
"loss": 16.0512,
"step": 299
},
{
"epoch": 0.21209844903009148,
"grad_norm": 26.15793800354004,
"learning_rate": 8.151422319474836e-05,
"loss": 16.7677,
"step": 300
},
{
"epoch": 0.21280544386019176,
"grad_norm": 23.193317413330078,
"learning_rate": 8.144201312910285e-05,
"loss": 17.2927,
"step": 301
},
{
"epoch": 0.21351243869029207,
"grad_norm": 22.430593490600586,
"learning_rate": 8.136980306345733e-05,
"loss": 15.1819,
"step": 302
},
{
"epoch": 0.21421943352039238,
"grad_norm": 18.35970115661621,
"learning_rate": 8.129759299781182e-05,
"loss": 15.0625,
"step": 303
},
{
"epoch": 0.2149264283504927,
"grad_norm": 23.23087501525879,
"learning_rate": 8.12253829321663e-05,
"loss": 15.07,
"step": 304
},
{
"epoch": 0.215633423180593,
"grad_norm": 19.622215270996094,
"learning_rate": 8.11531728665208e-05,
"loss": 17.0095,
"step": 305
},
{
"epoch": 0.2163404180106933,
"grad_norm": 19.727991104125977,
"learning_rate": 8.108096280087528e-05,
"loss": 15.6499,
"step": 306
},
{
"epoch": 0.2170474128407936,
"grad_norm": 24.34721565246582,
"learning_rate": 8.100875273522976e-05,
"loss": 15.6125,
"step": 307
},
{
"epoch": 0.2177544076708939,
"grad_norm": 20.973655700683594,
"learning_rate": 8.093654266958425e-05,
"loss": 14.6315,
"step": 308
},
{
"epoch": 0.21846140250099422,
"grad_norm": 24.30404281616211,
"learning_rate": 8.086433260393873e-05,
"loss": 12.6647,
"step": 309
},
{
"epoch": 0.21916839733109453,
"grad_norm": 21.908384323120117,
"learning_rate": 8.079212253829321e-05,
"loss": 14.0689,
"step": 310
},
{
"epoch": 0.2198753921611948,
"grad_norm": 19.879352569580078,
"learning_rate": 8.071991247264772e-05,
"loss": 17.1264,
"step": 311
},
{
"epoch": 0.22058238699129512,
"grad_norm": 18.12099266052246,
"learning_rate": 8.06477024070022e-05,
"loss": 15.1719,
"step": 312
},
{
"epoch": 0.22128938182139543,
"grad_norm": 19.50701141357422,
"learning_rate": 8.057549234135668e-05,
"loss": 15.1601,
"step": 313
},
{
"epoch": 0.22199637665149574,
"grad_norm": 19.107711791992188,
"learning_rate": 8.050328227571116e-05,
"loss": 16.1114,
"step": 314
},
{
"epoch": 0.22270337148159605,
"grad_norm": 21.994384765625,
"learning_rate": 8.043107221006565e-05,
"loss": 16.0056,
"step": 315
},
{
"epoch": 0.22341036631169633,
"grad_norm": 21.152990341186523,
"learning_rate": 8.035886214442014e-05,
"loss": 15.4405,
"step": 316
},
{
"epoch": 0.22411736114179664,
"grad_norm": 22.611316680908203,
"learning_rate": 8.028665207877463e-05,
"loss": 14.4719,
"step": 317
},
{
"epoch": 0.22482435597189696,
"grad_norm": 21.554346084594727,
"learning_rate": 8.02144420131291e-05,
"loss": 15.4078,
"step": 318
},
{
"epoch": 0.22553135080199727,
"grad_norm": 23.436702728271484,
"learning_rate": 8.014223194748359e-05,
"loss": 14.9444,
"step": 319
},
{
"epoch": 0.22623834563209758,
"grad_norm": 18.821659088134766,
"learning_rate": 8.007002188183808e-05,
"loss": 15.274,
"step": 320
},
{
"epoch": 0.22694534046219786,
"grad_norm": 21.983806610107422,
"learning_rate": 7.999781181619256e-05,
"loss": 14.9261,
"step": 321
},
{
"epoch": 0.22765233529229817,
"grad_norm": 21.53936004638672,
"learning_rate": 7.992560175054705e-05,
"loss": 16.714,
"step": 322
},
{
"epoch": 0.22835933012239848,
"grad_norm": 19.361005783081055,
"learning_rate": 7.985339168490154e-05,
"loss": 14.7466,
"step": 323
},
{
"epoch": 0.2290663249524988,
"grad_norm": 18.576501846313477,
"learning_rate": 7.978118161925603e-05,
"loss": 16.1151,
"step": 324
},
{
"epoch": 0.2297733197825991,
"grad_norm": 18.696819305419922,
"learning_rate": 7.97089715536105e-05,
"loss": 15.3406,
"step": 325
},
{
"epoch": 0.23048031461269938,
"grad_norm": 22.205421447753906,
"learning_rate": 7.963676148796499e-05,
"loss": 14.5233,
"step": 326
},
{
"epoch": 0.2311873094427997,
"grad_norm": 21.301319122314453,
"learning_rate": 7.956455142231948e-05,
"loss": 14.5574,
"step": 327
},
{
"epoch": 0.2318943042729,
"grad_norm": 19.592769622802734,
"learning_rate": 7.949234135667396e-05,
"loss": 15.653,
"step": 328
},
{
"epoch": 0.23260129910300031,
"grad_norm": 20.18784523010254,
"learning_rate": 7.942013129102846e-05,
"loss": 15.2137,
"step": 329
},
{
"epoch": 0.23330829393310062,
"grad_norm": 27.54798698425293,
"learning_rate": 7.934792122538294e-05,
"loss": 16.9681,
"step": 330
},
{
"epoch": 0.2340152887632009,
"grad_norm": 18.011117935180664,
"learning_rate": 7.927571115973742e-05,
"loss": 14.9344,
"step": 331
},
{
"epoch": 0.23472228359330122,
"grad_norm": 17.401601791381836,
"learning_rate": 7.92035010940919e-05,
"loss": 15.7647,
"step": 332
},
{
"epoch": 0.23542927842340153,
"grad_norm": 21.158891677856445,
"learning_rate": 7.913129102844639e-05,
"loss": 14.8844,
"step": 333
},
{
"epoch": 0.23613627325350184,
"grad_norm": 19.250036239624023,
"learning_rate": 7.905908096280088e-05,
"loss": 16.1713,
"step": 334
},
{
"epoch": 0.23684326808360215,
"grad_norm": 19.4791202545166,
"learning_rate": 7.898687089715537e-05,
"loss": 16.3514,
"step": 335
},
{
"epoch": 0.23755026291370243,
"grad_norm": 18.960033416748047,
"learning_rate": 7.891466083150986e-05,
"loss": 14.6703,
"step": 336
},
{
"epoch": 0.23825725774380274,
"grad_norm": 19.392698287963867,
"learning_rate": 7.884245076586433e-05,
"loss": 15.4453,
"step": 337
},
{
"epoch": 0.23896425257390305,
"grad_norm": 24.39886474609375,
"learning_rate": 7.877024070021882e-05,
"loss": 16.2057,
"step": 338
},
{
"epoch": 0.23967124740400336,
"grad_norm": 20.101104736328125,
"learning_rate": 7.86980306345733e-05,
"loss": 14.8668,
"step": 339
},
{
"epoch": 0.24037824223410367,
"grad_norm": 19.75140380859375,
"learning_rate": 7.862582056892779e-05,
"loss": 14.1417,
"step": 340
},
{
"epoch": 0.24108523706420396,
"grad_norm": 20.29349708557129,
"learning_rate": 7.855361050328228e-05,
"loss": 15.7627,
"step": 341
},
{
"epoch": 0.24179223189430427,
"grad_norm": 19.623151779174805,
"learning_rate": 7.848140043763677e-05,
"loss": 16.0984,
"step": 342
},
{
"epoch": 0.24249922672440458,
"grad_norm": 18.32193946838379,
"learning_rate": 7.840919037199126e-05,
"loss": 14.4583,
"step": 343
},
{
"epoch": 0.2432062215545049,
"grad_norm": 23.00566291809082,
"learning_rate": 7.833698030634573e-05,
"loss": 15.3754,
"step": 344
},
{
"epoch": 0.2439132163846052,
"grad_norm": 18.748294830322266,
"learning_rate": 7.826477024070022e-05,
"loss": 15.1513,
"step": 345
},
{
"epoch": 0.24462021121470548,
"grad_norm": 19.001272201538086,
"learning_rate": 7.81925601750547e-05,
"loss": 13.934,
"step": 346
},
{
"epoch": 0.2453272060448058,
"grad_norm": 18.504486083984375,
"learning_rate": 7.812035010940919e-05,
"loss": 14.2281,
"step": 347
},
{
"epoch": 0.2460342008749061,
"grad_norm": 18.790077209472656,
"learning_rate": 7.804814004376369e-05,
"loss": 14.1004,
"step": 348
},
{
"epoch": 0.2467411957050064,
"grad_norm": 18.546693801879883,
"learning_rate": 7.797592997811817e-05,
"loss": 15.2542,
"step": 349
},
{
"epoch": 0.24744819053510672,
"grad_norm": 17.351430892944336,
"learning_rate": 7.790371991247266e-05,
"loss": 15.9018,
"step": 350
},
{
"epoch": 0.248155185365207,
"grad_norm": 22.45965576171875,
"learning_rate": 7.783150984682713e-05,
"loss": 16.1705,
"step": 351
},
{
"epoch": 0.24886218019530731,
"grad_norm": 24.87619400024414,
"learning_rate": 7.775929978118162e-05,
"loss": 15.8143,
"step": 352
},
{
"epoch": 0.24956917502540762,
"grad_norm": 19.052438735961914,
"learning_rate": 7.768708971553611e-05,
"loss": 14.8811,
"step": 353
},
{
"epoch": 0.25027616985550794,
"grad_norm": 16.649532318115234,
"learning_rate": 7.76148796498906e-05,
"loss": 14.9621,
"step": 354
},
{
"epoch": 0.2509831646856082,
"grad_norm": 21.764617919921875,
"learning_rate": 7.754266958424508e-05,
"loss": 15.196,
"step": 355
},
{
"epoch": 0.25169015951570856,
"grad_norm": 17.580827713012695,
"learning_rate": 7.747045951859957e-05,
"loss": 14.2718,
"step": 356
},
{
"epoch": 0.25239715434580884,
"grad_norm": 22.757190704345703,
"learning_rate": 7.739824945295405e-05,
"loss": 16.2971,
"step": 357
},
{
"epoch": 0.2531041491759092,
"grad_norm": 23.23011016845703,
"learning_rate": 7.732603938730853e-05,
"loss": 15.0298,
"step": 358
},
{
"epoch": 0.25381114400600946,
"grad_norm": 18.563827514648438,
"learning_rate": 7.725382932166302e-05,
"loss": 13.9777,
"step": 359
},
{
"epoch": 0.25451813883610974,
"grad_norm": 18.0550479888916,
"learning_rate": 7.718161925601751e-05,
"loss": 15.8373,
"step": 360
},
{
"epoch": 0.2552251336662101,
"grad_norm": 18.1561279296875,
"learning_rate": 7.7109409190372e-05,
"loss": 14.195,
"step": 361
},
{
"epoch": 0.25593212849631036,
"grad_norm": 20.923843383789062,
"learning_rate": 7.703719912472647e-05,
"loss": 14.4304,
"step": 362
},
{
"epoch": 0.2566391233264107,
"grad_norm": 19.624542236328125,
"learning_rate": 7.696498905908096e-05,
"loss": 14.8553,
"step": 363
},
{
"epoch": 0.257346118156511,
"grad_norm": 31.558462142944336,
"learning_rate": 7.689277899343545e-05,
"loss": 15.293,
"step": 364
},
{
"epoch": 0.25805311298661127,
"grad_norm": 22.13149642944336,
"learning_rate": 7.682056892778993e-05,
"loss": 16.3153,
"step": 365
},
{
"epoch": 0.2587601078167116,
"grad_norm": 20.600019454956055,
"learning_rate": 7.674835886214443e-05,
"loss": 13.9211,
"step": 366
},
{
"epoch": 0.2594671026468119,
"grad_norm": 17.84665298461914,
"learning_rate": 7.667614879649891e-05,
"loss": 14.4503,
"step": 367
},
{
"epoch": 0.2601740974769122,
"grad_norm": 21.022363662719727,
"learning_rate": 7.66039387308534e-05,
"loss": 15.0906,
"step": 368
},
{
"epoch": 0.2608810923070125,
"grad_norm": 20.317344665527344,
"learning_rate": 7.653172866520787e-05,
"loss": 15.4168,
"step": 369
},
{
"epoch": 0.2615880871371128,
"grad_norm": 21.427474975585938,
"learning_rate": 7.645951859956236e-05,
"loss": 13.9196,
"step": 370
},
{
"epoch": 0.26229508196721313,
"grad_norm": 18.921598434448242,
"learning_rate": 7.638730853391685e-05,
"loss": 15.762,
"step": 371
},
{
"epoch": 0.2630020767973134,
"grad_norm": 18.84344482421875,
"learning_rate": 7.631509846827134e-05,
"loss": 13.6822,
"step": 372
},
{
"epoch": 0.26370907162741375,
"grad_norm": 24.704998016357422,
"learning_rate": 7.624288840262583e-05,
"loss": 15.9235,
"step": 373
},
{
"epoch": 0.26441606645751403,
"grad_norm": 20.00655746459961,
"learning_rate": 7.617067833698031e-05,
"loss": 14.3952,
"step": 374
},
{
"epoch": 0.2651230612876143,
"grad_norm": 24.082733154296875,
"learning_rate": 7.60984682713348e-05,
"loss": 15.3455,
"step": 375
},
{
"epoch": 0.26583005611771465,
"grad_norm": 26.383352279663086,
"learning_rate": 7.602625820568927e-05,
"loss": 14.619,
"step": 376
},
{
"epoch": 0.26653705094781494,
"grad_norm": 19.369657516479492,
"learning_rate": 7.595404814004376e-05,
"loss": 15.3833,
"step": 377
},
{
"epoch": 0.2672440457779153,
"grad_norm": 17.416828155517578,
"learning_rate": 7.588183807439825e-05,
"loss": 15.9244,
"step": 378
},
{
"epoch": 0.26795104060801556,
"grad_norm": 20.211816787719727,
"learning_rate": 7.580962800875274e-05,
"loss": 14.9873,
"step": 379
},
{
"epoch": 0.26865803543811584,
"grad_norm": 18.542783737182617,
"learning_rate": 7.573741794310723e-05,
"loss": 15.0154,
"step": 380
},
{
"epoch": 0.2693650302682162,
"grad_norm": 25.668907165527344,
"learning_rate": 7.56652078774617e-05,
"loss": 14.9847,
"step": 381
},
{
"epoch": 0.27007202509831646,
"grad_norm": 17.961862564086914,
"learning_rate": 7.55929978118162e-05,
"loss": 15.7291,
"step": 382
},
{
"epoch": 0.2707790199284168,
"grad_norm": 17.490516662597656,
"learning_rate": 7.552078774617067e-05,
"loss": 14.7697,
"step": 383
},
{
"epoch": 0.2714860147585171,
"grad_norm": 20.74443817138672,
"learning_rate": 7.544857768052517e-05,
"loss": 13.9046,
"step": 384
},
{
"epoch": 0.27219300958861736,
"grad_norm": 20.980619430541992,
"learning_rate": 7.537636761487966e-05,
"loss": 15.4803,
"step": 385
},
{
"epoch": 0.2729000044187177,
"grad_norm": 16.939481735229492,
"learning_rate": 7.530415754923414e-05,
"loss": 14.944,
"step": 386
},
{
"epoch": 0.273606999248818,
"grad_norm": 21.0670223236084,
"learning_rate": 7.523194748358863e-05,
"loss": 14.8068,
"step": 387
},
{
"epoch": 0.2743139940789183,
"grad_norm": 19.005022048950195,
"learning_rate": 7.51597374179431e-05,
"loss": 14.6864,
"step": 388
},
{
"epoch": 0.2750209889090186,
"grad_norm": 21.184141159057617,
"learning_rate": 7.50875273522976e-05,
"loss": 15.4141,
"step": 389
},
{
"epoch": 0.2757279837391189,
"grad_norm": 24.792299270629883,
"learning_rate": 7.501531728665208e-05,
"loss": 15.6437,
"step": 390
},
{
"epoch": 0.2764349785692192,
"grad_norm": 21.77752113342285,
"learning_rate": 7.494310722100657e-05,
"loss": 14.8935,
"step": 391
},
{
"epoch": 0.2771419733993195,
"grad_norm": 17.924402236938477,
"learning_rate": 7.487089715536105e-05,
"loss": 14.8585,
"step": 392
},
{
"epoch": 0.27784896822941985,
"grad_norm": 16.164682388305664,
"learning_rate": 7.479868708971554e-05,
"loss": 15.3977,
"step": 393
},
{
"epoch": 0.27855596305952013,
"grad_norm": 20.66676139831543,
"learning_rate": 7.472647702407003e-05,
"loss": 13.9999,
"step": 394
},
{
"epoch": 0.2792629578896204,
"grad_norm": 17.76395606994629,
"learning_rate": 7.46542669584245e-05,
"loss": 15.7318,
"step": 395
},
{
"epoch": 0.27996995271972075,
"grad_norm": 20.148448944091797,
"learning_rate": 7.458205689277899e-05,
"loss": 16.1327,
"step": 396
},
{
"epoch": 0.28067694754982103,
"grad_norm": 17.416706085205078,
"learning_rate": 7.450984682713348e-05,
"loss": 15.1453,
"step": 397
},
{
"epoch": 0.28138394237992137,
"grad_norm": 19.03015899658203,
"learning_rate": 7.443763676148797e-05,
"loss": 14.2771,
"step": 398
},
{
"epoch": 0.28209093721002165,
"grad_norm": 24.30375862121582,
"learning_rate": 7.436542669584245e-05,
"loss": 15.2704,
"step": 399
},
{
"epoch": 0.28279793204012194,
"grad_norm": 15.849617004394531,
"learning_rate": 7.429321663019694e-05,
"loss": 13.9025,
"step": 400
},
{
"epoch": 0.2835049268702223,
"grad_norm": 18.239795684814453,
"learning_rate": 7.422100656455143e-05,
"loss": 14.2617,
"step": 401
},
{
"epoch": 0.28421192170032256,
"grad_norm": 16.995162963867188,
"learning_rate": 7.414879649890592e-05,
"loss": 15.2838,
"step": 402
},
{
"epoch": 0.2849189165304229,
"grad_norm": 18.74176788330078,
"learning_rate": 7.40765864332604e-05,
"loss": 14.6324,
"step": 403
},
{
"epoch": 0.2856259113605232,
"grad_norm": 19.847698211669922,
"learning_rate": 7.400437636761488e-05,
"loss": 14.863,
"step": 404
},
{
"epoch": 0.28633290619062346,
"grad_norm": 22.41224479675293,
"learning_rate": 7.393216630196937e-05,
"loss": 15.3046,
"step": 405
},
{
"epoch": 0.2870399010207238,
"grad_norm": 19.6437931060791,
"learning_rate": 7.385995623632385e-05,
"loss": 15.9613,
"step": 406
},
{
"epoch": 0.2877468958508241,
"grad_norm": 18.622400283813477,
"learning_rate": 7.378774617067834e-05,
"loss": 14.8365,
"step": 407
},
{
"epoch": 0.2884538906809244,
"grad_norm": 15.845386505126953,
"learning_rate": 7.371553610503283e-05,
"loss": 14.7285,
"step": 408
},
{
"epoch": 0.2891608855110247,
"grad_norm": 17.305540084838867,
"learning_rate": 7.364332603938731e-05,
"loss": 12.5411,
"step": 409
},
{
"epoch": 0.289867880341125,
"grad_norm": 21.858407974243164,
"learning_rate": 7.35711159737418e-05,
"loss": 14.1674,
"step": 410
},
{
"epoch": 0.2905748751712253,
"grad_norm": 17.716867446899414,
"learning_rate": 7.349890590809628e-05,
"loss": 13.9771,
"step": 411
},
{
"epoch": 0.2912818700013256,
"grad_norm": 19.153947830200195,
"learning_rate": 7.342669584245077e-05,
"loss": 15.4464,
"step": 412
},
{
"epoch": 0.29198886483142594,
"grad_norm": 19.239585876464844,
"learning_rate": 7.335448577680525e-05,
"loss": 15.6705,
"step": 413
},
{
"epoch": 0.2926958596615262,
"grad_norm": 20.419544219970703,
"learning_rate": 7.328227571115973e-05,
"loss": 14.868,
"step": 414
},
{
"epoch": 0.2934028544916265,
"grad_norm": 24.518224716186523,
"learning_rate": 7.321006564551424e-05,
"loss": 13.4362,
"step": 415
},
{
"epoch": 0.29410984932172685,
"grad_norm": 20.21552276611328,
"learning_rate": 7.313785557986871e-05,
"loss": 15.6206,
"step": 416
},
{
"epoch": 0.29481684415182713,
"grad_norm": 20.633731842041016,
"learning_rate": 7.30656455142232e-05,
"loss": 14.4674,
"step": 417
},
{
"epoch": 0.29552383898192747,
"grad_norm": 20.492298126220703,
"learning_rate": 7.299343544857768e-05,
"loss": 13.6936,
"step": 418
},
{
"epoch": 0.29623083381202775,
"grad_norm": 16.64995765686035,
"learning_rate": 7.292122538293217e-05,
"loss": 14.5776,
"step": 419
},
{
"epoch": 0.29693782864212803,
"grad_norm": 16.90037727355957,
"learning_rate": 7.284901531728666e-05,
"loss": 14.7451,
"step": 420
},
{
"epoch": 0.29764482347222837,
"grad_norm": 22.07757568359375,
"learning_rate": 7.277680525164115e-05,
"loss": 15.1344,
"step": 421
},
{
"epoch": 0.29835181830232865,
"grad_norm": 20.070301055908203,
"learning_rate": 7.270459518599564e-05,
"loss": 15.2919,
"step": 422
},
{
"epoch": 0.299058813132429,
"grad_norm": 20.72273826599121,
"learning_rate": 7.263238512035011e-05,
"loss": 15.1353,
"step": 423
},
{
"epoch": 0.2997658079625293,
"grad_norm": 18.998750686645508,
"learning_rate": 7.25601750547046e-05,
"loss": 13.6777,
"step": 424
},
{
"epoch": 0.30047280279262956,
"grad_norm": 17.999582290649414,
"learning_rate": 7.248796498905908e-05,
"loss": 13.6477,
"step": 425
},
{
"epoch": 0.3011797976227299,
"grad_norm": 25.615734100341797,
"learning_rate": 7.241575492341357e-05,
"loss": 15.8751,
"step": 426
},
{
"epoch": 0.3018867924528302,
"grad_norm": 19.07990074157715,
"learning_rate": 7.234354485776806e-05,
"loss": 14.2716,
"step": 427
},
{
"epoch": 0.3025937872829305,
"grad_norm": 18.45189094543457,
"learning_rate": 7.227133479212255e-05,
"loss": 15.6696,
"step": 428
},
{
"epoch": 0.3033007821130308,
"grad_norm": 17.569032669067383,
"learning_rate": 7.219912472647702e-05,
"loss": 14.7068,
"step": 429
},
{
"epoch": 0.3040077769431311,
"grad_norm": 17.7779483795166,
"learning_rate": 7.212691466083151e-05,
"loss": 15.4694,
"step": 430
},
{
"epoch": 0.3047147717732314,
"grad_norm": 21.57255744934082,
"learning_rate": 7.2054704595186e-05,
"loss": 14.7624,
"step": 431
},
{
"epoch": 0.3054217666033317,
"grad_norm": 18.774274826049805,
"learning_rate": 7.198249452954048e-05,
"loss": 15.298,
"step": 432
},
{
"epoch": 0.30612876143343204,
"grad_norm": 19.423994064331055,
"learning_rate": 7.191028446389498e-05,
"loss": 15.2514,
"step": 433
},
{
"epoch": 0.3068357562635323,
"grad_norm": 21.010740280151367,
"learning_rate": 7.183807439824946e-05,
"loss": 16.4023,
"step": 434
},
{
"epoch": 0.3075427510936326,
"grad_norm": 18.57482147216797,
"learning_rate": 7.176586433260394e-05,
"loss": 14.6922,
"step": 435
},
{
"epoch": 0.30824974592373294,
"grad_norm": 21.362197875976562,
"learning_rate": 7.169365426695842e-05,
"loss": 15.3559,
"step": 436
},
{
"epoch": 0.3089567407538332,
"grad_norm": 18.94207763671875,
"learning_rate": 7.162144420131291e-05,
"loss": 15.7067,
"step": 437
},
{
"epoch": 0.30966373558393356,
"grad_norm": 20.90492820739746,
"learning_rate": 7.15492341356674e-05,
"loss": 13.3837,
"step": 438
},
{
"epoch": 0.31037073041403385,
"grad_norm": 19.057661056518555,
"learning_rate": 7.147702407002189e-05,
"loss": 16.2484,
"step": 439
},
{
"epoch": 0.31107772524413413,
"grad_norm": 19.009706497192383,
"learning_rate": 7.140481400437638e-05,
"loss": 13.5528,
"step": 440
},
{
"epoch": 0.31178472007423447,
"grad_norm": 16.633657455444336,
"learning_rate": 7.133260393873085e-05,
"loss": 14.4136,
"step": 441
},
{
"epoch": 0.31249171490433475,
"grad_norm": 23.301849365234375,
"learning_rate": 7.126039387308534e-05,
"loss": 15.5324,
"step": 442
},
{
"epoch": 0.3131987097344351,
"grad_norm": 19.16399574279785,
"learning_rate": 7.118818380743982e-05,
"loss": 14.37,
"step": 443
},
{
"epoch": 0.31390570456453537,
"grad_norm": 20.813108444213867,
"learning_rate": 7.111597374179431e-05,
"loss": 15.9513,
"step": 444
},
{
"epoch": 0.31461269939463565,
"grad_norm": 20.032018661499023,
"learning_rate": 7.10437636761488e-05,
"loss": 14.5105,
"step": 445
},
{
"epoch": 0.315319694224736,
"grad_norm": 20.41152572631836,
"learning_rate": 7.097155361050329e-05,
"loss": 14.2337,
"step": 446
},
{
"epoch": 0.3160266890548363,
"grad_norm": 18.890499114990234,
"learning_rate": 7.089934354485778e-05,
"loss": 14.4166,
"step": 447
},
{
"epoch": 0.3167336838849366,
"grad_norm": 22.742734909057617,
"learning_rate": 7.082713347921225e-05,
"loss": 14.1206,
"step": 448
},
{
"epoch": 0.3174406787150369,
"grad_norm": 19.84697723388672,
"learning_rate": 7.075492341356674e-05,
"loss": 13.6646,
"step": 449
},
{
"epoch": 0.3181476735451372,
"grad_norm": 15.977441787719727,
"learning_rate": 7.068271334792122e-05,
"loss": 14.1537,
"step": 450
},
{
"epoch": 0.3188546683752375,
"grad_norm": 20.601211547851562,
"learning_rate": 7.061050328227571e-05,
"loss": 16.1045,
"step": 451
},
{
"epoch": 0.3195616632053378,
"grad_norm": 20.70127296447754,
"learning_rate": 7.053829321663021e-05,
"loss": 15.5461,
"step": 452
},
{
"epoch": 0.32026865803543814,
"grad_norm": 20.077213287353516,
"learning_rate": 7.046608315098469e-05,
"loss": 15.0559,
"step": 453
},
{
"epoch": 0.3209756528655384,
"grad_norm": 17.75333023071289,
"learning_rate": 7.039387308533918e-05,
"loss": 15.5746,
"step": 454
},
{
"epoch": 0.3216826476956387,
"grad_norm": 19.279191970825195,
"learning_rate": 7.032166301969365e-05,
"loss": 14.1511,
"step": 455
},
{
"epoch": 0.32238964252573904,
"grad_norm": 23.31242561340332,
"learning_rate": 7.024945295404814e-05,
"loss": 14.4471,
"step": 456
},
{
"epoch": 0.3230966373558393,
"grad_norm": 20.544729232788086,
"learning_rate": 7.017724288840263e-05,
"loss": 14.2697,
"step": 457
},
{
"epoch": 0.32380363218593966,
"grad_norm": 20.453166961669922,
"learning_rate": 7.010503282275712e-05,
"loss": 16.2543,
"step": 458
},
{
"epoch": 0.32451062701603994,
"grad_norm": 17.394886016845703,
"learning_rate": 7.00328227571116e-05,
"loss": 15.5171,
"step": 459
},
{
"epoch": 0.3252176218461402,
"grad_norm": 20.17839813232422,
"learning_rate": 6.996061269146609e-05,
"loss": 14.7051,
"step": 460
},
{
"epoch": 0.32592461667624056,
"grad_norm": 17.22258758544922,
"learning_rate": 6.988840262582057e-05,
"loss": 14.1478,
"step": 461
},
{
"epoch": 0.32663161150634085,
"grad_norm": 19.707887649536133,
"learning_rate": 6.981619256017505e-05,
"loss": 14.2404,
"step": 462
},
{
"epoch": 0.3273386063364412,
"grad_norm": 21.5001163482666,
"learning_rate": 6.974398249452954e-05,
"loss": 14.6055,
"step": 463
},
{
"epoch": 0.32804560116654147,
"grad_norm": 16.11020851135254,
"learning_rate": 6.967177242888403e-05,
"loss": 14.7407,
"step": 464
},
{
"epoch": 0.32875259599664175,
"grad_norm": 17.926362991333008,
"learning_rate": 6.959956236323852e-05,
"loss": 13.7554,
"step": 465
},
{
"epoch": 0.3294595908267421,
"grad_norm": 20.194194793701172,
"learning_rate": 6.9527352297593e-05,
"loss": 15.0067,
"step": 466
},
{
"epoch": 0.33016658565684237,
"grad_norm": 20.20330810546875,
"learning_rate": 6.945514223194748e-05,
"loss": 14.5011,
"step": 467
},
{
"epoch": 0.3308735804869427,
"grad_norm": 22.408048629760742,
"learning_rate": 6.938293216630197e-05,
"loss": 14.1764,
"step": 468
},
{
"epoch": 0.331580575317043,
"grad_norm": 16.50922966003418,
"learning_rate": 6.931072210065645e-05,
"loss": 14.2467,
"step": 469
},
{
"epoch": 0.3322875701471433,
"grad_norm": 19.757509231567383,
"learning_rate": 6.923851203501095e-05,
"loss": 15.6154,
"step": 470
},
{
"epoch": 0.3329945649772436,
"grad_norm": 18.997314453125,
"learning_rate": 6.916630196936543e-05,
"loss": 14.8435,
"step": 471
},
{
"epoch": 0.3337015598073439,
"grad_norm": 19.006284713745117,
"learning_rate": 6.909409190371992e-05,
"loss": 14.3226,
"step": 472
},
{
"epoch": 0.33440855463744423,
"grad_norm": 20.591211318969727,
"learning_rate": 6.90218818380744e-05,
"loss": 14.5859,
"step": 473
},
{
"epoch": 0.3351155494675445,
"grad_norm": 21.25404930114746,
"learning_rate": 6.894967177242888e-05,
"loss": 15.4005,
"step": 474
},
{
"epoch": 0.3358225442976448,
"grad_norm": 19.690427780151367,
"learning_rate": 6.887746170678337e-05,
"loss": 13.4944,
"step": 475
},
{
"epoch": 0.33652953912774514,
"grad_norm": 18.97504997253418,
"learning_rate": 6.880525164113786e-05,
"loss": 15.8937,
"step": 476
},
{
"epoch": 0.3372365339578454,
"grad_norm": 20.747446060180664,
"learning_rate": 6.873304157549235e-05,
"loss": 15.5289,
"step": 477
},
{
"epoch": 0.33794352878794576,
"grad_norm": 47.02959442138672,
"learning_rate": 6.866083150984683e-05,
"loss": 12.8834,
"step": 478
},
{
"epoch": 0.33865052361804604,
"grad_norm": 19.15494728088379,
"learning_rate": 6.858862144420132e-05,
"loss": 14.4253,
"step": 479
},
{
"epoch": 0.3393575184481463,
"grad_norm": 20.28019905090332,
"learning_rate": 6.851641137855579e-05,
"loss": 14.4871,
"step": 480
},
{
"epoch": 0.34006451327824666,
"grad_norm": 21.266618728637695,
"learning_rate": 6.844420131291028e-05,
"loss": 16.2024,
"step": 481
},
{
"epoch": 0.34077150810834694,
"grad_norm": 16.92910385131836,
"learning_rate": 6.837199124726477e-05,
"loss": 14.8702,
"step": 482
},
{
"epoch": 0.3414785029384473,
"grad_norm": 17.67193603515625,
"learning_rate": 6.829978118161926e-05,
"loss": 14.5345,
"step": 483
},
{
"epoch": 0.34218549776854756,
"grad_norm": 18.977420806884766,
"learning_rate": 6.822757111597375e-05,
"loss": 14.5069,
"step": 484
},
{
"epoch": 0.34289249259864785,
"grad_norm": 21.56228256225586,
"learning_rate": 6.815536105032823e-05,
"loss": 14.5382,
"step": 485
},
{
"epoch": 0.3435994874287482,
"grad_norm": 21.39179039001465,
"learning_rate": 6.808315098468272e-05,
"loss": 15.6628,
"step": 486
},
{
"epoch": 0.34430648225884847,
"grad_norm": 19.451231002807617,
"learning_rate": 6.801094091903719e-05,
"loss": 14.3872,
"step": 487
},
{
"epoch": 0.3450134770889488,
"grad_norm": 16.918291091918945,
"learning_rate": 6.79387308533917e-05,
"loss": 12.4315,
"step": 488
},
{
"epoch": 0.3457204719190491,
"grad_norm": 15.275004386901855,
"learning_rate": 6.786652078774618e-05,
"loss": 13.0213,
"step": 489
},
{
"epoch": 0.34642746674914937,
"grad_norm": 16.77974510192871,
"learning_rate": 6.779431072210066e-05,
"loss": 15.1569,
"step": 490
},
{
"epoch": 0.3471344615792497,
"grad_norm": 17.43290901184082,
"learning_rate": 6.772210065645515e-05,
"loss": 13.2295,
"step": 491
},
{
"epoch": 0.34784145640935,
"grad_norm": 18.017515182495117,
"learning_rate": 6.764989059080962e-05,
"loss": 13.2887,
"step": 492
},
{
"epoch": 0.34854845123945033,
"grad_norm": 20.11395263671875,
"learning_rate": 6.757768052516411e-05,
"loss": 14.3822,
"step": 493
},
{
"epoch": 0.3492554460695506,
"grad_norm": 16.92176628112793,
"learning_rate": 6.75054704595186e-05,
"loss": 13.756,
"step": 494
},
{
"epoch": 0.3499624408996509,
"grad_norm": 17.108287811279297,
"learning_rate": 6.743326039387309e-05,
"loss": 14.2471,
"step": 495
},
{
"epoch": 0.35066943572975123,
"grad_norm": 17.316476821899414,
"learning_rate": 6.736105032822757e-05,
"loss": 14.6229,
"step": 496
},
{
"epoch": 0.3513764305598515,
"grad_norm": 17.58360481262207,
"learning_rate": 6.728884026258206e-05,
"loss": 15.1998,
"step": 497
},
{
"epoch": 0.35208342538995185,
"grad_norm": 25.55293846130371,
"learning_rate": 6.721663019693655e-05,
"loss": 16.3773,
"step": 498
},
{
"epoch": 0.35279042022005214,
"grad_norm": 17.25092124938965,
"learning_rate": 6.714442013129102e-05,
"loss": 13.9458,
"step": 499
},
{
"epoch": 0.3534974150501524,
"grad_norm": 21.322893142700195,
"learning_rate": 6.707221006564551e-05,
"loss": 13.8305,
"step": 500
},
{
"epoch": 0.35420440988025276,
"grad_norm": 22.437015533447266,
"learning_rate": 6.7e-05,
"loss": 14.1218,
"step": 501
},
{
"epoch": 0.35491140471035304,
"grad_norm": 25.787944793701172,
"learning_rate": 6.692778993435449e-05,
"loss": 14.9072,
"step": 502
},
{
"epoch": 0.3556183995404534,
"grad_norm": 20.496932983398438,
"learning_rate": 6.685557986870897e-05,
"loss": 15.5787,
"step": 503
},
{
"epoch": 0.35632539437055366,
"grad_norm": 17.342716217041016,
"learning_rate": 6.678336980306346e-05,
"loss": 14.5554,
"step": 504
},
{
"epoch": 0.35703238920065394,
"grad_norm": 17.56897735595703,
"learning_rate": 6.671115973741795e-05,
"loss": 15.3665,
"step": 505
},
{
"epoch": 0.3577393840307543,
"grad_norm": 19.746797561645508,
"learning_rate": 6.663894967177244e-05,
"loss": 14.08,
"step": 506
},
{
"epoch": 0.35844637886085456,
"grad_norm": 17.250167846679688,
"learning_rate": 6.656673960612693e-05,
"loss": 14.5058,
"step": 507
},
{
"epoch": 0.3591533736909549,
"grad_norm": 16.402482986450195,
"learning_rate": 6.64945295404814e-05,
"loss": 15.3058,
"step": 508
},
{
"epoch": 0.3598603685210552,
"grad_norm": 17.24100685119629,
"learning_rate": 6.642231947483589e-05,
"loss": 15.2186,
"step": 509
},
{
"epoch": 0.36056736335115547,
"grad_norm": 17.75218963623047,
"learning_rate": 6.635010940919037e-05,
"loss": 14.2236,
"step": 510
},
{
"epoch": 0.3612743581812558,
"grad_norm": 19.30592918395996,
"learning_rate": 6.627789934354486e-05,
"loss": 13.7971,
"step": 511
},
{
"epoch": 0.3619813530113561,
"grad_norm": 16.735496520996094,
"learning_rate": 6.620568927789935e-05,
"loss": 12.8338,
"step": 512
},
{
"epoch": 0.3626883478414564,
"grad_norm": 18.537858963012695,
"learning_rate": 6.613347921225383e-05,
"loss": 14.715,
"step": 513
},
{
"epoch": 0.3633953426715567,
"grad_norm": 17.93909454345703,
"learning_rate": 6.606126914660832e-05,
"loss": 15.0655,
"step": 514
},
{
"epoch": 0.364102337501657,
"grad_norm": 21.184032440185547,
"learning_rate": 6.59890590809628e-05,
"loss": 13.1322,
"step": 515
},
{
"epoch": 0.36480933233175733,
"grad_norm": 22.548582077026367,
"learning_rate": 6.591684901531729e-05,
"loss": 14.5715,
"step": 516
},
{
"epoch": 0.3655163271618576,
"grad_norm": 17.91143226623535,
"learning_rate": 6.584463894967177e-05,
"loss": 16.3511,
"step": 517
},
{
"epoch": 0.36622332199195795,
"grad_norm": 21.47669219970703,
"learning_rate": 6.577242888402625e-05,
"loss": 14.7748,
"step": 518
},
{
"epoch": 0.36693031682205823,
"grad_norm": 17.893421173095703,
"learning_rate": 6.570021881838076e-05,
"loss": 13.6889,
"step": 519
},
{
"epoch": 0.3676373116521585,
"grad_norm": 18.86720848083496,
"learning_rate": 6.562800875273523e-05,
"loss": 14.3464,
"step": 520
},
{
"epoch": 0.36834430648225885,
"grad_norm": 17.936094284057617,
"learning_rate": 6.555579868708972e-05,
"loss": 14.1864,
"step": 521
},
{
"epoch": 0.36905130131235914,
"grad_norm": 17.583314895629883,
"learning_rate": 6.54835886214442e-05,
"loss": 14.6255,
"step": 522
},
{
"epoch": 0.3697582961424595,
"grad_norm": 18.166719436645508,
"learning_rate": 6.541137855579869e-05,
"loss": 14.9096,
"step": 523
},
{
"epoch": 0.37046529097255976,
"grad_norm": 15.307825088500977,
"learning_rate": 6.533916849015316e-05,
"loss": 13.7143,
"step": 524
},
{
"epoch": 0.3711722858026601,
"grad_norm": 14.538102149963379,
"learning_rate": 6.526695842450767e-05,
"loss": 13.5959,
"step": 525
},
{
"epoch": 0.3718792806327604,
"grad_norm": 17.80873680114746,
"learning_rate": 6.519474835886216e-05,
"loss": 13.6337,
"step": 526
},
{
"epoch": 0.37258627546286066,
"grad_norm": 18.472633361816406,
"learning_rate": 6.512253829321663e-05,
"loss": 13.9751,
"step": 527
},
{
"epoch": 0.373293270292961,
"grad_norm": 22.643638610839844,
"learning_rate": 6.505032822757112e-05,
"loss": 14.0875,
"step": 528
},
{
"epoch": 0.3740002651230613,
"grad_norm": 17.975576400756836,
"learning_rate": 6.49781181619256e-05,
"loss": 14.5217,
"step": 529
},
{
"epoch": 0.3747072599531616,
"grad_norm": 17.388160705566406,
"learning_rate": 6.490590809628009e-05,
"loss": 14.6185,
"step": 530
},
{
"epoch": 0.3754142547832619,
"grad_norm": 20.172466278076172,
"learning_rate": 6.483369803063458e-05,
"loss": 14.3061,
"step": 531
},
{
"epoch": 0.3761212496133622,
"grad_norm": 17.396696090698242,
"learning_rate": 6.476148796498907e-05,
"loss": 12.8709,
"step": 532
},
{
"epoch": 0.3768282444434625,
"grad_norm": 160.06143188476562,
"learning_rate": 6.468927789934354e-05,
"loss": 15.2518,
"step": 533
},
{
"epoch": 0.3775352392735628,
"grad_norm": 18.92376136779785,
"learning_rate": 6.461706783369803e-05,
"loss": 13.0073,
"step": 534
},
{
"epoch": 0.37824223410366314,
"grad_norm": 19.5358943939209,
"learning_rate": 6.454485776805252e-05,
"loss": 14.3698,
"step": 535
},
{
"epoch": 0.3789492289337634,
"grad_norm": 17.032445907592773,
"learning_rate": 6.4472647702407e-05,
"loss": 15.5133,
"step": 536
},
{
"epoch": 0.3796562237638637,
"grad_norm": 18.888500213623047,
"learning_rate": 6.44004376367615e-05,
"loss": 13.7176,
"step": 537
},
{
"epoch": 0.38036321859396405,
"grad_norm": 16.944372177124023,
"learning_rate": 6.432822757111598e-05,
"loss": 14.7796,
"step": 538
},
{
"epoch": 0.38107021342406433,
"grad_norm": 17.16058921813965,
"learning_rate": 6.425601750547046e-05,
"loss": 15.3055,
"step": 539
},
{
"epoch": 0.38177720825416467,
"grad_norm": 16.601852416992188,
"learning_rate": 6.418380743982494e-05,
"loss": 15.0371,
"step": 540
},
{
"epoch": 0.38248420308426495,
"grad_norm": 17.420427322387695,
"learning_rate": 6.411159737417943e-05,
"loss": 14.7263,
"step": 541
},
{
"epoch": 0.38319119791436523,
"grad_norm": 15.347443580627441,
"learning_rate": 6.403938730853392e-05,
"loss": 13.7447,
"step": 542
},
{
"epoch": 0.38389819274446557,
"grad_norm": 16.60781478881836,
"learning_rate": 6.396717724288841e-05,
"loss": 15.4559,
"step": 543
},
{
"epoch": 0.38460518757456585,
"grad_norm": 16.276700973510742,
"learning_rate": 6.38949671772429e-05,
"loss": 13.3962,
"step": 544
},
{
"epoch": 0.3853121824046662,
"grad_norm": 22.067554473876953,
"learning_rate": 6.382275711159737e-05,
"loss": 13.5659,
"step": 545
},
{
"epoch": 0.3860191772347665,
"grad_norm": 18.238330841064453,
"learning_rate": 6.375054704595186e-05,
"loss": 13.4307,
"step": 546
},
{
"epoch": 0.38672617206486676,
"grad_norm": 18.11393165588379,
"learning_rate": 6.367833698030634e-05,
"loss": 13.1592,
"step": 547
},
{
"epoch": 0.3874331668949671,
"grad_norm": 22.48624610900879,
"learning_rate": 6.360612691466083e-05,
"loss": 14.6767,
"step": 548
},
{
"epoch": 0.3881401617250674,
"grad_norm": 20.461183547973633,
"learning_rate": 6.353391684901532e-05,
"loss": 13.5224,
"step": 549
},
{
"epoch": 0.3888471565551677,
"grad_norm": 17.46296501159668,
"learning_rate": 6.346170678336981e-05,
"loss": 14.4498,
"step": 550
},
{
"epoch": 0.389554151385268,
"grad_norm": 16.26678466796875,
"learning_rate": 6.33894967177243e-05,
"loss": 12.3649,
"step": 551
},
{
"epoch": 0.3902611462153683,
"grad_norm": 16.18904685974121,
"learning_rate": 6.331728665207877e-05,
"loss": 14.502,
"step": 552
},
{
"epoch": 0.3909681410454686,
"grad_norm": 15.722247123718262,
"learning_rate": 6.324507658643326e-05,
"loss": 15.1522,
"step": 553
},
{
"epoch": 0.3916751358755689,
"grad_norm": 18.695825576782227,
"learning_rate": 6.317286652078774e-05,
"loss": 14.2403,
"step": 554
},
{
"epoch": 0.39238213070566924,
"grad_norm": 17.641643524169922,
"learning_rate": 6.310065645514223e-05,
"loss": 12.1116,
"step": 555
},
{
"epoch": 0.3930891255357695,
"grad_norm": 18.540403366088867,
"learning_rate": 6.302844638949673e-05,
"loss": 15.2668,
"step": 556
},
{
"epoch": 0.3937961203658698,
"grad_norm": 19.0169620513916,
"learning_rate": 6.29562363238512e-05,
"loss": 16.3723,
"step": 557
},
{
"epoch": 0.39450311519597014,
"grad_norm": 15.761820793151855,
"learning_rate": 6.28840262582057e-05,
"loss": 14.452,
"step": 558
},
{
"epoch": 0.3952101100260704,
"grad_norm": 16.140188217163086,
"learning_rate": 6.281181619256017e-05,
"loss": 14.1948,
"step": 559
},
{
"epoch": 0.39591710485617077,
"grad_norm": 17.851272583007812,
"learning_rate": 6.273960612691466e-05,
"loss": 14.1573,
"step": 560
},
{
"epoch": 0.39662409968627105,
"grad_norm": 16.403974533081055,
"learning_rate": 6.266739606126915e-05,
"loss": 13.3192,
"step": 561
},
{
"epoch": 0.39733109451637133,
"grad_norm": 20.728595733642578,
"learning_rate": 6.259518599562364e-05,
"loss": 16.3781,
"step": 562
},
{
"epoch": 0.39803808934647167,
"grad_norm": 17.328998565673828,
"learning_rate": 6.252297592997813e-05,
"loss": 14.3546,
"step": 563
},
{
"epoch": 0.39874508417657195,
"grad_norm": 17.004539489746094,
"learning_rate": 6.24507658643326e-05,
"loss": 14.7815,
"step": 564
},
{
"epoch": 0.3994520790066723,
"grad_norm": 18.353988647460938,
"learning_rate": 6.23785557986871e-05,
"loss": 14.2092,
"step": 565
},
{
"epoch": 0.40015907383677257,
"grad_norm": 18.1688175201416,
"learning_rate": 6.230634573304157e-05,
"loss": 14.9211,
"step": 566
},
{
"epoch": 0.40086606866687285,
"grad_norm": 16.45387840270996,
"learning_rate": 6.223413566739606e-05,
"loss": 13.2728,
"step": 567
},
{
"epoch": 0.4015730634969732,
"grad_norm": 15.667755126953125,
"learning_rate": 6.216192560175055e-05,
"loss": 13.6101,
"step": 568
},
{
"epoch": 0.4022800583270735,
"grad_norm": 16.11159324645996,
"learning_rate": 6.208971553610504e-05,
"loss": 15.6139,
"step": 569
},
{
"epoch": 0.4029870531571738,
"grad_norm": 16.811134338378906,
"learning_rate": 6.201750547045951e-05,
"loss": 13.7581,
"step": 570
},
{
"epoch": 0.4036940479872741,
"grad_norm": 25.489770889282227,
"learning_rate": 6.1945295404814e-05,
"loss": 13.9163,
"step": 571
},
{
"epoch": 0.4044010428173744,
"grad_norm": 17.47158432006836,
"learning_rate": 6.18730853391685e-05,
"loss": 14.9882,
"step": 572
},
{
"epoch": 0.4051080376474747,
"grad_norm": 16.549089431762695,
"learning_rate": 6.180087527352298e-05,
"loss": 14.9192,
"step": 573
},
{
"epoch": 0.405815032477575,
"grad_norm": 18.736282348632812,
"learning_rate": 6.172866520787747e-05,
"loss": 13.5939,
"step": 574
},
{
"epoch": 0.40652202730767534,
"grad_norm": 15.82469367980957,
"learning_rate": 6.165645514223195e-05,
"loss": 15.7199,
"step": 575
},
{
"epoch": 0.4072290221377756,
"grad_norm": 16.66925048828125,
"learning_rate": 6.158424507658644e-05,
"loss": 13.5893,
"step": 576
},
{
"epoch": 0.4079360169678759,
"grad_norm": 16.197856903076172,
"learning_rate": 6.151203501094091e-05,
"loss": 12.5333,
"step": 577
},
{
"epoch": 0.40864301179797624,
"grad_norm": 18.60298728942871,
"learning_rate": 6.143982494529542e-05,
"loss": 13.3609,
"step": 578
},
{
"epoch": 0.4093500066280765,
"grad_norm": 18.529094696044922,
"learning_rate": 6.13676148796499e-05,
"loss": 14.9298,
"step": 579
},
{
"epoch": 0.41005700145817686,
"grad_norm": 19.059642791748047,
"learning_rate": 6.129540481400438e-05,
"loss": 13.1881,
"step": 580
},
{
"epoch": 0.41076399628827714,
"grad_norm": 17.82415199279785,
"learning_rate": 6.122319474835887e-05,
"loss": 12.5294,
"step": 581
},
{
"epoch": 0.4114709911183774,
"grad_norm": 15.246479988098145,
"learning_rate": 6.115098468271335e-05,
"loss": 13.1343,
"step": 582
},
{
"epoch": 0.41217798594847777,
"grad_norm": 18.232219696044922,
"learning_rate": 6.107877461706784e-05,
"loss": 14.1395,
"step": 583
},
{
"epoch": 0.41288498077857805,
"grad_norm": 25.185590744018555,
"learning_rate": 6.100656455142232e-05,
"loss": 15.0905,
"step": 584
},
{
"epoch": 0.4135919756086784,
"grad_norm": 25.201833724975586,
"learning_rate": 6.093435448577681e-05,
"loss": 15.0098,
"step": 585
},
{
"epoch": 0.41429897043877867,
"grad_norm": 20.78860855102539,
"learning_rate": 6.08621444201313e-05,
"loss": 14.0662,
"step": 586
},
{
"epoch": 0.41500596526887895,
"grad_norm": 32.28705978393555,
"learning_rate": 6.0789934354485774e-05,
"loss": 14.281,
"step": 587
},
{
"epoch": 0.4157129600989793,
"grad_norm": 18.067970275878906,
"learning_rate": 6.071772428884027e-05,
"loss": 14.2242,
"step": 588
},
{
"epoch": 0.41641995492907957,
"grad_norm": 21.324962615966797,
"learning_rate": 6.0645514223194746e-05,
"loss": 13.2133,
"step": 589
},
{
"epoch": 0.4171269497591799,
"grad_norm": 14.73901081085205,
"learning_rate": 6.0573304157549235e-05,
"loss": 14.2005,
"step": 590
},
{
"epoch": 0.4178339445892802,
"grad_norm": 20.540681838989258,
"learning_rate": 6.050109409190372e-05,
"loss": 14.2718,
"step": 591
},
{
"epoch": 0.4185409394193805,
"grad_norm": 17.47296905517578,
"learning_rate": 6.042888402625821e-05,
"loss": 15.0917,
"step": 592
},
{
"epoch": 0.4192479342494808,
"grad_norm": 17.222566604614258,
"learning_rate": 6.03566739606127e-05,
"loss": 13.2435,
"step": 593
},
{
"epoch": 0.4199549290795811,
"grad_norm": 17.993486404418945,
"learning_rate": 6.028446389496718e-05,
"loss": 14.8301,
"step": 594
},
{
"epoch": 0.42066192390968143,
"grad_norm": 15.780810356140137,
"learning_rate": 6.021225382932167e-05,
"loss": 13.1897,
"step": 595
},
{
"epoch": 0.4213689187397817,
"grad_norm": 16.930824279785156,
"learning_rate": 6.0140043763676145e-05,
"loss": 13.8884,
"step": 596
},
{
"epoch": 0.422075913569882,
"grad_norm": 17.779985427856445,
"learning_rate": 6.006783369803064e-05,
"loss": 13.1271,
"step": 597
},
{
"epoch": 0.42278290839998234,
"grad_norm": 22.52481460571289,
"learning_rate": 5.999562363238512e-05,
"loss": 15.1317,
"step": 598
},
{
"epoch": 0.4234899032300826,
"grad_norm": 17.568035125732422,
"learning_rate": 5.9923413566739606e-05,
"loss": 14.3153,
"step": 599
},
{
"epoch": 0.42419689806018296,
"grad_norm": 15.473036766052246,
"learning_rate": 5.9851203501094096e-05,
"loss": 14.0021,
"step": 600
},
{
"epoch": 0.42490389289028324,
"grad_norm": 19.565162658691406,
"learning_rate": 5.977899343544858e-05,
"loss": 15.8899,
"step": 601
},
{
"epoch": 0.4256108877203835,
"grad_norm": 15.546586990356445,
"learning_rate": 5.970678336980307e-05,
"loss": 14.276,
"step": 602
},
{
"epoch": 0.42631788255048386,
"grad_norm": 16.37678337097168,
"learning_rate": 5.963457330415755e-05,
"loss": 14.8217,
"step": 603
},
{
"epoch": 0.42702487738058414,
"grad_norm": 17.48716163635254,
"learning_rate": 5.956236323851204e-05,
"loss": 13.0474,
"step": 604
},
{
"epoch": 0.4277318722106845,
"grad_norm": 15.518293380737305,
"learning_rate": 5.9490153172866516e-05,
"loss": 13.5803,
"step": 605
},
{
"epoch": 0.42843886704078477,
"grad_norm": 17.393667221069336,
"learning_rate": 5.941794310722101e-05,
"loss": 13.1692,
"step": 606
},
{
"epoch": 0.42914586187088505,
"grad_norm": 17.597732543945312,
"learning_rate": 5.934573304157549e-05,
"loss": 14.37,
"step": 607
},
{
"epoch": 0.4298528567009854,
"grad_norm": 20.43920135498047,
"learning_rate": 5.927352297592998e-05,
"loss": 14.6832,
"step": 608
},
{
"epoch": 0.43055985153108567,
"grad_norm": 16.366361618041992,
"learning_rate": 5.9201312910284466e-05,
"loss": 13.5082,
"step": 609
},
{
"epoch": 0.431266846361186,
"grad_norm": 20.55727195739746,
"learning_rate": 5.912910284463895e-05,
"loss": 14.3706,
"step": 610
},
{
"epoch": 0.4319738411912863,
"grad_norm": 15.425431251525879,
"learning_rate": 5.905689277899344e-05,
"loss": 14.2575,
"step": 611
},
{
"epoch": 0.4326808360213866,
"grad_norm": 15.034273147583008,
"learning_rate": 5.898468271334792e-05,
"loss": 14.42,
"step": 612
},
{
"epoch": 0.4333878308514869,
"grad_norm": 15.592576026916504,
"learning_rate": 5.891247264770241e-05,
"loss": 14.2622,
"step": 613
},
{
"epoch": 0.4340948256815872,
"grad_norm": 16.532777786254883,
"learning_rate": 5.8840262582056886e-05,
"loss": 14.1881,
"step": 614
},
{
"epoch": 0.43480182051168753,
"grad_norm": 16.355937957763672,
"learning_rate": 5.876805251641138e-05,
"loss": 14.3817,
"step": 615
},
{
"epoch": 0.4355088153417878,
"grad_norm": 17.28464126586914,
"learning_rate": 5.869584245076587e-05,
"loss": 14.5989,
"step": 616
},
{
"epoch": 0.4362158101718881,
"grad_norm": 15.116608619689941,
"learning_rate": 5.862363238512035e-05,
"loss": 14.2519,
"step": 617
},
{
"epoch": 0.43692280500198843,
"grad_norm": 15.99145793914795,
"learning_rate": 5.855142231947484e-05,
"loss": 13.1821,
"step": 618
},
{
"epoch": 0.4376297998320887,
"grad_norm": 16.347198486328125,
"learning_rate": 5.847921225382932e-05,
"loss": 14.6225,
"step": 619
},
{
"epoch": 0.43833679466218906,
"grad_norm": 20.394723892211914,
"learning_rate": 5.840700218818381e-05,
"loss": 12.4356,
"step": 620
},
{
"epoch": 0.43904378949228934,
"grad_norm": 15.125354766845703,
"learning_rate": 5.833479212253829e-05,
"loss": 14.4433,
"step": 621
},
{
"epoch": 0.4397507843223896,
"grad_norm": 20.564420700073242,
"learning_rate": 5.826258205689278e-05,
"loss": 13.1734,
"step": 622
},
{
"epoch": 0.44045777915248996,
"grad_norm": 22.414443969726562,
"learning_rate": 5.819037199124727e-05,
"loss": 14.8843,
"step": 623
},
{
"epoch": 0.44116477398259024,
"grad_norm": 17.679035186767578,
"learning_rate": 5.8118161925601754e-05,
"loss": 13.2975,
"step": 624
},
{
"epoch": 0.4418717688126906,
"grad_norm": 17.960941314697266,
"learning_rate": 5.804595185995624e-05,
"loss": 13.4987,
"step": 625
},
{
"epoch": 0.44257876364279086,
"grad_norm": 18.424537658691406,
"learning_rate": 5.797374179431072e-05,
"loss": 14.5378,
"step": 626
},
{
"epoch": 0.44328575847289114,
"grad_norm": 21.95231819152832,
"learning_rate": 5.790153172866521e-05,
"loss": 14.2405,
"step": 627
},
{
"epoch": 0.4439927533029915,
"grad_norm": 17.942035675048828,
"learning_rate": 5.782932166301969e-05,
"loss": 13.6413,
"step": 628
},
{
"epoch": 0.44469974813309177,
"grad_norm": 16.451675415039062,
"learning_rate": 5.775711159737418e-05,
"loss": 13.9083,
"step": 629
},
{
"epoch": 0.4454067429631921,
"grad_norm": 20.17315101623535,
"learning_rate": 5.768490153172867e-05,
"loss": 14.6052,
"step": 630
},
{
"epoch": 0.4461137377932924,
"grad_norm": 19.259559631347656,
"learning_rate": 5.761269146608315e-05,
"loss": 14.11,
"step": 631
},
{
"epoch": 0.44682073262339267,
"grad_norm": 17.016231536865234,
"learning_rate": 5.754048140043764e-05,
"loss": 13.2856,
"step": 632
},
{
"epoch": 0.447527727453493,
"grad_norm": 18.402177810668945,
"learning_rate": 5.7468271334792124e-05,
"loss": 13.7489,
"step": 633
},
{
"epoch": 0.4482347222835933,
"grad_norm": 17.117433547973633,
"learning_rate": 5.7396061269146614e-05,
"loss": 12.8285,
"step": 634
},
{
"epoch": 0.44894171711369363,
"grad_norm": 17.498863220214844,
"learning_rate": 5.732385120350109e-05,
"loss": 13.031,
"step": 635
},
{
"epoch": 0.4496487119437939,
"grad_norm": 18.746173858642578,
"learning_rate": 5.725164113785558e-05,
"loss": 13.4874,
"step": 636
},
{
"epoch": 0.4503557067738942,
"grad_norm": 17.161046981811523,
"learning_rate": 5.7179431072210075e-05,
"loss": 12.6915,
"step": 637
},
{
"epoch": 0.45106270160399453,
"grad_norm": 16.37310028076172,
"learning_rate": 5.710722100656455e-05,
"loss": 14.3346,
"step": 638
},
{
"epoch": 0.4517696964340948,
"grad_norm": 15.53888988494873,
"learning_rate": 5.703501094091904e-05,
"loss": 13.9628,
"step": 639
},
{
"epoch": 0.45247669126419515,
"grad_norm": 15.410018920898438,
"learning_rate": 5.696280087527352e-05,
"loss": 12.9064,
"step": 640
},
{
"epoch": 0.45318368609429543,
"grad_norm": 14.989131927490234,
"learning_rate": 5.689059080962801e-05,
"loss": 12.2063,
"step": 641
},
{
"epoch": 0.4538906809243957,
"grad_norm": 17.183135986328125,
"learning_rate": 5.6818380743982495e-05,
"loss": 14.2211,
"step": 642
},
{
"epoch": 0.45459767575449606,
"grad_norm": 15.444555282592773,
"learning_rate": 5.6746170678336985e-05,
"loss": 12.9035,
"step": 643
},
{
"epoch": 0.45530467058459634,
"grad_norm": 21.209993362426758,
"learning_rate": 5.667396061269146e-05,
"loss": 14.7551,
"step": 644
},
{
"epoch": 0.4560116654146967,
"grad_norm": 17.231727600097656,
"learning_rate": 5.660175054704595e-05,
"loss": 14.44,
"step": 645
},
{
"epoch": 0.45671866024479696,
"grad_norm": 14.78808879852295,
"learning_rate": 5.6529540481400446e-05,
"loss": 13.0656,
"step": 646
},
{
"epoch": 0.45742565507489724,
"grad_norm": 16.139951705932617,
"learning_rate": 5.645733041575492e-05,
"loss": 13.5085,
"step": 647
},
{
"epoch": 0.4581326499049976,
"grad_norm": 16.530834197998047,
"learning_rate": 5.638512035010941e-05,
"loss": 13.8217,
"step": 648
},
{
"epoch": 0.45883964473509786,
"grad_norm": 15.445442199707031,
"learning_rate": 5.6312910284463894e-05,
"loss": 13.6222,
"step": 649
},
{
"epoch": 0.4595466395651982,
"grad_norm": 18.647668838500977,
"learning_rate": 5.6240700218818384e-05,
"loss": 12.5635,
"step": 650
},
{
"epoch": 0.4602536343952985,
"grad_norm": 16.22862434387207,
"learning_rate": 5.6168490153172866e-05,
"loss": 13.4586,
"step": 651
},
{
"epoch": 0.46096062922539877,
"grad_norm": 14.387091636657715,
"learning_rate": 5.6096280087527356e-05,
"loss": 14.6621,
"step": 652
},
{
"epoch": 0.4616676240554991,
"grad_norm": 18.270614624023438,
"learning_rate": 5.6024070021881845e-05,
"loss": 13.5773,
"step": 653
},
{
"epoch": 0.4623746188855994,
"grad_norm": 14.584330558776855,
"learning_rate": 5.595185995623632e-05,
"loss": 12.9477,
"step": 654
},
{
"epoch": 0.4630816137156997,
"grad_norm": 17.457223892211914,
"learning_rate": 5.587964989059082e-05,
"loss": 13.4352,
"step": 655
},
{
"epoch": 0.4637886085458,
"grad_norm": 16.300119400024414,
"learning_rate": 5.580743982494529e-05,
"loss": 14.6178,
"step": 656
},
{
"epoch": 0.4644956033759003,
"grad_norm": 16.14145278930664,
"learning_rate": 5.573522975929978e-05,
"loss": 14.3801,
"step": 657
},
{
"epoch": 0.46520259820600063,
"grad_norm": 17.065845489501953,
"learning_rate": 5.5663019693654265e-05,
"loss": 14.7923,
"step": 658
},
{
"epoch": 0.4659095930361009,
"grad_norm": 18.83047103881836,
"learning_rate": 5.5590809628008754e-05,
"loss": 15.783,
"step": 659
},
{
"epoch": 0.46661658786620125,
"grad_norm": 17.221378326416016,
"learning_rate": 5.5518599562363244e-05,
"loss": 15.0286,
"step": 660
},
{
"epoch": 0.46732358269630153,
"grad_norm": 16.273569107055664,
"learning_rate": 5.5446389496717727e-05,
"loss": 13.78,
"step": 661
},
{
"epoch": 0.4680305775264018,
"grad_norm": 22.21025276184082,
"learning_rate": 5.5374179431072216e-05,
"loss": 14.7329,
"step": 662
},
{
"epoch": 0.46873757235650215,
"grad_norm": 15.67103385925293,
"learning_rate": 5.530196936542669e-05,
"loss": 13.9941,
"step": 663
},
{
"epoch": 0.46944456718660243,
"grad_norm": 18.95551872253418,
"learning_rate": 5.522975929978119e-05,
"loss": 13.7779,
"step": 664
},
{
"epoch": 0.4701515620167028,
"grad_norm": 17.02660369873047,
"learning_rate": 5.5157549234135664e-05,
"loss": 13.7453,
"step": 665
},
{
"epoch": 0.47085855684680306,
"grad_norm": 15.895587921142578,
"learning_rate": 5.508533916849015e-05,
"loss": 14.482,
"step": 666
},
{
"epoch": 0.47156555167690334,
"grad_norm": 15.587700843811035,
"learning_rate": 5.501312910284464e-05,
"loss": 13.6505,
"step": 667
},
{
"epoch": 0.4722725465070037,
"grad_norm": 15.575263977050781,
"learning_rate": 5.4940919037199125e-05,
"loss": 13.2174,
"step": 668
},
{
"epoch": 0.47297954133710396,
"grad_norm": 16.528423309326172,
"learning_rate": 5.4868708971553615e-05,
"loss": 14.258,
"step": 669
},
{
"epoch": 0.4736865361672043,
"grad_norm": 18.095470428466797,
"learning_rate": 5.47964989059081e-05,
"loss": 13.604,
"step": 670
},
{
"epoch": 0.4743935309973046,
"grad_norm": 19.37974739074707,
"learning_rate": 5.472428884026259e-05,
"loss": 14.3852,
"step": 671
},
{
"epoch": 0.47510052582740486,
"grad_norm": 17.618635177612305,
"learning_rate": 5.465207877461706e-05,
"loss": 13.0826,
"step": 672
},
{
"epoch": 0.4758075206575052,
"grad_norm": 15.718649864196777,
"learning_rate": 5.457986870897156e-05,
"loss": 14.0187,
"step": 673
},
{
"epoch": 0.4765145154876055,
"grad_norm": 15.932500839233398,
"learning_rate": 5.450765864332605e-05,
"loss": 13.1078,
"step": 674
},
{
"epoch": 0.4772215103177058,
"grad_norm": 16.64781951904297,
"learning_rate": 5.4435448577680524e-05,
"loss": 13.7169,
"step": 675
},
{
"epoch": 0.4779285051478061,
"grad_norm": 16.83970069885254,
"learning_rate": 5.4363238512035014e-05,
"loss": 12.9768,
"step": 676
},
{
"epoch": 0.4786354999779064,
"grad_norm": 17.58168601989746,
"learning_rate": 5.4291028446389496e-05,
"loss": 13.2531,
"step": 677
},
{
"epoch": 0.4793424948080067,
"grad_norm": 16.762788772583008,
"learning_rate": 5.4218818380743986e-05,
"loss": 12.9015,
"step": 678
},
{
"epoch": 0.480049489638107,
"grad_norm": 17.409942626953125,
"learning_rate": 5.414660831509847e-05,
"loss": 14.3195,
"step": 679
},
{
"epoch": 0.48075648446820735,
"grad_norm": 17.278255462646484,
"learning_rate": 5.407439824945296e-05,
"loss": 13.7849,
"step": 680
},
{
"epoch": 0.48146347929830763,
"grad_norm": 15.23432731628418,
"learning_rate": 5.4002188183807434e-05,
"loss": 13.4574,
"step": 681
},
{
"epoch": 0.4821704741284079,
"grad_norm": 15.967668533325195,
"learning_rate": 5.392997811816192e-05,
"loss": 14.7562,
"step": 682
},
{
"epoch": 0.48287746895850825,
"grad_norm": 17.900205612182617,
"learning_rate": 5.385776805251642e-05,
"loss": 14.5042,
"step": 683
},
{
"epoch": 0.48358446378860853,
"grad_norm": 20.11735725402832,
"learning_rate": 5.3785557986870895e-05,
"loss": 14.3391,
"step": 684
},
{
"epoch": 0.48429145861870887,
"grad_norm": 18.700807571411133,
"learning_rate": 5.3713347921225384e-05,
"loss": 14.819,
"step": 685
},
{
"epoch": 0.48499845344880915,
"grad_norm": 18.93809700012207,
"learning_rate": 5.364113785557987e-05,
"loss": 13.8124,
"step": 686
},
{
"epoch": 0.48570544827890944,
"grad_norm": 16.458322525024414,
"learning_rate": 5.3568927789934357e-05,
"loss": 12.7255,
"step": 687
},
{
"epoch": 0.4864124431090098,
"grad_norm": 18.722389221191406,
"learning_rate": 5.349671772428884e-05,
"loss": 12.7826,
"step": 688
},
{
"epoch": 0.48711943793911006,
"grad_norm": 16.748310089111328,
"learning_rate": 5.342450765864333e-05,
"loss": 15.2382,
"step": 689
},
{
"epoch": 0.4878264327692104,
"grad_norm": 17.008487701416016,
"learning_rate": 5.335229759299782e-05,
"loss": 13.056,
"step": 690
},
{
"epoch": 0.4885334275993107,
"grad_norm": 18.254348754882812,
"learning_rate": 5.3280087527352294e-05,
"loss": 13.7845,
"step": 691
},
{
"epoch": 0.48924042242941096,
"grad_norm": 18.541841506958008,
"learning_rate": 5.320787746170679e-05,
"loss": 12.7892,
"step": 692
},
{
"epoch": 0.4899474172595113,
"grad_norm": 19.059993743896484,
"learning_rate": 5.3135667396061266e-05,
"loss": 13.726,
"step": 693
},
{
"epoch": 0.4906544120896116,
"grad_norm": 17.068769454956055,
"learning_rate": 5.3063457330415755e-05,
"loss": 14.1355,
"step": 694
},
{
"epoch": 0.4913614069197119,
"grad_norm": 16.517131805419922,
"learning_rate": 5.299124726477024e-05,
"loss": 13.899,
"step": 695
},
{
"epoch": 0.4920684017498122,
"grad_norm": 18.925899505615234,
"learning_rate": 5.291903719912473e-05,
"loss": 13.7766,
"step": 696
},
{
"epoch": 0.4927753965799125,
"grad_norm": 17.37291717529297,
"learning_rate": 5.284682713347922e-05,
"loss": 15.1303,
"step": 697
},
{
"epoch": 0.4934823914100128,
"grad_norm": 17.28036117553711,
"learning_rate": 5.27746170678337e-05,
"loss": 13.8902,
"step": 698
},
{
"epoch": 0.4941893862401131,
"grad_norm": 16.070369720458984,
"learning_rate": 5.270240700218819e-05,
"loss": 13.8348,
"step": 699
},
{
"epoch": 0.49489638107021344,
"grad_norm": 15.887845039367676,
"learning_rate": 5.2630196936542665e-05,
"loss": 13.2942,
"step": 700
},
{
"epoch": 0.4956033759003137,
"grad_norm": 18.914541244506836,
"learning_rate": 5.255798687089716e-05,
"loss": 12.6842,
"step": 701
},
{
"epoch": 0.496310370730414,
"grad_norm": 19.618558883666992,
"learning_rate": 5.248577680525164e-05,
"loss": 14.2209,
"step": 702
},
{
"epoch": 0.49701736556051435,
"grad_norm": 19.322799682617188,
"learning_rate": 5.2413566739606126e-05,
"loss": 13.0114,
"step": 703
},
{
"epoch": 0.49772436039061463,
"grad_norm": 17.97751235961914,
"learning_rate": 5.2341356673960616e-05,
"loss": 15.5439,
"step": 704
},
{
"epoch": 0.49843135522071497,
"grad_norm": 16.378734588623047,
"learning_rate": 5.22691466083151e-05,
"loss": 13.4201,
"step": 705
},
{
"epoch": 0.49913835005081525,
"grad_norm": 19.820524215698242,
"learning_rate": 5.219693654266959e-05,
"loss": 13.839,
"step": 706
},
{
"epoch": 0.49984534488091553,
"grad_norm": 16.8355655670166,
"learning_rate": 5.212472647702407e-05,
"loss": 15.2506,
"step": 707
},
{
"epoch": 0.5005523397110159,
"grad_norm": 16.595901489257812,
"learning_rate": 5.205251641137856e-05,
"loss": 15.0275,
"step": 708
},
{
"epoch": 0.5012593345411162,
"grad_norm": 15.100963592529297,
"learning_rate": 5.1980306345733036e-05,
"loss": 12.7207,
"step": 709
},
{
"epoch": 0.5019663293712164,
"grad_norm": 19.29062271118164,
"learning_rate": 5.190809628008753e-05,
"loss": 14.1738,
"step": 710
},
{
"epoch": 0.5026733242013168,
"grad_norm": 14.703215599060059,
"learning_rate": 5.183588621444202e-05,
"loss": 12.9587,
"step": 711
},
{
"epoch": 0.5033803190314171,
"grad_norm": 20.79590606689453,
"learning_rate": 5.17636761487965e-05,
"loss": 13.2206,
"step": 712
},
{
"epoch": 0.5040873138615174,
"grad_norm": 20.551342010498047,
"learning_rate": 5.1691466083150987e-05,
"loss": 14.2096,
"step": 713
},
{
"epoch": 0.5047943086916177,
"grad_norm": 16.27815055847168,
"learning_rate": 5.161925601750547e-05,
"loss": 13.7653,
"step": 714
},
{
"epoch": 0.505501303521718,
"grad_norm": 15.842145919799805,
"learning_rate": 5.154704595185996e-05,
"loss": 13.535,
"step": 715
},
{
"epoch": 0.5062082983518184,
"grad_norm": 15.07165241241455,
"learning_rate": 5.147483588621444e-05,
"loss": 13.2622,
"step": 716
},
{
"epoch": 0.5069152931819186,
"grad_norm": 17.643245697021484,
"learning_rate": 5.140262582056893e-05,
"loss": 13.8885,
"step": 717
},
{
"epoch": 0.5076222880120189,
"grad_norm": 19.250755310058594,
"learning_rate": 5.1330415754923407e-05,
"loss": 13.8963,
"step": 718
},
{
"epoch": 0.5083292828421192,
"grad_norm": 19.744308471679688,
"learning_rate": 5.12582056892779e-05,
"loss": 13.7372,
"step": 719
},
{
"epoch": 0.5090362776722195,
"grad_norm": 18.203929901123047,
"learning_rate": 5.118599562363239e-05,
"loss": 13.5853,
"step": 720
},
{
"epoch": 0.5097432725023199,
"grad_norm": 16.226526260375977,
"learning_rate": 5.111378555798687e-05,
"loss": 12.8439,
"step": 721
},
{
"epoch": 0.5104502673324202,
"grad_norm": 15.358694076538086,
"learning_rate": 5.104157549234136e-05,
"loss": 13.431,
"step": 722
},
{
"epoch": 0.5111572621625204,
"grad_norm": 17.199031829833984,
"learning_rate": 5.096936542669584e-05,
"loss": 13.7108,
"step": 723
},
{
"epoch": 0.5118642569926207,
"grad_norm": 25.309284210205078,
"learning_rate": 5.089715536105033e-05,
"loss": 11.9808,
"step": 724
},
{
"epoch": 0.512571251822721,
"grad_norm": 14.532613754272461,
"learning_rate": 5.082494529540481e-05,
"loss": 13.4445,
"step": 725
},
{
"epoch": 0.5132782466528214,
"grad_norm": 15.828657150268555,
"learning_rate": 5.07527352297593e-05,
"loss": 13.6649,
"step": 726
},
{
"epoch": 0.5139852414829217,
"grad_norm": 17.062015533447266,
"learning_rate": 5.068052516411379e-05,
"loss": 13.5041,
"step": 727
},
{
"epoch": 0.514692236313022,
"grad_norm": 17.339509963989258,
"learning_rate": 5.0608315098468274e-05,
"loss": 14.0583,
"step": 728
},
{
"epoch": 0.5153992311431222,
"grad_norm": 18.00756072998047,
"learning_rate": 5.053610503282276e-05,
"loss": 14.3492,
"step": 729
},
{
"epoch": 0.5161062259732225,
"grad_norm": 15.429224967956543,
"learning_rate": 5.0463894967177246e-05,
"loss": 12.9663,
"step": 730
},
{
"epoch": 0.5168132208033229,
"grad_norm": 20.52001190185547,
"learning_rate": 5.039168490153173e-05,
"loss": 12.6214,
"step": 731
},
{
"epoch": 0.5175202156334232,
"grad_norm": 18.917030334472656,
"learning_rate": 5.031947483588622e-05,
"loss": 14.0433,
"step": 732
},
{
"epoch": 0.5182272104635235,
"grad_norm": 18.37959098815918,
"learning_rate": 5.02472647702407e-05,
"loss": 13.7528,
"step": 733
},
{
"epoch": 0.5189342052936238,
"grad_norm": 16.279067993164062,
"learning_rate": 5.017505470459518e-05,
"loss": 13.9975,
"step": 734
},
{
"epoch": 0.5196412001237241,
"grad_norm": 15.994711875915527,
"learning_rate": 5.010284463894967e-05,
"loss": 12.3636,
"step": 735
},
{
"epoch": 0.5203481949538244,
"grad_norm": 17.147029876708984,
"learning_rate": 5.0030634573304155e-05,
"loss": 13.8101,
"step": 736
},
{
"epoch": 0.5210551897839247,
"grad_norm": 17.469881057739258,
"learning_rate": 4.9958424507658645e-05,
"loss": 12.6908,
"step": 737
},
{
"epoch": 0.521762184614025,
"grad_norm": 17.51668930053711,
"learning_rate": 4.9886214442013134e-05,
"loss": 13.9891,
"step": 738
},
{
"epoch": 0.5224691794441253,
"grad_norm": 15.285223007202148,
"learning_rate": 4.9814004376367617e-05,
"loss": 13.1897,
"step": 739
},
{
"epoch": 0.5231761742742256,
"grad_norm": 16.82007598876953,
"learning_rate": 4.97417943107221e-05,
"loss": 13.274,
"step": 740
},
{
"epoch": 0.523883169104326,
"grad_norm": 16.825483322143555,
"learning_rate": 4.966958424507659e-05,
"loss": 13.3952,
"step": 741
},
{
"epoch": 0.5245901639344263,
"grad_norm": 14.855537414550781,
"learning_rate": 4.959737417943107e-05,
"loss": 11.8834,
"step": 742
},
{
"epoch": 0.5252971587645265,
"grad_norm": 17.55827522277832,
"learning_rate": 4.9525164113785554e-05,
"loss": 13.2329,
"step": 743
},
{
"epoch": 0.5260041535946268,
"grad_norm": 18.297609329223633,
"learning_rate": 4.945295404814004e-05,
"loss": 12.6993,
"step": 744
},
{
"epoch": 0.5267111484247271,
"grad_norm": 16.5467472076416,
"learning_rate": 4.938074398249453e-05,
"loss": 13.5798,
"step": 745
},
{
"epoch": 0.5274181432548275,
"grad_norm": 22.284879684448242,
"learning_rate": 4.9308533916849015e-05,
"loss": 13.9687,
"step": 746
},
{
"epoch": 0.5281251380849278,
"grad_norm": 17.486873626708984,
"learning_rate": 4.9236323851203505e-05,
"loss": 14.5825,
"step": 747
},
{
"epoch": 0.5288321329150281,
"grad_norm": 16.19643783569336,
"learning_rate": 4.916411378555799e-05,
"loss": 13.7465,
"step": 748
},
{
"epoch": 0.5295391277451283,
"grad_norm": 18.479318618774414,
"learning_rate": 4.909190371991247e-05,
"loss": 12.8109,
"step": 749
},
{
"epoch": 0.5302461225752286,
"grad_norm": 17.267440795898438,
"learning_rate": 4.901969365426696e-05,
"loss": 13.4236,
"step": 750
},
{
"epoch": 0.530953117405329,
"grad_norm": 16.307403564453125,
"learning_rate": 4.894748358862144e-05,
"loss": 12.6858,
"step": 751
},
{
"epoch": 0.5316601122354293,
"grad_norm": 14.105761528015137,
"learning_rate": 4.887527352297593e-05,
"loss": 14.5454,
"step": 752
},
{
"epoch": 0.5323671070655296,
"grad_norm": 19.452232360839844,
"learning_rate": 4.880306345733042e-05,
"loss": 12.3663,
"step": 753
},
{
"epoch": 0.5330741018956299,
"grad_norm": 16.186485290527344,
"learning_rate": 4.8730853391684904e-05,
"loss": 13.1737,
"step": 754
},
{
"epoch": 0.5337810967257302,
"grad_norm": 15.852377891540527,
"learning_rate": 4.8658643326039386e-05,
"loss": 12.8635,
"step": 755
},
{
"epoch": 0.5344880915558305,
"grad_norm": 16.142702102661133,
"learning_rate": 4.8586433260393876e-05,
"loss": 13.1662,
"step": 756
},
{
"epoch": 0.5351950863859308,
"grad_norm": 17.95094871520996,
"learning_rate": 4.851422319474836e-05,
"loss": 12.9467,
"step": 757
},
{
"epoch": 0.5359020812160311,
"grad_norm": 15.025976181030273,
"learning_rate": 4.844201312910284e-05,
"loss": 11.8638,
"step": 758
},
{
"epoch": 0.5366090760461314,
"grad_norm": 16.33597183227539,
"learning_rate": 4.836980306345733e-05,
"loss": 14.2383,
"step": 759
},
{
"epoch": 0.5373160708762317,
"grad_norm": 18.008317947387695,
"learning_rate": 4.829759299781182e-05,
"loss": 13.0342,
"step": 760
},
{
"epoch": 0.5380230657063321,
"grad_norm": 18.021818161010742,
"learning_rate": 4.82253829321663e-05,
"loss": 13.9605,
"step": 761
},
{
"epoch": 0.5387300605364324,
"grad_norm": 17.876670837402344,
"learning_rate": 4.815317286652079e-05,
"loss": 12.6084,
"step": 762
},
{
"epoch": 0.5394370553665326,
"grad_norm": 17.02657699584961,
"learning_rate": 4.8080962800875275e-05,
"loss": 14.9515,
"step": 763
},
{
"epoch": 0.5401440501966329,
"grad_norm": 19.892004013061523,
"learning_rate": 4.800875273522976e-05,
"loss": 14.049,
"step": 764
},
{
"epoch": 0.5408510450267332,
"grad_norm": 15.48623275756836,
"learning_rate": 4.793654266958425e-05,
"loss": 12.996,
"step": 765
},
{
"epoch": 0.5415580398568336,
"grad_norm": 15.053301811218262,
"learning_rate": 4.786433260393873e-05,
"loss": 11.4286,
"step": 766
},
{
"epoch": 0.5422650346869339,
"grad_norm": 18.168964385986328,
"learning_rate": 4.779212253829322e-05,
"loss": 14.5376,
"step": 767
},
{
"epoch": 0.5429720295170342,
"grad_norm": 15.425690650939941,
"learning_rate": 4.771991247264771e-05,
"loss": 12.4469,
"step": 768
},
{
"epoch": 0.5436790243471344,
"grad_norm": 14.769143104553223,
"learning_rate": 4.764770240700219e-05,
"loss": 13.4108,
"step": 769
},
{
"epoch": 0.5443860191772347,
"grad_norm": 15.65718936920166,
"learning_rate": 4.757549234135667e-05,
"loss": 12.8447,
"step": 770
},
{
"epoch": 0.5450930140073351,
"grad_norm": 14.275728225708008,
"learning_rate": 4.750328227571116e-05,
"loss": 12.6218,
"step": 771
},
{
"epoch": 0.5458000088374354,
"grad_norm": 15.515166282653809,
"learning_rate": 4.7431072210065645e-05,
"loss": 13.6347,
"step": 772
},
{
"epoch": 0.5465070036675357,
"grad_norm": 17.69280242919922,
"learning_rate": 4.735886214442013e-05,
"loss": 12.6965,
"step": 773
},
{
"epoch": 0.547213998497636,
"grad_norm": 16.940570831298828,
"learning_rate": 4.728665207877462e-05,
"loss": 14.2508,
"step": 774
},
{
"epoch": 0.5479209933277362,
"grad_norm": 14.886067390441895,
"learning_rate": 4.721444201312911e-05,
"loss": 12.0344,
"step": 775
},
{
"epoch": 0.5486279881578366,
"grad_norm": 14.371471405029297,
"learning_rate": 4.714223194748359e-05,
"loss": 12.4236,
"step": 776
},
{
"epoch": 0.5493349829879369,
"grad_norm": 15.021502494812012,
"learning_rate": 4.707002188183808e-05,
"loss": 12.5939,
"step": 777
},
{
"epoch": 0.5500419778180372,
"grad_norm": 16.001462936401367,
"learning_rate": 4.699781181619256e-05,
"loss": 12.8323,
"step": 778
},
{
"epoch": 0.5507489726481375,
"grad_norm": 16.28582191467285,
"learning_rate": 4.6925601750547044e-05,
"loss": 14.4682,
"step": 779
},
{
"epoch": 0.5514559674782378,
"grad_norm": 16.198036193847656,
"learning_rate": 4.6853391684901534e-05,
"loss": 12.7082,
"step": 780
},
{
"epoch": 0.5521629623083382,
"grad_norm": 15.891390800476074,
"learning_rate": 4.6781181619256016e-05,
"loss": 12.4412,
"step": 781
},
{
"epoch": 0.5528699571384385,
"grad_norm": 19.185029983520508,
"learning_rate": 4.6708971553610506e-05,
"loss": 13.2326,
"step": 782
},
{
"epoch": 0.5535769519685387,
"grad_norm": 16.253828048706055,
"learning_rate": 4.663676148796499e-05,
"loss": 13.2036,
"step": 783
},
{
"epoch": 0.554283946798639,
"grad_norm": 19.45000648498535,
"learning_rate": 4.656455142231948e-05,
"loss": 14.3846,
"step": 784
},
{
"epoch": 0.5549909416287393,
"grad_norm": 15.610483169555664,
"learning_rate": 4.649234135667396e-05,
"loss": 11.9146,
"step": 785
},
{
"epoch": 0.5556979364588397,
"grad_norm": 15.691834449768066,
"learning_rate": 4.642013129102844e-05,
"loss": 14.6285,
"step": 786
},
{
"epoch": 0.55640493128894,
"grad_norm": 15.287580490112305,
"learning_rate": 4.634792122538293e-05,
"loss": 13.8478,
"step": 787
},
{
"epoch": 0.5571119261190403,
"grad_norm": 16.664316177368164,
"learning_rate": 4.6275711159737415e-05,
"loss": 13.5122,
"step": 788
},
{
"epoch": 0.5578189209491405,
"grad_norm": 17.661788940429688,
"learning_rate": 4.6203501094091905e-05,
"loss": 13.331,
"step": 789
},
{
"epoch": 0.5585259157792408,
"grad_norm": 16.202898025512695,
"learning_rate": 4.6131291028446394e-05,
"loss": 12.4447,
"step": 790
},
{
"epoch": 0.5592329106093412,
"grad_norm": 18.325153350830078,
"learning_rate": 4.605908096280088e-05,
"loss": 14.4831,
"step": 791
},
{
"epoch": 0.5599399054394415,
"grad_norm": 17.563846588134766,
"learning_rate": 4.598687089715536e-05,
"loss": 12.4172,
"step": 792
},
{
"epoch": 0.5606469002695418,
"grad_norm": 17.11208724975586,
"learning_rate": 4.591466083150985e-05,
"loss": 13.6366,
"step": 793
},
{
"epoch": 0.5613538950996421,
"grad_norm": 16.057981491088867,
"learning_rate": 4.584245076586433e-05,
"loss": 12.2651,
"step": 794
},
{
"epoch": 0.5620608899297423,
"grad_norm": 19.833967208862305,
"learning_rate": 4.5770240700218814e-05,
"loss": 15.0997,
"step": 795
},
{
"epoch": 0.5627678847598427,
"grad_norm": 15.77823257446289,
"learning_rate": 4.5698030634573303e-05,
"loss": 12.1512,
"step": 796
},
{
"epoch": 0.563474879589943,
"grad_norm": 18.81294059753418,
"learning_rate": 4.562582056892779e-05,
"loss": 12.9248,
"step": 797
},
{
"epoch": 0.5641818744200433,
"grad_norm": 15.902440071105957,
"learning_rate": 4.5553610503282275e-05,
"loss": 12.3549,
"step": 798
},
{
"epoch": 0.5648888692501436,
"grad_norm": 15.62425422668457,
"learning_rate": 4.5481400437636765e-05,
"loss": 13.4902,
"step": 799
},
{
"epoch": 0.5655958640802439,
"grad_norm": 17.2721004486084,
"learning_rate": 4.540919037199125e-05,
"loss": 14.88,
"step": 800
},
{
"epoch": 0.5663028589103443,
"grad_norm": 17.061033248901367,
"learning_rate": 4.533698030634573e-05,
"loss": 13.0052,
"step": 801
},
{
"epoch": 0.5670098537404445,
"grad_norm": 15.536235809326172,
"learning_rate": 4.526477024070022e-05,
"loss": 13.4578,
"step": 802
},
{
"epoch": 0.5677168485705448,
"grad_norm": 17.601093292236328,
"learning_rate": 4.51925601750547e-05,
"loss": 14.9889,
"step": 803
},
{
"epoch": 0.5684238434006451,
"grad_norm": 15.94250202178955,
"learning_rate": 4.512035010940919e-05,
"loss": 13.9519,
"step": 804
},
{
"epoch": 0.5691308382307454,
"grad_norm": 17.229337692260742,
"learning_rate": 4.504814004376368e-05,
"loss": 14.2852,
"step": 805
},
{
"epoch": 0.5698378330608458,
"grad_norm": 19.297306060791016,
"learning_rate": 4.4975929978118164e-05,
"loss": 12.9865,
"step": 806
},
{
"epoch": 0.5705448278909461,
"grad_norm": 17.727935791015625,
"learning_rate": 4.4903719912472646e-05,
"loss": 12.9064,
"step": 807
},
{
"epoch": 0.5712518227210464,
"grad_norm": 15.566085815429688,
"learning_rate": 4.4831509846827136e-05,
"loss": 13.8597,
"step": 808
},
{
"epoch": 0.5719588175511466,
"grad_norm": 17.039579391479492,
"learning_rate": 4.475929978118162e-05,
"loss": 12.2828,
"step": 809
},
{
"epoch": 0.5726658123812469,
"grad_norm": 17.201379776000977,
"learning_rate": 4.46870897155361e-05,
"loss": 13.0608,
"step": 810
},
{
"epoch": 0.5733728072113473,
"grad_norm": 17.841808319091797,
"learning_rate": 4.461487964989059e-05,
"loss": 14.4647,
"step": 811
},
{
"epoch": 0.5740798020414476,
"grad_norm": 17.14201545715332,
"learning_rate": 4.454266958424508e-05,
"loss": 14.1853,
"step": 812
},
{
"epoch": 0.5747867968715479,
"grad_norm": 14.405500411987305,
"learning_rate": 4.447045951859956e-05,
"loss": 13.2748,
"step": 813
},
{
"epoch": 0.5754937917016482,
"grad_norm": 15.193861961364746,
"learning_rate": 4.439824945295405e-05,
"loss": 12.3332,
"step": 814
},
{
"epoch": 0.5762007865317484,
"grad_norm": 14.01460075378418,
"learning_rate": 4.4326039387308535e-05,
"loss": 13.8787,
"step": 815
},
{
"epoch": 0.5769077813618488,
"grad_norm": 15.624826431274414,
"learning_rate": 4.425382932166302e-05,
"loss": 13.7502,
"step": 816
},
{
"epoch": 0.5776147761919491,
"grad_norm": 18.81705093383789,
"learning_rate": 4.418161925601751e-05,
"loss": 12.6278,
"step": 817
},
{
"epoch": 0.5783217710220494,
"grad_norm": 15.476762771606445,
"learning_rate": 4.410940919037199e-05,
"loss": 12.5758,
"step": 818
},
{
"epoch": 0.5790287658521497,
"grad_norm": 16.792394638061523,
"learning_rate": 4.403719912472648e-05,
"loss": 13.3397,
"step": 819
},
{
"epoch": 0.57973576068225,
"grad_norm": 15.57129192352295,
"learning_rate": 4.396498905908097e-05,
"loss": 14.78,
"step": 820
},
{
"epoch": 0.5804427555123504,
"grad_norm": 15.04116153717041,
"learning_rate": 4.389277899343545e-05,
"loss": 15.0815,
"step": 821
},
{
"epoch": 0.5811497503424506,
"grad_norm": 14.832640647888184,
"learning_rate": 4.3820568927789933e-05,
"loss": 12.3045,
"step": 822
},
{
"epoch": 0.5818567451725509,
"grad_norm": 16.435665130615234,
"learning_rate": 4.374835886214442e-05,
"loss": 13.0494,
"step": 823
},
{
"epoch": 0.5825637400026512,
"grad_norm": 15.572066307067871,
"learning_rate": 4.3676148796498905e-05,
"loss": 12.4681,
"step": 824
},
{
"epoch": 0.5832707348327515,
"grad_norm": 16.7429141998291,
"learning_rate": 4.360393873085339e-05,
"loss": 12.9876,
"step": 825
},
{
"epoch": 0.5839777296628519,
"grad_norm": 15.126506805419922,
"learning_rate": 4.353172866520788e-05,
"loss": 14.0403,
"step": 826
},
{
"epoch": 0.5846847244929522,
"grad_norm": 16.73342514038086,
"learning_rate": 4.345951859956237e-05,
"loss": 13.5968,
"step": 827
},
{
"epoch": 0.5853917193230525,
"grad_norm": 16.196666717529297,
"learning_rate": 4.338730853391685e-05,
"loss": 12.7303,
"step": 828
},
{
"epoch": 0.5860987141531527,
"grad_norm": 15.55926513671875,
"learning_rate": 4.331509846827133e-05,
"loss": 13.6608,
"step": 829
},
{
"epoch": 0.586805708983253,
"grad_norm": 16.581199645996094,
"learning_rate": 4.324288840262582e-05,
"loss": 13.3497,
"step": 830
},
{
"epoch": 0.5875127038133534,
"grad_norm": 18.875598907470703,
"learning_rate": 4.3170678336980304e-05,
"loss": 14.7135,
"step": 831
},
{
"epoch": 0.5882196986434537,
"grad_norm": 17.710857391357422,
"learning_rate": 4.3098468271334794e-05,
"loss": 12.4087,
"step": 832
},
{
"epoch": 0.588926693473554,
"grad_norm": 18.93647575378418,
"learning_rate": 4.3026258205689276e-05,
"loss": 14.614,
"step": 833
},
{
"epoch": 0.5896336883036543,
"grad_norm": 17.00315284729004,
"learning_rate": 4.2954048140043766e-05,
"loss": 13.7616,
"step": 834
},
{
"epoch": 0.5903406831337545,
"grad_norm": 19.230234146118164,
"learning_rate": 4.288183807439825e-05,
"loss": 13.9937,
"step": 835
},
{
"epoch": 0.5910476779638549,
"grad_norm": 18.548187255859375,
"learning_rate": 4.280962800875274e-05,
"loss": 12.0333,
"step": 836
},
{
"epoch": 0.5917546727939552,
"grad_norm": 20.600353240966797,
"learning_rate": 4.273741794310722e-05,
"loss": 13.4433,
"step": 837
},
{
"epoch": 0.5924616676240555,
"grad_norm": 17.80040740966797,
"learning_rate": 4.26652078774617e-05,
"loss": 13.8373,
"step": 838
},
{
"epoch": 0.5931686624541558,
"grad_norm": 15.60269832611084,
"learning_rate": 4.259299781181619e-05,
"loss": 13.4195,
"step": 839
},
{
"epoch": 0.5938756572842561,
"grad_norm": 24.892202377319336,
"learning_rate": 4.2520787746170675e-05,
"loss": 12.2826,
"step": 840
},
{
"epoch": 0.5945826521143565,
"grad_norm": 16.01949119567871,
"learning_rate": 4.2448577680525165e-05,
"loss": 14.2682,
"step": 841
},
{
"epoch": 0.5952896469444567,
"grad_norm": 17.417011260986328,
"learning_rate": 4.2376367614879654e-05,
"loss": 14.0882,
"step": 842
},
{
"epoch": 0.595996641774557,
"grad_norm": 17.576231002807617,
"learning_rate": 4.230415754923414e-05,
"loss": 13.4012,
"step": 843
},
{
"epoch": 0.5967036366046573,
"grad_norm": 16.846078872680664,
"learning_rate": 4.223194748358862e-05,
"loss": 13.3299,
"step": 844
},
{
"epoch": 0.5974106314347576,
"grad_norm": 18.340309143066406,
"learning_rate": 4.215973741794311e-05,
"loss": 13.7724,
"step": 845
},
{
"epoch": 0.598117626264858,
"grad_norm": 16.44016456604004,
"learning_rate": 4.208752735229759e-05,
"loss": 13.7315,
"step": 846
},
{
"epoch": 0.5988246210949583,
"grad_norm": 19.964340209960938,
"learning_rate": 4.2015317286652074e-05,
"loss": 15.7788,
"step": 847
},
{
"epoch": 0.5995316159250585,
"grad_norm": 14.742883682250977,
"learning_rate": 4.1943107221006563e-05,
"loss": 13.6036,
"step": 848
},
{
"epoch": 0.6002386107551588,
"grad_norm": 15.614455223083496,
"learning_rate": 4.187089715536105e-05,
"loss": 14.4348,
"step": 849
},
{
"epoch": 0.6009456055852591,
"grad_norm": 20.77018928527832,
"learning_rate": 4.1798687089715536e-05,
"loss": 12.572,
"step": 850
},
{
"epoch": 0.6016526004153595,
"grad_norm": 16.104507446289062,
"learning_rate": 4.1726477024070025e-05,
"loss": 14.0879,
"step": 851
},
{
"epoch": 0.6023595952454598,
"grad_norm": 18.730669021606445,
"learning_rate": 4.165426695842451e-05,
"loss": 14.3006,
"step": 852
},
{
"epoch": 0.6030665900755601,
"grad_norm": 16.253700256347656,
"learning_rate": 4.158205689277899e-05,
"loss": 12.9148,
"step": 853
},
{
"epoch": 0.6037735849056604,
"grad_norm": 15.1681547164917,
"learning_rate": 4.150984682713348e-05,
"loss": 13.2516,
"step": 854
},
{
"epoch": 0.6044805797357606,
"grad_norm": 16.01544952392578,
"learning_rate": 4.143763676148796e-05,
"loss": 13.3555,
"step": 855
},
{
"epoch": 0.605187574565861,
"grad_norm": 18.961162567138672,
"learning_rate": 4.136542669584245e-05,
"loss": 11.8692,
"step": 856
},
{
"epoch": 0.6058945693959613,
"grad_norm": 16.640958786010742,
"learning_rate": 4.129321663019694e-05,
"loss": 11.0722,
"step": 857
},
{
"epoch": 0.6066015642260616,
"grad_norm": 17.749479293823242,
"learning_rate": 4.1221006564551424e-05,
"loss": 14.4611,
"step": 858
},
{
"epoch": 0.6073085590561619,
"grad_norm": 18.945444107055664,
"learning_rate": 4.1148796498905906e-05,
"loss": 14.5662,
"step": 859
},
{
"epoch": 0.6080155538862622,
"grad_norm": 16.346588134765625,
"learning_rate": 4.1076586433260396e-05,
"loss": 13.4941,
"step": 860
},
{
"epoch": 0.6087225487163626,
"grad_norm": 18.323013305664062,
"learning_rate": 4.100437636761488e-05,
"loss": 12.2041,
"step": 861
},
{
"epoch": 0.6094295435464628,
"grad_norm": 18.375028610229492,
"learning_rate": 4.093216630196936e-05,
"loss": 13.5977,
"step": 862
},
{
"epoch": 0.6101365383765631,
"grad_norm": 16.597517013549805,
"learning_rate": 4.085995623632386e-05,
"loss": 13.6122,
"step": 863
},
{
"epoch": 0.6108435332066634,
"grad_norm": 15.910663604736328,
"learning_rate": 4.078774617067834e-05,
"loss": 13.3603,
"step": 864
},
{
"epoch": 0.6115505280367637,
"grad_norm": 20.180830001831055,
"learning_rate": 4.071553610503282e-05,
"loss": 12.6904,
"step": 865
},
{
"epoch": 0.6122575228668641,
"grad_norm": 17.37298011779785,
"learning_rate": 4.064332603938731e-05,
"loss": 13.7586,
"step": 866
},
{
"epoch": 0.6129645176969644,
"grad_norm": 19.180036544799805,
"learning_rate": 4.0571115973741795e-05,
"loss": 13.3099,
"step": 867
},
{
"epoch": 0.6136715125270646,
"grad_norm": 16.653715133666992,
"learning_rate": 4.049890590809628e-05,
"loss": 14.2502,
"step": 868
},
{
"epoch": 0.6143785073571649,
"grad_norm": 16.960859298706055,
"learning_rate": 4.042669584245077e-05,
"loss": 14.3287,
"step": 869
},
{
"epoch": 0.6150855021872652,
"grad_norm": 16.48638153076172,
"learning_rate": 4.035448577680525e-05,
"loss": 12.0678,
"step": 870
},
{
"epoch": 0.6157924970173656,
"grad_norm": 14.477529525756836,
"learning_rate": 4.028227571115974e-05,
"loss": 14.2217,
"step": 871
},
{
"epoch": 0.6164994918474659,
"grad_norm": 16.016817092895508,
"learning_rate": 4.021006564551423e-05,
"loss": 13.3089,
"step": 872
},
{
"epoch": 0.6172064866775662,
"grad_norm": 15.40858268737793,
"learning_rate": 4.013785557986871e-05,
"loss": 14.2675,
"step": 873
},
{
"epoch": 0.6179134815076665,
"grad_norm": 17.52202796936035,
"learning_rate": 4.0065645514223193e-05,
"loss": 13.6666,
"step": 874
},
{
"epoch": 0.6186204763377667,
"grad_norm": 16.316030502319336,
"learning_rate": 3.999343544857768e-05,
"loss": 13.3957,
"step": 875
},
{
"epoch": 0.6193274711678671,
"grad_norm": 16.079437255859375,
"learning_rate": 3.9921225382932166e-05,
"loss": 12.3198,
"step": 876
},
{
"epoch": 0.6200344659979674,
"grad_norm": 15.243290901184082,
"learning_rate": 3.984901531728665e-05,
"loss": 12.659,
"step": 877
},
{
"epoch": 0.6207414608280677,
"grad_norm": 17.72751808166504,
"learning_rate": 3.977680525164114e-05,
"loss": 13.4509,
"step": 878
},
{
"epoch": 0.621448455658168,
"grad_norm": 16.8057918548584,
"learning_rate": 3.970459518599563e-05,
"loss": 11.7917,
"step": 879
},
{
"epoch": 0.6221554504882683,
"grad_norm": 16.497268676757812,
"learning_rate": 3.963238512035011e-05,
"loss": 13.8699,
"step": 880
},
{
"epoch": 0.6228624453183687,
"grad_norm": 16.132577896118164,
"learning_rate": 3.956017505470459e-05,
"loss": 13.2419,
"step": 881
},
{
"epoch": 0.6235694401484689,
"grad_norm": 17.70035171508789,
"learning_rate": 3.948796498905908e-05,
"loss": 13.3996,
"step": 882
},
{
"epoch": 0.6242764349785692,
"grad_norm": 16.91880989074707,
"learning_rate": 3.9415754923413564e-05,
"loss": 13.4762,
"step": 883
},
{
"epoch": 0.6249834298086695,
"grad_norm": 15.370857238769531,
"learning_rate": 3.9343544857768054e-05,
"loss": 13.7917,
"step": 884
},
{
"epoch": 0.6256904246387698,
"grad_norm": 15.553352355957031,
"learning_rate": 3.9271334792122536e-05,
"loss": 11.7293,
"step": 885
},
{
"epoch": 0.6263974194688702,
"grad_norm": 16.794897079467773,
"learning_rate": 3.9199124726477026e-05,
"loss": 14.1652,
"step": 886
},
{
"epoch": 0.6271044142989705,
"grad_norm": 15.572064399719238,
"learning_rate": 3.912691466083151e-05,
"loss": 12.5518,
"step": 887
},
{
"epoch": 0.6278114091290707,
"grad_norm": 15.595902442932129,
"learning_rate": 3.9054704595186e-05,
"loss": 13.406,
"step": 888
},
{
"epoch": 0.628518403959171,
"grad_norm": 17.84473991394043,
"learning_rate": 3.898249452954048e-05,
"loss": 12.8387,
"step": 889
},
{
"epoch": 0.6292253987892713,
"grad_norm": 17.220897674560547,
"learning_rate": 3.891028446389496e-05,
"loss": 14.2772,
"step": 890
},
{
"epoch": 0.6299323936193717,
"grad_norm": 17.487550735473633,
"learning_rate": 3.883807439824945e-05,
"loss": 13.5513,
"step": 891
},
{
"epoch": 0.630639388449472,
"grad_norm": 14.752918243408203,
"learning_rate": 3.8765864332603935e-05,
"loss": 13.3264,
"step": 892
},
{
"epoch": 0.6313463832795723,
"grad_norm": 17.476144790649414,
"learning_rate": 3.8693654266958425e-05,
"loss": 13.0027,
"step": 893
},
{
"epoch": 0.6320533781096725,
"grad_norm": 19.83499526977539,
"learning_rate": 3.8621444201312914e-05,
"loss": 13.8219,
"step": 894
},
{
"epoch": 0.6327603729397728,
"grad_norm": 16.396671295166016,
"learning_rate": 3.85492341356674e-05,
"loss": 12.0832,
"step": 895
},
{
"epoch": 0.6334673677698732,
"grad_norm": 17.362865447998047,
"learning_rate": 3.847702407002188e-05,
"loss": 13.1703,
"step": 896
},
{
"epoch": 0.6341743625999735,
"grad_norm": 18.174884796142578,
"learning_rate": 3.840481400437637e-05,
"loss": 12.7365,
"step": 897
},
{
"epoch": 0.6348813574300738,
"grad_norm": 15.047250747680664,
"learning_rate": 3.833260393873085e-05,
"loss": 12.9793,
"step": 898
},
{
"epoch": 0.6355883522601741,
"grad_norm": 18.287412643432617,
"learning_rate": 3.8260393873085334e-05,
"loss": 13.649,
"step": 899
},
{
"epoch": 0.6362953470902744,
"grad_norm": 15.475159645080566,
"learning_rate": 3.818818380743983e-05,
"loss": 13.8619,
"step": 900
},
{
"epoch": 0.6370023419203747,
"grad_norm": 17.742231369018555,
"learning_rate": 3.811597374179431e-05,
"loss": 13.5366,
"step": 901
},
{
"epoch": 0.637709336750475,
"grad_norm": 15.232536315917969,
"learning_rate": 3.8043763676148796e-05,
"loss": 12.6299,
"step": 902
},
{
"epoch": 0.6384163315805753,
"grad_norm": 15.838523864746094,
"learning_rate": 3.7971553610503285e-05,
"loss": 12.1131,
"step": 903
},
{
"epoch": 0.6391233264106756,
"grad_norm": 15.444711685180664,
"learning_rate": 3.789934354485777e-05,
"loss": 13.2765,
"step": 904
},
{
"epoch": 0.6398303212407759,
"grad_norm": 15.970074653625488,
"learning_rate": 3.782713347921225e-05,
"loss": 12.0993,
"step": 905
},
{
"epoch": 0.6405373160708763,
"grad_norm": 14.429159164428711,
"learning_rate": 3.775492341356674e-05,
"loss": 12.4137,
"step": 906
},
{
"epoch": 0.6412443109009766,
"grad_norm": 17.08701515197754,
"learning_rate": 3.768271334792122e-05,
"loss": 12.2321,
"step": 907
},
{
"epoch": 0.6419513057310768,
"grad_norm": 17.066822052001953,
"learning_rate": 3.761050328227571e-05,
"loss": 13.6998,
"step": 908
},
{
"epoch": 0.6426583005611771,
"grad_norm": 15.362284660339355,
"learning_rate": 3.75382932166302e-05,
"loss": 11.0648,
"step": 909
},
{
"epoch": 0.6433652953912774,
"grad_norm": 17.06339454650879,
"learning_rate": 3.7466083150984684e-05,
"loss": 13.2253,
"step": 910
},
{
"epoch": 0.6440722902213778,
"grad_norm": 14.42895221710205,
"learning_rate": 3.7393873085339166e-05,
"loss": 12.1794,
"step": 911
},
{
"epoch": 0.6447792850514781,
"grad_norm": 15.388317108154297,
"learning_rate": 3.7321663019693656e-05,
"loss": 12.4478,
"step": 912
},
{
"epoch": 0.6454862798815784,
"grad_norm": 14.884230613708496,
"learning_rate": 3.724945295404814e-05,
"loss": 12.9629,
"step": 913
},
{
"epoch": 0.6461932747116786,
"grad_norm": 14.791316986083984,
"learning_rate": 3.717724288840262e-05,
"loss": 13.0299,
"step": 914
},
{
"epoch": 0.6469002695417789,
"grad_norm": 17.605379104614258,
"learning_rate": 3.710503282275712e-05,
"loss": 13.2825,
"step": 915
},
{
"epoch": 0.6476072643718793,
"grad_norm": 17.050025939941406,
"learning_rate": 3.70328227571116e-05,
"loss": 13.9545,
"step": 916
},
{
"epoch": 0.6483142592019796,
"grad_norm": 13.724760055541992,
"learning_rate": 3.696061269146608e-05,
"loss": 12.4339,
"step": 917
},
{
"epoch": 0.6490212540320799,
"grad_norm": 15.219618797302246,
"learning_rate": 3.688840262582057e-05,
"loss": 12.0493,
"step": 918
},
{
"epoch": 0.6497282488621802,
"grad_norm": 15.740148544311523,
"learning_rate": 3.6816192560175055e-05,
"loss": 12.2087,
"step": 919
},
{
"epoch": 0.6504352436922805,
"grad_norm": 15.154293060302734,
"learning_rate": 3.674398249452954e-05,
"loss": 11.8399,
"step": 920
},
{
"epoch": 0.6511422385223808,
"grad_norm": 21.17293930053711,
"learning_rate": 3.667177242888403e-05,
"loss": 13.5817,
"step": 921
},
{
"epoch": 0.6518492333524811,
"grad_norm": 17.151151657104492,
"learning_rate": 3.659956236323851e-05,
"loss": 14.2616,
"step": 922
},
{
"epoch": 0.6525562281825814,
"grad_norm": 15.818124771118164,
"learning_rate": 3.6527352297593e-05,
"loss": 12.9145,
"step": 923
},
{
"epoch": 0.6532632230126817,
"grad_norm": 17.215805053710938,
"learning_rate": 3.645514223194749e-05,
"loss": 12.8664,
"step": 924
},
{
"epoch": 0.653970217842782,
"grad_norm": 15.398895263671875,
"learning_rate": 3.638293216630197e-05,
"loss": 13.5099,
"step": 925
},
{
"epoch": 0.6546772126728824,
"grad_norm": 16.104904174804688,
"learning_rate": 3.6310722100656454e-05,
"loss": 13.3627,
"step": 926
},
{
"epoch": 0.6553842075029827,
"grad_norm": 16.67665672302246,
"learning_rate": 3.623851203501094e-05,
"loss": 13.7105,
"step": 927
},
{
"epoch": 0.6560912023330829,
"grad_norm": 14.506566047668457,
"learning_rate": 3.6166301969365426e-05,
"loss": 12.5536,
"step": 928
},
{
"epoch": 0.6567981971631832,
"grad_norm": 15.468367576599121,
"learning_rate": 3.609409190371991e-05,
"loss": 13.3451,
"step": 929
},
{
"epoch": 0.6575051919932835,
"grad_norm": 17.965709686279297,
"learning_rate": 3.60218818380744e-05,
"loss": 14.0162,
"step": 930
},
{
"epoch": 0.6582121868233839,
"grad_norm": 14.338950157165527,
"learning_rate": 3.594967177242889e-05,
"loss": 12.1867,
"step": 931
},
{
"epoch": 0.6589191816534842,
"grad_norm": 14.590508460998535,
"learning_rate": 3.587746170678337e-05,
"loss": 13.2619,
"step": 932
},
{
"epoch": 0.6596261764835845,
"grad_norm": 15.523008346557617,
"learning_rate": 3.580525164113785e-05,
"loss": 13.6548,
"step": 933
},
{
"epoch": 0.6603331713136847,
"grad_norm": 17.361780166625977,
"learning_rate": 3.573304157549234e-05,
"loss": 11.8496,
"step": 934
},
{
"epoch": 0.661040166143785,
"grad_norm": 14.874725341796875,
"learning_rate": 3.5660831509846824e-05,
"loss": 12.9375,
"step": 935
},
{
"epoch": 0.6617471609738854,
"grad_norm": 15.033089637756348,
"learning_rate": 3.5588621444201314e-05,
"loss": 11.8268,
"step": 936
},
{
"epoch": 0.6624541558039857,
"grad_norm": 16.188312530517578,
"learning_rate": 3.5516411378555796e-05,
"loss": 13.1623,
"step": 937
},
{
"epoch": 0.663161150634086,
"grad_norm": 14.005878448486328,
"learning_rate": 3.5444201312910286e-05,
"loss": 13.1447,
"step": 938
},
{
"epoch": 0.6638681454641863,
"grad_norm": 16.06510353088379,
"learning_rate": 3.537199124726477e-05,
"loss": 11.294,
"step": 939
},
{
"epoch": 0.6645751402942865,
"grad_norm": 15.923260688781738,
"learning_rate": 3.529978118161926e-05,
"loss": 13.9193,
"step": 940
},
{
"epoch": 0.6652821351243869,
"grad_norm": 16.1591796875,
"learning_rate": 3.522757111597374e-05,
"loss": 12.5801,
"step": 941
},
{
"epoch": 0.6659891299544872,
"grad_norm": 18.498062133789062,
"learning_rate": 3.515536105032822e-05,
"loss": 13.0376,
"step": 942
},
{
"epoch": 0.6666961247845875,
"grad_norm": 16.772415161132812,
"learning_rate": 3.508315098468271e-05,
"loss": 13.0875,
"step": 943
},
{
"epoch": 0.6674031196146878,
"grad_norm": 15.799750328063965,
"learning_rate": 3.5010940919037195e-05,
"loss": 12.7625,
"step": 944
},
{
"epoch": 0.6681101144447881,
"grad_norm": 14.897823333740234,
"learning_rate": 3.4938730853391685e-05,
"loss": 14.2135,
"step": 945
},
{
"epoch": 0.6688171092748885,
"grad_norm": 13.589593887329102,
"learning_rate": 3.4866520787746174e-05,
"loss": 13.0602,
"step": 946
},
{
"epoch": 0.6695241041049887,
"grad_norm": 15.410765647888184,
"learning_rate": 3.479431072210066e-05,
"loss": 12.3363,
"step": 947
},
{
"epoch": 0.670231098935089,
"grad_norm": 13.951040267944336,
"learning_rate": 3.472210065645514e-05,
"loss": 13.05,
"step": 948
},
{
"epoch": 0.6709380937651893,
"grad_norm": 18.58137321472168,
"learning_rate": 3.464989059080963e-05,
"loss": 12.2522,
"step": 949
},
{
"epoch": 0.6716450885952896,
"grad_norm": 17.50958251953125,
"learning_rate": 3.457768052516411e-05,
"loss": 12.5067,
"step": 950
},
{
"epoch": 0.67235208342539,
"grad_norm": 17.609472274780273,
"learning_rate": 3.4505470459518594e-05,
"loss": 12.2965,
"step": 951
},
{
"epoch": 0.6730590782554903,
"grad_norm": 14.07121753692627,
"learning_rate": 3.443326039387309e-05,
"loss": 11.6223,
"step": 952
},
{
"epoch": 0.6737660730855906,
"grad_norm": 17.91677474975586,
"learning_rate": 3.436105032822757e-05,
"loss": 13.2656,
"step": 953
},
{
"epoch": 0.6744730679156908,
"grad_norm": 17.619646072387695,
"learning_rate": 3.4288840262582056e-05,
"loss": 12.0533,
"step": 954
},
{
"epoch": 0.6751800627457911,
"grad_norm": 15.973097801208496,
"learning_rate": 3.4216630196936545e-05,
"loss": 12.6224,
"step": 955
},
{
"epoch": 0.6758870575758915,
"grad_norm": 16.727569580078125,
"learning_rate": 3.414442013129103e-05,
"loss": 10.9571,
"step": 956
},
{
"epoch": 0.6765940524059918,
"grad_norm": 15.502710342407227,
"learning_rate": 3.407221006564551e-05,
"loss": 11.7565,
"step": 957
},
{
"epoch": 0.6773010472360921,
"grad_norm": 15.650954246520996,
"learning_rate": 3.4e-05,
"loss": 13.8277,
"step": 958
},
{
"epoch": 0.6780080420661924,
"grad_norm": 15.571489334106445,
"learning_rate": 3.392778993435448e-05,
"loss": 13.1988,
"step": 959
},
{
"epoch": 0.6787150368962926,
"grad_norm": 15.313315391540527,
"learning_rate": 3.385557986870897e-05,
"loss": 13.5607,
"step": 960
},
{
"epoch": 0.679422031726393,
"grad_norm": 15.986994743347168,
"learning_rate": 3.378336980306346e-05,
"loss": 13.3168,
"step": 961
},
{
"epoch": 0.6801290265564933,
"grad_norm": 16.067049026489258,
"learning_rate": 3.3711159737417944e-05,
"loss": 12.8027,
"step": 962
},
{
"epoch": 0.6808360213865936,
"grad_norm": 19.818334579467773,
"learning_rate": 3.3638949671772426e-05,
"loss": 11.4662,
"step": 963
},
{
"epoch": 0.6815430162166939,
"grad_norm": 17.01699447631836,
"learning_rate": 3.3566739606126916e-05,
"loss": 13.2136,
"step": 964
},
{
"epoch": 0.6822500110467942,
"grad_norm": 15.616992950439453,
"learning_rate": 3.34945295404814e-05,
"loss": 12.6188,
"step": 965
},
{
"epoch": 0.6829570058768946,
"grad_norm": 13.695746421813965,
"learning_rate": 3.342231947483588e-05,
"loss": 11.3794,
"step": 966
},
{
"epoch": 0.6836640007069948,
"grad_norm": 15.043867111206055,
"learning_rate": 3.335010940919038e-05,
"loss": 11.3308,
"step": 967
},
{
"epoch": 0.6843709955370951,
"grad_norm": 14.187246322631836,
"learning_rate": 3.327789934354486e-05,
"loss": 13.3129,
"step": 968
},
{
"epoch": 0.6850779903671954,
"grad_norm": 16.36101722717285,
"learning_rate": 3.320568927789934e-05,
"loss": 14.5099,
"step": 969
},
{
"epoch": 0.6857849851972957,
"grad_norm": 16.514877319335938,
"learning_rate": 3.313347921225383e-05,
"loss": 11.603,
"step": 970
},
{
"epoch": 0.6864919800273961,
"grad_norm": 17.427112579345703,
"learning_rate": 3.3061269146608315e-05,
"loss": 12.8277,
"step": 971
},
{
"epoch": 0.6871989748574964,
"grad_norm": 16.44808006286621,
"learning_rate": 3.29890590809628e-05,
"loss": 12.8675,
"step": 972
},
{
"epoch": 0.6879059696875967,
"grad_norm": 16.135822296142578,
"learning_rate": 3.291684901531729e-05,
"loss": 11.6851,
"step": 973
},
{
"epoch": 0.6886129645176969,
"grad_norm": 16.02741241455078,
"learning_rate": 3.284463894967177e-05,
"loss": 13.8546,
"step": 974
},
{
"epoch": 0.6893199593477972,
"grad_norm": 17.733104705810547,
"learning_rate": 3.277242888402626e-05,
"loss": 13.0497,
"step": 975
},
{
"epoch": 0.6900269541778976,
"grad_norm": 16.3909969329834,
"learning_rate": 3.270021881838075e-05,
"loss": 11.9032,
"step": 976
},
{
"epoch": 0.6907339490079979,
"grad_norm": 15.314737319946289,
"learning_rate": 3.262800875273523e-05,
"loss": 13.2063,
"step": 977
},
{
"epoch": 0.6914409438380982,
"grad_norm": 16.040050506591797,
"learning_rate": 3.2555798687089714e-05,
"loss": 12.6159,
"step": 978
},
{
"epoch": 0.6921479386681985,
"grad_norm": 15.172967910766602,
"learning_rate": 3.24835886214442e-05,
"loss": 12.5432,
"step": 979
},
{
"epoch": 0.6928549334982987,
"grad_norm": 14.832226753234863,
"learning_rate": 3.2411378555798686e-05,
"loss": 12.2323,
"step": 980
},
{
"epoch": 0.6935619283283991,
"grad_norm": 16.011014938354492,
"learning_rate": 3.233916849015317e-05,
"loss": 12.7174,
"step": 981
},
{
"epoch": 0.6942689231584994,
"grad_norm": 15.02815055847168,
"learning_rate": 3.226695842450766e-05,
"loss": 13.0848,
"step": 982
},
{
"epoch": 0.6949759179885997,
"grad_norm": 16.430984497070312,
"learning_rate": 3.219474835886215e-05,
"loss": 12.228,
"step": 983
},
{
"epoch": 0.6956829128187,
"grad_norm": 15.233168601989746,
"learning_rate": 3.212253829321663e-05,
"loss": 14.0058,
"step": 984
},
{
"epoch": 0.6963899076488003,
"grad_norm": 15.990872383117676,
"learning_rate": 3.205032822757111e-05,
"loss": 13.9359,
"step": 985
},
{
"epoch": 0.6970969024789007,
"grad_norm": 14.508731842041016,
"learning_rate": 3.19781181619256e-05,
"loss": 13.4266,
"step": 986
},
{
"epoch": 0.6978038973090009,
"grad_norm": 15.588955879211426,
"learning_rate": 3.1905908096280084e-05,
"loss": 12.4986,
"step": 987
},
{
"epoch": 0.6985108921391012,
"grad_norm": 16.39841651916504,
"learning_rate": 3.1833698030634574e-05,
"loss": 12.6671,
"step": 988
},
{
"epoch": 0.6992178869692015,
"grad_norm": 15.241385459899902,
"learning_rate": 3.176148796498906e-05,
"loss": 14.0819,
"step": 989
},
{
"epoch": 0.6999248817993018,
"grad_norm": 17.080127716064453,
"learning_rate": 3.1689277899343546e-05,
"loss": 12.5114,
"step": 990
},
{
"epoch": 0.7006318766294022,
"grad_norm": 16.860857009887695,
"learning_rate": 3.161706783369803e-05,
"loss": 13.9591,
"step": 991
},
{
"epoch": 0.7013388714595025,
"grad_norm": 16.064014434814453,
"learning_rate": 3.154485776805252e-05,
"loss": 13.746,
"step": 992
},
{
"epoch": 0.7020458662896027,
"grad_norm": 16.772647857666016,
"learning_rate": 3.1472647702407e-05,
"loss": 12.8039,
"step": 993
},
{
"epoch": 0.702752861119703,
"grad_norm": 14.531611442565918,
"learning_rate": 3.140043763676149e-05,
"loss": 11.8692,
"step": 994
},
{
"epoch": 0.7034598559498033,
"grad_norm": 17.939926147460938,
"learning_rate": 3.132822757111597e-05,
"loss": 13.6893,
"step": 995
},
{
"epoch": 0.7041668507799037,
"grad_norm": 16.083425521850586,
"learning_rate": 3.1256017505470455e-05,
"loss": 12.779,
"step": 996
},
{
"epoch": 0.704873845610004,
"grad_norm": 15.42416000366211,
"learning_rate": 3.118380743982495e-05,
"loss": 12.7348,
"step": 997
},
{
"epoch": 0.7055808404401043,
"grad_norm": 14.658026695251465,
"learning_rate": 3.1111597374179434e-05,
"loss": 11.8572,
"step": 998
},
{
"epoch": 0.7062878352702046,
"grad_norm": 16.61526107788086,
"learning_rate": 3.103938730853392e-05,
"loss": 14.1659,
"step": 999
},
{
"epoch": 0.7069948301003048,
"grad_norm": 16.476240158081055,
"learning_rate": 3.0967177242888406e-05,
"loss": 12.8316,
"step": 1000
},
{
"epoch": 0.7077018249304052,
"grad_norm": 15.872008323669434,
"learning_rate": 3.089496717724289e-05,
"loss": 11.1954,
"step": 1001
},
{
"epoch": 0.7084088197605055,
"grad_norm": 15.403485298156738,
"learning_rate": 3.082275711159737e-05,
"loss": 12.8882,
"step": 1002
},
{
"epoch": 0.7091158145906058,
"grad_norm": 16.67386817932129,
"learning_rate": 3.075054704595186e-05,
"loss": 12.7755,
"step": 1003
},
{
"epoch": 0.7098228094207061,
"grad_norm": 15.99714183807373,
"learning_rate": 3.067833698030635e-05,
"loss": 12.6526,
"step": 1004
},
{
"epoch": 0.7105298042508064,
"grad_norm": 13.916271209716797,
"learning_rate": 3.060612691466083e-05,
"loss": 13.3718,
"step": 1005
},
{
"epoch": 0.7112367990809068,
"grad_norm": 15.985695838928223,
"learning_rate": 3.053391684901532e-05,
"loss": 12.9525,
"step": 1006
},
{
"epoch": 0.711943793911007,
"grad_norm": 17.183103561401367,
"learning_rate": 3.0461706783369805e-05,
"loss": 12.8737,
"step": 1007
},
{
"epoch": 0.7126507887411073,
"grad_norm": 14.209635734558105,
"learning_rate": 3.038949671772429e-05,
"loss": 11.7585,
"step": 1008
},
{
"epoch": 0.7133577835712076,
"grad_norm": 15.675783157348633,
"learning_rate": 3.0317286652078777e-05,
"loss": 13.5172,
"step": 1009
},
{
"epoch": 0.7140647784013079,
"grad_norm": 15.367440223693848,
"learning_rate": 3.024507658643326e-05,
"loss": 11.7943,
"step": 1010
},
{
"epoch": 0.7147717732314083,
"grad_norm": 18.714014053344727,
"learning_rate": 3.0172866520787746e-05,
"loss": 12.9202,
"step": 1011
},
{
"epoch": 0.7154787680615086,
"grad_norm": 15.120142936706543,
"learning_rate": 3.0100656455142235e-05,
"loss": 13.5838,
"step": 1012
},
{
"epoch": 0.7161857628916088,
"grad_norm": 19.008926391601562,
"learning_rate": 3.002844638949672e-05,
"loss": 14.1862,
"step": 1013
},
{
"epoch": 0.7168927577217091,
"grad_norm": 14.798474311828613,
"learning_rate": 2.9956236323851207e-05,
"loss": 12.3924,
"step": 1014
},
{
"epoch": 0.7175997525518094,
"grad_norm": 16.070415496826172,
"learning_rate": 2.9884026258205693e-05,
"loss": 13.5143,
"step": 1015
},
{
"epoch": 0.7183067473819098,
"grad_norm": 17.342212677001953,
"learning_rate": 2.9811816192560176e-05,
"loss": 12.2586,
"step": 1016
},
{
"epoch": 0.7190137422120101,
"grad_norm": 16.9810791015625,
"learning_rate": 2.9739606126914662e-05,
"loss": 13.5872,
"step": 1017
},
{
"epoch": 0.7197207370421104,
"grad_norm": 15.211740493774414,
"learning_rate": 2.9667396061269148e-05,
"loss": 12.3832,
"step": 1018
},
{
"epoch": 0.7204277318722107,
"grad_norm": 14.690281867980957,
"learning_rate": 2.9595185995623637e-05,
"loss": 12.2162,
"step": 1019
},
{
"epoch": 0.7211347267023109,
"grad_norm": 14.952332496643066,
"learning_rate": 2.9522975929978123e-05,
"loss": 13.2556,
"step": 1020
},
{
"epoch": 0.7218417215324113,
"grad_norm": 14.745388984680176,
"learning_rate": 2.9450765864332606e-05,
"loss": 11.7899,
"step": 1021
},
{
"epoch": 0.7225487163625116,
"grad_norm": 16.4384822845459,
"learning_rate": 2.9378555798687092e-05,
"loss": 12.6083,
"step": 1022
},
{
"epoch": 0.7232557111926119,
"grad_norm": 15.447677612304688,
"learning_rate": 2.9306345733041578e-05,
"loss": 12.1423,
"step": 1023
},
{
"epoch": 0.7239627060227122,
"grad_norm": 15.585071563720703,
"learning_rate": 2.9234135667396064e-05,
"loss": 12.0851,
"step": 1024
},
{
"epoch": 0.7246697008528125,
"grad_norm": 14.9990873336792,
"learning_rate": 2.9161925601750547e-05,
"loss": 12.8785,
"step": 1025
},
{
"epoch": 0.7253766956829129,
"grad_norm": 19.603099822998047,
"learning_rate": 2.908971553610504e-05,
"loss": 12.1019,
"step": 1026
},
{
"epoch": 0.7260836905130131,
"grad_norm": 13.465499877929688,
"learning_rate": 2.9017505470459522e-05,
"loss": 13.0492,
"step": 1027
},
{
"epoch": 0.7267906853431134,
"grad_norm": 14.473004341125488,
"learning_rate": 2.894529540481401e-05,
"loss": 14.0964,
"step": 1028
},
{
"epoch": 0.7274976801732137,
"grad_norm": 14.698629379272461,
"learning_rate": 2.8873085339168494e-05,
"loss": 11.0322,
"step": 1029
},
{
"epoch": 0.728204675003314,
"grad_norm": 16.44681739807129,
"learning_rate": 2.8800875273522977e-05,
"loss": 13.5584,
"step": 1030
},
{
"epoch": 0.7289116698334144,
"grad_norm": 13.492003440856934,
"learning_rate": 2.8728665207877463e-05,
"loss": 13.2963,
"step": 1031
},
{
"epoch": 0.7296186646635147,
"grad_norm": 15.496600151062012,
"learning_rate": 2.865645514223195e-05,
"loss": 12.6854,
"step": 1032
},
{
"epoch": 0.7303256594936149,
"grad_norm": 19.080289840698242,
"learning_rate": 2.8584245076586435e-05,
"loss": 13.4954,
"step": 1033
},
{
"epoch": 0.7310326543237152,
"grad_norm": 14.68902587890625,
"learning_rate": 2.8512035010940925e-05,
"loss": 12.0387,
"step": 1034
},
{
"epoch": 0.7317396491538155,
"grad_norm": 14.314245223999023,
"learning_rate": 2.843982494529541e-05,
"loss": 13.6249,
"step": 1035
},
{
"epoch": 0.7324466439839159,
"grad_norm": 16.663728713989258,
"learning_rate": 2.8367614879649893e-05,
"loss": 13.3793,
"step": 1036
},
{
"epoch": 0.7331536388140162,
"grad_norm": 14.021472930908203,
"learning_rate": 2.829540481400438e-05,
"loss": 12.4353,
"step": 1037
},
{
"epoch": 0.7338606336441165,
"grad_norm": 14.659539222717285,
"learning_rate": 2.8223194748358865e-05,
"loss": 11.4481,
"step": 1038
},
{
"epoch": 0.7345676284742167,
"grad_norm": 14.972705841064453,
"learning_rate": 2.8150984682713348e-05,
"loss": 12.917,
"step": 1039
},
{
"epoch": 0.735274623304317,
"grad_norm": 14.24467945098877,
"learning_rate": 2.8078774617067834e-05,
"loss": 13.0419,
"step": 1040
},
{
"epoch": 0.7359816181344174,
"grad_norm": 17.544336318969727,
"learning_rate": 2.8006564551422323e-05,
"loss": 12.9449,
"step": 1041
},
{
"epoch": 0.7366886129645177,
"grad_norm": 15.078471183776855,
"learning_rate": 2.793435448577681e-05,
"loss": 13.5799,
"step": 1042
},
{
"epoch": 0.737395607794618,
"grad_norm": 15.030024528503418,
"learning_rate": 2.7862144420131295e-05,
"loss": 11.3935,
"step": 1043
},
{
"epoch": 0.7381026026247183,
"grad_norm": 17.29486846923828,
"learning_rate": 2.778993435448578e-05,
"loss": 12.6275,
"step": 1044
},
{
"epoch": 0.7388095974548186,
"grad_norm": 15.28783893585205,
"learning_rate": 2.7717724288840264e-05,
"loss": 11.2874,
"step": 1045
},
{
"epoch": 0.739516592284919,
"grad_norm": 20.103883743286133,
"learning_rate": 2.764551422319475e-05,
"loss": 11.5156,
"step": 1046
},
{
"epoch": 0.7402235871150192,
"grad_norm": 15.663703918457031,
"learning_rate": 2.7573304157549236e-05,
"loss": 13.4875,
"step": 1047
},
{
"epoch": 0.7409305819451195,
"grad_norm": 16.604524612426758,
"learning_rate": 2.750109409190372e-05,
"loss": 12.9352,
"step": 1048
},
{
"epoch": 0.7416375767752198,
"grad_norm": 13.528607368469238,
"learning_rate": 2.742888402625821e-05,
"loss": 12.3269,
"step": 1049
},
{
"epoch": 0.7423445716053202,
"grad_norm": 15.969796180725098,
"learning_rate": 2.7356673960612694e-05,
"loss": 13.7365,
"step": 1050
},
{
"epoch": 0.7430515664354205,
"grad_norm": 15.28542709350586,
"learning_rate": 2.728446389496718e-05,
"loss": 12.0605,
"step": 1051
},
{
"epoch": 0.7437585612655208,
"grad_norm": 14.971050262451172,
"learning_rate": 2.7212253829321666e-05,
"loss": 12.715,
"step": 1052
},
{
"epoch": 0.744465556095621,
"grad_norm": 14.354373931884766,
"learning_rate": 2.714004376367615e-05,
"loss": 12.9591,
"step": 1053
},
{
"epoch": 0.7451725509257213,
"grad_norm": 16.311098098754883,
"learning_rate": 2.7067833698030635e-05,
"loss": 11.784,
"step": 1054
},
{
"epoch": 0.7458795457558217,
"grad_norm": 15.502252578735352,
"learning_rate": 2.699562363238512e-05,
"loss": 11.6811,
"step": 1055
},
{
"epoch": 0.746586540585922,
"grad_norm": 16.668073654174805,
"learning_rate": 2.692341356673961e-05,
"loss": 12.4734,
"step": 1056
},
{
"epoch": 0.7472935354160223,
"grad_norm": 21.601903915405273,
"learning_rate": 2.6851203501094096e-05,
"loss": 13.3575,
"step": 1057
},
{
"epoch": 0.7480005302461226,
"grad_norm": 15.497323989868164,
"learning_rate": 2.6778993435448582e-05,
"loss": 13.4247,
"step": 1058
},
{
"epoch": 0.7487075250762228,
"grad_norm": 17.963159561157227,
"learning_rate": 2.6706783369803065e-05,
"loss": 11.312,
"step": 1059
},
{
"epoch": 0.7494145199063232,
"grad_norm": 17.467559814453125,
"learning_rate": 2.663457330415755e-05,
"loss": 12.6942,
"step": 1060
},
{
"epoch": 0.7501215147364235,
"grad_norm": 15.068920135498047,
"learning_rate": 2.6562363238512037e-05,
"loss": 13.5305,
"step": 1061
},
{
"epoch": 0.7508285095665238,
"grad_norm": 15.089822769165039,
"learning_rate": 2.649015317286652e-05,
"loss": 13.7907,
"step": 1062
},
{
"epoch": 0.7515355043966241,
"grad_norm": 14.680115699768066,
"learning_rate": 2.6417943107221013e-05,
"loss": 10.9175,
"step": 1063
},
{
"epoch": 0.7522424992267244,
"grad_norm": 16.606077194213867,
"learning_rate": 2.6345733041575495e-05,
"loss": 13.6108,
"step": 1064
},
{
"epoch": 0.7529494940568248,
"grad_norm": 16.429758071899414,
"learning_rate": 2.627352297592998e-05,
"loss": 13.0759,
"step": 1065
},
{
"epoch": 0.753656488886925,
"grad_norm": 14.46728515625,
"learning_rate": 2.6201312910284467e-05,
"loss": 11.7365,
"step": 1066
},
{
"epoch": 0.7543634837170253,
"grad_norm": 15.043597221374512,
"learning_rate": 2.6129102844638953e-05,
"loss": 13.0305,
"step": 1067
},
{
"epoch": 0.7550704785471256,
"grad_norm": 16.372493743896484,
"learning_rate": 2.6056892778993436e-05,
"loss": 12.9214,
"step": 1068
},
{
"epoch": 0.7557774733772259,
"grad_norm": 13.306395530700684,
"learning_rate": 2.5984682713347922e-05,
"loss": 11.733,
"step": 1069
},
{
"epoch": 0.7564844682073263,
"grad_norm": 17.534255981445312,
"learning_rate": 2.5912472647702408e-05,
"loss": 12.869,
"step": 1070
},
{
"epoch": 0.7571914630374266,
"grad_norm": 16.98424530029297,
"learning_rate": 2.5840262582056897e-05,
"loss": 12.9999,
"step": 1071
},
{
"epoch": 0.7578984578675269,
"grad_norm": 13.720797538757324,
"learning_rate": 2.5768052516411384e-05,
"loss": 10.7562,
"step": 1072
},
{
"epoch": 0.7586054526976271,
"grad_norm": 15.344558715820312,
"learning_rate": 2.5695842450765866e-05,
"loss": 12.3965,
"step": 1073
},
{
"epoch": 0.7593124475277274,
"grad_norm": 14.723435401916504,
"learning_rate": 2.5623632385120352e-05,
"loss": 12.8219,
"step": 1074
},
{
"epoch": 0.7600194423578278,
"grad_norm": 15.032986640930176,
"learning_rate": 2.5551422319474838e-05,
"loss": 12.2727,
"step": 1075
},
{
"epoch": 0.7607264371879281,
"grad_norm": 15.117292404174805,
"learning_rate": 2.5479212253829324e-05,
"loss": 13.233,
"step": 1076
},
{
"epoch": 0.7614334320180284,
"grad_norm": 15.652196884155273,
"learning_rate": 2.540700218818381e-05,
"loss": 12.6491,
"step": 1077
},
{
"epoch": 0.7621404268481287,
"grad_norm": 15.632040023803711,
"learning_rate": 2.5334792122538296e-05,
"loss": 12.8196,
"step": 1078
},
{
"epoch": 0.7628474216782289,
"grad_norm": 16.843076705932617,
"learning_rate": 2.526258205689278e-05,
"loss": 12.47,
"step": 1079
},
{
"epoch": 0.7635544165083293,
"grad_norm": 14.941967964172363,
"learning_rate": 2.519037199124727e-05,
"loss": 11.7957,
"step": 1080
},
{
"epoch": 0.7642614113384296,
"grad_norm": 15.734443664550781,
"learning_rate": 2.5118161925601754e-05,
"loss": 12.8278,
"step": 1081
},
{
"epoch": 0.7649684061685299,
"grad_norm": 13.94361400604248,
"learning_rate": 2.5045951859956237e-05,
"loss": 12.3736,
"step": 1082
},
{
"epoch": 0.7656754009986302,
"grad_norm": 18.560409545898438,
"learning_rate": 2.4973741794310723e-05,
"loss": 12.0621,
"step": 1083
},
{
"epoch": 0.7663823958287305,
"grad_norm": 16.558231353759766,
"learning_rate": 2.4901531728665213e-05,
"loss": 13.4814,
"step": 1084
},
{
"epoch": 0.7670893906588309,
"grad_norm": 19.616657257080078,
"learning_rate": 2.4829321663019695e-05,
"loss": 13.0437,
"step": 1085
},
{
"epoch": 0.7677963854889311,
"grad_norm": 14.448963165283203,
"learning_rate": 2.475711159737418e-05,
"loss": 12.7444,
"step": 1086
},
{
"epoch": 0.7685033803190314,
"grad_norm": 14.808650016784668,
"learning_rate": 2.4684901531728667e-05,
"loss": 13.2923,
"step": 1087
},
{
"epoch": 0.7692103751491317,
"grad_norm": 16.270593643188477,
"learning_rate": 2.4612691466083153e-05,
"loss": 12.4998,
"step": 1088
},
{
"epoch": 0.769917369979232,
"grad_norm": 16.274654388427734,
"learning_rate": 2.454048140043764e-05,
"loss": 11.7857,
"step": 1089
},
{
"epoch": 0.7706243648093324,
"grad_norm": 20.461088180541992,
"learning_rate": 2.4468271334792125e-05,
"loss": 13.5802,
"step": 1090
},
{
"epoch": 0.7713313596394327,
"grad_norm": 14.93079662322998,
"learning_rate": 2.439606126914661e-05,
"loss": 12.1022,
"step": 1091
},
{
"epoch": 0.772038354469533,
"grad_norm": 13.616817474365234,
"learning_rate": 2.4323851203501097e-05,
"loss": 11.7097,
"step": 1092
},
{
"epoch": 0.7727453492996332,
"grad_norm": 16.070207595825195,
"learning_rate": 2.4251641137855583e-05,
"loss": 13.3532,
"step": 1093
},
{
"epoch": 0.7734523441297335,
"grad_norm": 15.457805633544922,
"learning_rate": 2.4179431072210066e-05,
"loss": 13.265,
"step": 1094
},
{
"epoch": 0.7741593389598339,
"grad_norm": 15.124858856201172,
"learning_rate": 2.4107221006564555e-05,
"loss": 13.2484,
"step": 1095
},
{
"epoch": 0.7748663337899342,
"grad_norm": 15.762398719787598,
"learning_rate": 2.403501094091904e-05,
"loss": 12.2059,
"step": 1096
},
{
"epoch": 0.7755733286200345,
"grad_norm": 15.04523754119873,
"learning_rate": 2.3962800875273524e-05,
"loss": 13.4343,
"step": 1097
},
{
"epoch": 0.7762803234501348,
"grad_norm": 14.77971363067627,
"learning_rate": 2.389059080962801e-05,
"loss": 11.309,
"step": 1098
},
{
"epoch": 0.776987318280235,
"grad_norm": 17.487321853637695,
"learning_rate": 2.3818380743982496e-05,
"loss": 12.3208,
"step": 1099
},
{
"epoch": 0.7776943131103354,
"grad_norm": 16.295358657836914,
"learning_rate": 2.3746170678336982e-05,
"loss": 10.7923,
"step": 1100
},
{
"epoch": 0.7784013079404357,
"grad_norm": 14.476552963256836,
"learning_rate": 2.3673960612691468e-05,
"loss": 13.6797,
"step": 1101
},
{
"epoch": 0.779108302770536,
"grad_norm": 16.841829299926758,
"learning_rate": 2.3601750547045954e-05,
"loss": 12.7085,
"step": 1102
},
{
"epoch": 0.7798152976006363,
"grad_norm": 15.268396377563477,
"learning_rate": 2.352954048140044e-05,
"loss": 13.7208,
"step": 1103
},
{
"epoch": 0.7805222924307366,
"grad_norm": 19.5039119720459,
"learning_rate": 2.3457330415754926e-05,
"loss": 12.752,
"step": 1104
},
{
"epoch": 0.781229287260837,
"grad_norm": 17.3958740234375,
"learning_rate": 2.338512035010941e-05,
"loss": 13.2237,
"step": 1105
},
{
"epoch": 0.7819362820909372,
"grad_norm": 14.796465873718262,
"learning_rate": 2.33129102844639e-05,
"loss": 12.6552,
"step": 1106
},
{
"epoch": 0.7826432769210375,
"grad_norm": 15.115665435791016,
"learning_rate": 2.3240700218818384e-05,
"loss": 13.2209,
"step": 1107
},
{
"epoch": 0.7833502717511378,
"grad_norm": 16.425796508789062,
"learning_rate": 2.3168490153172867e-05,
"loss": 13.3238,
"step": 1108
},
{
"epoch": 0.7840572665812381,
"grad_norm": 16.225072860717773,
"learning_rate": 2.3096280087527353e-05,
"loss": 11.3711,
"step": 1109
},
{
"epoch": 0.7847642614113385,
"grad_norm": 14.445027351379395,
"learning_rate": 2.3024070021881843e-05,
"loss": 13.2492,
"step": 1110
},
{
"epoch": 0.7854712562414388,
"grad_norm": 18.562454223632812,
"learning_rate": 2.2951859956236325e-05,
"loss": 13.406,
"step": 1111
},
{
"epoch": 0.786178251071539,
"grad_norm": 15.524725914001465,
"learning_rate": 2.287964989059081e-05,
"loss": 14.6405,
"step": 1112
},
{
"epoch": 0.7868852459016393,
"grad_norm": 16.006057739257812,
"learning_rate": 2.2807439824945297e-05,
"loss": 12.8032,
"step": 1113
},
{
"epoch": 0.7875922407317396,
"grad_norm": 14.516286849975586,
"learning_rate": 2.2735229759299783e-05,
"loss": 12.244,
"step": 1114
},
{
"epoch": 0.78829923556184,
"grad_norm": 14.554705619812012,
"learning_rate": 2.266301969365427e-05,
"loss": 11.4593,
"step": 1115
},
{
"epoch": 0.7890062303919403,
"grad_norm": 16.04813003540039,
"learning_rate": 2.2590809628008755e-05,
"loss": 12.6374,
"step": 1116
},
{
"epoch": 0.7897132252220406,
"grad_norm": 16.26800537109375,
"learning_rate": 2.251859956236324e-05,
"loss": 12.4228,
"step": 1117
},
{
"epoch": 0.7904202200521409,
"grad_norm": 14.735860824584961,
"learning_rate": 2.2446389496717727e-05,
"loss": 12.6189,
"step": 1118
},
{
"epoch": 0.7911272148822411,
"grad_norm": 16.355247497558594,
"learning_rate": 2.2374179431072213e-05,
"loss": 12.1257,
"step": 1119
},
{
"epoch": 0.7918342097123415,
"grad_norm": 16.468690872192383,
"learning_rate": 2.2301969365426696e-05,
"loss": 11.1213,
"step": 1120
},
{
"epoch": 0.7925412045424418,
"grad_norm": 17.71234703063965,
"learning_rate": 2.2229759299781185e-05,
"loss": 12.2162,
"step": 1121
},
{
"epoch": 0.7932481993725421,
"grad_norm": 16.555984497070312,
"learning_rate": 2.2157549234135668e-05,
"loss": 12.296,
"step": 1122
},
{
"epoch": 0.7939551942026424,
"grad_norm": 12.559003829956055,
"learning_rate": 2.2085339168490154e-05,
"loss": 12.0311,
"step": 1123
},
{
"epoch": 0.7946621890327427,
"grad_norm": 18.56354331970215,
"learning_rate": 2.201312910284464e-05,
"loss": 13.0041,
"step": 1124
},
{
"epoch": 0.795369183862843,
"grad_norm": 15.111113548278809,
"learning_rate": 2.1940919037199126e-05,
"loss": 13.2092,
"step": 1125
},
{
"epoch": 0.7960761786929433,
"grad_norm": 17.982934951782227,
"learning_rate": 2.1868708971553612e-05,
"loss": 14.1011,
"step": 1126
},
{
"epoch": 0.7967831735230436,
"grad_norm": 14.821832656860352,
"learning_rate": 2.1796498905908098e-05,
"loss": 13.0946,
"step": 1127
},
{
"epoch": 0.7974901683531439,
"grad_norm": 14.307646751403809,
"learning_rate": 2.1724288840262584e-05,
"loss": 11.4379,
"step": 1128
},
{
"epoch": 0.7981971631832442,
"grad_norm": 15.424671173095703,
"learning_rate": 2.165207877461707e-05,
"loss": 12.6333,
"step": 1129
},
{
"epoch": 0.7989041580133446,
"grad_norm": 16.159860610961914,
"learning_rate": 2.1579868708971556e-05,
"loss": 11.3861,
"step": 1130
},
{
"epoch": 0.7996111528434449,
"grad_norm": 17.140806198120117,
"learning_rate": 2.150765864332604e-05,
"loss": 12.7795,
"step": 1131
},
{
"epoch": 0.8003181476735451,
"grad_norm": 15.116321563720703,
"learning_rate": 2.143544857768053e-05,
"loss": 13.9064,
"step": 1132
},
{
"epoch": 0.8010251425036454,
"grad_norm": 15.199187278747559,
"learning_rate": 2.1363238512035014e-05,
"loss": 10.8875,
"step": 1133
},
{
"epoch": 0.8017321373337457,
"grad_norm": 14.66556167602539,
"learning_rate": 2.1291028446389497e-05,
"loss": 12.2029,
"step": 1134
},
{
"epoch": 0.8024391321638461,
"grad_norm": 15.456762313842773,
"learning_rate": 2.1218818380743983e-05,
"loss": 13.1825,
"step": 1135
},
{
"epoch": 0.8031461269939464,
"grad_norm": 18.5980167388916,
"learning_rate": 2.1146608315098473e-05,
"loss": 13.0355,
"step": 1136
},
{
"epoch": 0.8038531218240467,
"grad_norm": 15.780075073242188,
"learning_rate": 2.1074398249452955e-05,
"loss": 12.517,
"step": 1137
},
{
"epoch": 0.804560116654147,
"grad_norm": 17.828128814697266,
"learning_rate": 2.100218818380744e-05,
"loss": 11.9271,
"step": 1138
},
{
"epoch": 0.8052671114842472,
"grad_norm": 13.664030075073242,
"learning_rate": 2.092997811816193e-05,
"loss": 12.1154,
"step": 1139
},
{
"epoch": 0.8059741063143476,
"grad_norm": 21.724184036254883,
"learning_rate": 2.0857768052516413e-05,
"loss": 11.6643,
"step": 1140
},
{
"epoch": 0.8066811011444479,
"grad_norm": 15.66214370727539,
"learning_rate": 2.07855579868709e-05,
"loss": 12.9989,
"step": 1141
},
{
"epoch": 0.8073880959745482,
"grad_norm": 14.08765697479248,
"learning_rate": 2.0713347921225385e-05,
"loss": 12.1448,
"step": 1142
},
{
"epoch": 0.8080950908046485,
"grad_norm": 13.611175537109375,
"learning_rate": 2.064113785557987e-05,
"loss": 11.826,
"step": 1143
},
{
"epoch": 0.8088020856347488,
"grad_norm": 13.81747817993164,
"learning_rate": 2.0568927789934357e-05,
"loss": 11.613,
"step": 1144
},
{
"epoch": 0.8095090804648492,
"grad_norm": 13.661465644836426,
"learning_rate": 2.0496717724288843e-05,
"loss": 12.3789,
"step": 1145
},
{
"epoch": 0.8102160752949494,
"grad_norm": 17.0196590423584,
"learning_rate": 2.0424507658643326e-05,
"loss": 11.9157,
"step": 1146
},
{
"epoch": 0.8109230701250497,
"grad_norm": 14.75942325592041,
"learning_rate": 2.0352297592997815e-05,
"loss": 12.264,
"step": 1147
},
{
"epoch": 0.81163006495515,
"grad_norm": 15.581911087036133,
"learning_rate": 2.0280087527352298e-05,
"loss": 12.006,
"step": 1148
},
{
"epoch": 0.8123370597852503,
"grad_norm": 16.459049224853516,
"learning_rate": 2.0207877461706784e-05,
"loss": 11.139,
"step": 1149
},
{
"epoch": 0.8130440546153507,
"grad_norm": 14.035103797912598,
"learning_rate": 2.013566739606127e-05,
"loss": 12.8145,
"step": 1150
},
{
"epoch": 0.813751049445451,
"grad_norm": 13.479121208190918,
"learning_rate": 2.0063457330415756e-05,
"loss": 11.6135,
"step": 1151
},
{
"epoch": 0.8144580442755512,
"grad_norm": 13.676473617553711,
"learning_rate": 1.9991247264770242e-05,
"loss": 12.0856,
"step": 1152
},
{
"epoch": 0.8151650391056515,
"grad_norm": 16.638296127319336,
"learning_rate": 1.9919037199124728e-05,
"loss": 12.1883,
"step": 1153
},
{
"epoch": 0.8158720339357518,
"grad_norm": 14.359823226928711,
"learning_rate": 1.9846827133479214e-05,
"loss": 12.3571,
"step": 1154
},
{
"epoch": 0.8165790287658522,
"grad_norm": 15.50848388671875,
"learning_rate": 1.97746170678337e-05,
"loss": 11.2356,
"step": 1155
},
{
"epoch": 0.8172860235959525,
"grad_norm": 16.52731704711914,
"learning_rate": 1.9702407002188186e-05,
"loss": 12.2212,
"step": 1156
},
{
"epoch": 0.8179930184260528,
"grad_norm": 15.665827751159668,
"learning_rate": 1.963019693654267e-05,
"loss": 11.4575,
"step": 1157
},
{
"epoch": 0.818700013256153,
"grad_norm": 13.586274147033691,
"learning_rate": 1.955798687089716e-05,
"loss": 10.5227,
"step": 1158
},
{
"epoch": 0.8194070080862533,
"grad_norm": 13.108797073364258,
"learning_rate": 1.9485776805251644e-05,
"loss": 11.4108,
"step": 1159
},
{
"epoch": 0.8201140029163537,
"grad_norm": 15.360386848449707,
"learning_rate": 1.9413566739606127e-05,
"loss": 11.9431,
"step": 1160
},
{
"epoch": 0.820820997746454,
"grad_norm": 15.90074348449707,
"learning_rate": 1.9341356673960613e-05,
"loss": 13.6355,
"step": 1161
},
{
"epoch": 0.8215279925765543,
"grad_norm": 15.171931266784668,
"learning_rate": 1.9269146608315103e-05,
"loss": 10.6912,
"step": 1162
},
{
"epoch": 0.8222349874066546,
"grad_norm": 14.833294868469238,
"learning_rate": 1.9196936542669585e-05,
"loss": 13.6622,
"step": 1163
},
{
"epoch": 0.8229419822367549,
"grad_norm": 14.614151954650879,
"learning_rate": 1.912472647702407e-05,
"loss": 13.8156,
"step": 1164
},
{
"epoch": 0.8236489770668552,
"grad_norm": 14.428544044494629,
"learning_rate": 1.905251641137856e-05,
"loss": 11.7255,
"step": 1165
},
{
"epoch": 0.8243559718969555,
"grad_norm": 14.791826248168945,
"learning_rate": 1.8980306345733043e-05,
"loss": 11.8131,
"step": 1166
},
{
"epoch": 0.8250629667270558,
"grad_norm": 15.477858543395996,
"learning_rate": 1.890809628008753e-05,
"loss": 12.7562,
"step": 1167
},
{
"epoch": 0.8257699615571561,
"grad_norm": 14.780648231506348,
"learning_rate": 1.8835886214442015e-05,
"loss": 11.6997,
"step": 1168
},
{
"epoch": 0.8264769563872564,
"grad_norm": 15.95986557006836,
"learning_rate": 1.87636761487965e-05,
"loss": 12.5469,
"step": 1169
},
{
"epoch": 0.8271839512173568,
"grad_norm": 16.474977493286133,
"learning_rate": 1.8691466083150987e-05,
"loss": 12.6622,
"step": 1170
},
{
"epoch": 0.827890946047457,
"grad_norm": 14.389117240905762,
"learning_rate": 1.8619256017505473e-05,
"loss": 11.4078,
"step": 1171
},
{
"epoch": 0.8285979408775573,
"grad_norm": 21.426359176635742,
"learning_rate": 1.8547045951859956e-05,
"loss": 13.2855,
"step": 1172
},
{
"epoch": 0.8293049357076576,
"grad_norm": 15.901067733764648,
"learning_rate": 1.8474835886214446e-05,
"loss": 12.2708,
"step": 1173
},
{
"epoch": 0.8300119305377579,
"grad_norm": 15.242277145385742,
"learning_rate": 1.8402625820568928e-05,
"loss": 13.9031,
"step": 1174
},
{
"epoch": 0.8307189253678583,
"grad_norm": 14.417196273803711,
"learning_rate": 1.8330415754923414e-05,
"loss": 12.5818,
"step": 1175
},
{
"epoch": 0.8314259201979586,
"grad_norm": 15.016650199890137,
"learning_rate": 1.82582056892779e-05,
"loss": 12.4367,
"step": 1176
},
{
"epoch": 0.8321329150280589,
"grad_norm": 15.281384468078613,
"learning_rate": 1.8185995623632386e-05,
"loss": 12.3963,
"step": 1177
},
{
"epoch": 0.8328399098581591,
"grad_norm": 14.031335830688477,
"learning_rate": 1.8113785557986872e-05,
"loss": 12.2232,
"step": 1178
},
{
"epoch": 0.8335469046882594,
"grad_norm": 15.91395092010498,
"learning_rate": 1.804157549234136e-05,
"loss": 13.4726,
"step": 1179
},
{
"epoch": 0.8342538995183598,
"grad_norm": 18.70294952392578,
"learning_rate": 1.7969365426695844e-05,
"loss": 13.1052,
"step": 1180
},
{
"epoch": 0.8349608943484601,
"grad_norm": 16.646923065185547,
"learning_rate": 1.789715536105033e-05,
"loss": 11.53,
"step": 1181
},
{
"epoch": 0.8356678891785604,
"grad_norm": 15.27785873413086,
"learning_rate": 1.7824945295404816e-05,
"loss": 12.2457,
"step": 1182
},
{
"epoch": 0.8363748840086607,
"grad_norm": 14.599854469299316,
"learning_rate": 1.77527352297593e-05,
"loss": 11.63,
"step": 1183
},
{
"epoch": 0.837081878838761,
"grad_norm": 15.050758361816406,
"learning_rate": 1.768052516411379e-05,
"loss": 11.4395,
"step": 1184
},
{
"epoch": 0.8377888736688613,
"grad_norm": 12.898660659790039,
"learning_rate": 1.7608315098468275e-05,
"loss": 10.9921,
"step": 1185
},
{
"epoch": 0.8384958684989616,
"grad_norm": 16.362031936645508,
"learning_rate": 1.7536105032822757e-05,
"loss": 11.6573,
"step": 1186
},
{
"epoch": 0.8392028633290619,
"grad_norm": 16.562292098999023,
"learning_rate": 1.7463894967177243e-05,
"loss": 13.4289,
"step": 1187
},
{
"epoch": 0.8399098581591622,
"grad_norm": 15.194472312927246,
"learning_rate": 1.7391684901531733e-05,
"loss": 12.5519,
"step": 1188
},
{
"epoch": 0.8406168529892625,
"grad_norm": 15.846282958984375,
"learning_rate": 1.7319474835886215e-05,
"loss": 12.8095,
"step": 1189
},
{
"epoch": 0.8413238478193629,
"grad_norm": 15.602988243103027,
"learning_rate": 1.72472647702407e-05,
"loss": 11.5179,
"step": 1190
},
{
"epoch": 0.8420308426494632,
"grad_norm": 15.674636840820312,
"learning_rate": 1.717505470459519e-05,
"loss": 13.1862,
"step": 1191
},
{
"epoch": 0.8427378374795634,
"grad_norm": 13.627044677734375,
"learning_rate": 1.7102844638949673e-05,
"loss": 11.5302,
"step": 1192
},
{
"epoch": 0.8434448323096637,
"grad_norm": 16.106996536254883,
"learning_rate": 1.703063457330416e-05,
"loss": 12.02,
"step": 1193
},
{
"epoch": 0.844151827139764,
"grad_norm": 15.40173053741455,
"learning_rate": 1.6958424507658645e-05,
"loss": 13.0771,
"step": 1194
},
{
"epoch": 0.8448588219698644,
"grad_norm": 13.60293197631836,
"learning_rate": 1.688621444201313e-05,
"loss": 11.2656,
"step": 1195
},
{
"epoch": 0.8455658167999647,
"grad_norm": 15.298815727233887,
"learning_rate": 1.6814004376367617e-05,
"loss": 12.9284,
"step": 1196
},
{
"epoch": 0.846272811630065,
"grad_norm": 14.378645896911621,
"learning_rate": 1.6741794310722103e-05,
"loss": 11.3607,
"step": 1197
},
{
"epoch": 0.8469798064601652,
"grad_norm": 17.42547035217285,
"learning_rate": 1.6669584245076586e-05,
"loss": 12.823,
"step": 1198
},
{
"epoch": 0.8476868012902655,
"grad_norm": 18.732912063598633,
"learning_rate": 1.6597374179431076e-05,
"loss": 12.9339,
"step": 1199
},
{
"epoch": 0.8483937961203659,
"grad_norm": 15.936331748962402,
"learning_rate": 1.6525164113785558e-05,
"loss": 13.8907,
"step": 1200
},
{
"epoch": 0.8491007909504662,
"grad_norm": 17.328195571899414,
"learning_rate": 1.6452954048140044e-05,
"loss": 12.764,
"step": 1201
},
{
"epoch": 0.8498077857805665,
"grad_norm": 15.351299285888672,
"learning_rate": 1.6380743982494534e-05,
"loss": 12.9472,
"step": 1202
},
{
"epoch": 0.8505147806106668,
"grad_norm": 16.695545196533203,
"learning_rate": 1.6308533916849016e-05,
"loss": 11.2182,
"step": 1203
},
{
"epoch": 0.851221775440767,
"grad_norm": 15.395929336547852,
"learning_rate": 1.6236323851203502e-05,
"loss": 13.5168,
"step": 1204
},
{
"epoch": 0.8519287702708674,
"grad_norm": 14.865396499633789,
"learning_rate": 1.616411378555799e-05,
"loss": 12.2607,
"step": 1205
},
{
"epoch": 0.8526357651009677,
"grad_norm": 14.395364761352539,
"learning_rate": 1.6091903719912474e-05,
"loss": 11.1264,
"step": 1206
},
{
"epoch": 0.853342759931068,
"grad_norm": 15.32667064666748,
"learning_rate": 1.601969365426696e-05,
"loss": 13.1768,
"step": 1207
},
{
"epoch": 0.8540497547611683,
"grad_norm": 18.79914093017578,
"learning_rate": 1.5947483588621443e-05,
"loss": 13.2718,
"step": 1208
},
{
"epoch": 0.8547567495912686,
"grad_norm": 14.390812873840332,
"learning_rate": 1.587527352297593e-05,
"loss": 11.8755,
"step": 1209
},
{
"epoch": 0.855463744421369,
"grad_norm": 15.825724601745605,
"learning_rate": 1.580306345733042e-05,
"loss": 11.9834,
"step": 1210
},
{
"epoch": 0.8561707392514692,
"grad_norm": 14.854561805725098,
"learning_rate": 1.57308533916849e-05,
"loss": 13.5324,
"step": 1211
},
{
"epoch": 0.8568777340815695,
"grad_norm": 16.236326217651367,
"learning_rate": 1.5658643326039387e-05,
"loss": 12.7902,
"step": 1212
},
{
"epoch": 0.8575847289116698,
"grad_norm": 14.993921279907227,
"learning_rate": 1.5586433260393873e-05,
"loss": 12.5322,
"step": 1213
},
{
"epoch": 0.8582917237417701,
"grad_norm": 13.216548919677734,
"learning_rate": 1.551422319474836e-05,
"loss": 12.2942,
"step": 1214
},
{
"epoch": 0.8589987185718705,
"grad_norm": 15.003395080566406,
"learning_rate": 1.5442013129102845e-05,
"loss": 14.2319,
"step": 1215
},
{
"epoch": 0.8597057134019708,
"grad_norm": 15.051974296569824,
"learning_rate": 1.536980306345733e-05,
"loss": 12.4653,
"step": 1216
},
{
"epoch": 0.860412708232071,
"grad_norm": 16.103477478027344,
"learning_rate": 1.5297592997811817e-05,
"loss": 12.6086,
"step": 1217
},
{
"epoch": 0.8611197030621713,
"grad_norm": 15.790787696838379,
"learning_rate": 1.5225382932166303e-05,
"loss": 11.7837,
"step": 1218
},
{
"epoch": 0.8618266978922716,
"grad_norm": 14.117712020874023,
"learning_rate": 1.515317286652079e-05,
"loss": 13.6153,
"step": 1219
},
{
"epoch": 0.862533692722372,
"grad_norm": 13.054887771606445,
"learning_rate": 1.5080962800875274e-05,
"loss": 12.3808,
"step": 1220
},
{
"epoch": 0.8632406875524723,
"grad_norm": 14.148192405700684,
"learning_rate": 1.5008752735229761e-05,
"loss": 12.9498,
"step": 1221
},
{
"epoch": 0.8639476823825726,
"grad_norm": 15.982940673828125,
"learning_rate": 1.4936542669584247e-05,
"loss": 13.6622,
"step": 1222
},
{
"epoch": 0.8646546772126729,
"grad_norm": 14.222307205200195,
"learning_rate": 1.4864332603938732e-05,
"loss": 12.2357,
"step": 1223
},
{
"epoch": 0.8653616720427731,
"grad_norm": 14.06337833404541,
"learning_rate": 1.4792122538293218e-05,
"loss": 11.9331,
"step": 1224
},
{
"epoch": 0.8660686668728735,
"grad_norm": 18.584463119506836,
"learning_rate": 1.4719912472647706e-05,
"loss": 12.0133,
"step": 1225
},
{
"epoch": 0.8667756617029738,
"grad_norm": 14.86030101776123,
"learning_rate": 1.464770240700219e-05,
"loss": 13.2664,
"step": 1226
},
{
"epoch": 0.8674826565330741,
"grad_norm": 14.793909072875977,
"learning_rate": 1.4575492341356674e-05,
"loss": 12.9823,
"step": 1227
},
{
"epoch": 0.8681896513631744,
"grad_norm": 15.59019947052002,
"learning_rate": 1.4503282275711162e-05,
"loss": 12.4917,
"step": 1228
},
{
"epoch": 0.8688966461932747,
"grad_norm": 13.743398666381836,
"learning_rate": 1.4431072210065648e-05,
"loss": 11.7762,
"step": 1229
},
{
"epoch": 0.8696036410233751,
"grad_norm": 16.514175415039062,
"learning_rate": 1.4358862144420132e-05,
"loss": 11.7951,
"step": 1230
},
{
"epoch": 0.8703106358534753,
"grad_norm": 14.362953186035156,
"learning_rate": 1.4286652078774618e-05,
"loss": 10.9248,
"step": 1231
},
{
"epoch": 0.8710176306835756,
"grad_norm": 16.149871826171875,
"learning_rate": 1.4214442013129106e-05,
"loss": 12.5736,
"step": 1232
},
{
"epoch": 0.8717246255136759,
"grad_norm": 15.02379322052002,
"learning_rate": 1.414223194748359e-05,
"loss": 12.475,
"step": 1233
},
{
"epoch": 0.8724316203437762,
"grad_norm": 15.37895679473877,
"learning_rate": 1.4070021881838075e-05,
"loss": 11.1181,
"step": 1234
},
{
"epoch": 0.8731386151738766,
"grad_norm": 14.252886772155762,
"learning_rate": 1.399781181619256e-05,
"loss": 13.0672,
"step": 1235
},
{
"epoch": 0.8738456100039769,
"grad_norm": 14.250099182128906,
"learning_rate": 1.3925601750547049e-05,
"loss": 12.467,
"step": 1236
},
{
"epoch": 0.8745526048340772,
"grad_norm": 15.835908889770508,
"learning_rate": 1.3853391684901533e-05,
"loss": 14.3098,
"step": 1237
},
{
"epoch": 0.8752595996641774,
"grad_norm": 15.358033180236816,
"learning_rate": 1.3781181619256019e-05,
"loss": 12.1643,
"step": 1238
},
{
"epoch": 0.8759665944942777,
"grad_norm": 14.915874481201172,
"learning_rate": 1.3708971553610507e-05,
"loss": 12.093,
"step": 1239
},
{
"epoch": 0.8766735893243781,
"grad_norm": 17.685922622680664,
"learning_rate": 1.3636761487964991e-05,
"loss": 11.1443,
"step": 1240
},
{
"epoch": 0.8773805841544784,
"grad_norm": 14.141484260559082,
"learning_rate": 1.3564551422319477e-05,
"loss": 11.1317,
"step": 1241
},
{
"epoch": 0.8780875789845787,
"grad_norm": 14.050369262695312,
"learning_rate": 1.3492341356673961e-05,
"loss": 12.0695,
"step": 1242
},
{
"epoch": 0.878794573814679,
"grad_norm": 14.12048053741455,
"learning_rate": 1.3420131291028449e-05,
"loss": 10.7497,
"step": 1243
},
{
"epoch": 0.8795015686447792,
"grad_norm": 15.856414794921875,
"learning_rate": 1.3347921225382933e-05,
"loss": 11.6056,
"step": 1244
},
{
"epoch": 0.8802085634748796,
"grad_norm": 15.296753883361816,
"learning_rate": 1.327571115973742e-05,
"loss": 12.4171,
"step": 1245
},
{
"epoch": 0.8809155583049799,
"grad_norm": 14.189363479614258,
"learning_rate": 1.3203501094091905e-05,
"loss": 11.1442,
"step": 1246
},
{
"epoch": 0.8816225531350802,
"grad_norm": 13.78650188446045,
"learning_rate": 1.313129102844639e-05,
"loss": 12.2327,
"step": 1247
},
{
"epoch": 0.8823295479651805,
"grad_norm": 14.706070899963379,
"learning_rate": 1.3059080962800877e-05,
"loss": 12.6943,
"step": 1248
},
{
"epoch": 0.8830365427952808,
"grad_norm": 17.17374038696289,
"learning_rate": 1.2986870897155362e-05,
"loss": 13.3784,
"step": 1249
},
{
"epoch": 0.8837435376253812,
"grad_norm": 15.617532730102539,
"learning_rate": 1.2914660831509848e-05,
"loss": 12.0446,
"step": 1250
},
{
"epoch": 0.8844505324554814,
"grad_norm": 15.952999114990234,
"learning_rate": 1.2842450765864334e-05,
"loss": 11.5787,
"step": 1251
},
{
"epoch": 0.8851575272855817,
"grad_norm": 14.995427131652832,
"learning_rate": 1.277024070021882e-05,
"loss": 11.7415,
"step": 1252
},
{
"epoch": 0.885864522115682,
"grad_norm": 17.298742294311523,
"learning_rate": 1.2698030634573306e-05,
"loss": 12.72,
"step": 1253
},
{
"epoch": 0.8865715169457823,
"grad_norm": 14.929863929748535,
"learning_rate": 1.2625820568927792e-05,
"loss": 12.582,
"step": 1254
},
{
"epoch": 0.8872785117758827,
"grad_norm": 12.553905487060547,
"learning_rate": 1.2553610503282278e-05,
"loss": 10.473,
"step": 1255
},
{
"epoch": 0.887985506605983,
"grad_norm": 14.32827377319336,
"learning_rate": 1.2481400437636762e-05,
"loss": 13.7437,
"step": 1256
},
{
"epoch": 0.8886925014360832,
"grad_norm": 15.110925674438477,
"learning_rate": 1.2409190371991248e-05,
"loss": 13.3395,
"step": 1257
},
{
"epoch": 0.8893994962661835,
"grad_norm": 13.540974617004395,
"learning_rate": 1.2336980306345734e-05,
"loss": 10.6814,
"step": 1258
},
{
"epoch": 0.8901064910962838,
"grad_norm": 13.67444896697998,
"learning_rate": 1.226477024070022e-05,
"loss": 12.2924,
"step": 1259
},
{
"epoch": 0.8908134859263842,
"grad_norm": 14.55646800994873,
"learning_rate": 1.2192560175054705e-05,
"loss": 12.3313,
"step": 1260
},
{
"epoch": 0.8915204807564845,
"grad_norm": 16.98577880859375,
"learning_rate": 1.2120350109409193e-05,
"loss": 12.2591,
"step": 1261
},
{
"epoch": 0.8922274755865848,
"grad_norm": 15.357362747192383,
"learning_rate": 1.2048140043763677e-05,
"loss": 11.7354,
"step": 1262
},
{
"epoch": 0.892934470416685,
"grad_norm": 14.358413696289062,
"learning_rate": 1.1975929978118163e-05,
"loss": 12.9059,
"step": 1263
},
{
"epoch": 0.8936414652467853,
"grad_norm": 14.79417896270752,
"learning_rate": 1.1903719912472649e-05,
"loss": 12.4349,
"step": 1264
},
{
"epoch": 0.8943484600768857,
"grad_norm": 15.485979080200195,
"learning_rate": 1.1831509846827135e-05,
"loss": 11.7835,
"step": 1265
},
{
"epoch": 0.895055454906986,
"grad_norm": 13.867639541625977,
"learning_rate": 1.1759299781181621e-05,
"loss": 10.9385,
"step": 1266
},
{
"epoch": 0.8957624497370863,
"grad_norm": 17.667858123779297,
"learning_rate": 1.1687089715536107e-05,
"loss": 11.9884,
"step": 1267
},
{
"epoch": 0.8964694445671866,
"grad_norm": 16.6583251953125,
"learning_rate": 1.1614879649890593e-05,
"loss": 12.0298,
"step": 1268
},
{
"epoch": 0.8971764393972869,
"grad_norm": 18.523691177368164,
"learning_rate": 1.1542669584245077e-05,
"loss": 11.216,
"step": 1269
},
{
"epoch": 0.8978834342273873,
"grad_norm": 19.020814895629883,
"learning_rate": 1.1470459518599563e-05,
"loss": 13.4564,
"step": 1270
},
{
"epoch": 0.8985904290574875,
"grad_norm": 15.264403343200684,
"learning_rate": 1.139824945295405e-05,
"loss": 11.9713,
"step": 1271
},
{
"epoch": 0.8992974238875878,
"grad_norm": 14.84862995147705,
"learning_rate": 1.1326039387308535e-05,
"loss": 11.2933,
"step": 1272
},
{
"epoch": 0.9000044187176881,
"grad_norm": 14.617576599121094,
"learning_rate": 1.125382932166302e-05,
"loss": 13.196,
"step": 1273
},
{
"epoch": 0.9007114135477884,
"grad_norm": 17.44744873046875,
"learning_rate": 1.1181619256017508e-05,
"loss": 12.2643,
"step": 1274
},
{
"epoch": 0.9014184083778888,
"grad_norm": 15.30611515045166,
"learning_rate": 1.1109409190371992e-05,
"loss": 11.58,
"step": 1275
},
{
"epoch": 0.9021254032079891,
"grad_norm": 15.377030372619629,
"learning_rate": 1.1037199124726478e-05,
"loss": 12.4231,
"step": 1276
},
{
"epoch": 0.9028323980380893,
"grad_norm": 13.975566864013672,
"learning_rate": 1.0964989059080966e-05,
"loss": 12.0694,
"step": 1277
},
{
"epoch": 0.9035393928681896,
"grad_norm": 16.075166702270508,
"learning_rate": 1.089277899343545e-05,
"loss": 12.8699,
"step": 1278
},
{
"epoch": 0.9042463876982899,
"grad_norm": 13.96810245513916,
"learning_rate": 1.0820568927789936e-05,
"loss": 11.9998,
"step": 1279
},
{
"epoch": 0.9049533825283903,
"grad_norm": 14.134039878845215,
"learning_rate": 1.0748358862144422e-05,
"loss": 11.1477,
"step": 1280
},
{
"epoch": 0.9056603773584906,
"grad_norm": 13.151305198669434,
"learning_rate": 1.0676148796498908e-05,
"loss": 11.9197,
"step": 1281
},
{
"epoch": 0.9063673721885909,
"grad_norm": 13.498329162597656,
"learning_rate": 1.0603938730853392e-05,
"loss": 12.8619,
"step": 1282
},
{
"epoch": 0.9070743670186912,
"grad_norm": 14.909584045410156,
"learning_rate": 1.0531728665207878e-05,
"loss": 11.9246,
"step": 1283
},
{
"epoch": 0.9077813618487914,
"grad_norm": 14.063223838806152,
"learning_rate": 1.0459518599562364e-05,
"loss": 12.7957,
"step": 1284
},
{
"epoch": 0.9084883566788918,
"grad_norm": 13.720608711242676,
"learning_rate": 1.038730853391685e-05,
"loss": 12.7551,
"step": 1285
},
{
"epoch": 0.9091953515089921,
"grad_norm": 14.793830871582031,
"learning_rate": 1.0315098468271335e-05,
"loss": 12.179,
"step": 1286
},
{
"epoch": 0.9099023463390924,
"grad_norm": 12.872452735900879,
"learning_rate": 1.0242888402625823e-05,
"loss": 11.0132,
"step": 1287
},
{
"epoch": 0.9106093411691927,
"grad_norm": 16.042049407958984,
"learning_rate": 1.0170678336980307e-05,
"loss": 11.8582,
"step": 1288
},
{
"epoch": 0.911316335999293,
"grad_norm": 21.02182388305664,
"learning_rate": 1.0098468271334793e-05,
"loss": 13.052,
"step": 1289
},
{
"epoch": 0.9120233308293934,
"grad_norm": 13.702282905578613,
"learning_rate": 1.002625820568928e-05,
"loss": 11.8454,
"step": 1290
},
{
"epoch": 0.9127303256594936,
"grad_norm": 13.618666648864746,
"learning_rate": 9.954048140043765e-06,
"loss": 11.8811,
"step": 1291
},
{
"epoch": 0.9134373204895939,
"grad_norm": 16.58145523071289,
"learning_rate": 9.881838074398251e-06,
"loss": 12.1977,
"step": 1292
},
{
"epoch": 0.9141443153196942,
"grad_norm": 12.983065605163574,
"learning_rate": 9.809628008752737e-06,
"loss": 10.1651,
"step": 1293
},
{
"epoch": 0.9148513101497945,
"grad_norm": 15.553836822509766,
"learning_rate": 9.737417943107223e-06,
"loss": 12.1087,
"step": 1294
},
{
"epoch": 0.9155583049798949,
"grad_norm": 12.707442283630371,
"learning_rate": 9.665207877461707e-06,
"loss": 11.662,
"step": 1295
},
{
"epoch": 0.9162652998099952,
"grad_norm": 16.751333236694336,
"learning_rate": 9.592997811816193e-06,
"loss": 11.397,
"step": 1296
},
{
"epoch": 0.9169722946400954,
"grad_norm": 14.83692455291748,
"learning_rate": 9.52078774617068e-06,
"loss": 10.7658,
"step": 1297
},
{
"epoch": 0.9176792894701957,
"grad_norm": 12.964903831481934,
"learning_rate": 9.448577680525165e-06,
"loss": 11.8422,
"step": 1298
},
{
"epoch": 0.918386284300296,
"grad_norm": 13.346761703491211,
"learning_rate": 9.37636761487965e-06,
"loss": 11.7921,
"step": 1299
},
{
"epoch": 0.9190932791303964,
"grad_norm": 15.598644256591797,
"learning_rate": 9.304157549234138e-06,
"loss": 11.7002,
"step": 1300
},
{
"epoch": 0.9198002739604967,
"grad_norm": 14.48563003540039,
"learning_rate": 9.231947483588622e-06,
"loss": 11.5056,
"step": 1301
},
{
"epoch": 0.920507268790597,
"grad_norm": 13.817610740661621,
"learning_rate": 9.159737417943108e-06,
"loss": 11.8712,
"step": 1302
},
{
"epoch": 0.9212142636206972,
"grad_norm": 16.314889907836914,
"learning_rate": 9.087527352297596e-06,
"loss": 12.5035,
"step": 1303
},
{
"epoch": 0.9219212584507975,
"grad_norm": 16.28759765625,
"learning_rate": 9.01531728665208e-06,
"loss": 12.0567,
"step": 1304
},
{
"epoch": 0.9226282532808979,
"grad_norm": 14.504642486572266,
"learning_rate": 8.943107221006566e-06,
"loss": 11.6086,
"step": 1305
},
{
"epoch": 0.9233352481109982,
"grad_norm": 15.890057563781738,
"learning_rate": 8.870897155361052e-06,
"loss": 13.1966,
"step": 1306
},
{
"epoch": 0.9240422429410985,
"grad_norm": 15.602300643920898,
"learning_rate": 8.798687089715538e-06,
"loss": 11.283,
"step": 1307
},
{
"epoch": 0.9247492377711988,
"grad_norm": 14.971783638000488,
"learning_rate": 8.726477024070022e-06,
"loss": 10.8825,
"step": 1308
},
{
"epoch": 0.9254562326012991,
"grad_norm": 14.330822944641113,
"learning_rate": 8.654266958424508e-06,
"loss": 10.4557,
"step": 1309
},
{
"epoch": 0.9261632274313994,
"grad_norm": 14.569314002990723,
"learning_rate": 8.582056892778994e-06,
"loss": 11.6277,
"step": 1310
},
{
"epoch": 0.9268702222614997,
"grad_norm": 16.473087310791016,
"learning_rate": 8.50984682713348e-06,
"loss": 12.8317,
"step": 1311
},
{
"epoch": 0.9275772170916,
"grad_norm": 14.927108764648438,
"learning_rate": 8.437636761487965e-06,
"loss": 11.2278,
"step": 1312
},
{
"epoch": 0.9282842119217003,
"grad_norm": 14.588191986083984,
"learning_rate": 8.365426695842451e-06,
"loss": 11.8304,
"step": 1313
},
{
"epoch": 0.9289912067518006,
"grad_norm": 14.352690696716309,
"learning_rate": 8.293216630196937e-06,
"loss": 12.0047,
"step": 1314
},
{
"epoch": 0.929698201581901,
"grad_norm": 13.969871520996094,
"learning_rate": 8.221006564551423e-06,
"loss": 11.8746,
"step": 1315
},
{
"epoch": 0.9304051964120013,
"grad_norm": 14.721321105957031,
"learning_rate": 8.148796498905909e-06,
"loss": 11.3543,
"step": 1316
},
{
"epoch": 0.9311121912421015,
"grad_norm": 13.538402557373047,
"learning_rate": 8.076586433260395e-06,
"loss": 12.1419,
"step": 1317
},
{
"epoch": 0.9318191860722018,
"grad_norm": 16.33662986755371,
"learning_rate": 8.004376367614881e-06,
"loss": 13.1276,
"step": 1318
},
{
"epoch": 0.9325261809023021,
"grad_norm": 14.316162109375,
"learning_rate": 7.932166301969365e-06,
"loss": 12.8216,
"step": 1319
},
{
"epoch": 0.9332331757324025,
"grad_norm": 17.847124099731445,
"learning_rate": 7.859956236323853e-06,
"loss": 11.2353,
"step": 1320
},
{
"epoch": 0.9339401705625028,
"grad_norm": 14.48947525024414,
"learning_rate": 7.787746170678337e-06,
"loss": 11.4874,
"step": 1321
},
{
"epoch": 0.9346471653926031,
"grad_norm": 14.466111183166504,
"learning_rate": 7.715536105032823e-06,
"loss": 12.4364,
"step": 1322
},
{
"epoch": 0.9353541602227033,
"grad_norm": 14.350505828857422,
"learning_rate": 7.64332603938731e-06,
"loss": 12.3382,
"step": 1323
},
{
"epoch": 0.9360611550528036,
"grad_norm": 17.1221866607666,
"learning_rate": 7.571115973741795e-06,
"loss": 13.8187,
"step": 1324
},
{
"epoch": 0.936768149882904,
"grad_norm": 14.196785926818848,
"learning_rate": 7.49890590809628e-06,
"loss": 11.4155,
"step": 1325
},
{
"epoch": 0.9374751447130043,
"grad_norm": 14.461777687072754,
"learning_rate": 7.426695842450766e-06,
"loss": 12.296,
"step": 1326
},
{
"epoch": 0.9381821395431046,
"grad_norm": 14.26052474975586,
"learning_rate": 7.354485776805253e-06,
"loss": 11.1558,
"step": 1327
},
{
"epoch": 0.9388891343732049,
"grad_norm": 15.175439834594727,
"learning_rate": 7.282275711159738e-06,
"loss": 12.9709,
"step": 1328
},
{
"epoch": 0.9395961292033052,
"grad_norm": 14.596419334411621,
"learning_rate": 7.210065645514224e-06,
"loss": 12.8197,
"step": 1329
},
{
"epoch": 0.9403031240334055,
"grad_norm": 14.158638000488281,
"learning_rate": 7.137855579868709e-06,
"loss": 12.6762,
"step": 1330
},
{
"epoch": 0.9410101188635058,
"grad_norm": 16.99709701538086,
"learning_rate": 7.065645514223194e-06,
"loss": 10.9333,
"step": 1331
},
{
"epoch": 0.9417171136936061,
"grad_norm": 15.491727828979492,
"learning_rate": 6.99343544857768e-06,
"loss": 11.3055,
"step": 1332
},
{
"epoch": 0.9424241085237064,
"grad_norm": 16.94435691833496,
"learning_rate": 6.921225382932166e-06,
"loss": 13.25,
"step": 1333
},
{
"epoch": 0.9431311033538067,
"grad_norm": 14.477254867553711,
"learning_rate": 6.8490153172866524e-06,
"loss": 11.0744,
"step": 1334
},
{
"epoch": 0.9438380981839071,
"grad_norm": 15.165593147277832,
"learning_rate": 6.7768052516411385e-06,
"loss": 10.9101,
"step": 1335
},
{
"epoch": 0.9445450930140074,
"grad_norm": 14.337298393249512,
"learning_rate": 6.704595185995624e-06,
"loss": 11.1425,
"step": 1336
},
{
"epoch": 0.9452520878441076,
"grad_norm": 15.68160629272461,
"learning_rate": 6.63238512035011e-06,
"loss": 12.3696,
"step": 1337
},
{
"epoch": 0.9459590826742079,
"grad_norm": 13.884684562683105,
"learning_rate": 6.560175054704596e-06,
"loss": 10.9927,
"step": 1338
},
{
"epoch": 0.9466660775043082,
"grad_norm": 16.531490325927734,
"learning_rate": 6.487964989059081e-06,
"loss": 11.8801,
"step": 1339
},
{
"epoch": 0.9473730723344086,
"grad_norm": 18.980274200439453,
"learning_rate": 6.415754923413567e-06,
"loss": 13.2553,
"step": 1340
},
{
"epoch": 0.9480800671645089,
"grad_norm": 14.567325592041016,
"learning_rate": 6.343544857768053e-06,
"loss": 12.6151,
"step": 1341
},
{
"epoch": 0.9487870619946092,
"grad_norm": 15.460344314575195,
"learning_rate": 6.271334792122538e-06,
"loss": 10.808,
"step": 1342
},
{
"epoch": 0.9494940568247094,
"grad_norm": 13.62412166595459,
"learning_rate": 6.199124726477024e-06,
"loss": 12.6647,
"step": 1343
},
{
"epoch": 0.9502010516548097,
"grad_norm": 14.015798568725586,
"learning_rate": 6.126914660831509e-06,
"loss": 11.7514,
"step": 1344
},
{
"epoch": 0.9509080464849101,
"grad_norm": 15.991093635559082,
"learning_rate": 6.054704595185995e-06,
"loss": 12.831,
"step": 1345
},
{
"epoch": 0.9516150413150104,
"grad_norm": 14.038092613220215,
"learning_rate": 5.982494529540482e-06,
"loss": 11.2131,
"step": 1346
},
{
"epoch": 0.9523220361451107,
"grad_norm": 19.56915855407715,
"learning_rate": 5.9102844638949674e-06,
"loss": 12.5214,
"step": 1347
},
{
"epoch": 0.953029030975211,
"grad_norm": 13.879122734069824,
"learning_rate": 5.8380743982494535e-06,
"loss": 11.75,
"step": 1348
},
{
"epoch": 0.9537360258053112,
"grad_norm": 14.805500984191895,
"learning_rate": 5.765864332603939e-06,
"loss": 12.5357,
"step": 1349
},
{
"epoch": 0.9544430206354116,
"grad_norm": 12.95010757446289,
"learning_rate": 5.693654266958425e-06,
"loss": 10.1176,
"step": 1350
},
{
"epoch": 0.9551500154655119,
"grad_norm": 14.349968910217285,
"learning_rate": 5.621444201312911e-06,
"loss": 10.9632,
"step": 1351
},
{
"epoch": 0.9558570102956122,
"grad_norm": 16.88902473449707,
"learning_rate": 5.549234135667396e-06,
"loss": 11.671,
"step": 1352
},
{
"epoch": 0.9565640051257125,
"grad_norm": 14.569315910339355,
"learning_rate": 5.477024070021882e-06,
"loss": 11.9923,
"step": 1353
},
{
"epoch": 0.9572709999558128,
"grad_norm": 14.235941886901855,
"learning_rate": 5.404814004376368e-06,
"loss": 11.1376,
"step": 1354
},
{
"epoch": 0.9579779947859132,
"grad_norm": 16.43010711669922,
"learning_rate": 5.332603938730853e-06,
"loss": 11.198,
"step": 1355
},
{
"epoch": 0.9586849896160134,
"grad_norm": 14.069321632385254,
"learning_rate": 5.260393873085339e-06,
"loss": 10.3045,
"step": 1356
},
{
"epoch": 0.9593919844461137,
"grad_norm": 15.589603424072266,
"learning_rate": 5.188183807439824e-06,
"loss": 12.1372,
"step": 1357
},
{
"epoch": 0.960098979276214,
"grad_norm": 15.514388084411621,
"learning_rate": 5.11597374179431e-06,
"loss": 12.2847,
"step": 1358
},
{
"epoch": 0.9608059741063143,
"grad_norm": 13.620744705200195,
"learning_rate": 5.043763676148797e-06,
"loss": 12.6984,
"step": 1359
},
{
"epoch": 0.9615129689364147,
"grad_norm": 14.314813613891602,
"learning_rate": 4.9715536105032825e-06,
"loss": 12.9155,
"step": 1360
},
{
"epoch": 0.962219963766515,
"grad_norm": 14.13175106048584,
"learning_rate": 4.8993435448577685e-06,
"loss": 10.4515,
"step": 1361
},
{
"epoch": 0.9629269585966153,
"grad_norm": 13.860292434692383,
"learning_rate": 4.827133479212254e-06,
"loss": 10.8381,
"step": 1362
},
{
"epoch": 0.9636339534267155,
"grad_norm": 15.983375549316406,
"learning_rate": 4.75492341356674e-06,
"loss": 12.8587,
"step": 1363
},
{
"epoch": 0.9643409482568158,
"grad_norm": 14.781173706054688,
"learning_rate": 4.682713347921226e-06,
"loss": 11.3576,
"step": 1364
},
{
"epoch": 0.9650479430869162,
"grad_norm": 14.570517539978027,
"learning_rate": 4.610503282275712e-06,
"loss": 11.6258,
"step": 1365
},
{
"epoch": 0.9657549379170165,
"grad_norm": 14.185210227966309,
"learning_rate": 4.538293216630197e-06,
"loss": 12.2246,
"step": 1366
},
{
"epoch": 0.9664619327471168,
"grad_norm": 13.819804191589355,
"learning_rate": 4.466083150984683e-06,
"loss": 12.2629,
"step": 1367
},
{
"epoch": 0.9671689275772171,
"grad_norm": 15.09450912475586,
"learning_rate": 4.393873085339168e-06,
"loss": 11.8637,
"step": 1368
},
{
"epoch": 0.9678759224073173,
"grad_norm": 12.672414779663086,
"learning_rate": 4.321663019693654e-06,
"loss": 10.5733,
"step": 1369
},
{
"epoch": 0.9685829172374177,
"grad_norm": 15.714422225952148,
"learning_rate": 4.24945295404814e-06,
"loss": 12.4989,
"step": 1370
},
{
"epoch": 0.969289912067518,
"grad_norm": 16.135835647583008,
"learning_rate": 4.177242888402626e-06,
"loss": 11.4435,
"step": 1371
},
{
"epoch": 0.9699969068976183,
"grad_norm": 15.989006996154785,
"learning_rate": 4.105032822757112e-06,
"loss": 12.7837,
"step": 1372
},
{
"epoch": 0.9707039017277186,
"grad_norm": 14.65039348602295,
"learning_rate": 4.0328227571115975e-06,
"loss": 11.1327,
"step": 1373
},
{
"epoch": 0.9714108965578189,
"grad_norm": 14.098581314086914,
"learning_rate": 3.9606126914660835e-06,
"loss": 13.5108,
"step": 1374
},
{
"epoch": 0.9721178913879193,
"grad_norm": 14.409170150756836,
"learning_rate": 3.8884026258205695e-06,
"loss": 11.501,
"step": 1375
},
{
"epoch": 0.9728248862180195,
"grad_norm": 15.440765380859375,
"learning_rate": 3.816192560175055e-06,
"loss": 11.2453,
"step": 1376
},
{
"epoch": 0.9735318810481198,
"grad_norm": 13.261435508728027,
"learning_rate": 3.7439824945295407e-06,
"loss": 11.9026,
"step": 1377
},
{
"epoch": 0.9742388758782201,
"grad_norm": 13.441767692565918,
"learning_rate": 3.6717724288840268e-06,
"loss": 12.8835,
"step": 1378
},
{
"epoch": 0.9749458707083204,
"grad_norm": 16.43163299560547,
"learning_rate": 3.599562363238512e-06,
"loss": 12.0741,
"step": 1379
},
{
"epoch": 0.9756528655384208,
"grad_norm": 13.344820022583008,
"learning_rate": 3.527352297592998e-06,
"loss": 11.7269,
"step": 1380
},
{
"epoch": 0.9763598603685211,
"grad_norm": 16.93538475036621,
"learning_rate": 3.455142231947484e-06,
"loss": 12.4897,
"step": 1381
},
{
"epoch": 0.9770668551986214,
"grad_norm": 15.433717727661133,
"learning_rate": 3.382932166301969e-06,
"loss": 12.7013,
"step": 1382
},
{
"epoch": 0.9777738500287216,
"grad_norm": 16.81320571899414,
"learning_rate": 3.3107221006564552e-06,
"loss": 12.175,
"step": 1383
},
{
"epoch": 0.9784808448588219,
"grad_norm": 14.076522827148438,
"learning_rate": 3.2385120350109413e-06,
"loss": 11.0918,
"step": 1384
},
{
"epoch": 0.9791878396889223,
"grad_norm": 15.008047103881836,
"learning_rate": 3.1663019693654264e-06,
"loss": 11.1541,
"step": 1385
},
{
"epoch": 0.9798948345190226,
"grad_norm": 16.03972053527832,
"learning_rate": 3.0940919037199125e-06,
"loss": 11.2622,
"step": 1386
},
{
"epoch": 0.9806018293491229,
"grad_norm": 12.36665153503418,
"learning_rate": 3.0218818380743985e-06,
"loss": 11.854,
"step": 1387
},
{
"epoch": 0.9813088241792232,
"grad_norm": 15.743560791015625,
"learning_rate": 2.9496717724288845e-06,
"loss": 11.1556,
"step": 1388
},
{
"epoch": 0.9820158190093234,
"grad_norm": 13.828715324401855,
"learning_rate": 2.8774617067833697e-06,
"loss": 12.3923,
"step": 1389
},
{
"epoch": 0.9827228138394238,
"grad_norm": 15.561683654785156,
"learning_rate": 2.8052516411378558e-06,
"loss": 12.3997,
"step": 1390
},
{
"epoch": 0.9834298086695241,
"grad_norm": 15.786792755126953,
"learning_rate": 2.7330415754923414e-06,
"loss": 12.4365,
"step": 1391
},
{
"epoch": 0.9841368034996244,
"grad_norm": 14.015902519226074,
"learning_rate": 2.660831509846827e-06,
"loss": 12.3683,
"step": 1392
},
{
"epoch": 0.9848437983297247,
"grad_norm": 12.718472480773926,
"learning_rate": 2.588621444201313e-06,
"loss": 11.3825,
"step": 1393
},
{
"epoch": 0.985550793159825,
"grad_norm": 14.474007606506348,
"learning_rate": 2.5164113785557986e-06,
"loss": 11.0584,
"step": 1394
},
{
"epoch": 0.9862577879899254,
"grad_norm": 15.664868354797363,
"learning_rate": 2.4442013129102846e-06,
"loss": 12.1448,
"step": 1395
},
{
"epoch": 0.9869647828200256,
"grad_norm": 13.61034107208252,
"learning_rate": 2.3719912472647702e-06,
"loss": 11.5551,
"step": 1396
},
{
"epoch": 0.9876717776501259,
"grad_norm": 14.565057754516602,
"learning_rate": 2.2997811816192563e-06,
"loss": 11.7505,
"step": 1397
},
{
"epoch": 0.9883787724802262,
"grad_norm": 16.881675720214844,
"learning_rate": 2.227571115973742e-06,
"loss": 11.632,
"step": 1398
},
{
"epoch": 0.9890857673103265,
"grad_norm": 17.384876251220703,
"learning_rate": 2.1553610503282275e-06,
"loss": 12.2454,
"step": 1399
},
{
"epoch": 0.9897927621404269,
"grad_norm": 14.5512056350708,
"learning_rate": 2.0831509846827135e-06,
"loss": 12.0764,
"step": 1400
},
{
"epoch": 0.9904997569705272,
"grad_norm": 15.20163345336914,
"learning_rate": 2.0109409190371995e-06,
"loss": 11.3819,
"step": 1401
},
{
"epoch": 0.9912067518006275,
"grad_norm": 13.974905967712402,
"learning_rate": 1.9387308533916847e-06,
"loss": 12.4306,
"step": 1402
},
{
"epoch": 0.9919137466307277,
"grad_norm": 13.457085609436035,
"learning_rate": 1.8665207877461708e-06,
"loss": 11.195,
"step": 1403
},
{
"epoch": 0.992620741460828,
"grad_norm": 12.736709594726562,
"learning_rate": 1.7943107221006566e-06,
"loss": 10.1447,
"step": 1404
},
{
"epoch": 0.9933277362909284,
"grad_norm": 14.890710830688477,
"learning_rate": 1.7221006564551424e-06,
"loss": 11.5008,
"step": 1405
},
{
"epoch": 0.9940347311210287,
"grad_norm": 14.965149879455566,
"learning_rate": 1.649890590809628e-06,
"loss": 10.8517,
"step": 1406
},
{
"epoch": 0.994741725951129,
"grad_norm": 13.528691291809082,
"learning_rate": 1.5776805251641138e-06,
"loss": 11.7516,
"step": 1407
},
{
"epoch": 0.9954487207812293,
"grad_norm": 13.84748363494873,
"learning_rate": 1.5054704595185996e-06,
"loss": 11.2027,
"step": 1408
},
{
"epoch": 0.9961557156113295,
"grad_norm": 14.555986404418945,
"learning_rate": 1.4332603938730853e-06,
"loss": 12.3957,
"step": 1409
},
{
"epoch": 0.9968627104414299,
"grad_norm": 15.274754524230957,
"learning_rate": 1.361050328227571e-06,
"loss": 12.8592,
"step": 1410
},
{
"epoch": 0.9975697052715302,
"grad_norm": 15.99559497833252,
"learning_rate": 1.2888402625820569e-06,
"loss": 11.2982,
"step": 1411
},
{
"epoch": 0.9982767001016305,
"grad_norm": 14.77381420135498,
"learning_rate": 1.2166301969365427e-06,
"loss": 12.3372,
"step": 1412
},
{
"epoch": 0.9989836949317308,
"grad_norm": 13.04557991027832,
"learning_rate": 1.1444201312910285e-06,
"loss": 11.7988,
"step": 1413
},
{
"epoch": 0.9996906897618311,
"grad_norm": 14.513769149780273,
"learning_rate": 1.0722100656455141e-06,
"loss": 11.4676,
"step": 1414
},
{
"epoch": 0.9996906897618311,
"step": 1414,
"total_flos": 0.0,
"train_loss": 15.723444231160123,
"train_runtime": 27373.0059,
"train_samples_per_second": 13.228,
"train_steps_per_second": 0.052
}
],
"logging_steps": 1.0,
"max_steps": 1414,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": false,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}