rootxhacker's picture
Training in progress, step 39000, checkpoint
3035fd0 verified
{
"best_global_step": 38000,
"best_metric": 0.98431396484375,
"best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-38000",
"epoch": 2.999769248519345,
"eval_steps": 250,
"global_step": 39000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.003845858010922237,
"grad_norm": 10.111166000366211,
"learning_rate": 1.76e-05,
"loss": 13.1594,
"step": 50
},
{
"epoch": 0.007691716021844474,
"grad_norm": 5.981163501739502,
"learning_rate": 3.76e-05,
"loss": 7.7407,
"step": 100
},
{
"epoch": 0.01153757403276671,
"grad_norm": 7.169628143310547,
"learning_rate": 5.76e-05,
"loss": 6.5841,
"step": 150
},
{
"epoch": 0.015383432043688947,
"grad_norm": 9.481369972229004,
"learning_rate": 7.76e-05,
"loss": 5.8929,
"step": 200
},
{
"epoch": 0.019229290054611183,
"grad_norm": 5.348191261291504,
"learning_rate": 9.76e-05,
"loss": 5.8651,
"step": 250
},
{
"epoch": 0.019229290054611183,
"eval_loss": 5.691133975982666,
"eval_runtime": 19.1859,
"eval_samples_per_second": 52.122,
"eval_steps_per_second": 13.03,
"step": 250
},
{
"epoch": 0.02307514806553342,
"grad_norm": 9.455660820007324,
"learning_rate": 0.0001176,
"loss": 5.3783,
"step": 300
},
{
"epoch": 0.02692100607645566,
"grad_norm": 11.51925277709961,
"learning_rate": 0.00013759999999999998,
"loss": 5.0764,
"step": 350
},
{
"epoch": 0.030766864087377895,
"grad_norm": 5.797102928161621,
"learning_rate": 0.0001576,
"loss": 5.2967,
"step": 400
},
{
"epoch": 0.03461272209830013,
"grad_norm": 3.018831253051758,
"learning_rate": 0.0001776,
"loss": 5.1947,
"step": 450
},
{
"epoch": 0.038458580109222366,
"grad_norm": 5.181191921234131,
"learning_rate": 0.0001976,
"loss": 5.2159,
"step": 500
},
{
"epoch": 0.038458580109222366,
"eval_loss": 5.163902759552002,
"eval_runtime": 19.7141,
"eval_samples_per_second": 50.725,
"eval_steps_per_second": 12.681,
"step": 500
},
{
"epoch": 0.0423044381201446,
"grad_norm": 6.368561744689941,
"learning_rate": 0.0001997714463808015,
"loss": 5.0723,
"step": 550
},
{
"epoch": 0.04615029613106684,
"grad_norm": 4.666464328765869,
"learning_rate": 0.00019951172635898504,
"loss": 5.0754,
"step": 600
},
{
"epoch": 0.04999615414198908,
"grad_norm": 9.395013809204102,
"learning_rate": 0.00019925200633716855,
"loss": 4.7718,
"step": 650
},
{
"epoch": 0.05384201215291132,
"grad_norm": 4.279386520385742,
"learning_rate": 0.00019899228631535204,
"loss": 4.781,
"step": 700
},
{
"epoch": 0.05768787016383355,
"grad_norm": 4.274901866912842,
"learning_rate": 0.00019873256629353558,
"loss": 4.9529,
"step": 750
},
{
"epoch": 0.05768787016383355,
"eval_loss": 4.844481945037842,
"eval_runtime": 18.9171,
"eval_samples_per_second": 52.862,
"eval_steps_per_second": 13.216,
"step": 750
},
{
"epoch": 0.06153372817475579,
"grad_norm": 3.241445541381836,
"learning_rate": 0.0001984728462717191,
"loss": 5.0586,
"step": 800
},
{
"epoch": 0.06537958618567802,
"grad_norm": 5.880845546722412,
"learning_rate": 0.00019821312624990263,
"loss": 4.952,
"step": 850
},
{
"epoch": 0.06922544419660026,
"grad_norm": 3.81998610496521,
"learning_rate": 0.00019795340622808614,
"loss": 5.0012,
"step": 900
},
{
"epoch": 0.0730713022075225,
"grad_norm": 4.52741003036499,
"learning_rate": 0.00019769368620626966,
"loss": 4.7775,
"step": 950
},
{
"epoch": 0.07691716021844473,
"grad_norm": 27.68866729736328,
"learning_rate": 0.00019743916058488948,
"loss": 4.9852,
"step": 1000
},
{
"epoch": 0.07691716021844473,
"eval_loss": 4.836514472961426,
"eval_runtime": 19.7103,
"eval_samples_per_second": 50.735,
"eval_steps_per_second": 12.684,
"step": 1000
},
{
"epoch": 0.08076301822936698,
"grad_norm": 5.79191255569458,
"learning_rate": 0.00019717944056307302,
"loss": 4.6731,
"step": 1050
},
{
"epoch": 0.0846088762402892,
"grad_norm": 4.957877159118652,
"learning_rate": 0.00019691972054125653,
"loss": 4.7966,
"step": 1100
},
{
"epoch": 0.08845473425121145,
"grad_norm": 3.2968597412109375,
"learning_rate": 0.00019666000051944005,
"loss": 4.752,
"step": 1150
},
{
"epoch": 0.09230059226213368,
"grad_norm": 6.059363842010498,
"learning_rate": 0.00019640028049762359,
"loss": 4.7368,
"step": 1200
},
{
"epoch": 0.09614645027305592,
"grad_norm": 5.9793171882629395,
"learning_rate": 0.0001961405604758071,
"loss": 4.9613,
"step": 1250
},
{
"epoch": 0.09614645027305592,
"eval_loss": 4.76948356628418,
"eval_runtime": 18.6703,
"eval_samples_per_second": 53.561,
"eval_steps_per_second": 13.39,
"step": 1250
},
{
"epoch": 0.09999230828397816,
"grad_norm": 7.604544162750244,
"learning_rate": 0.0001958808404539906,
"loss": 4.8365,
"step": 1300
},
{
"epoch": 0.10383816629490039,
"grad_norm": 6.4756083488464355,
"learning_rate": 0.00019562112043217412,
"loss": 4.8381,
"step": 1350
},
{
"epoch": 0.10768402430582263,
"grad_norm": 3.605341672897339,
"learning_rate": 0.00019536140041035764,
"loss": 4.7004,
"step": 1400
},
{
"epoch": 0.11152988231674486,
"grad_norm": 3.4805853366851807,
"learning_rate": 0.00019510168038854115,
"loss": 4.7293,
"step": 1450
},
{
"epoch": 0.1153757403276671,
"grad_norm": 4.5123796463012695,
"learning_rate": 0.0001948419603667247,
"loss": 4.8315,
"step": 1500
},
{
"epoch": 0.1153757403276671,
"eval_loss": 4.784451961517334,
"eval_runtime": 18.5259,
"eval_samples_per_second": 53.979,
"eval_steps_per_second": 13.495,
"step": 1500
},
{
"epoch": 0.11922159833858934,
"grad_norm": 3.392902135848999,
"learning_rate": 0.0001945822403449082,
"loss": 4.859,
"step": 1550
},
{
"epoch": 0.12306745634951158,
"grad_norm": 4.783381938934326,
"learning_rate": 0.00019432252032309172,
"loss": 4.7791,
"step": 1600
},
{
"epoch": 0.12691331436043382,
"grad_norm": 3.96708345413208,
"learning_rate": 0.00019406280030127523,
"loss": 4.7743,
"step": 1650
},
{
"epoch": 0.13075917237135604,
"grad_norm": 5.381892681121826,
"learning_rate": 0.00019380308027945874,
"loss": 4.8915,
"step": 1700
},
{
"epoch": 0.13460503038227828,
"grad_norm": 3.269899606704712,
"learning_rate": 0.00019354336025764228,
"loss": 4.8823,
"step": 1750
},
{
"epoch": 0.13460503038227828,
"eval_loss": 4.689135551452637,
"eval_runtime": 18.4665,
"eval_samples_per_second": 54.152,
"eval_steps_per_second": 13.538,
"step": 1750
},
{
"epoch": 0.13845088839320052,
"grad_norm": 2.5163233280181885,
"learning_rate": 0.0001932836402358258,
"loss": 4.8513,
"step": 1800
},
{
"epoch": 0.14229674640412276,
"grad_norm": 4.244402885437012,
"learning_rate": 0.0001930239202140093,
"loss": 4.6855,
"step": 1850
},
{
"epoch": 0.146142604415045,
"grad_norm": 6.551025867462158,
"learning_rate": 0.00019276420019219285,
"loss": 4.8173,
"step": 1900
},
{
"epoch": 0.14998846242596722,
"grad_norm": 2.791435956954956,
"learning_rate": 0.00019250448017037633,
"loss": 4.6983,
"step": 1950
},
{
"epoch": 0.15383432043688947,
"grad_norm": 3.4012036323547363,
"learning_rate": 0.00019224476014855984,
"loss": 4.8178,
"step": 2000
},
{
"epoch": 0.15383432043688947,
"eval_loss": 4.666446208953857,
"eval_runtime": 18.3775,
"eval_samples_per_second": 54.414,
"eval_steps_per_second": 13.604,
"step": 2000
},
{
"epoch": 0.1576801784478117,
"grad_norm": 6.46567964553833,
"learning_rate": 0.00019198504012674338,
"loss": 4.8124,
"step": 2050
},
{
"epoch": 0.16152603645873395,
"grad_norm": 4.300732135772705,
"learning_rate": 0.0001917253201049269,
"loss": 4.7302,
"step": 2100
},
{
"epoch": 0.16537189446965617,
"grad_norm": 4.140190601348877,
"learning_rate": 0.0001914656000831104,
"loss": 4.6498,
"step": 2150
},
{
"epoch": 0.1692177524805784,
"grad_norm": 3.221662998199463,
"learning_rate": 0.00019120588006129395,
"loss": 4.6428,
"step": 2200
},
{
"epoch": 0.17306361049150065,
"grad_norm": 2.6079111099243164,
"learning_rate": 0.00019094616003947746,
"loss": 4.6115,
"step": 2250
},
{
"epoch": 0.17306361049150065,
"eval_loss": 4.661706924438477,
"eval_runtime": 18.7413,
"eval_samples_per_second": 53.358,
"eval_steps_per_second": 13.34,
"step": 2250
},
{
"epoch": 0.1769094685024229,
"grad_norm": 4.190133094787598,
"learning_rate": 0.00019068644001766095,
"loss": 4.7041,
"step": 2300
},
{
"epoch": 0.18075532651334514,
"grad_norm": 5.242035388946533,
"learning_rate": 0.0001904267199958445,
"loss": 4.6656,
"step": 2350
},
{
"epoch": 0.18460118452426735,
"grad_norm": 6.203541278839111,
"learning_rate": 0.000190166999974028,
"loss": 4.6835,
"step": 2400
},
{
"epoch": 0.1884470425351896,
"grad_norm": 2.8591034412384033,
"learning_rate": 0.00018990727995221151,
"loss": 4.6751,
"step": 2450
},
{
"epoch": 0.19229290054611184,
"grad_norm": 5.204433441162109,
"learning_rate": 0.00018964755993039505,
"loss": 4.7301,
"step": 2500
},
{
"epoch": 0.19229290054611184,
"eval_loss": 4.597048282623291,
"eval_runtime": 18.5098,
"eval_samples_per_second": 54.025,
"eval_steps_per_second": 13.506,
"step": 2500
},
{
"epoch": 0.19613875855703408,
"grad_norm": 2.59525465965271,
"learning_rate": 0.00018938783990857857,
"loss": 4.4364,
"step": 2550
},
{
"epoch": 0.19998461656795632,
"grad_norm": 3.844686985015869,
"learning_rate": 0.00018912811988676208,
"loss": 4.6437,
"step": 2600
},
{
"epoch": 0.20383047457887854,
"grad_norm": 3.4633946418762207,
"learning_rate": 0.0001888683998649456,
"loss": 4.5016,
"step": 2650
},
{
"epoch": 0.20767633258980078,
"grad_norm": 3.7852296829223633,
"learning_rate": 0.0001886086798431291,
"loss": 4.5845,
"step": 2700
},
{
"epoch": 0.21152219060072303,
"grad_norm": 3.8716065883636475,
"learning_rate": 0.00018834895982131265,
"loss": 4.3669,
"step": 2750
},
{
"epoch": 0.21152219060072303,
"eval_loss": 4.602295875549316,
"eval_runtime": 18.4747,
"eval_samples_per_second": 54.128,
"eval_steps_per_second": 13.532,
"step": 2750
},
{
"epoch": 0.21536804861164527,
"grad_norm": 3.9932167530059814,
"learning_rate": 0.00018808923979949616,
"loss": 4.4545,
"step": 2800
},
{
"epoch": 0.21921390662256748,
"grad_norm": 4.182176113128662,
"learning_rate": 0.00018782951977767967,
"loss": 4.5355,
"step": 2850
},
{
"epoch": 0.22305976463348973,
"grad_norm": 3.776895523071289,
"learning_rate": 0.00018756979975586318,
"loss": 4.6859,
"step": 2900
},
{
"epoch": 0.22690562264441197,
"grad_norm": 3.9219324588775635,
"learning_rate": 0.0001873100797340467,
"loss": 4.5186,
"step": 2950
},
{
"epoch": 0.2307514806553342,
"grad_norm": 3.7879323959350586,
"learning_rate": 0.0001870503597122302,
"loss": 4.612,
"step": 3000
},
{
"epoch": 0.2307514806553342,
"eval_loss": 4.579595565795898,
"eval_runtime": 18.476,
"eval_samples_per_second": 54.124,
"eval_steps_per_second": 13.531,
"step": 3000
},
{
"epoch": 0.23459733866625646,
"grad_norm": 6.408226490020752,
"learning_rate": 0.00018679063969041375,
"loss": 4.6614,
"step": 3050
},
{
"epoch": 0.23844319667717867,
"grad_norm": 6.529899597167969,
"learning_rate": 0.00018653091966859726,
"loss": 4.5576,
"step": 3100
},
{
"epoch": 0.2422890546881009,
"grad_norm": 4.417705059051514,
"learning_rate": 0.00018627119964678078,
"loss": 4.5466,
"step": 3150
},
{
"epoch": 0.24613491269902316,
"grad_norm": 3.052746295928955,
"learning_rate": 0.0001860114796249643,
"loss": 4.6959,
"step": 3200
},
{
"epoch": 0.2499807707099454,
"grad_norm": 5.121955871582031,
"learning_rate": 0.0001857517596031478,
"loss": 4.6796,
"step": 3250
},
{
"epoch": 0.2499807707099454,
"eval_loss": 4.576270580291748,
"eval_runtime": 18.8896,
"eval_samples_per_second": 52.939,
"eval_steps_per_second": 13.235,
"step": 3250
},
{
"epoch": 0.25382662872086764,
"grad_norm": 2.772292375564575,
"learning_rate": 0.00018549203958133131,
"loss": 4.534,
"step": 3300
},
{
"epoch": 0.2576724867317899,
"grad_norm": 6.105401039123535,
"learning_rate": 0.00018523231955951485,
"loss": 4.4958,
"step": 3350
},
{
"epoch": 0.26151834474271207,
"grad_norm": 2.742037057876587,
"learning_rate": 0.00018497259953769837,
"loss": 4.5835,
"step": 3400
},
{
"epoch": 0.2653642027536343,
"grad_norm": 3.6125235557556152,
"learning_rate": 0.0001847128795158819,
"loss": 4.5917,
"step": 3450
},
{
"epoch": 0.26921006076455656,
"grad_norm": 3.8596699237823486,
"learning_rate": 0.00018445315949406542,
"loss": 4.4909,
"step": 3500
},
{
"epoch": 0.26921006076455656,
"eval_loss": 4.520543098449707,
"eval_runtime": 18.7111,
"eval_samples_per_second": 53.444,
"eval_steps_per_second": 13.361,
"step": 3500
},
{
"epoch": 0.2730559187754788,
"grad_norm": 3.639690399169922,
"learning_rate": 0.0001841934394722489,
"loss": 4.5127,
"step": 3550
},
{
"epoch": 0.27690177678640104,
"grad_norm": 3.864473342895508,
"learning_rate": 0.00018393371945043245,
"loss": 4.599,
"step": 3600
},
{
"epoch": 0.2807476347973233,
"grad_norm": 4.662705421447754,
"learning_rate": 0.00018367399942861596,
"loss": 4.5951,
"step": 3650
},
{
"epoch": 0.28459349280824553,
"grad_norm": 3.066333532333374,
"learning_rate": 0.00018341427940679947,
"loss": 4.5583,
"step": 3700
},
{
"epoch": 0.2884393508191678,
"grad_norm": 4.054274082183838,
"learning_rate": 0.000183154559384983,
"loss": 4.5732,
"step": 3750
},
{
"epoch": 0.2884393508191678,
"eval_loss": 4.569947242736816,
"eval_runtime": 18.5242,
"eval_samples_per_second": 53.983,
"eval_steps_per_second": 13.496,
"step": 3750
},
{
"epoch": 0.29228520883009,
"grad_norm": 6.440130710601807,
"learning_rate": 0.00018289483936316652,
"loss": 4.6025,
"step": 3800
},
{
"epoch": 0.2961310668410122,
"grad_norm": 5.849060535430908,
"learning_rate": 0.00018263511934135004,
"loss": 4.4932,
"step": 3850
},
{
"epoch": 0.29997692485193445,
"grad_norm": 7.0537800788879395,
"learning_rate": 0.00018237539931953355,
"loss": 4.6986,
"step": 3900
},
{
"epoch": 0.3038227828628567,
"grad_norm": 5.134806156158447,
"learning_rate": 0.00018211567929771706,
"loss": 4.5074,
"step": 3950
},
{
"epoch": 0.30766864087377893,
"grad_norm": 2.872307777404785,
"learning_rate": 0.00018185595927590058,
"loss": 4.547,
"step": 4000
},
{
"epoch": 0.30766864087377893,
"eval_loss": 4.5459160804748535,
"eval_runtime": 18.4232,
"eval_samples_per_second": 54.279,
"eval_steps_per_second": 13.57,
"step": 4000
},
{
"epoch": 0.3115144988847012,
"grad_norm": 5.374965190887451,
"learning_rate": 0.00018159623925408412,
"loss": 4.5327,
"step": 4050
},
{
"epoch": 0.3153603568956234,
"grad_norm": 4.99652624130249,
"learning_rate": 0.00018133651923226763,
"loss": 4.3966,
"step": 4100
},
{
"epoch": 0.31920621490654566,
"grad_norm": 6.257124423980713,
"learning_rate": 0.00018107679921045114,
"loss": 4.5792,
"step": 4150
},
{
"epoch": 0.3230520729174679,
"grad_norm": 8.056533813476562,
"learning_rate": 0.00018081707918863465,
"loss": 4.6401,
"step": 4200
},
{
"epoch": 0.32689793092839015,
"grad_norm": 4.024567127227783,
"learning_rate": 0.00018055735916681817,
"loss": 4.5516,
"step": 4250
},
{
"epoch": 0.32689793092839015,
"eval_loss": 4.558110237121582,
"eval_runtime": 18.5343,
"eval_samples_per_second": 53.954,
"eval_steps_per_second": 13.488,
"step": 4250
},
{
"epoch": 0.33074378893931233,
"grad_norm": 2.4326066970825195,
"learning_rate": 0.0001802976391450017,
"loss": 4.5761,
"step": 4300
},
{
"epoch": 0.3345896469502346,
"grad_norm": 2.7615299224853516,
"learning_rate": 0.00018003791912318522,
"loss": 4.4291,
"step": 4350
},
{
"epoch": 0.3384355049611568,
"grad_norm": 3.9387362003326416,
"learning_rate": 0.00017977819910136873,
"loss": 4.5422,
"step": 4400
},
{
"epoch": 0.34228136297207906,
"grad_norm": 23.72602653503418,
"learning_rate": 0.00017951847907955227,
"loss": 4.4379,
"step": 4450
},
{
"epoch": 0.3461272209830013,
"grad_norm": 2.968930959701538,
"learning_rate": 0.0001792639534581721,
"loss": 4.433,
"step": 4500
},
{
"epoch": 0.3461272209830013,
"eval_loss": 4.5359063148498535,
"eval_runtime": 18.5422,
"eval_samples_per_second": 53.931,
"eval_steps_per_second": 13.483,
"step": 4500
},
{
"epoch": 0.34997307899392355,
"grad_norm": 6.404330730438232,
"learning_rate": 0.0001790042334363556,
"loss": 4.5673,
"step": 4550
},
{
"epoch": 0.3538189370048458,
"grad_norm": 10.212136268615723,
"learning_rate": 0.00017874451341453912,
"loss": 4.6249,
"step": 4600
},
{
"epoch": 0.35766479501576803,
"grad_norm": 4.401816368103027,
"learning_rate": 0.00017848479339272266,
"loss": 4.6305,
"step": 4650
},
{
"epoch": 0.3615106530266903,
"grad_norm": 4.710996150970459,
"learning_rate": 0.00017822507337090617,
"loss": 4.3731,
"step": 4700
},
{
"epoch": 0.36535651103761246,
"grad_norm": 3.150613307952881,
"learning_rate": 0.0001779653533490897,
"loss": 4.4491,
"step": 4750
},
{
"epoch": 0.36535651103761246,
"eval_loss": 4.564510345458984,
"eval_runtime": 18.5575,
"eval_samples_per_second": 53.886,
"eval_steps_per_second": 13.472,
"step": 4750
},
{
"epoch": 0.3692023690485347,
"grad_norm": 4.828207492828369,
"learning_rate": 0.0001777056333272732,
"loss": 4.5923,
"step": 4800
},
{
"epoch": 0.37304822705945695,
"grad_norm": 3.780848264694214,
"learning_rate": 0.0001774459133054567,
"loss": 4.5544,
"step": 4850
},
{
"epoch": 0.3768940850703792,
"grad_norm": 4.04913854598999,
"learning_rate": 0.00017718619328364023,
"loss": 4.5271,
"step": 4900
},
{
"epoch": 0.38073994308130144,
"grad_norm": 4.097137451171875,
"learning_rate": 0.00017692647326182377,
"loss": 4.4929,
"step": 4950
},
{
"epoch": 0.3845858010922237,
"grad_norm": 4.65788459777832,
"learning_rate": 0.00017666675324000728,
"loss": 4.5888,
"step": 5000
},
{
"epoch": 0.3845858010922237,
"eval_loss": 4.465761661529541,
"eval_runtime": 18.6518,
"eval_samples_per_second": 53.614,
"eval_steps_per_second": 13.404,
"step": 5000
},
{
"epoch": 0.3884316591031459,
"grad_norm": 3.0376453399658203,
"learning_rate": 0.00017640703321819082,
"loss": 4.574,
"step": 5050
},
{
"epoch": 0.39227751711406816,
"grad_norm": 2.6457693576812744,
"learning_rate": 0.00017614731319637433,
"loss": 4.5782,
"step": 5100
},
{
"epoch": 0.3961233751249904,
"grad_norm": 4.438416957855225,
"learning_rate": 0.00017588759317455782,
"loss": 4.5586,
"step": 5150
},
{
"epoch": 0.39996923313591265,
"grad_norm": 5.325882911682129,
"learning_rate": 0.00017562787315274136,
"loss": 4.5136,
"step": 5200
},
{
"epoch": 0.40381509114683484,
"grad_norm": 1.816029667854309,
"learning_rate": 0.00017536815313092487,
"loss": 4.4819,
"step": 5250
},
{
"epoch": 0.40381509114683484,
"eval_loss": 4.561609268188477,
"eval_runtime": 18.5623,
"eval_samples_per_second": 53.873,
"eval_steps_per_second": 13.468,
"step": 5250
},
{
"epoch": 0.4076609491577571,
"grad_norm": 4.011863708496094,
"learning_rate": 0.00017510843310910838,
"loss": 4.4284,
"step": 5300
},
{
"epoch": 0.4115068071686793,
"grad_norm": 10.169037818908691,
"learning_rate": 0.00017484871308729192,
"loss": 4.5414,
"step": 5350
},
{
"epoch": 0.41535266517960157,
"grad_norm": 3.9185502529144287,
"learning_rate": 0.00017458899306547544,
"loss": 4.6659,
"step": 5400
},
{
"epoch": 0.4191985231905238,
"grad_norm": 2.873530864715576,
"learning_rate": 0.00017432927304365895,
"loss": 4.4482,
"step": 5450
},
{
"epoch": 0.42304438120144605,
"grad_norm": 4.379590034484863,
"learning_rate": 0.00017406955302184246,
"loss": 4.5501,
"step": 5500
},
{
"epoch": 0.42304438120144605,
"eval_loss": 4.513929843902588,
"eval_runtime": 18.6196,
"eval_samples_per_second": 53.707,
"eval_steps_per_second": 13.427,
"step": 5500
},
{
"epoch": 0.4268902392123683,
"grad_norm": 4.2113752365112305,
"learning_rate": 0.00017380983300002597,
"loss": 4.5454,
"step": 5550
},
{
"epoch": 0.43073609722329054,
"grad_norm": 4.782048225402832,
"learning_rate": 0.0001735501129782095,
"loss": 4.4705,
"step": 5600
},
{
"epoch": 0.4345819552342128,
"grad_norm": 2.6753036975860596,
"learning_rate": 0.00017329039295639303,
"loss": 4.4592,
"step": 5650
},
{
"epoch": 0.43842781324513497,
"grad_norm": 5.012415885925293,
"learning_rate": 0.00017303067293457654,
"loss": 4.623,
"step": 5700
},
{
"epoch": 0.4422736712560572,
"grad_norm": 3.0666699409484863,
"learning_rate": 0.00017277095291276005,
"loss": 4.4325,
"step": 5750
},
{
"epoch": 0.4422736712560572,
"eval_loss": 4.489352703094482,
"eval_runtime": 18.4948,
"eval_samples_per_second": 54.069,
"eval_steps_per_second": 13.517,
"step": 5750
},
{
"epoch": 0.44611952926697945,
"grad_norm": 6.8570876121521,
"learning_rate": 0.00017251123289094357,
"loss": 4.5179,
"step": 5800
},
{
"epoch": 0.4499653872779017,
"grad_norm": 7.190755844116211,
"learning_rate": 0.00017225151286912708,
"loss": 4.5877,
"step": 5850
},
{
"epoch": 0.45381124528882394,
"grad_norm": 4.404886722564697,
"learning_rate": 0.00017199179284731062,
"loss": 4.4072,
"step": 5900
},
{
"epoch": 0.4576571032997462,
"grad_norm": 3.0543084144592285,
"learning_rate": 0.00017173207282549413,
"loss": 4.4656,
"step": 5950
},
{
"epoch": 0.4615029613106684,
"grad_norm": 6.7454514503479,
"learning_rate": 0.00017147235280367764,
"loss": 4.5688,
"step": 6000
},
{
"epoch": 0.4615029613106684,
"eval_loss": 4.480144023895264,
"eval_runtime": 18.5584,
"eval_samples_per_second": 53.884,
"eval_steps_per_second": 13.471,
"step": 6000
},
{
"epoch": 0.46534881932159067,
"grad_norm": 5.5196661949157715,
"learning_rate": 0.00017121263278186118,
"loss": 4.5703,
"step": 6050
},
{
"epoch": 0.4691946773325129,
"grad_norm": 4.253966331481934,
"learning_rate": 0.00017095291276004467,
"loss": 4.6012,
"step": 6100
},
{
"epoch": 0.4730405353434351,
"grad_norm": 2.459376096725464,
"learning_rate": 0.00017069319273822818,
"loss": 4.5002,
"step": 6150
},
{
"epoch": 0.47688639335435734,
"grad_norm": 4.933450698852539,
"learning_rate": 0.00017043347271641172,
"loss": 4.5703,
"step": 6200
},
{
"epoch": 0.4807322513652796,
"grad_norm": 4.511186599731445,
"learning_rate": 0.00017017375269459524,
"loss": 4.4665,
"step": 6250
},
{
"epoch": 0.4807322513652796,
"eval_loss": 4.485811233520508,
"eval_runtime": 18.522,
"eval_samples_per_second": 53.99,
"eval_steps_per_second": 13.497,
"step": 6250
},
{
"epoch": 0.4845781093762018,
"grad_norm": 5.634074687957764,
"learning_rate": 0.00016991403267277875,
"loss": 4.4616,
"step": 6300
},
{
"epoch": 0.48842396738712407,
"grad_norm": 3.319650650024414,
"learning_rate": 0.0001696543126509623,
"loss": 4.4836,
"step": 6350
},
{
"epoch": 0.4922698253980463,
"grad_norm": 3.306976079940796,
"learning_rate": 0.00016939459262914577,
"loss": 4.5256,
"step": 6400
},
{
"epoch": 0.49611568340896856,
"grad_norm": 4.1797308921813965,
"learning_rate": 0.0001691348726073293,
"loss": 4.3822,
"step": 6450
},
{
"epoch": 0.4999615414198908,
"grad_norm": 3.2349929809570312,
"learning_rate": 0.00016887515258551283,
"loss": 4.4384,
"step": 6500
},
{
"epoch": 0.4999615414198908,
"eval_loss": 4.485826015472412,
"eval_runtime": 18.614,
"eval_samples_per_second": 53.723,
"eval_steps_per_second": 13.431,
"step": 6500
},
{
"epoch": 0.503807399430813,
"grad_norm": 3.325056791305542,
"learning_rate": 0.00016861543256369634,
"loss": 4.618,
"step": 6550
},
{
"epoch": 0.5076532574417353,
"grad_norm": 4.026259899139404,
"learning_rate": 0.0001683609069423162,
"loss": 4.5485,
"step": 6600
},
{
"epoch": 0.5114991154526575,
"grad_norm": 3.1270413398742676,
"learning_rate": 0.00016810118692049973,
"loss": 4.4426,
"step": 6650
},
{
"epoch": 0.5153449734635798,
"grad_norm": 5.264435768127441,
"learning_rate": 0.00016784146689868324,
"loss": 4.3544,
"step": 6700
},
{
"epoch": 0.519190831474502,
"grad_norm": 4.349465847015381,
"learning_rate": 0.00016758694127730307,
"loss": 4.5912,
"step": 6750
},
{
"epoch": 0.519190831474502,
"eval_loss": 4.503914833068848,
"eval_runtime": 18.5932,
"eval_samples_per_second": 53.783,
"eval_steps_per_second": 13.446,
"step": 6750
},
{
"epoch": 0.5230366894854241,
"grad_norm": 2.5264992713928223,
"learning_rate": 0.00016732722125548658,
"loss": 4.3703,
"step": 6800
},
{
"epoch": 0.5268825474963464,
"grad_norm": 4.934820175170898,
"learning_rate": 0.00016706750123367012,
"loss": 4.5652,
"step": 6850
},
{
"epoch": 0.5307284055072686,
"grad_norm": 4.21425724029541,
"learning_rate": 0.00016680778121185363,
"loss": 4.4684,
"step": 6900
},
{
"epoch": 0.5345742635181909,
"grad_norm": 5.111146926879883,
"learning_rate": 0.00016654806119003715,
"loss": 4.5119,
"step": 6950
},
{
"epoch": 0.5384201215291131,
"grad_norm": 4.563775062561035,
"learning_rate": 0.00016628834116822069,
"loss": 4.489,
"step": 7000
},
{
"epoch": 0.5384201215291131,
"eval_loss": 4.429446220397949,
"eval_runtime": 18.6345,
"eval_samples_per_second": 53.664,
"eval_steps_per_second": 13.416,
"step": 7000
},
{
"epoch": 0.5422659795400354,
"grad_norm": 6.055607795715332,
"learning_rate": 0.00016602862114640417,
"loss": 4.3805,
"step": 7050
},
{
"epoch": 0.5461118375509576,
"grad_norm": 3.190605878829956,
"learning_rate": 0.00016576890112458768,
"loss": 4.561,
"step": 7100
},
{
"epoch": 0.5499576955618799,
"grad_norm": 2.95857834815979,
"learning_rate": 0.00016550918110277122,
"loss": 4.5957,
"step": 7150
},
{
"epoch": 0.5538035535728021,
"grad_norm": 5.055838108062744,
"learning_rate": 0.00016524946108095474,
"loss": 4.3916,
"step": 7200
},
{
"epoch": 0.5576494115837243,
"grad_norm": 6.25083589553833,
"learning_rate": 0.00016498974105913825,
"loss": 4.6649,
"step": 7250
},
{
"epoch": 0.5576494115837243,
"eval_loss": 4.42822265625,
"eval_runtime": 18.5876,
"eval_samples_per_second": 53.799,
"eval_steps_per_second": 13.45,
"step": 7250
},
{
"epoch": 0.5614952695946466,
"grad_norm": 3.6657636165618896,
"learning_rate": 0.0001647300210373218,
"loss": 4.5059,
"step": 7300
},
{
"epoch": 0.5653411276055688,
"grad_norm": 3.6100645065307617,
"learning_rate": 0.0001644703010155053,
"loss": 4.4379,
"step": 7350
},
{
"epoch": 0.5691869856164911,
"grad_norm": 3.535804271697998,
"learning_rate": 0.0001642105809936888,
"loss": 4.3881,
"step": 7400
},
{
"epoch": 0.5730328436274132,
"grad_norm": 2.9636013507843018,
"learning_rate": 0.00016395086097187233,
"loss": 4.491,
"step": 7450
},
{
"epoch": 0.5768787016383355,
"grad_norm": 2.9023678302764893,
"learning_rate": 0.00016369114095005584,
"loss": 4.5567,
"step": 7500
},
{
"epoch": 0.5768787016383355,
"eval_loss": 4.459234237670898,
"eval_runtime": 18.5868,
"eval_samples_per_second": 53.802,
"eval_steps_per_second": 13.45,
"step": 7500
},
{
"epoch": 0.5807245596492577,
"grad_norm": 4.381130218505859,
"learning_rate": 0.00016343142092823938,
"loss": 4.521,
"step": 7550
},
{
"epoch": 0.58457041766018,
"grad_norm": 2.533957004547119,
"learning_rate": 0.0001631717009064229,
"loss": 4.3233,
"step": 7600
},
{
"epoch": 0.5884162756711022,
"grad_norm": 3.277646541595459,
"learning_rate": 0.0001629119808846064,
"loss": 4.5503,
"step": 7650
},
{
"epoch": 0.5922621336820244,
"grad_norm": 3.871952772140503,
"learning_rate": 0.00016265226086278992,
"loss": 4.4196,
"step": 7700
},
{
"epoch": 0.5961079916929467,
"grad_norm": 3.191589832305908,
"learning_rate": 0.00016239254084097343,
"loss": 4.479,
"step": 7750
},
{
"epoch": 0.5961079916929467,
"eval_loss": 4.44810152053833,
"eval_runtime": 18.425,
"eval_samples_per_second": 54.274,
"eval_steps_per_second": 13.569,
"step": 7750
},
{
"epoch": 0.5999538497038689,
"grad_norm": 3.7022933959960938,
"learning_rate": 0.00016213282081915695,
"loss": 4.5402,
"step": 7800
},
{
"epoch": 0.6037997077147912,
"grad_norm": 2.883859395980835,
"learning_rate": 0.00016187310079734049,
"loss": 4.5457,
"step": 7850
},
{
"epoch": 0.6076455657257134,
"grad_norm": 3.5229415893554688,
"learning_rate": 0.000161613380775524,
"loss": 4.5185,
"step": 7900
},
{
"epoch": 0.6114914237366357,
"grad_norm": 4.529599666595459,
"learning_rate": 0.0001613536607537075,
"loss": 4.4489,
"step": 7950
},
{
"epoch": 0.6153372817475579,
"grad_norm": 5.1023850440979,
"learning_rate": 0.00016109394073189102,
"loss": 4.6043,
"step": 8000
},
{
"epoch": 0.6153372817475579,
"eval_loss": 4.478011608123779,
"eval_runtime": 18.576,
"eval_samples_per_second": 53.833,
"eval_steps_per_second": 13.458,
"step": 8000
},
{
"epoch": 0.6191831397584802,
"grad_norm": 4.5992255210876465,
"learning_rate": 0.00016083422071007454,
"loss": 4.3146,
"step": 8050
},
{
"epoch": 0.6230289977694023,
"grad_norm": 5.412031650543213,
"learning_rate": 0.00016057450068825805,
"loss": 4.5841,
"step": 8100
},
{
"epoch": 0.6268748557803245,
"grad_norm": 4.4531779289245605,
"learning_rate": 0.0001603147806664416,
"loss": 4.5084,
"step": 8150
},
{
"epoch": 0.6307207137912468,
"grad_norm": 3.913174867630005,
"learning_rate": 0.0001600550606446251,
"loss": 4.4183,
"step": 8200
},
{
"epoch": 0.634566571802169,
"grad_norm": 3.3952407836914062,
"learning_rate": 0.00015979534062280862,
"loss": 4.3496,
"step": 8250
},
{
"epoch": 0.634566571802169,
"eval_loss": 4.46174430847168,
"eval_runtime": 18.5381,
"eval_samples_per_second": 53.943,
"eval_steps_per_second": 13.486,
"step": 8250
},
{
"epoch": 0.6384124298130913,
"grad_norm": 4.156221389770508,
"learning_rate": 0.00015953562060099213,
"loss": 4.361,
"step": 8300
},
{
"epoch": 0.6422582878240135,
"grad_norm": 3.2021920680999756,
"learning_rate": 0.00015927590057917564,
"loss": 4.473,
"step": 8350
},
{
"epoch": 0.6461041458349358,
"grad_norm": 5.048036575317383,
"learning_rate": 0.00015901618055735918,
"loss": 4.4142,
"step": 8400
},
{
"epoch": 0.649950003845858,
"grad_norm": 3.91768217086792,
"learning_rate": 0.0001587564605355427,
"loss": 4.4672,
"step": 8450
},
{
"epoch": 0.6537958618567803,
"grad_norm": 9.229452133178711,
"learning_rate": 0.0001584967405137262,
"loss": 4.7837,
"step": 8500
},
{
"epoch": 0.6537958618567803,
"eval_loss": 4.409055709838867,
"eval_runtime": 18.5943,
"eval_samples_per_second": 53.78,
"eval_steps_per_second": 13.445,
"step": 8500
},
{
"epoch": 0.6576417198677025,
"grad_norm": 2.7313661575317383,
"learning_rate": 0.00015823702049190975,
"loss": 4.4721,
"step": 8550
},
{
"epoch": 0.6614875778786247,
"grad_norm": 4.160475730895996,
"learning_rate": 0.00015797730047009326,
"loss": 4.501,
"step": 8600
},
{
"epoch": 0.665333435889547,
"grad_norm": 11.54045581817627,
"learning_rate": 0.00015771758044827675,
"loss": 4.4433,
"step": 8650
},
{
"epoch": 0.6691792939004692,
"grad_norm": 4.087617874145508,
"learning_rate": 0.00015745786042646029,
"loss": 4.4981,
"step": 8700
},
{
"epoch": 0.6730251519113915,
"grad_norm": 4.155121803283691,
"learning_rate": 0.0001571981404046438,
"loss": 4.3874,
"step": 8750
},
{
"epoch": 0.6730251519113915,
"eval_loss": 4.418811321258545,
"eval_runtime": 18.6306,
"eval_samples_per_second": 53.675,
"eval_steps_per_second": 13.419,
"step": 8750
},
{
"epoch": 0.6768710099223136,
"grad_norm": 4.071916580200195,
"learning_rate": 0.0001569384203828273,
"loss": 4.5531,
"step": 8800
},
{
"epoch": 0.6807168679332359,
"grad_norm": 3.395460605621338,
"learning_rate": 0.00015667870036101085,
"loss": 4.4609,
"step": 8850
},
{
"epoch": 0.6845627259441581,
"grad_norm": 3.4933230876922607,
"learning_rate": 0.00015641898033919436,
"loss": 4.4536,
"step": 8900
},
{
"epoch": 0.6884085839550804,
"grad_norm": 6.921072483062744,
"learning_rate": 0.00015615926031737788,
"loss": 4.3478,
"step": 8950
},
{
"epoch": 0.6922544419660026,
"grad_norm": 3.920626401901245,
"learning_rate": 0.0001558995402955614,
"loss": 4.3761,
"step": 9000
},
{
"epoch": 0.6922544419660026,
"eval_loss": 4.415992259979248,
"eval_runtime": 18.5147,
"eval_samples_per_second": 54.011,
"eval_steps_per_second": 13.503,
"step": 9000
},
{
"epoch": 0.6961002999769248,
"grad_norm": 7.213745594024658,
"learning_rate": 0.0001556398202737449,
"loss": 4.323,
"step": 9050
},
{
"epoch": 0.6999461579878471,
"grad_norm": 3.2426984310150146,
"learning_rate": 0.00015538010025192842,
"loss": 4.3922,
"step": 9100
},
{
"epoch": 0.7037920159987693,
"grad_norm": 3.256950855255127,
"learning_rate": 0.00015512038023011196,
"loss": 4.2602,
"step": 9150
},
{
"epoch": 0.7076378740096916,
"grad_norm": 6.132264614105225,
"learning_rate": 0.00015486066020829547,
"loss": 4.3734,
"step": 9200
},
{
"epoch": 0.7114837320206138,
"grad_norm": 3.921595573425293,
"learning_rate": 0.00015460094018647898,
"loss": 4.4776,
"step": 9250
},
{
"epoch": 0.7114837320206138,
"eval_loss": 4.3921356201171875,
"eval_runtime": 18.5511,
"eval_samples_per_second": 53.905,
"eval_steps_per_second": 13.476,
"step": 9250
},
{
"epoch": 0.7153295900315361,
"grad_norm": 5.416064739227295,
"learning_rate": 0.0001543412201646625,
"loss": 4.5659,
"step": 9300
},
{
"epoch": 0.7191754480424583,
"grad_norm": 4.542217254638672,
"learning_rate": 0.000154081500142846,
"loss": 4.2557,
"step": 9350
},
{
"epoch": 0.7230213060533806,
"grad_norm": 3.7075681686401367,
"learning_rate": 0.00015382178012102955,
"loss": 4.453,
"step": 9400
},
{
"epoch": 0.7268671640643027,
"grad_norm": 4.457496166229248,
"learning_rate": 0.00015356206009921306,
"loss": 4.3861,
"step": 9450
},
{
"epoch": 0.7307130220752249,
"grad_norm": 1.7784981727600098,
"learning_rate": 0.00015330234007739657,
"loss": 4.3258,
"step": 9500
},
{
"epoch": 0.7307130220752249,
"eval_loss": 4.357193470001221,
"eval_runtime": 18.6372,
"eval_samples_per_second": 53.656,
"eval_steps_per_second": 13.414,
"step": 9500
},
{
"epoch": 0.7345588800861472,
"grad_norm": 3.8532371520996094,
"learning_rate": 0.0001530426200555801,
"loss": 4.3473,
"step": 9550
},
{
"epoch": 0.7384047380970694,
"grad_norm": 4.654659271240234,
"learning_rate": 0.0001527829000337636,
"loss": 4.457,
"step": 9600
},
{
"epoch": 0.7422505961079917,
"grad_norm": 2.420182228088379,
"learning_rate": 0.0001525231800119471,
"loss": 4.4521,
"step": 9650
},
{
"epoch": 0.7460964541189139,
"grad_norm": 4.189414978027344,
"learning_rate": 0.00015226345999013065,
"loss": 4.2569,
"step": 9700
},
{
"epoch": 0.7499423121298362,
"grad_norm": 2.824084997177124,
"learning_rate": 0.00015200373996831416,
"loss": 4.3409,
"step": 9750
},
{
"epoch": 0.7499423121298362,
"eval_loss": 4.378731727600098,
"eval_runtime": 18.4857,
"eval_samples_per_second": 54.096,
"eval_steps_per_second": 13.524,
"step": 9750
},
{
"epoch": 0.7537881701407584,
"grad_norm": 6.379781723022461,
"learning_rate": 0.00015174401994649768,
"loss": 4.3041,
"step": 9800
},
{
"epoch": 0.7576340281516807,
"grad_norm": 1.7334113121032715,
"learning_rate": 0.00015148429992468122,
"loss": 4.3096,
"step": 9850
},
{
"epoch": 0.7614798861626029,
"grad_norm": 4.287415027618408,
"learning_rate": 0.00015122457990286473,
"loss": 4.4411,
"step": 9900
},
{
"epoch": 0.7653257441735252,
"grad_norm": 3.3184821605682373,
"learning_rate": 0.00015096485988104821,
"loss": 4.3992,
"step": 9950
},
{
"epoch": 0.7691716021844474,
"grad_norm": 4.698968887329102,
"learning_rate": 0.00015070513985923175,
"loss": 4.4726,
"step": 10000
},
{
"epoch": 0.7691716021844474,
"eval_loss": 4.408615589141846,
"eval_runtime": 18.5408,
"eval_samples_per_second": 53.935,
"eval_steps_per_second": 13.484,
"step": 10000
},
{
"epoch": 0.7730174601953695,
"grad_norm": 3.882775068283081,
"learning_rate": 0.00015044541983741527,
"loss": 4.5207,
"step": 10050
},
{
"epoch": 0.7768633182062918,
"grad_norm": 5.814795017242432,
"learning_rate": 0.0001501856998155988,
"loss": 4.2462,
"step": 10100
},
{
"epoch": 0.780709176217214,
"grad_norm": 4.733581066131592,
"learning_rate": 0.00014992597979378232,
"loss": 4.5563,
"step": 10150
},
{
"epoch": 0.7845550342281363,
"grad_norm": 4.805403232574463,
"learning_rate": 0.00014966625977196583,
"loss": 4.4353,
"step": 10200
},
{
"epoch": 0.7884008922390585,
"grad_norm": 5.814332008361816,
"learning_rate": 0.00014940653975014935,
"loss": 4.4004,
"step": 10250
},
{
"epoch": 0.7884008922390585,
"eval_loss": 4.41144323348999,
"eval_runtime": 18.5402,
"eval_samples_per_second": 53.937,
"eval_steps_per_second": 13.484,
"step": 10250
},
{
"epoch": 0.7922467502499808,
"grad_norm": 5.321393013000488,
"learning_rate": 0.00014914681972833286,
"loss": 4.4383,
"step": 10300
},
{
"epoch": 0.796092608260903,
"grad_norm": 3.681452751159668,
"learning_rate": 0.00014888709970651637,
"loss": 4.5094,
"step": 10350
},
{
"epoch": 0.7999384662718253,
"grad_norm": 4.766401767730713,
"learning_rate": 0.0001486273796846999,
"loss": 4.3104,
"step": 10400
},
{
"epoch": 0.8037843242827475,
"grad_norm": 4.676774024963379,
"learning_rate": 0.00014836765966288342,
"loss": 4.516,
"step": 10450
},
{
"epoch": 0.8076301822936697,
"grad_norm": 3.623643159866333,
"learning_rate": 0.00014810793964106694,
"loss": 4.5146,
"step": 10500
},
{
"epoch": 0.8076301822936697,
"eval_loss": 4.398375034332275,
"eval_runtime": 18.4969,
"eval_samples_per_second": 54.063,
"eval_steps_per_second": 13.516,
"step": 10500
},
{
"epoch": 0.811476040304592,
"grad_norm": 4.379317760467529,
"learning_rate": 0.00014784821961925045,
"loss": 4.6715,
"step": 10550
},
{
"epoch": 0.8153218983155142,
"grad_norm": 3.034796714782715,
"learning_rate": 0.00014758849959743396,
"loss": 4.4511,
"step": 10600
},
{
"epoch": 0.8191677563264365,
"grad_norm": 3.8016927242279053,
"learning_rate": 0.00014732877957561748,
"loss": 4.2966,
"step": 10650
},
{
"epoch": 0.8230136143373586,
"grad_norm": 4.330080509185791,
"learning_rate": 0.00014706905955380102,
"loss": 4.6496,
"step": 10700
},
{
"epoch": 0.8268594723482809,
"grad_norm": 8.032389640808105,
"learning_rate": 0.00014680933953198453,
"loss": 4.2758,
"step": 10750
},
{
"epoch": 0.8268594723482809,
"eval_loss": 4.388455867767334,
"eval_runtime": 18.481,
"eval_samples_per_second": 54.11,
"eval_steps_per_second": 13.527,
"step": 10750
},
{
"epoch": 0.8307053303592031,
"grad_norm": 3.1724319458007812,
"learning_rate": 0.00014654961951016807,
"loss": 4.405,
"step": 10800
},
{
"epoch": 0.8345511883701254,
"grad_norm": 3.529196262359619,
"learning_rate": 0.00014628989948835155,
"loss": 4.4433,
"step": 10850
},
{
"epoch": 0.8383970463810476,
"grad_norm": 4.758362293243408,
"learning_rate": 0.00014603017946653507,
"loss": 4.5747,
"step": 10900
},
{
"epoch": 0.8422429043919698,
"grad_norm": 3.524068832397461,
"learning_rate": 0.0001457704594447186,
"loss": 4.3469,
"step": 10950
},
{
"epoch": 0.8460887624028921,
"grad_norm": 4.452401161193848,
"learning_rate": 0.00014551073942290212,
"loss": 4.3767,
"step": 11000
},
{
"epoch": 0.8460887624028921,
"eval_loss": 4.352676868438721,
"eval_runtime": 18.5817,
"eval_samples_per_second": 53.816,
"eval_steps_per_second": 13.454,
"step": 11000
},
{
"epoch": 0.8499346204138143,
"grad_norm": 6.128251075744629,
"learning_rate": 0.00014525101940108563,
"loss": 4.3993,
"step": 11050
},
{
"epoch": 0.8537804784247366,
"grad_norm": 3.9961323738098145,
"learning_rate": 0.00014499129937926917,
"loss": 4.406,
"step": 11100
},
{
"epoch": 0.8576263364356588,
"grad_norm": 3.889711856842041,
"learning_rate": 0.00014473157935745269,
"loss": 4.2991,
"step": 11150
},
{
"epoch": 0.8614721944465811,
"grad_norm": 3.5852463245391846,
"learning_rate": 0.00014447185933563617,
"loss": 4.2967,
"step": 11200
},
{
"epoch": 0.8653180524575033,
"grad_norm": 3.343247652053833,
"learning_rate": 0.0001442121393138197,
"loss": 4.314,
"step": 11250
},
{
"epoch": 0.8653180524575033,
"eval_loss": 4.387504577636719,
"eval_runtime": 18.554,
"eval_samples_per_second": 53.897,
"eval_steps_per_second": 13.474,
"step": 11250
},
{
"epoch": 0.8691639104684256,
"grad_norm": 3.5600407123565674,
"learning_rate": 0.00014395241929200322,
"loss": 4.2958,
"step": 11300
},
{
"epoch": 0.8730097684793477,
"grad_norm": 4.299932956695557,
"learning_rate": 0.00014369269927018674,
"loss": 4.2966,
"step": 11350
},
{
"epoch": 0.8768556264902699,
"grad_norm": 2.5763466358184814,
"learning_rate": 0.00014343297924837028,
"loss": 4.3739,
"step": 11400
},
{
"epoch": 0.8807014845011922,
"grad_norm": 3.670653820037842,
"learning_rate": 0.0001431732592265538,
"loss": 4.3932,
"step": 11450
},
{
"epoch": 0.8845473425121144,
"grad_norm": 5.472078800201416,
"learning_rate": 0.0001429135392047373,
"loss": 4.3183,
"step": 11500
},
{
"epoch": 0.8845473425121144,
"eval_loss": 4.377117156982422,
"eval_runtime": 18.5574,
"eval_samples_per_second": 53.887,
"eval_steps_per_second": 13.472,
"step": 11500
},
{
"epoch": 0.8883932005230367,
"grad_norm": 4.711415767669678,
"learning_rate": 0.00014265381918292082,
"loss": 4.5701,
"step": 11550
},
{
"epoch": 0.8922390585339589,
"grad_norm": 3.1737523078918457,
"learning_rate": 0.00014239409916110433,
"loss": 4.4522,
"step": 11600
},
{
"epoch": 0.8960849165448812,
"grad_norm": 4.876018047332764,
"learning_rate": 0.00014213437913928787,
"loss": 4.3937,
"step": 11650
},
{
"epoch": 0.8999307745558034,
"grad_norm": 7.117967128753662,
"learning_rate": 0.00014187465911747138,
"loss": 4.3585,
"step": 11700
},
{
"epoch": 0.9037766325667257,
"grad_norm": 2.587160587310791,
"learning_rate": 0.0001416149390956549,
"loss": 1.5054,
"step": 11750
},
{
"epoch": 0.9037766325667257,
"eval_loss": 1.5468424558639526,
"eval_runtime": 17.9563,
"eval_samples_per_second": 55.691,
"eval_steps_per_second": 13.923,
"step": 11750
},
{
"epoch": 0.9076224905776479,
"grad_norm": 1.4935526847839355,
"learning_rate": 0.0001413552190738384,
"loss": 1.5719,
"step": 11800
},
{
"epoch": 0.9114683485885701,
"grad_norm": 1.4879201650619507,
"learning_rate": 0.00014109549905202192,
"loss": 1.4862,
"step": 11850
},
{
"epoch": 0.9153142065994924,
"grad_norm": 1.7936193943023682,
"learning_rate": 0.00014083577903020543,
"loss": 1.488,
"step": 11900
},
{
"epoch": 0.9191600646104146,
"grad_norm": 2.14953875541687,
"learning_rate": 0.00014057605900838897,
"loss": 1.5383,
"step": 11950
},
{
"epoch": 0.9230059226213368,
"grad_norm": 1.5575013160705566,
"learning_rate": 0.00014031633898657249,
"loss": 1.4814,
"step": 12000
},
{
"epoch": 0.9230059226213368,
"eval_loss": 1.5273067951202393,
"eval_runtime": 17.7259,
"eval_samples_per_second": 56.414,
"eval_steps_per_second": 14.104,
"step": 12000
},
{
"epoch": 0.926851780632259,
"grad_norm": 1.9142653942108154,
"learning_rate": 0.000140056618964756,
"loss": 1.5436,
"step": 12050
},
{
"epoch": 0.9306976386431813,
"grad_norm": 1.9730989933013916,
"learning_rate": 0.00013979689894293954,
"loss": 1.5175,
"step": 12100
},
{
"epoch": 0.9345434966541035,
"grad_norm": 0.9956797361373901,
"learning_rate": 0.00013953717892112302,
"loss": 1.4491,
"step": 12150
},
{
"epoch": 0.9383893546650258,
"grad_norm": 1.0358608961105347,
"learning_rate": 0.00013927745889930654,
"loss": 1.4488,
"step": 12200
},
{
"epoch": 0.942235212675948,
"grad_norm": 1.5184404850006104,
"learning_rate": 0.00013901773887749008,
"loss": 1.5037,
"step": 12250
},
{
"epoch": 0.942235212675948,
"eval_loss": 1.5220181941986084,
"eval_runtime": 17.7647,
"eval_samples_per_second": 56.291,
"eval_steps_per_second": 14.073,
"step": 12250
},
{
"epoch": 0.9460810706868702,
"grad_norm": 1.3881593942642212,
"learning_rate": 0.0001387580188556736,
"loss": 1.4952,
"step": 12300
},
{
"epoch": 0.9499269286977925,
"grad_norm": 2.319173574447632,
"learning_rate": 0.0001384982988338571,
"loss": 1.4667,
"step": 12350
},
{
"epoch": 0.9537727867087147,
"grad_norm": 1.9042879343032837,
"learning_rate": 0.00013823857881204064,
"loss": 1.6038,
"step": 12400
},
{
"epoch": 0.957618644719637,
"grad_norm": 1.9162698984146118,
"learning_rate": 0.00013797885879022415,
"loss": 1.4904,
"step": 12450
},
{
"epoch": 0.9614645027305592,
"grad_norm": 2.0601863861083984,
"learning_rate": 0.00013771913876840767,
"loss": 1.4758,
"step": 12500
},
{
"epoch": 0.9614645027305592,
"eval_loss": 1.5356587171554565,
"eval_runtime": 17.7404,
"eval_samples_per_second": 56.369,
"eval_steps_per_second": 14.092,
"step": 12500
},
{
"epoch": 0.9653103607414815,
"grad_norm": 1.888836145401001,
"learning_rate": 0.00013745941874659118,
"loss": 1.5681,
"step": 12550
},
{
"epoch": 0.9691562187524037,
"grad_norm": 1.4329860210418701,
"learning_rate": 0.0001371996987247747,
"loss": 1.5018,
"step": 12600
},
{
"epoch": 0.973002076763326,
"grad_norm": 1.969533920288086,
"learning_rate": 0.00013693997870295823,
"loss": 1.5494,
"step": 12650
},
{
"epoch": 0.9768479347742481,
"grad_norm": 2.1219890117645264,
"learning_rate": 0.00013668025868114175,
"loss": 1.5386,
"step": 12700
},
{
"epoch": 0.9806937927851703,
"grad_norm": 1.6632941961288452,
"learning_rate": 0.00013642053865932526,
"loss": 1.5253,
"step": 12750
},
{
"epoch": 0.9806937927851703,
"eval_loss": 1.5147372484207153,
"eval_runtime": 17.8072,
"eval_samples_per_second": 56.157,
"eval_steps_per_second": 14.039,
"step": 12750
},
{
"epoch": 0.9845396507960926,
"grad_norm": 1.292913794517517,
"learning_rate": 0.00013616081863750877,
"loss": 1.4413,
"step": 12800
},
{
"epoch": 0.9883855088070148,
"grad_norm": 1.1377824544906616,
"learning_rate": 0.00013590109861569228,
"loss": 1.5705,
"step": 12850
},
{
"epoch": 0.9922313668179371,
"grad_norm": 1.656996726989746,
"learning_rate": 0.0001356413785938758,
"loss": 1.5311,
"step": 12900
},
{
"epoch": 0.9960772248288593,
"grad_norm": 1.6639357805252075,
"learning_rate": 0.00013538165857205934,
"loss": 1.5676,
"step": 12950
},
{
"epoch": 0.9999230828397816,
"grad_norm": 1.0893466472625732,
"learning_rate": 0.00013512193855024285,
"loss": 1.5215,
"step": 13000
},
{
"epoch": 0.9999230828397816,
"eval_loss": 1.5119102001190186,
"eval_runtime": 17.7291,
"eval_samples_per_second": 56.404,
"eval_steps_per_second": 14.101,
"step": 13000
},
{
"epoch": 1.0037689408507038,
"grad_norm": 1.4789248704910278,
"learning_rate": 0.00013486221852842636,
"loss": 1.4551,
"step": 13050
},
{
"epoch": 1.007614798861626,
"grad_norm": 1.3270663022994995,
"learning_rate": 0.00013460249850660988,
"loss": 1.4776,
"step": 13100
},
{
"epoch": 1.0114606568725482,
"grad_norm": 1.3546854257583618,
"learning_rate": 0.0001343427784847934,
"loss": 1.5807,
"step": 13150
},
{
"epoch": 1.0153065148834706,
"grad_norm": 1.303915023803711,
"learning_rate": 0.0001340830584629769,
"loss": 1.4309,
"step": 13200
},
{
"epoch": 1.0191523728943928,
"grad_norm": 0.8854748606681824,
"learning_rate": 0.00013382333844116044,
"loss": 1.4395,
"step": 13250
},
{
"epoch": 1.0191523728943928,
"eval_loss": 1.5238608121871948,
"eval_runtime": 17.7588,
"eval_samples_per_second": 56.31,
"eval_steps_per_second": 14.077,
"step": 13250
},
{
"epoch": 1.022998230905315,
"grad_norm": 1.5649653673171997,
"learning_rate": 0.00013356361841934395,
"loss": 1.5022,
"step": 13300
},
{
"epoch": 1.0268440889162371,
"grad_norm": 1.6031616926193237,
"learning_rate": 0.0001333038983975275,
"loss": 1.4315,
"step": 13350
},
{
"epoch": 1.0306899469271595,
"grad_norm": 1.3788844347000122,
"learning_rate": 0.00013304417837571098,
"loss": 1.578,
"step": 13400
},
{
"epoch": 1.0345358049380817,
"grad_norm": 1.4347171783447266,
"learning_rate": 0.0001327844583538945,
"loss": 1.3993,
"step": 13450
},
{
"epoch": 1.038381662949004,
"grad_norm": 1.9777193069458008,
"learning_rate": 0.00013252473833207803,
"loss": 1.4331,
"step": 13500
},
{
"epoch": 1.038381662949004,
"eval_loss": 1.5231057405471802,
"eval_runtime": 17.8114,
"eval_samples_per_second": 56.144,
"eval_steps_per_second": 14.036,
"step": 13500
},
{
"epoch": 1.042227520959926,
"grad_norm": 2.056574583053589,
"learning_rate": 0.00013226501831026155,
"loss": 1.5387,
"step": 13550
},
{
"epoch": 1.0460733789708483,
"grad_norm": 1.41805899143219,
"learning_rate": 0.00013200529828844506,
"loss": 1.4194,
"step": 13600
},
{
"epoch": 1.0499192369817707,
"grad_norm": 1.5727626085281372,
"learning_rate": 0.0001317455782666286,
"loss": 1.4763,
"step": 13650
},
{
"epoch": 1.0537650949926929,
"grad_norm": 1.8175796270370483,
"learning_rate": 0.0001314858582448121,
"loss": 1.5232,
"step": 13700
},
{
"epoch": 1.057610953003615,
"grad_norm": 1.459721565246582,
"learning_rate": 0.0001312261382229956,
"loss": 1.4926,
"step": 13750
},
{
"epoch": 1.057610953003615,
"eval_loss": 1.5073590278625488,
"eval_runtime": 17.8208,
"eval_samples_per_second": 56.114,
"eval_steps_per_second": 14.029,
"step": 13750
},
{
"epoch": 1.0614568110145373,
"grad_norm": 1.7236889600753784,
"learning_rate": 0.00013096641820117914,
"loss": 1.4485,
"step": 13800
},
{
"epoch": 1.0653026690254597,
"grad_norm": 1.1652172803878784,
"learning_rate": 0.00013070669817936265,
"loss": 1.4706,
"step": 13850
},
{
"epoch": 1.0691485270363819,
"grad_norm": 1.1279985904693604,
"learning_rate": 0.00013044697815754616,
"loss": 1.5507,
"step": 13900
},
{
"epoch": 1.072994385047304,
"grad_norm": 2.2368061542510986,
"learning_rate": 0.0001301872581357297,
"loss": 1.5184,
"step": 13950
},
{
"epoch": 1.0768402430582262,
"grad_norm": 1.1515541076660156,
"learning_rate": 0.00012992753811391322,
"loss": 1.5184,
"step": 14000
},
{
"epoch": 1.0768402430582262,
"eval_loss": 1.5123000144958496,
"eval_runtime": 17.8325,
"eval_samples_per_second": 56.077,
"eval_steps_per_second": 14.019,
"step": 14000
},
{
"epoch": 1.0806861010691486,
"grad_norm": 2.1172475814819336,
"learning_rate": 0.00012966781809209673,
"loss": 1.4055,
"step": 14050
},
{
"epoch": 1.0845319590800708,
"grad_norm": 1.196999430656433,
"learning_rate": 0.00012940809807028024,
"loss": 1.464,
"step": 14100
},
{
"epoch": 1.088377817090993,
"grad_norm": 1.3582040071487427,
"learning_rate": 0.00012914837804846375,
"loss": 1.4502,
"step": 14150
},
{
"epoch": 1.0922236751019152,
"grad_norm": 1.6588162183761597,
"learning_rate": 0.0001288886580266473,
"loss": 1.5174,
"step": 14200
},
{
"epoch": 1.0960695331128374,
"grad_norm": 1.7531650066375732,
"learning_rate": 0.0001286289380048308,
"loss": 1.505,
"step": 14250
},
{
"epoch": 1.0960695331128374,
"eval_loss": 1.5160688161849976,
"eval_runtime": 17.764,
"eval_samples_per_second": 56.294,
"eval_steps_per_second": 14.073,
"step": 14250
},
{
"epoch": 1.0999153911237598,
"grad_norm": 1.868784785270691,
"learning_rate": 0.00012836921798301432,
"loss": 1.5544,
"step": 14300
},
{
"epoch": 1.103761249134682,
"grad_norm": 1.9493080377578735,
"learning_rate": 0.00012810949796119783,
"loss": 1.533,
"step": 14350
},
{
"epoch": 1.1076071071456042,
"grad_norm": 0.7309526801109314,
"learning_rate": 0.00012784977793938135,
"loss": 1.4672,
"step": 14400
},
{
"epoch": 1.1114529651565264,
"grad_norm": 1.3281447887420654,
"learning_rate": 0.00012759005791756486,
"loss": 1.3874,
"step": 14450
},
{
"epoch": 1.1152988231674485,
"grad_norm": 1.0158611536026,
"learning_rate": 0.0001273303378957484,
"loss": 1.4966,
"step": 14500
},
{
"epoch": 1.1152988231674485,
"eval_loss": 1.5160739421844482,
"eval_runtime": 17.6672,
"eval_samples_per_second": 56.602,
"eval_steps_per_second": 14.151,
"step": 14500
},
{
"epoch": 1.119144681178371,
"grad_norm": 1.6422228813171387,
"learning_rate": 0.0001270706178739319,
"loss": 1.5494,
"step": 14550
},
{
"epoch": 1.1229905391892931,
"grad_norm": 0.7187716960906982,
"learning_rate": 0.00012681089785211542,
"loss": 1.4221,
"step": 14600
},
{
"epoch": 1.1268363972002153,
"grad_norm": 1.2605098485946655,
"learning_rate": 0.00012655117783029896,
"loss": 1.4299,
"step": 14650
},
{
"epoch": 1.1306822552111375,
"grad_norm": 2.598015069961548,
"learning_rate": 0.00012629145780848245,
"loss": 1.5757,
"step": 14700
},
{
"epoch": 1.13452811322206,
"grad_norm": 1.4004614353179932,
"learning_rate": 0.00012603173778666596,
"loss": 1.482,
"step": 14750
},
{
"epoch": 1.13452811322206,
"eval_loss": 1.5089725255966187,
"eval_runtime": 17.9036,
"eval_samples_per_second": 55.855,
"eval_steps_per_second": 13.964,
"step": 14750
},
{
"epoch": 1.1383739712329821,
"grad_norm": 1.3800735473632812,
"learning_rate": 0.0001257720177648495,
"loss": 1.5285,
"step": 14800
},
{
"epoch": 1.1422198292439043,
"grad_norm": 1.3741459846496582,
"learning_rate": 0.00012551229774303301,
"loss": 1.5242,
"step": 14850
},
{
"epoch": 1.1460656872548265,
"grad_norm": 2.232680559158325,
"learning_rate": 0.00012525257772121653,
"loss": 1.4483,
"step": 14900
},
{
"epoch": 1.149911545265749,
"grad_norm": 1.4408409595489502,
"learning_rate": 0.00012499285769940007,
"loss": 1.5414,
"step": 14950
},
{
"epoch": 1.153757403276671,
"grad_norm": 1.5221819877624512,
"learning_rate": 0.00012473313767758355,
"loss": 1.5246,
"step": 15000
},
{
"epoch": 1.153757403276671,
"eval_loss": 1.516871452331543,
"eval_runtime": 17.9308,
"eval_samples_per_second": 55.77,
"eval_steps_per_second": 13.943,
"step": 15000
},
{
"epoch": 1.1576032612875933,
"grad_norm": 1.5970553159713745,
"learning_rate": 0.0001244734176557671,
"loss": 1.4515,
"step": 15050
},
{
"epoch": 1.1614491192985155,
"grad_norm": 1.0201988220214844,
"learning_rate": 0.0001242136976339506,
"loss": 1.5431,
"step": 15100
},
{
"epoch": 1.1652949773094377,
"grad_norm": 0.9830596446990967,
"learning_rate": 0.00012395397761213412,
"loss": 1.4581,
"step": 15150
},
{
"epoch": 1.16914083532036,
"grad_norm": 2.3363943099975586,
"learning_rate": 0.00012369425759031766,
"loss": 1.5147,
"step": 15200
},
{
"epoch": 1.1729866933312822,
"grad_norm": 1.8213731050491333,
"learning_rate": 0.00012343453756850117,
"loss": 1.5497,
"step": 15250
},
{
"epoch": 1.1729866933312822,
"eval_loss": 1.5053696632385254,
"eval_runtime": 17.7358,
"eval_samples_per_second": 56.383,
"eval_steps_per_second": 14.096,
"step": 15250
},
{
"epoch": 1.1768325513422044,
"grad_norm": 1.1468195915222168,
"learning_rate": 0.00012317481754668468,
"loss": 1.4292,
"step": 15300
},
{
"epoch": 1.1806784093531266,
"grad_norm": 1.2235878705978394,
"learning_rate": 0.0001229150975248682,
"loss": 1.4428,
"step": 15350
},
{
"epoch": 1.1845242673640488,
"grad_norm": 1.5518691539764404,
"learning_rate": 0.0001226553775030517,
"loss": 1.4889,
"step": 15400
},
{
"epoch": 1.1883701253749712,
"grad_norm": 1.217755913734436,
"learning_rate": 0.00012239565748123522,
"loss": 1.4841,
"step": 15450
},
{
"epoch": 1.1922159833858934,
"grad_norm": 1.6093647480010986,
"learning_rate": 0.00012213593745941876,
"loss": 1.5461,
"step": 15500
},
{
"epoch": 1.1922159833858934,
"eval_loss": 1.5013692378997803,
"eval_runtime": 18.0052,
"eval_samples_per_second": 55.539,
"eval_steps_per_second": 13.885,
"step": 15500
},
{
"epoch": 1.1960618413968156,
"grad_norm": 3.367638349533081,
"learning_rate": 0.00012187621743760226,
"loss": 1.4261,
"step": 15550
},
{
"epoch": 1.1999076994077378,
"grad_norm": 1.6329169273376465,
"learning_rate": 0.00012161649741578578,
"loss": 1.4475,
"step": 15600
},
{
"epoch": 1.2037535574186602,
"grad_norm": 1.397910475730896,
"learning_rate": 0.00012135677739396932,
"loss": 1.4747,
"step": 15650
},
{
"epoch": 1.2075994154295824,
"grad_norm": 1.7463736534118652,
"learning_rate": 0.00012109705737215283,
"loss": 1.4745,
"step": 15700
},
{
"epoch": 1.2114452734405046,
"grad_norm": 1.8097542524337769,
"learning_rate": 0.00012083733735033633,
"loss": 1.5276,
"step": 15750
},
{
"epoch": 1.2114452734405046,
"eval_loss": 1.4955236911773682,
"eval_runtime": 17.9054,
"eval_samples_per_second": 55.849,
"eval_steps_per_second": 13.962,
"step": 15750
},
{
"epoch": 1.2152911314514268,
"grad_norm": 1.745730996131897,
"learning_rate": 0.00012057761732851987,
"loss": 1.5282,
"step": 15800
},
{
"epoch": 1.2191369894623492,
"grad_norm": 1.3859128952026367,
"learning_rate": 0.00012031789730670338,
"loss": 1.4795,
"step": 15850
},
{
"epoch": 1.2229828474732714,
"grad_norm": 1.5086127519607544,
"learning_rate": 0.0001200581772848869,
"loss": 1.437,
"step": 15900
},
{
"epoch": 1.2268287054841935,
"grad_norm": 2.088292121887207,
"learning_rate": 0.00011979845726307042,
"loss": 1.5011,
"step": 15950
},
{
"epoch": 1.2306745634951157,
"grad_norm": 1.2746011018753052,
"learning_rate": 0.00011953873724125393,
"loss": 1.4596,
"step": 16000
},
{
"epoch": 1.2306745634951157,
"eval_loss": 1.4984314441680908,
"eval_runtime": 17.8066,
"eval_samples_per_second": 56.159,
"eval_steps_per_second": 14.04,
"step": 16000
},
{
"epoch": 1.234520421506038,
"grad_norm": 6.292486667633057,
"learning_rate": 0.00011927901721943746,
"loss": 1.3934,
"step": 16050
},
{
"epoch": 1.2383662795169603,
"grad_norm": 1.6574532985687256,
"learning_rate": 0.00011901929719762097,
"loss": 1.4821,
"step": 16100
},
{
"epoch": 1.2422121375278825,
"grad_norm": 1.8651037216186523,
"learning_rate": 0.00011875957717580448,
"loss": 1.49,
"step": 16150
},
{
"epoch": 1.2460579955388047,
"grad_norm": 1.3768175840377808,
"learning_rate": 0.00011849985715398801,
"loss": 1.4579,
"step": 16200
},
{
"epoch": 1.2499038535497269,
"grad_norm": 1.1569020748138428,
"learning_rate": 0.00011824013713217152,
"loss": 1.4029,
"step": 16250
},
{
"epoch": 1.2499038535497269,
"eval_loss": 1.4893407821655273,
"eval_runtime": 17.9552,
"eval_samples_per_second": 55.694,
"eval_steps_per_second": 13.924,
"step": 16250
},
{
"epoch": 1.253749711560649,
"grad_norm": 1.8632296323776245,
"learning_rate": 0.00011798041711035504,
"loss": 1.4592,
"step": 16300
},
{
"epoch": 1.2575955695715715,
"grad_norm": 1.8080470561981201,
"learning_rate": 0.00011772069708853856,
"loss": 1.4678,
"step": 16350
},
{
"epoch": 1.2614414275824937,
"grad_norm": 1.4193981885910034,
"learning_rate": 0.00011746097706672208,
"loss": 1.5031,
"step": 16400
},
{
"epoch": 1.2652872855934159,
"grad_norm": 1.5050238370895386,
"learning_rate": 0.00011720125704490559,
"loss": 1.4107,
"step": 16450
},
{
"epoch": 1.2691331436043383,
"grad_norm": 1.12454092502594,
"learning_rate": 0.00011694153702308911,
"loss": 1.4572,
"step": 16500
},
{
"epoch": 1.2691331436043383,
"eval_loss": 1.4972718954086304,
"eval_runtime": 17.8677,
"eval_samples_per_second": 55.967,
"eval_steps_per_second": 13.992,
"step": 16500
},
{
"epoch": 1.2729790016152602,
"grad_norm": 1.3523976802825928,
"learning_rate": 0.00011668181700127263,
"loss": 1.4796,
"step": 16550
},
{
"epoch": 1.2768248596261826,
"grad_norm": 1.9770869016647339,
"learning_rate": 0.00011642209697945614,
"loss": 1.4756,
"step": 16600
},
{
"epoch": 1.2806707176371048,
"grad_norm": 1.7973159551620483,
"learning_rate": 0.00011616237695763967,
"loss": 1.4163,
"step": 16650
},
{
"epoch": 1.284516575648027,
"grad_norm": 1.3054739236831665,
"learning_rate": 0.00011590265693582318,
"loss": 1.4105,
"step": 16700
},
{
"epoch": 1.2883624336589494,
"grad_norm": 1.457047939300537,
"learning_rate": 0.00011564293691400672,
"loss": 1.4698,
"step": 16750
},
{
"epoch": 1.2883624336589494,
"eval_loss": 1.4798808097839355,
"eval_runtime": 17.8696,
"eval_samples_per_second": 55.961,
"eval_steps_per_second": 13.99,
"step": 16750
},
{
"epoch": 1.2922082916698716,
"grad_norm": 2.117663860321045,
"learning_rate": 0.00011538841129262656,
"loss": 1.4252,
"step": 16800
},
{
"epoch": 1.2960541496807938,
"grad_norm": 3.7321341037750244,
"learning_rate": 0.00011512869127081007,
"loss": 1.4818,
"step": 16850
},
{
"epoch": 1.299900007691716,
"grad_norm": 2.497528553009033,
"learning_rate": 0.00011486897124899358,
"loss": 1.4176,
"step": 16900
},
{
"epoch": 1.3037458657026382,
"grad_norm": 1.6085398197174072,
"learning_rate": 0.00011460925122717712,
"loss": 1.4373,
"step": 16950
},
{
"epoch": 1.3075917237135606,
"grad_norm": 1.8937525749206543,
"learning_rate": 0.00011434953120536062,
"loss": 1.4823,
"step": 17000
},
{
"epoch": 1.3075917237135606,
"eval_loss": 1.4854488372802734,
"eval_runtime": 17.8761,
"eval_samples_per_second": 55.941,
"eval_steps_per_second": 13.985,
"step": 17000
},
{
"epoch": 1.3114375817244828,
"grad_norm": 1.6637665033340454,
"learning_rate": 0.00011408981118354414,
"loss": 1.5071,
"step": 17050
},
{
"epoch": 1.315283439735405,
"grad_norm": 2.0815582275390625,
"learning_rate": 0.00011383009116172768,
"loss": 1.4606,
"step": 17100
},
{
"epoch": 1.3191292977463271,
"grad_norm": 1.6492595672607422,
"learning_rate": 0.00011357037113991117,
"loss": 1.4408,
"step": 17150
},
{
"epoch": 1.3229751557572493,
"grad_norm": 0.8617509603500366,
"learning_rate": 0.00011331065111809469,
"loss": 1.497,
"step": 17200
},
{
"epoch": 1.3268210137681717,
"grad_norm": 1.6395294666290283,
"learning_rate": 0.00011305093109627823,
"loss": 1.4774,
"step": 17250
},
{
"epoch": 1.3268210137681717,
"eval_loss": 1.4835026264190674,
"eval_runtime": 18.0136,
"eval_samples_per_second": 55.514,
"eval_steps_per_second": 13.878,
"step": 17250
},
{
"epoch": 1.330666871779094,
"grad_norm": 2.7765560150146484,
"learning_rate": 0.00011279121107446174,
"loss": 1.4864,
"step": 17300
},
{
"epoch": 1.3345127297900161,
"grad_norm": 1.2104064226150513,
"learning_rate": 0.00011253149105264524,
"loss": 1.4075,
"step": 17350
},
{
"epoch": 1.3383585878009385,
"grad_norm": 1.6772801876068115,
"learning_rate": 0.00011227177103082878,
"loss": 1.4629,
"step": 17400
},
{
"epoch": 1.3422044458118605,
"grad_norm": 2.254371404647827,
"learning_rate": 0.00011201205100901229,
"loss": 1.4447,
"step": 17450
},
{
"epoch": 1.346050303822783,
"grad_norm": 2.2015669345855713,
"learning_rate": 0.00011175233098719582,
"loss": 1.4664,
"step": 17500
},
{
"epoch": 1.346050303822783,
"eval_loss": 1.496685266494751,
"eval_runtime": 17.7963,
"eval_samples_per_second": 56.191,
"eval_steps_per_second": 14.048,
"step": 17500
},
{
"epoch": 1.349896161833705,
"grad_norm": 1.740045428276062,
"learning_rate": 0.00011149261096537933,
"loss": 1.4486,
"step": 17550
},
{
"epoch": 1.3537420198446273,
"grad_norm": 1.299919605255127,
"learning_rate": 0.00011123289094356284,
"loss": 1.4868,
"step": 17600
},
{
"epoch": 1.3575878778555497,
"grad_norm": 1.6288009881973267,
"learning_rate": 0.00011097317092174637,
"loss": 1.4595,
"step": 17650
},
{
"epoch": 1.3614337358664719,
"grad_norm": 0.8747851252555847,
"learning_rate": 0.00011071345089992988,
"loss": 1.4241,
"step": 17700
},
{
"epoch": 1.365279593877394,
"grad_norm": 1.9510573148727417,
"learning_rate": 0.0001104537308781134,
"loss": 1.496,
"step": 17750
},
{
"epoch": 1.365279593877394,
"eval_loss": 1.4812238216400146,
"eval_runtime": 17.8955,
"eval_samples_per_second": 55.88,
"eval_steps_per_second": 13.97,
"step": 17750
},
{
"epoch": 1.3691254518883162,
"grad_norm": 1.4853876829147339,
"learning_rate": 0.00011019401085629692,
"loss": 1.4645,
"step": 17800
},
{
"epoch": 1.3729713098992384,
"grad_norm": 1.5125057697296143,
"learning_rate": 0.00010993429083448044,
"loss": 1.4052,
"step": 17850
},
{
"epoch": 1.3768171679101608,
"grad_norm": 0.7320863008499146,
"learning_rate": 0.00010967457081266395,
"loss": 1.5016,
"step": 17900
},
{
"epoch": 1.380663025921083,
"grad_norm": 1.9995285272598267,
"learning_rate": 0.00010941485079084747,
"loss": 1.4234,
"step": 17950
},
{
"epoch": 1.3845088839320052,
"grad_norm": 0.8304823637008667,
"learning_rate": 0.00010915513076903099,
"loss": 1.4068,
"step": 18000
},
{
"epoch": 1.3845088839320052,
"eval_loss": 1.4845945835113525,
"eval_runtime": 17.7402,
"eval_samples_per_second": 56.369,
"eval_steps_per_second": 14.092,
"step": 18000
},
{
"epoch": 1.3883547419429274,
"grad_norm": 1.8189436197280884,
"learning_rate": 0.0001088954107472145,
"loss": 1.4872,
"step": 18050
},
{
"epoch": 1.3922005999538496,
"grad_norm": 1.4212762117385864,
"learning_rate": 0.00010863569072539803,
"loss": 1.4067,
"step": 18100
},
{
"epoch": 1.396046457964772,
"grad_norm": 1.9733264446258545,
"learning_rate": 0.00010837597070358154,
"loss": 1.4787,
"step": 18150
},
{
"epoch": 1.3998923159756942,
"grad_norm": 1.320064663887024,
"learning_rate": 0.00010811625068176505,
"loss": 1.5242,
"step": 18200
},
{
"epoch": 1.4037381739866164,
"grad_norm": 1.4177141189575195,
"learning_rate": 0.00010785653065994858,
"loss": 1.4719,
"step": 18250
},
{
"epoch": 1.4037381739866164,
"eval_loss": 1.4992233514785767,
"eval_runtime": 17.8574,
"eval_samples_per_second": 55.999,
"eval_steps_per_second": 14.0,
"step": 18250
},
{
"epoch": 1.4075840319975388,
"grad_norm": 1.8219791650772095,
"learning_rate": 0.00010759681063813209,
"loss": 1.4392,
"step": 18300
},
{
"epoch": 1.4114298900084608,
"grad_norm": 1.3196603059768677,
"learning_rate": 0.00010733709061631563,
"loss": 1.4837,
"step": 18350
},
{
"epoch": 1.4152757480193832,
"grad_norm": 1.01405668258667,
"learning_rate": 0.00010707737059449914,
"loss": 1.455,
"step": 18400
},
{
"epoch": 1.4191216060303053,
"grad_norm": 1.8538917303085327,
"learning_rate": 0.00010681765057268264,
"loss": 1.4209,
"step": 18450
},
{
"epoch": 1.4229674640412275,
"grad_norm": 0.8785907030105591,
"learning_rate": 0.00010655793055086618,
"loss": 1.4716,
"step": 18500
},
{
"epoch": 1.4229674640412275,
"eval_loss": 1.485592246055603,
"eval_runtime": 17.7246,
"eval_samples_per_second": 56.419,
"eval_steps_per_second": 14.105,
"step": 18500
},
{
"epoch": 1.42681332205215,
"grad_norm": 1.0700381994247437,
"learning_rate": 0.0001062982105290497,
"loss": 1.3846,
"step": 18550
},
{
"epoch": 1.4306591800630721,
"grad_norm": 1.2848351001739502,
"learning_rate": 0.0001060384905072332,
"loss": 1.4874,
"step": 18600
},
{
"epoch": 1.4345050380739943,
"grad_norm": 1.4261386394500732,
"learning_rate": 0.00010577877048541674,
"loss": 1.5771,
"step": 18650
},
{
"epoch": 1.4383508960849165,
"grad_norm": 1.3613426685333252,
"learning_rate": 0.00010551905046360025,
"loss": 1.4445,
"step": 18700
},
{
"epoch": 1.4421967540958387,
"grad_norm": 0.8625685572624207,
"learning_rate": 0.00010525933044178375,
"loss": 1.3845,
"step": 18750
},
{
"epoch": 1.4421967540958387,
"eval_loss": 1.4757392406463623,
"eval_runtime": 17.8903,
"eval_samples_per_second": 55.896,
"eval_steps_per_second": 13.974,
"step": 18750
},
{
"epoch": 1.446042612106761,
"grad_norm": 1.5575672388076782,
"learning_rate": 0.00010499961041996729,
"loss": 1.5387,
"step": 18800
},
{
"epoch": 1.4498884701176833,
"grad_norm": 2.2173306941986084,
"learning_rate": 0.0001047398903981508,
"loss": 1.459,
"step": 18850
},
{
"epoch": 1.4537343281286055,
"grad_norm": 1.904809832572937,
"learning_rate": 0.00010448017037633431,
"loss": 1.5229,
"step": 18900
},
{
"epoch": 1.4575801861395277,
"grad_norm": 1.8832893371582031,
"learning_rate": 0.00010422045035451784,
"loss": 1.4093,
"step": 18950
},
{
"epoch": 1.4614260441504499,
"grad_norm": 1.852971076965332,
"learning_rate": 0.00010396073033270135,
"loss": 1.389,
"step": 19000
},
{
"epoch": 1.4614260441504499,
"eval_loss": 1.4788576364517212,
"eval_runtime": 17.7751,
"eval_samples_per_second": 56.258,
"eval_steps_per_second": 14.065,
"step": 19000
},
{
"epoch": 1.4652719021613723,
"grad_norm": 1.2875189781188965,
"learning_rate": 0.00010370101031088487,
"loss": 1.4826,
"step": 19050
},
{
"epoch": 1.4691177601722945,
"grad_norm": 1.7036223411560059,
"learning_rate": 0.00010344129028906839,
"loss": 1.4392,
"step": 19100
},
{
"epoch": 1.4729636181832166,
"grad_norm": 1.537514328956604,
"learning_rate": 0.0001031815702672519,
"loss": 1.4846,
"step": 19150
},
{
"epoch": 1.476809476194139,
"grad_norm": 0.9159242510795593,
"learning_rate": 0.00010292185024543543,
"loss": 1.4668,
"step": 19200
},
{
"epoch": 1.480655334205061,
"grad_norm": 3.47868013381958,
"learning_rate": 0.00010266213022361894,
"loss": 1.493,
"step": 19250
},
{
"epoch": 1.480655334205061,
"eval_loss": 1.4858986139297485,
"eval_runtime": 17.7738,
"eval_samples_per_second": 56.263,
"eval_steps_per_second": 14.066,
"step": 19250
},
{
"epoch": 1.4845011922159834,
"grad_norm": 1.467437505722046,
"learning_rate": 0.00010240241020180246,
"loss": 1.4639,
"step": 19300
},
{
"epoch": 1.4883470502269056,
"grad_norm": 1.2710049152374268,
"learning_rate": 0.00010214269017998598,
"loss": 1.4244,
"step": 19350
},
{
"epoch": 1.4921929082378278,
"grad_norm": 2.0059661865234375,
"learning_rate": 0.0001018829701581695,
"loss": 1.4258,
"step": 19400
},
{
"epoch": 1.4960387662487502,
"grad_norm": 1.7536308765411377,
"learning_rate": 0.00010162325013635301,
"loss": 1.396,
"step": 19450
},
{
"epoch": 1.4998846242596724,
"grad_norm": 0.9684279561042786,
"learning_rate": 0.00010136353011453655,
"loss": 1.4598,
"step": 19500
},
{
"epoch": 1.4998846242596724,
"eval_loss": 1.4841110706329346,
"eval_runtime": 17.9151,
"eval_samples_per_second": 55.819,
"eval_steps_per_second": 13.955,
"step": 19500
},
{
"epoch": 1.5037304822705946,
"grad_norm": 0.9244908690452576,
"learning_rate": 0.00010110381009272005,
"loss": 1.3834,
"step": 19550
},
{
"epoch": 1.5075763402815168,
"grad_norm": 1.6488862037658691,
"learning_rate": 0.00010084409007090356,
"loss": 1.5311,
"step": 19600
},
{
"epoch": 1.511422198292439,
"grad_norm": 1.9130067825317383,
"learning_rate": 0.0001005843700490871,
"loss": 1.5212,
"step": 19650
},
{
"epoch": 1.5152680563033614,
"grad_norm": 1.326277256011963,
"learning_rate": 0.0001003246500272706,
"loss": 1.4067,
"step": 19700
},
{
"epoch": 1.5191139143142836,
"grad_norm": 1.7258195877075195,
"learning_rate": 0.00010006493000545411,
"loss": 1.4844,
"step": 19750
},
{
"epoch": 1.5191139143142836,
"eval_loss": 1.4816969633102417,
"eval_runtime": 17.826,
"eval_samples_per_second": 56.098,
"eval_steps_per_second": 14.024,
"step": 19750
},
{
"epoch": 1.5229597723252057,
"grad_norm": 1.8164838552474976,
"learning_rate": 9.980520998363765e-05,
"loss": 1.4233,
"step": 19800
},
{
"epoch": 1.5268056303361282,
"grad_norm": 1.5884016752243042,
"learning_rate": 9.954548996182115e-05,
"loss": 1.4313,
"step": 19850
},
{
"epoch": 1.5306514883470501,
"grad_norm": 1.5381648540496826,
"learning_rate": 9.928576994000468e-05,
"loss": 1.4789,
"step": 19900
},
{
"epoch": 1.5344973463579725,
"grad_norm": 1.6448626518249512,
"learning_rate": 9.90260499181882e-05,
"loss": 1.3756,
"step": 19950
},
{
"epoch": 1.5383432043688947,
"grad_norm": 1.6137230396270752,
"learning_rate": 9.876632989637172e-05,
"loss": 1.4171,
"step": 20000
},
{
"epoch": 1.5383432043688947,
"eval_loss": 1.4770597219467163,
"eval_runtime": 17.9365,
"eval_samples_per_second": 55.752,
"eval_steps_per_second": 13.938,
"step": 20000
},
{
"epoch": 1.542189062379817,
"grad_norm": 1.7058050632476807,
"learning_rate": 9.850660987455523e-05,
"loss": 1.4646,
"step": 20050
},
{
"epoch": 1.5460349203907393,
"grad_norm": 2.2624917030334473,
"learning_rate": 9.824688985273876e-05,
"loss": 1.4586,
"step": 20100
},
{
"epoch": 1.5498807784016613,
"grad_norm": 2.216883420944214,
"learning_rate": 9.798716983092227e-05,
"loss": 1.4376,
"step": 20150
},
{
"epoch": 1.5537266364125837,
"grad_norm": 1.9749584197998047,
"learning_rate": 9.772744980910578e-05,
"loss": 1.47,
"step": 20200
},
{
"epoch": 1.5575724944235059,
"grad_norm": 2.185480833053589,
"learning_rate": 9.746772978728931e-05,
"loss": 1.39,
"step": 20250
},
{
"epoch": 1.5575724944235059,
"eval_loss": 1.4724150896072388,
"eval_runtime": 17.8191,
"eval_samples_per_second": 56.12,
"eval_steps_per_second": 14.03,
"step": 20250
},
{
"epoch": 1.561418352434428,
"grad_norm": 1.5810290575027466,
"learning_rate": 9.720800976547284e-05,
"loss": 1.4267,
"step": 20300
},
{
"epoch": 1.5652642104453505,
"grad_norm": 2.0329344272613525,
"learning_rate": 9.694828974365633e-05,
"loss": 1.5066,
"step": 20350
},
{
"epoch": 1.5691100684562724,
"grad_norm": 2.7335126399993896,
"learning_rate": 9.668856972183986e-05,
"loss": 1.4749,
"step": 20400
},
{
"epoch": 1.5729559264671948,
"grad_norm": 1.0576220750808716,
"learning_rate": 9.642884970002339e-05,
"loss": 1.4318,
"step": 20450
},
{
"epoch": 1.576801784478117,
"grad_norm": 0.6857870817184448,
"learning_rate": 9.616912967820689e-05,
"loss": 1.3567,
"step": 20500
},
{
"epoch": 1.576801784478117,
"eval_loss": 1.479669213294983,
"eval_runtime": 17.8378,
"eval_samples_per_second": 56.061,
"eval_steps_per_second": 14.015,
"step": 20500
},
{
"epoch": 1.5806476424890392,
"grad_norm": 1.430114507675171,
"learning_rate": 9.590940965639041e-05,
"loss": 1.3714,
"step": 20550
},
{
"epoch": 1.5844935004999616,
"grad_norm": 1.7613717317581177,
"learning_rate": 9.564968963457394e-05,
"loss": 1.4285,
"step": 20600
},
{
"epoch": 1.5883393585108838,
"grad_norm": 1.3678529262542725,
"learning_rate": 9.538996961275745e-05,
"loss": 1.4463,
"step": 20650
},
{
"epoch": 1.592185216521806,
"grad_norm": 3.1135504245758057,
"learning_rate": 9.513024959094097e-05,
"loss": 1.4287,
"step": 20700
},
{
"epoch": 1.5960310745327284,
"grad_norm": 1.7866570949554443,
"learning_rate": 9.487052956912449e-05,
"loss": 1.4515,
"step": 20750
},
{
"epoch": 1.5960310745327284,
"eval_loss": 1.4817472696304321,
"eval_runtime": 17.9267,
"eval_samples_per_second": 55.783,
"eval_steps_per_second": 13.946,
"step": 20750
},
{
"epoch": 1.5998769325436504,
"grad_norm": 1.2406786680221558,
"learning_rate": 9.4610809547308e-05,
"loss": 1.5364,
"step": 20800
},
{
"epoch": 1.6037227905545728,
"grad_norm": 1.9183951616287231,
"learning_rate": 9.435108952549152e-05,
"loss": 1.5169,
"step": 20850
},
{
"epoch": 1.607568648565495,
"grad_norm": 1.852089524269104,
"learning_rate": 9.409136950367504e-05,
"loss": 1.4668,
"step": 20900
},
{
"epoch": 1.6114145065764172,
"grad_norm": 2.3430335521698,
"learning_rate": 9.383164948185856e-05,
"loss": 1.4297,
"step": 20950
},
{
"epoch": 1.6152603645873396,
"grad_norm": 1.9646743535995483,
"learning_rate": 9.357192946004207e-05,
"loss": 1.5143,
"step": 21000
},
{
"epoch": 1.6152603645873396,
"eval_loss": 1.4690666198730469,
"eval_runtime": 17.7563,
"eval_samples_per_second": 56.318,
"eval_steps_per_second": 14.079,
"step": 21000
},
{
"epoch": 1.6191062225982615,
"grad_norm": 1.8204776048660278,
"learning_rate": 9.33122094382256e-05,
"loss": 1.4423,
"step": 21050
},
{
"epoch": 1.622952080609184,
"grad_norm": 1.6150448322296143,
"learning_rate": 9.305248941640912e-05,
"loss": 1.4368,
"step": 21100
},
{
"epoch": 1.6267979386201061,
"grad_norm": 2.4393765926361084,
"learning_rate": 9.279276939459264e-05,
"loss": 1.6115,
"step": 21150
},
{
"epoch": 1.6306437966310283,
"grad_norm": 1.217399001121521,
"learning_rate": 9.253304937277615e-05,
"loss": 1.5792,
"step": 21200
},
{
"epoch": 1.6344896546419507,
"grad_norm": 1.4467514753341675,
"learning_rate": 9.227332935095967e-05,
"loss": 1.4838,
"step": 21250
},
{
"epoch": 1.6344896546419507,
"eval_loss": 1.4741238355636597,
"eval_runtime": 17.7656,
"eval_samples_per_second": 56.289,
"eval_steps_per_second": 14.072,
"step": 21250
},
{
"epoch": 1.6383355126528727,
"grad_norm": 1.4274511337280273,
"learning_rate": 9.201360932914319e-05,
"loss": 1.4747,
"step": 21300
},
{
"epoch": 1.642181370663795,
"grad_norm": 1.5080102682113647,
"learning_rate": 9.175908370776304e-05,
"loss": 1.4262,
"step": 21350
},
{
"epoch": 1.6460272286747173,
"grad_norm": 1.2717032432556152,
"learning_rate": 9.149936368594655e-05,
"loss": 1.4418,
"step": 21400
},
{
"epoch": 1.6498730866856395,
"grad_norm": 1.3594080209732056,
"learning_rate": 9.123964366413006e-05,
"loss": 1.5184,
"step": 21450
},
{
"epoch": 1.6537189446965619,
"grad_norm": 1.896607518196106,
"learning_rate": 9.097992364231359e-05,
"loss": 1.4702,
"step": 21500
},
{
"epoch": 1.6537189446965619,
"eval_loss": 1.4777735471725464,
"eval_runtime": 17.8161,
"eval_samples_per_second": 56.129,
"eval_steps_per_second": 14.032,
"step": 21500
},
{
"epoch": 1.657564802707484,
"grad_norm": 1.3839844465255737,
"learning_rate": 9.072020362049712e-05,
"loss": 1.522,
"step": 21550
},
{
"epoch": 1.6614106607184063,
"grad_norm": 1.187853455543518,
"learning_rate": 9.046048359868063e-05,
"loss": 1.4469,
"step": 21600
},
{
"epoch": 1.6652565187293287,
"grad_norm": 2.8398866653442383,
"learning_rate": 9.020076357686414e-05,
"loss": 1.4875,
"step": 21650
},
{
"epoch": 1.6691023767402506,
"grad_norm": 1.801963448524475,
"learning_rate": 8.994104355504767e-05,
"loss": 1.3892,
"step": 21700
},
{
"epoch": 1.672948234751173,
"grad_norm": 1.7727116346359253,
"learning_rate": 8.968132353323118e-05,
"loss": 1.3886,
"step": 21750
},
{
"epoch": 1.672948234751173,
"eval_loss": 1.4662117958068848,
"eval_runtime": 17.7809,
"eval_samples_per_second": 56.24,
"eval_steps_per_second": 14.06,
"step": 21750
},
{
"epoch": 1.6767940927620952,
"grad_norm": 1.7052053213119507,
"learning_rate": 8.94216035114147e-05,
"loss": 1.4483,
"step": 21800
},
{
"epoch": 1.6806399507730174,
"grad_norm": 1.9812465906143188,
"learning_rate": 8.916188348959822e-05,
"loss": 1.4325,
"step": 21850
},
{
"epoch": 1.6844858087839398,
"grad_norm": 1.2499721050262451,
"learning_rate": 8.890216346778173e-05,
"loss": 1.5246,
"step": 21900
},
{
"epoch": 1.6883316667948618,
"grad_norm": 1.2503259181976318,
"learning_rate": 8.864244344596525e-05,
"loss": 1.4661,
"step": 21950
},
{
"epoch": 1.6921775248057842,
"grad_norm": 2.201223134994507,
"learning_rate": 8.838272342414877e-05,
"loss": 1.5103,
"step": 22000
},
{
"epoch": 1.6921775248057842,
"eval_loss": 1.470120906829834,
"eval_runtime": 17.7715,
"eval_samples_per_second": 56.27,
"eval_steps_per_second": 14.067,
"step": 22000
},
{
"epoch": 1.6960233828167064,
"grad_norm": 0.8225556015968323,
"learning_rate": 8.81230034023323e-05,
"loss": 1.4372,
"step": 22050
},
{
"epoch": 1.6998692408276286,
"grad_norm": 1.495335578918457,
"learning_rate": 8.78632833805158e-05,
"loss": 1.4231,
"step": 22100
},
{
"epoch": 1.703715098838551,
"grad_norm": 1.6913652420043945,
"learning_rate": 8.760356335869933e-05,
"loss": 1.4859,
"step": 22150
},
{
"epoch": 1.707560956849473,
"grad_norm": 1.9825598001480103,
"learning_rate": 8.734384333688285e-05,
"loss": 1.4354,
"step": 22200
},
{
"epoch": 1.7114068148603954,
"grad_norm": 2.212759017944336,
"learning_rate": 8.708412331506635e-05,
"loss": 1.3861,
"step": 22250
},
{
"epoch": 1.7114068148603954,
"eval_loss": 1.4596961736679077,
"eval_runtime": 17.8176,
"eval_samples_per_second": 56.124,
"eval_steps_per_second": 14.031,
"step": 22250
},
{
"epoch": 1.7152526728713176,
"grad_norm": 1.4325975179672241,
"learning_rate": 8.682440329324988e-05,
"loss": 1.4805,
"step": 22300
},
{
"epoch": 1.7190985308822397,
"grad_norm": 1.9796292781829834,
"learning_rate": 8.65646832714334e-05,
"loss": 1.3836,
"step": 22350
},
{
"epoch": 1.7229443888931621,
"grad_norm": 1.6221562623977661,
"learning_rate": 8.630496324961692e-05,
"loss": 1.4176,
"step": 22400
},
{
"epoch": 1.7267902469040843,
"grad_norm": 1.1431959867477417,
"learning_rate": 8.604524322780043e-05,
"loss": 1.4553,
"step": 22450
},
{
"epoch": 1.7306361049150065,
"grad_norm": 1.1562083959579468,
"learning_rate": 8.578552320598396e-05,
"loss": 1.4489,
"step": 22500
},
{
"epoch": 1.7306361049150065,
"eval_loss": 1.464021921157837,
"eval_runtime": 17.8166,
"eval_samples_per_second": 56.127,
"eval_steps_per_second": 14.032,
"step": 22500
},
{
"epoch": 1.734481962925929,
"grad_norm": 0.8531803488731384,
"learning_rate": 8.552580318416747e-05,
"loss": 1.4298,
"step": 22550
},
{
"epoch": 1.738327820936851,
"grad_norm": 1.3987632989883423,
"learning_rate": 8.526608316235098e-05,
"loss": 1.4629,
"step": 22600
},
{
"epoch": 1.7421736789477733,
"grad_norm": 1.4521870613098145,
"learning_rate": 8.500636314053451e-05,
"loss": 1.4005,
"step": 22650
},
{
"epoch": 1.7460195369586955,
"grad_norm": 1.0557054281234741,
"learning_rate": 8.474664311871803e-05,
"loss": 1.4079,
"step": 22700
},
{
"epoch": 1.7498653949696177,
"grad_norm": 1.5067927837371826,
"learning_rate": 8.448692309690155e-05,
"loss": 1.4664,
"step": 22750
},
{
"epoch": 1.7498653949696177,
"eval_loss": 1.4728831052780151,
"eval_runtime": 17.7456,
"eval_samples_per_second": 56.352,
"eval_steps_per_second": 14.088,
"step": 22750
},
{
"epoch": 1.75371125298054,
"grad_norm": 1.3237221240997314,
"learning_rate": 8.422720307508506e-05,
"loss": 1.5266,
"step": 22800
},
{
"epoch": 1.757557110991462,
"grad_norm": 1.4342700242996216,
"learning_rate": 8.396748305326859e-05,
"loss": 1.4929,
"step": 22850
},
{
"epoch": 1.7614029690023845,
"grad_norm": 5.2202534675598145,
"learning_rate": 8.37077630314521e-05,
"loss": 1.457,
"step": 22900
},
{
"epoch": 1.7652488270133067,
"grad_norm": 2.7584545612335205,
"learning_rate": 8.344804300963561e-05,
"loss": 1.4523,
"step": 22950
},
{
"epoch": 1.7690946850242288,
"grad_norm": 1.8208624124526978,
"learning_rate": 8.318832298781914e-05,
"loss": 1.4746,
"step": 23000
},
{
"epoch": 1.7690946850242288,
"eval_loss": 1.467396855354309,
"eval_runtime": 17.663,
"eval_samples_per_second": 56.616,
"eval_steps_per_second": 14.154,
"step": 23000
},
{
"epoch": 1.7729405430351513,
"grad_norm": 1.6651790142059326,
"learning_rate": 8.292860296600265e-05,
"loss": 1.4537,
"step": 23050
},
{
"epoch": 1.7767864010460732,
"grad_norm": 1.5910587310791016,
"learning_rate": 8.266888294418618e-05,
"loss": 1.5305,
"step": 23100
},
{
"epoch": 1.7806322590569956,
"grad_norm": 1.2711199522018433,
"learning_rate": 8.240916292236969e-05,
"loss": 1.4877,
"step": 23150
},
{
"epoch": 1.7844781170679178,
"grad_norm": 1.7133463621139526,
"learning_rate": 8.21494429005532e-05,
"loss": 1.4547,
"step": 23200
},
{
"epoch": 1.78832397507884,
"grad_norm": 1.6850301027297974,
"learning_rate": 8.188972287873673e-05,
"loss": 1.4228,
"step": 23250
},
{
"epoch": 1.78832397507884,
"eval_loss": 1.4735645055770874,
"eval_runtime": 17.7983,
"eval_samples_per_second": 56.185,
"eval_steps_per_second": 14.046,
"step": 23250
},
{
"epoch": 1.7921698330897624,
"grad_norm": 1.1047898530960083,
"learning_rate": 8.163000285692024e-05,
"loss": 1.4817,
"step": 23300
},
{
"epoch": 1.7960156911006846,
"grad_norm": 1.8970743417739868,
"learning_rate": 8.137028283510376e-05,
"loss": 1.4884,
"step": 23350
},
{
"epoch": 1.7998615491116068,
"grad_norm": 1.4730740785598755,
"learning_rate": 8.111056281328728e-05,
"loss": 1.4436,
"step": 23400
},
{
"epoch": 1.8037074071225292,
"grad_norm": 1.8288697004318237,
"learning_rate": 8.08508427914708e-05,
"loss": 1.4614,
"step": 23450
},
{
"epoch": 1.8075532651334512,
"grad_norm": 1.2339516878128052,
"learning_rate": 8.059112276965432e-05,
"loss": 1.4276,
"step": 23500
},
{
"epoch": 1.8075532651334512,
"eval_loss": 1.4655163288116455,
"eval_runtime": 17.9781,
"eval_samples_per_second": 55.623,
"eval_steps_per_second": 13.906,
"step": 23500
},
{
"epoch": 1.8113991231443736,
"grad_norm": 1.8125578165054321,
"learning_rate": 8.033140274783783e-05,
"loss": 1.4318,
"step": 23550
},
{
"epoch": 1.8152449811552958,
"grad_norm": 1.941846489906311,
"learning_rate": 8.007168272602136e-05,
"loss": 1.4569,
"step": 23600
},
{
"epoch": 1.819090839166218,
"grad_norm": 2.054161787033081,
"learning_rate": 7.981196270420487e-05,
"loss": 1.4426,
"step": 23650
},
{
"epoch": 1.8229366971771404,
"grad_norm": 1.6102700233459473,
"learning_rate": 7.955224268238839e-05,
"loss": 1.4407,
"step": 23700
},
{
"epoch": 1.8267825551880623,
"grad_norm": 1.7302616834640503,
"learning_rate": 7.929252266057191e-05,
"loss": 1.4432,
"step": 23750
},
{
"epoch": 1.8267825551880623,
"eval_loss": 1.4570631980895996,
"eval_runtime": 17.8893,
"eval_samples_per_second": 55.899,
"eval_steps_per_second": 13.975,
"step": 23750
},
{
"epoch": 1.8306284131989847,
"grad_norm": 9.521604537963867,
"learning_rate": 7.903280263875543e-05,
"loss": 1.3995,
"step": 23800
},
{
"epoch": 1.834474271209907,
"grad_norm": 1.9546847343444824,
"learning_rate": 7.877308261693894e-05,
"loss": 1.4957,
"step": 23850
},
{
"epoch": 1.838320129220829,
"grad_norm": 1.173722505569458,
"learning_rate": 7.851336259512246e-05,
"loss": 1.4157,
"step": 23900
},
{
"epoch": 1.8421659872317515,
"grad_norm": 1.5322128534317017,
"learning_rate": 7.825364257330599e-05,
"loss": 1.4453,
"step": 23950
},
{
"epoch": 1.8460118452426735,
"grad_norm": 1.0376055240631104,
"learning_rate": 7.799392255148949e-05,
"loss": 1.4646,
"step": 24000
},
{
"epoch": 1.8460118452426735,
"eval_loss": 1.4612687826156616,
"eval_runtime": 17.984,
"eval_samples_per_second": 55.605,
"eval_steps_per_second": 13.901,
"step": 24000
},
{
"epoch": 1.8498577032535959,
"grad_norm": 1.5047483444213867,
"learning_rate": 7.773420252967302e-05,
"loss": 1.4546,
"step": 24050
},
{
"epoch": 1.853703561264518,
"grad_norm": 1.0463405847549438,
"learning_rate": 7.747448250785654e-05,
"loss": 1.5014,
"step": 24100
},
{
"epoch": 1.8575494192754403,
"grad_norm": 1.8368524312973022,
"learning_rate": 7.721476248604004e-05,
"loss": 1.4616,
"step": 24150
},
{
"epoch": 1.8613952772863627,
"grad_norm": 1.4084677696228027,
"learning_rate": 7.695504246422357e-05,
"loss": 1.4255,
"step": 24200
},
{
"epoch": 1.8652411352972849,
"grad_norm": 1.2279951572418213,
"learning_rate": 7.66953224424071e-05,
"loss": 1.4254,
"step": 24250
},
{
"epoch": 1.8652411352972849,
"eval_loss": 1.460336685180664,
"eval_runtime": 17.7185,
"eval_samples_per_second": 56.438,
"eval_steps_per_second": 14.11,
"step": 24250
},
{
"epoch": 1.869086993308207,
"grad_norm": 1.9729641675949097,
"learning_rate": 7.643560242059061e-05,
"loss": 1.4656,
"step": 24300
},
{
"epoch": 1.8729328513191295,
"grad_norm": 0.9121168255805969,
"learning_rate": 7.617588239877412e-05,
"loss": 1.3949,
"step": 24350
},
{
"epoch": 1.8767787093300514,
"grad_norm": 1.8953206539154053,
"learning_rate": 7.591616237695765e-05,
"loss": 1.4006,
"step": 24400
},
{
"epoch": 1.8806245673409738,
"grad_norm": 1.5828944444656372,
"learning_rate": 7.565644235514116e-05,
"loss": 1.5085,
"step": 24450
},
{
"epoch": 1.884470425351896,
"grad_norm": 2.027841329574585,
"learning_rate": 7.539672233332467e-05,
"loss": 1.3978,
"step": 24500
},
{
"epoch": 1.884470425351896,
"eval_loss": 1.445096731185913,
"eval_runtime": 18.4795,
"eval_samples_per_second": 54.114,
"eval_steps_per_second": 13.528,
"step": 24500
},
{
"epoch": 1.8883162833628182,
"grad_norm": 3.0785481929779053,
"learning_rate": 7.51370023115082e-05,
"loss": 1.493,
"step": 24550
},
{
"epoch": 1.8921621413737406,
"grad_norm": 1.2793898582458496,
"learning_rate": 7.487728228969173e-05,
"loss": 1.3865,
"step": 24600
},
{
"epoch": 1.8960079993846626,
"grad_norm": 1.474137544631958,
"learning_rate": 7.461756226787522e-05,
"loss": 1.4316,
"step": 24650
},
{
"epoch": 1.899853857395585,
"grad_norm": 1.270415186882019,
"learning_rate": 7.435784224605875e-05,
"loss": 1.3844,
"step": 24700
},
{
"epoch": 1.9036997154065072,
"grad_norm": 1.3681602478027344,
"learning_rate": 7.409812222424228e-05,
"loss": 1.4202,
"step": 24750
},
{
"epoch": 1.9036997154065072,
"eval_loss": 1.4705748558044434,
"eval_runtime": 18.5933,
"eval_samples_per_second": 53.783,
"eval_steps_per_second": 13.446,
"step": 24750
},
{
"epoch": 1.9075455734174294,
"grad_norm": 1.028225302696228,
"learning_rate": 7.384359660286212e-05,
"loss": 1.4658,
"step": 24800
},
{
"epoch": 1.9113914314283518,
"grad_norm": 1.1178765296936035,
"learning_rate": 7.358387658104564e-05,
"loss": 1.4616,
"step": 24850
},
{
"epoch": 1.9152372894392737,
"grad_norm": 2.1088242530822754,
"learning_rate": 7.332415655922915e-05,
"loss": 1.4916,
"step": 24900
},
{
"epoch": 1.9190831474501961,
"grad_norm": 0.7096924781799316,
"learning_rate": 7.306443653741267e-05,
"loss": 1.3687,
"step": 24950
},
{
"epoch": 1.9229290054611183,
"grad_norm": 1.4435713291168213,
"learning_rate": 7.28047165155962e-05,
"loss": 1.5273,
"step": 25000
},
{
"epoch": 1.9229290054611183,
"eval_loss": 1.4638206958770752,
"eval_runtime": 18.4856,
"eval_samples_per_second": 54.096,
"eval_steps_per_second": 13.524,
"step": 25000
},
{
"epoch": 1.9267748634720405,
"grad_norm": 1.4893878698349,
"learning_rate": 7.25449964937797e-05,
"loss": 1.5067,
"step": 25050
},
{
"epoch": 1.930620721482963,
"grad_norm": 0.8735935091972351,
"learning_rate": 7.228527647196322e-05,
"loss": 1.4671,
"step": 25100
},
{
"epoch": 1.9344665794938851,
"grad_norm": 1.6086535453796387,
"learning_rate": 7.202555645014675e-05,
"loss": 1.4551,
"step": 25150
},
{
"epoch": 1.9383124375048073,
"grad_norm": 0.683675229549408,
"learning_rate": 7.176583642833027e-05,
"loss": 1.4673,
"step": 25200
},
{
"epoch": 1.9421582955157297,
"grad_norm": 1.9318158626556396,
"learning_rate": 7.150611640651378e-05,
"loss": 1.4199,
"step": 25250
},
{
"epoch": 1.9421582955157297,
"eval_loss": 1.4574114084243774,
"eval_runtime": 18.5222,
"eval_samples_per_second": 53.989,
"eval_steps_per_second": 13.497,
"step": 25250
},
{
"epoch": 1.9460041535266517,
"grad_norm": 1.9871971607208252,
"learning_rate": 7.12463963846973e-05,
"loss": 1.5002,
"step": 25300
},
{
"epoch": 1.949850011537574,
"grad_norm": 1.4302830696105957,
"learning_rate": 7.098667636288082e-05,
"loss": 1.46,
"step": 25350
},
{
"epoch": 1.9536958695484963,
"grad_norm": 1.8389050960540771,
"learning_rate": 7.072695634106434e-05,
"loss": 1.4025,
"step": 25400
},
{
"epoch": 1.9575417275594185,
"grad_norm": 1.7089191675186157,
"learning_rate": 7.046723631924785e-05,
"loss": 1.4507,
"step": 25450
},
{
"epoch": 1.9613875855703409,
"grad_norm": 1.3698766231536865,
"learning_rate": 7.020751629743138e-05,
"loss": 1.4954,
"step": 25500
},
{
"epoch": 1.9613875855703409,
"eval_loss": 1.454710841178894,
"eval_runtime": 18.5708,
"eval_samples_per_second": 53.848,
"eval_steps_per_second": 13.462,
"step": 25500
},
{
"epoch": 1.9652334435812628,
"grad_norm": 1.808030128479004,
"learning_rate": 6.994779627561489e-05,
"loss": 1.4254,
"step": 25550
},
{
"epoch": 1.9690793015921852,
"grad_norm": 1.634099006652832,
"learning_rate": 6.96880762537984e-05,
"loss": 1.3325,
"step": 25600
},
{
"epoch": 1.9729251596031074,
"grad_norm": 0.8857108354568481,
"learning_rate": 6.942835623198193e-05,
"loss": 1.3983,
"step": 25650
},
{
"epoch": 1.9767710176140296,
"grad_norm": 1.6114498376846313,
"learning_rate": 6.916863621016545e-05,
"loss": 1.3348,
"step": 25700
},
{
"epoch": 1.980616875624952,
"grad_norm": 1.4415462017059326,
"learning_rate": 6.890891618834895e-05,
"loss": 1.4617,
"step": 25750
},
{
"epoch": 1.980616875624952,
"eval_loss": 1.4457746744155884,
"eval_runtime": 17.9567,
"eval_samples_per_second": 55.689,
"eval_steps_per_second": 13.922,
"step": 25750
},
{
"epoch": 1.984462733635874,
"grad_norm": 1.9614554643630981,
"learning_rate": 6.864919616653248e-05,
"loss": 1.3813,
"step": 25800
},
{
"epoch": 1.9883085916467964,
"grad_norm": 1.2938437461853027,
"learning_rate": 6.8389476144716e-05,
"loss": 1.4047,
"step": 25850
},
{
"epoch": 1.9921544496577186,
"grad_norm": 2.1129326820373535,
"learning_rate": 6.812975612289952e-05,
"loss": 1.4362,
"step": 25900
},
{
"epoch": 1.9960003076686408,
"grad_norm": 0.8634279370307922,
"learning_rate": 6.787003610108303e-05,
"loss": 1.4805,
"step": 25950
},
{
"epoch": 1.9998461656795632,
"grad_norm": 2.995699405670166,
"learning_rate": 6.761031607926656e-05,
"loss": 1.5073,
"step": 26000
},
{
"epoch": 1.9998461656795632,
"eval_loss": 1.453719973564148,
"eval_runtime": 18.0595,
"eval_samples_per_second": 55.373,
"eval_steps_per_second": 13.843,
"step": 26000
},
{
"epoch": 2.003692023690485,
"grad_norm": 1.621793508529663,
"learning_rate": 6.73557904578864e-05,
"loss": 1.369,
"step": 26050
},
{
"epoch": 2.0075378817014076,
"grad_norm": 2.223520278930664,
"learning_rate": 6.709607043606992e-05,
"loss": 1.4269,
"step": 26100
},
{
"epoch": 2.01138373971233,
"grad_norm": 1.4860827922821045,
"learning_rate": 6.683635041425344e-05,
"loss": 1.3634,
"step": 26150
},
{
"epoch": 2.015229597723252,
"grad_norm": 2.0796148777008057,
"learning_rate": 6.657663039243696e-05,
"loss": 1.4233,
"step": 26200
},
{
"epoch": 2.0190754557341744,
"grad_norm": 1.6398444175720215,
"learning_rate": 6.631691037062047e-05,
"loss": 1.4058,
"step": 26250
},
{
"epoch": 2.0190754557341744,
"eval_loss": 1.4733901023864746,
"eval_runtime": 18.0349,
"eval_samples_per_second": 55.448,
"eval_steps_per_second": 13.862,
"step": 26250
},
{
"epoch": 2.0229213137450963,
"grad_norm": 1.7550077438354492,
"learning_rate": 6.605719034880399e-05,
"loss": 1.4436,
"step": 26300
},
{
"epoch": 2.0267671717560187,
"grad_norm": 2.3273561000823975,
"learning_rate": 6.579747032698751e-05,
"loss": 1.356,
"step": 26350
},
{
"epoch": 2.030613029766941,
"grad_norm": 1.1432509422302246,
"learning_rate": 6.553775030517103e-05,
"loss": 1.4116,
"step": 26400
},
{
"epoch": 2.034458887777863,
"grad_norm": 1.2345376014709473,
"learning_rate": 6.527803028335455e-05,
"loss": 1.4465,
"step": 26450
},
{
"epoch": 2.0383047457887855,
"grad_norm": 1.485564112663269,
"learning_rate": 6.501831026153807e-05,
"loss": 1.3896,
"step": 26500
},
{
"epoch": 2.0383047457887855,
"eval_loss": 1.4492217302322388,
"eval_runtime": 17.9114,
"eval_samples_per_second": 55.83,
"eval_steps_per_second": 13.958,
"step": 26500
},
{
"epoch": 2.042150603799708,
"grad_norm": 1.0810669660568237,
"learning_rate": 6.475859023972158e-05,
"loss": 1.2936,
"step": 26550
},
{
"epoch": 2.04599646181063,
"grad_norm": 1.23382568359375,
"learning_rate": 6.44988702179051e-05,
"loss": 1.3922,
"step": 26600
},
{
"epoch": 2.0498423198215523,
"grad_norm": 1.8218950033187866,
"learning_rate": 6.423915019608862e-05,
"loss": 1.4041,
"step": 26650
},
{
"epoch": 2.0536881778324743,
"grad_norm": 0.6482899785041809,
"learning_rate": 6.397943017427213e-05,
"loss": 1.3609,
"step": 26700
},
{
"epoch": 2.0575340358433967,
"grad_norm": 1.9538156986236572,
"learning_rate": 6.371971015245566e-05,
"loss": 1.4558,
"step": 26750
},
{
"epoch": 2.0575340358433967,
"eval_loss": 1.454428791999817,
"eval_runtime": 18.2441,
"eval_samples_per_second": 54.812,
"eval_steps_per_second": 13.703,
"step": 26750
},
{
"epoch": 2.061379893854319,
"grad_norm": 1.087234377861023,
"learning_rate": 6.345999013063917e-05,
"loss": 1.3967,
"step": 26800
},
{
"epoch": 2.065225751865241,
"grad_norm": 1.638533592224121,
"learning_rate": 6.32002701088227e-05,
"loss": 1.4558,
"step": 26850
},
{
"epoch": 2.0690716098761635,
"grad_norm": 1.4552900791168213,
"learning_rate": 6.294055008700621e-05,
"loss": 1.4553,
"step": 26900
},
{
"epoch": 2.0729174678870854,
"grad_norm": 1.6330054998397827,
"learning_rate": 6.268083006518974e-05,
"loss": 1.4977,
"step": 26950
},
{
"epoch": 2.076763325898008,
"grad_norm": 1.497938632965088,
"learning_rate": 6.242111004337325e-05,
"loss": 1.4449,
"step": 27000
},
{
"epoch": 2.076763325898008,
"eval_loss": 1.4610888957977295,
"eval_runtime": 17.9657,
"eval_samples_per_second": 55.662,
"eval_steps_per_second": 13.915,
"step": 27000
},
{
"epoch": 2.0806091839089302,
"grad_norm": 0.9695401191711426,
"learning_rate": 6.216139002155676e-05,
"loss": 1.4943,
"step": 27050
},
{
"epoch": 2.084455041919852,
"grad_norm": 1.243717074394226,
"learning_rate": 6.190166999974029e-05,
"loss": 1.4199,
"step": 27100
},
{
"epoch": 2.0883008999307746,
"grad_norm": 1.6151024103164673,
"learning_rate": 6.16419499779238e-05,
"loss": 1.4248,
"step": 27150
},
{
"epoch": 2.0921467579416966,
"grad_norm": 1.7448607683181763,
"learning_rate": 6.138222995610731e-05,
"loss": 1.3944,
"step": 27200
},
{
"epoch": 2.095992615952619,
"grad_norm": 3.0997345447540283,
"learning_rate": 6.112250993429084e-05,
"loss": 1.4174,
"step": 27250
},
{
"epoch": 2.095992615952619,
"eval_loss": 1.4659229516983032,
"eval_runtime": 18.0653,
"eval_samples_per_second": 55.355,
"eval_steps_per_second": 13.839,
"step": 27250
},
{
"epoch": 2.0998384739635414,
"grad_norm": 1.637845754623413,
"learning_rate": 6.086278991247436e-05,
"loss": 1.446,
"step": 27300
},
{
"epoch": 2.1036843319744634,
"grad_norm": 1.4263664484024048,
"learning_rate": 6.060306989065787e-05,
"loss": 0.9793,
"step": 27350
},
{
"epoch": 2.1075301899853858,
"grad_norm": 0.8709418773651123,
"learning_rate": 6.034334986884139e-05,
"loss": 0.9715,
"step": 27400
},
{
"epoch": 2.111376047996308,
"grad_norm": 0.8483341336250305,
"learning_rate": 6.008362984702491e-05,
"loss": 0.9163,
"step": 27450
},
{
"epoch": 2.11522190600723,
"grad_norm": 2.070937156677246,
"learning_rate": 5.9823909825208425e-05,
"loss": 1.0047,
"step": 27500
},
{
"epoch": 2.11522190600723,
"eval_loss": 1.0159448385238647,
"eval_runtime": 17.7021,
"eval_samples_per_second": 56.49,
"eval_steps_per_second": 14.123,
"step": 27500
},
{
"epoch": 2.1190677640181526,
"grad_norm": 0.9638277292251587,
"learning_rate": 5.9564189803391944e-05,
"loss": 0.9753,
"step": 27550
},
{
"epoch": 2.1229136220290745,
"grad_norm": 1.1322181224822998,
"learning_rate": 5.9304469781575464e-05,
"loss": 0.948,
"step": 27600
},
{
"epoch": 2.126759480039997,
"grad_norm": 1.144047737121582,
"learning_rate": 5.904474975975898e-05,
"loss": 1.009,
"step": 27650
},
{
"epoch": 2.1306053380509193,
"grad_norm": 2.713625431060791,
"learning_rate": 5.8785029737942496e-05,
"loss": 1.0069,
"step": 27700
},
{
"epoch": 2.1344511960618413,
"grad_norm": 1.1035822629928589,
"learning_rate": 5.852530971612602e-05,
"loss": 1.0013,
"step": 27750
},
{
"epoch": 2.1344511960618413,
"eval_loss": 1.0144418478012085,
"eval_runtime": 17.6881,
"eval_samples_per_second": 56.535,
"eval_steps_per_second": 14.134,
"step": 27750
},
{
"epoch": 2.1382970540727637,
"grad_norm": 1.2658100128173828,
"learning_rate": 5.826558969430954e-05,
"loss": 1.0185,
"step": 27800
},
{
"epoch": 2.1421429120836857,
"grad_norm": 0.9421238303184509,
"learning_rate": 5.800586967249305e-05,
"loss": 0.9892,
"step": 27850
},
{
"epoch": 2.145988770094608,
"grad_norm": 0.9409565925598145,
"learning_rate": 5.7746149650676575e-05,
"loss": 0.9985,
"step": 27900
},
{
"epoch": 2.1498346281055305,
"grad_norm": 1.445890188217163,
"learning_rate": 5.7486429628860094e-05,
"loss": 1.0147,
"step": 27950
},
{
"epoch": 2.1536804861164525,
"grad_norm": 1.109020709991455,
"learning_rate": 5.722670960704361e-05,
"loss": 1.0093,
"step": 28000
},
{
"epoch": 2.1536804861164525,
"eval_loss": 1.008616328239441,
"eval_runtime": 17.6489,
"eval_samples_per_second": 56.661,
"eval_steps_per_second": 14.165,
"step": 28000
},
{
"epoch": 2.157526344127375,
"grad_norm": 1.3012685775756836,
"learning_rate": 5.696698958522713e-05,
"loss": 1.0703,
"step": 28050
},
{
"epoch": 2.1613722021382973,
"grad_norm": 0.8459142446517944,
"learning_rate": 5.6707269563410646e-05,
"loss": 1.0507,
"step": 28100
},
{
"epoch": 2.1652180601492192,
"grad_norm": 0.6889505982398987,
"learning_rate": 5.6447549541594166e-05,
"loss": 1.0056,
"step": 28150
},
{
"epoch": 2.1690639181601417,
"grad_norm": 1.2236456871032715,
"learning_rate": 5.618782951977768e-05,
"loss": 0.988,
"step": 28200
},
{
"epoch": 2.1729097761710636,
"grad_norm": 1.3419203758239746,
"learning_rate": 5.59281094979612e-05,
"loss": 0.9467,
"step": 28250
},
{
"epoch": 2.1729097761710636,
"eval_loss": 1.0140153169631958,
"eval_runtime": 17.7299,
"eval_samples_per_second": 56.402,
"eval_steps_per_second": 14.1,
"step": 28250
},
{
"epoch": 2.176755634181986,
"grad_norm": 0.8394871354103088,
"learning_rate": 5.5668389476144725e-05,
"loss": 0.9824,
"step": 28300
},
{
"epoch": 2.1806014921929084,
"grad_norm": 0.7906908392906189,
"learning_rate": 5.540866945432823e-05,
"loss": 0.9915,
"step": 28350
},
{
"epoch": 2.1844473502038304,
"grad_norm": 1.0852785110473633,
"learning_rate": 5.514894943251175e-05,
"loss": 0.9473,
"step": 28400
},
{
"epoch": 2.188293208214753,
"grad_norm": 0.9187583923339844,
"learning_rate": 5.488922941069528e-05,
"loss": 0.9418,
"step": 28450
},
{
"epoch": 2.192139066225675,
"grad_norm": 0.7200838923454285,
"learning_rate": 5.4629509388878797e-05,
"loss": 0.9776,
"step": 28500
},
{
"epoch": 2.192139066225675,
"eval_loss": 1.0052642822265625,
"eval_runtime": 17.5069,
"eval_samples_per_second": 57.12,
"eval_steps_per_second": 14.28,
"step": 28500
},
{
"epoch": 2.195984924236597,
"grad_norm": 1.00751793384552,
"learning_rate": 5.436978936706231e-05,
"loss": 0.9479,
"step": 28550
},
{
"epoch": 2.1998307822475196,
"grad_norm": 0.9239784479141235,
"learning_rate": 5.411006934524583e-05,
"loss": 0.9896,
"step": 28600
},
{
"epoch": 2.2036766402584416,
"grad_norm": 0.9650816917419434,
"learning_rate": 5.385034932342935e-05,
"loss": 1.0081,
"step": 28650
},
{
"epoch": 2.207522498269364,
"grad_norm": 1.455723524093628,
"learning_rate": 5.359062930161286e-05,
"loss": 0.9717,
"step": 28700
},
{
"epoch": 2.211368356280286,
"grad_norm": 0.609380304813385,
"learning_rate": 5.333090927979638e-05,
"loss": 0.9774,
"step": 28750
},
{
"epoch": 2.211368356280286,
"eval_loss": 1.0018259286880493,
"eval_runtime": 17.3475,
"eval_samples_per_second": 57.645,
"eval_steps_per_second": 14.411,
"step": 28750
},
{
"epoch": 2.2152142142912084,
"grad_norm": 0.9981700778007507,
"learning_rate": 5.30711892579799e-05,
"loss": 0.9873,
"step": 28800
},
{
"epoch": 2.2190600723021308,
"grad_norm": 0.5452422499656677,
"learning_rate": 5.2811469236163414e-05,
"loss": 1.0078,
"step": 28850
},
{
"epoch": 2.2229059303130527,
"grad_norm": 1.4739840030670166,
"learning_rate": 5.255174921434693e-05,
"loss": 0.9739,
"step": 28900
},
{
"epoch": 2.226751788323975,
"grad_norm": 0.9250359535217285,
"learning_rate": 5.229202919253045e-05,
"loss": 0.9723,
"step": 28950
},
{
"epoch": 2.230597646334897,
"grad_norm": 1.0843122005462646,
"learning_rate": 5.203230917071398e-05,
"loss": 0.989,
"step": 29000
},
{
"epoch": 2.230597646334897,
"eval_loss": 0.9999537467956543,
"eval_runtime": 17.2417,
"eval_samples_per_second": 57.999,
"eval_steps_per_second": 14.5,
"step": 29000
},
{
"epoch": 2.2344435043458195,
"grad_norm": 0.5559306740760803,
"learning_rate": 5.1772589148897485e-05,
"loss": 0.9749,
"step": 29050
},
{
"epoch": 2.238289362356742,
"grad_norm": 0.6720598936080933,
"learning_rate": 5.151286912708101e-05,
"loss": 0.9662,
"step": 29100
},
{
"epoch": 2.242135220367664,
"grad_norm": 1.117200493812561,
"learning_rate": 5.125314910526453e-05,
"loss": 0.9552,
"step": 29150
},
{
"epoch": 2.2459810783785863,
"grad_norm": 0.6818645000457764,
"learning_rate": 5.0993429083448044e-05,
"loss": 0.9032,
"step": 29200
},
{
"epoch": 2.2498269363895087,
"grad_norm": 0.9796412587165833,
"learning_rate": 5.0733709061631564e-05,
"loss": 0.9429,
"step": 29250
},
{
"epoch": 2.2498269363895087,
"eval_loss": 1.0121312141418457,
"eval_runtime": 17.3211,
"eval_samples_per_second": 57.733,
"eval_steps_per_second": 14.433,
"step": 29250
},
{
"epoch": 2.2536727944004307,
"grad_norm": 1.021713137626648,
"learning_rate": 5.047398903981508e-05,
"loss": 0.979,
"step": 29300
},
{
"epoch": 2.257518652411353,
"grad_norm": 1.1321250200271606,
"learning_rate": 5.02142690179986e-05,
"loss": 1.0327,
"step": 29350
},
{
"epoch": 2.261364510422275,
"grad_norm": 0.7670277953147888,
"learning_rate": 4.9954548996182116e-05,
"loss": 0.9668,
"step": 29400
},
{
"epoch": 2.2652103684331975,
"grad_norm": 1.447698712348938,
"learning_rate": 4.9694828974365635e-05,
"loss": 0.9288,
"step": 29450
},
{
"epoch": 2.26905622644412,
"grad_norm": 1.0438776016235352,
"learning_rate": 4.9435108952549155e-05,
"loss": 1.0154,
"step": 29500
},
{
"epoch": 2.26905622644412,
"eval_loss": 1.0100510120391846,
"eval_runtime": 17.4293,
"eval_samples_per_second": 57.375,
"eval_steps_per_second": 14.344,
"step": 29500
},
{
"epoch": 2.272902084455042,
"grad_norm": 0.6814424991607666,
"learning_rate": 4.9175388930732675e-05,
"loss": 0.9563,
"step": 29550
},
{
"epoch": 2.2767479424659642,
"grad_norm": 0.85086989402771,
"learning_rate": 4.891566890891619e-05,
"loss": 0.9778,
"step": 29600
},
{
"epoch": 2.280593800476886,
"grad_norm": 1.1364527940750122,
"learning_rate": 4.8655948887099714e-05,
"loss": 0.996,
"step": 29650
},
{
"epoch": 2.2844396584878086,
"grad_norm": 0.7810873985290527,
"learning_rate": 4.839622886528323e-05,
"loss": 1.0141,
"step": 29700
},
{
"epoch": 2.288285516498731,
"grad_norm": 1.02638578414917,
"learning_rate": 4.8136508843466746e-05,
"loss": 0.97,
"step": 29750
},
{
"epoch": 2.288285516498731,
"eval_loss": 1.0048705339431763,
"eval_runtime": 17.487,
"eval_samples_per_second": 57.185,
"eval_steps_per_second": 14.296,
"step": 29750
},
{
"epoch": 2.292131374509653,
"grad_norm": 1.334876298904419,
"learning_rate": 4.7876788821650266e-05,
"loss": 0.9533,
"step": 29800
},
{
"epoch": 2.2959772325205754,
"grad_norm": 1.3690305948257446,
"learning_rate": 4.761706879983378e-05,
"loss": 0.992,
"step": 29850
},
{
"epoch": 2.299823090531498,
"grad_norm": 1.525981068611145,
"learning_rate": 4.7357348778017305e-05,
"loss": 0.949,
"step": 29900
},
{
"epoch": 2.3036689485424198,
"grad_norm": 0.6612236499786377,
"learning_rate": 4.709762875620082e-05,
"loss": 1.0112,
"step": 29950
},
{
"epoch": 2.307514806553342,
"grad_norm": 0.8452871441841125,
"learning_rate": 4.683790873438433e-05,
"loss": 0.9858,
"step": 30000
},
{
"epoch": 2.307514806553342,
"eval_loss": 0.9945911169052124,
"eval_runtime": 17.2501,
"eval_samples_per_second": 57.971,
"eval_steps_per_second": 14.493,
"step": 30000
},
{
"epoch": 2.311360664564264,
"grad_norm": 1.3450604677200317,
"learning_rate": 4.657818871256786e-05,
"loss": 0.973,
"step": 30050
},
{
"epoch": 2.3152065225751866,
"grad_norm": 1.1462957859039307,
"learning_rate": 4.631846869075137e-05,
"loss": 0.9789,
"step": 30100
},
{
"epoch": 2.3190523805861085,
"grad_norm": 1.0691301822662354,
"learning_rate": 4.605874866893489e-05,
"loss": 0.9937,
"step": 30150
},
{
"epoch": 2.322898238597031,
"grad_norm": 1.527723789215088,
"learning_rate": 4.579902864711841e-05,
"loss": 0.9276,
"step": 30200
},
{
"epoch": 2.3267440966079533,
"grad_norm": 0.7396986484527588,
"learning_rate": 4.553930862530192e-05,
"loss": 0.9648,
"step": 30250
},
{
"epoch": 2.3267440966079533,
"eval_loss": 1.0029717683792114,
"eval_runtime": 17.2662,
"eval_samples_per_second": 57.917,
"eval_steps_per_second": 14.479,
"step": 30250
},
{
"epoch": 2.3305899546188753,
"grad_norm": 0.9694539308547974,
"learning_rate": 4.527958860348545e-05,
"loss": 0.9378,
"step": 30300
},
{
"epoch": 2.3344358126297977,
"grad_norm": 1.161028504371643,
"learning_rate": 4.501986858166896e-05,
"loss": 0.9807,
"step": 30350
},
{
"epoch": 2.33828167064072,
"grad_norm": 1.6975845098495483,
"learning_rate": 4.476014855985248e-05,
"loss": 0.9809,
"step": 30400
},
{
"epoch": 2.342127528651642,
"grad_norm": 1.3228962421417236,
"learning_rate": 4.4500428538036e-05,
"loss": 1.0339,
"step": 30450
},
{
"epoch": 2.3459733866625645,
"grad_norm": 1.0168397426605225,
"learning_rate": 4.424070851621952e-05,
"loss": 0.9544,
"step": 30500
},
{
"epoch": 2.3459733866625645,
"eval_loss": 1.002646565437317,
"eval_runtime": 17.3363,
"eval_samples_per_second": 57.682,
"eval_steps_per_second": 14.421,
"step": 30500
},
{
"epoch": 2.3498192446734865,
"grad_norm": 0.9951680302619934,
"learning_rate": 4.398098849440303e-05,
"loss": 1.0062,
"step": 30550
},
{
"epoch": 2.353665102684409,
"grad_norm": 0.5752933025360107,
"learning_rate": 4.372126847258655e-05,
"loss": 0.9649,
"step": 30600
},
{
"epoch": 2.3575109606953313,
"grad_norm": 1.0051320791244507,
"learning_rate": 4.346154845077007e-05,
"loss": 0.9958,
"step": 30650
},
{
"epoch": 2.3613568187062532,
"grad_norm": 0.7760717868804932,
"learning_rate": 4.320182842895359e-05,
"loss": 1.0125,
"step": 30700
},
{
"epoch": 2.3652026767171757,
"grad_norm": 0.852301836013794,
"learning_rate": 4.294210840713711e-05,
"loss": 0.9523,
"step": 30750
},
{
"epoch": 2.3652026767171757,
"eval_loss": 1.003655195236206,
"eval_runtime": 17.3241,
"eval_samples_per_second": 57.723,
"eval_steps_per_second": 14.431,
"step": 30750
},
{
"epoch": 2.3690485347280976,
"grad_norm": 0.9062100648880005,
"learning_rate": 4.2682388385320624e-05,
"loss": 0.9735,
"step": 30800
},
{
"epoch": 2.37289439273902,
"grad_norm": 1.309615969657898,
"learning_rate": 4.2427862763940476e-05,
"loss": 0.9668,
"step": 30850
},
{
"epoch": 2.3767402507499424,
"grad_norm": 1.0907591581344604,
"learning_rate": 4.2168142742123995e-05,
"loss": 0.9502,
"step": 30900
},
{
"epoch": 2.3805861087608644,
"grad_norm": 1.0946288108825684,
"learning_rate": 4.190842272030751e-05,
"loss": 0.9545,
"step": 30950
},
{
"epoch": 2.384431966771787,
"grad_norm": 1.225540280342102,
"learning_rate": 4.164870269849103e-05,
"loss": 0.9635,
"step": 31000
},
{
"epoch": 2.384431966771787,
"eval_loss": 1.0031681060791016,
"eval_runtime": 17.2062,
"eval_samples_per_second": 58.119,
"eval_steps_per_second": 14.53,
"step": 31000
},
{
"epoch": 2.3882778247827092,
"grad_norm": 0.8717153668403625,
"learning_rate": 4.138898267667455e-05,
"loss": 1.0038,
"step": 31050
},
{
"epoch": 2.392123682793631,
"grad_norm": 1.007270097732544,
"learning_rate": 4.112926265485807e-05,
"loss": 0.9858,
"step": 31100
},
{
"epoch": 2.3959695408045536,
"grad_norm": 1.9409807920455933,
"learning_rate": 4.0869542633041587e-05,
"loss": 1.004,
"step": 31150
},
{
"epoch": 2.3998153988154756,
"grad_norm": 0.6027572154998779,
"learning_rate": 4.06098226112251e-05,
"loss": 0.9958,
"step": 31200
},
{
"epoch": 2.403661256826398,
"grad_norm": 0.8274515867233276,
"learning_rate": 4.035010258940862e-05,
"loss": 0.9852,
"step": 31250
},
{
"epoch": 2.403661256826398,
"eval_loss": 1.0001976490020752,
"eval_runtime": 17.2233,
"eval_samples_per_second": 58.061,
"eval_steps_per_second": 14.515,
"step": 31250
},
{
"epoch": 2.4075071148373204,
"grad_norm": 0.7238942384719849,
"learning_rate": 4.009038256759214e-05,
"loss": 1.0062,
"step": 31300
},
{
"epoch": 2.4113529728482423,
"grad_norm": 0.8912849426269531,
"learning_rate": 3.983066254577565e-05,
"loss": 0.9802,
"step": 31350
},
{
"epoch": 2.4151988308591648,
"grad_norm": 1.1922829151153564,
"learning_rate": 3.957094252395918e-05,
"loss": 1.0317,
"step": 31400
},
{
"epoch": 2.4190446888700867,
"grad_norm": 1.3773999214172363,
"learning_rate": 3.931122250214269e-05,
"loss": 0.9773,
"step": 31450
},
{
"epoch": 2.422890546881009,
"grad_norm": 1.0747745037078857,
"learning_rate": 3.905150248032621e-05,
"loss": 1.0208,
"step": 31500
},
{
"epoch": 2.422890546881009,
"eval_loss": 1.0009056329727173,
"eval_runtime": 17.1069,
"eval_samples_per_second": 58.456,
"eval_steps_per_second": 14.614,
"step": 31500
},
{
"epoch": 2.4267364048919315,
"grad_norm": 1.3409994840621948,
"learning_rate": 3.879178245850973e-05,
"loss": 0.9546,
"step": 31550
},
{
"epoch": 2.4305822629028535,
"grad_norm": 1.482633352279663,
"learning_rate": 3.853206243669324e-05,
"loss": 1.0155,
"step": 31600
},
{
"epoch": 2.434428120913776,
"grad_norm": 1.0458152294158936,
"learning_rate": 3.827234241487677e-05,
"loss": 0.989,
"step": 31650
},
{
"epoch": 2.4382739789246983,
"grad_norm": 0.9805555939674377,
"learning_rate": 3.801262239306028e-05,
"loss": 0.9125,
"step": 31700
},
{
"epoch": 2.4421198369356203,
"grad_norm": 1.5680670738220215,
"learning_rate": 3.77529023712438e-05,
"loss": 0.9713,
"step": 31750
},
{
"epoch": 2.4421198369356203,
"eval_loss": 0.9981379508972168,
"eval_runtime": 17.3284,
"eval_samples_per_second": 57.709,
"eval_steps_per_second": 14.427,
"step": 31750
},
{
"epoch": 2.4459656949465427,
"grad_norm": 1.0316798686981201,
"learning_rate": 3.749318234942732e-05,
"loss": 0.9689,
"step": 31800
},
{
"epoch": 2.4498115529574647,
"grad_norm": 0.6630721092224121,
"learning_rate": 3.7233462327610834e-05,
"loss": 0.9772,
"step": 31850
},
{
"epoch": 2.453657410968387,
"grad_norm": 1.1662702560424805,
"learning_rate": 3.6973742305794354e-05,
"loss": 1.0165,
"step": 31900
},
{
"epoch": 2.457503268979309,
"grad_norm": 1.0451244115829468,
"learning_rate": 3.671402228397787e-05,
"loss": 1.0233,
"step": 31950
},
{
"epoch": 2.4613491269902315,
"grad_norm": 0.9077771306037903,
"learning_rate": 3.645430226216139e-05,
"loss": 0.9797,
"step": 32000
},
{
"epoch": 2.4613491269902315,
"eval_loss": 0.9961766004562378,
"eval_runtime": 17.2799,
"eval_samples_per_second": 57.871,
"eval_steps_per_second": 14.468,
"step": 32000
},
{
"epoch": 2.465194985001154,
"grad_norm": 0.999718964099884,
"learning_rate": 3.619458224034491e-05,
"loss": 0.9725,
"step": 32050
},
{
"epoch": 2.469040843012076,
"grad_norm": 2.740297794342041,
"learning_rate": 3.5934862218528425e-05,
"loss": 0.9838,
"step": 32100
},
{
"epoch": 2.4728867010229982,
"grad_norm": 1.199425458908081,
"learning_rate": 3.5675142196711945e-05,
"loss": 0.9807,
"step": 32150
},
{
"epoch": 2.4767325590339206,
"grad_norm": 0.7113758325576782,
"learning_rate": 3.5415422174895465e-05,
"loss": 1.0046,
"step": 32200
},
{
"epoch": 2.4805784170448426,
"grad_norm": 0.9929390549659729,
"learning_rate": 3.5155702153078984e-05,
"loss": 0.9692,
"step": 32250
},
{
"epoch": 2.4805784170448426,
"eval_loss": 1.002519965171814,
"eval_runtime": 17.3042,
"eval_samples_per_second": 57.789,
"eval_steps_per_second": 14.447,
"step": 32250
},
{
"epoch": 2.484424275055765,
"grad_norm": 0.8170703649520874,
"learning_rate": 3.48959821312625e-05,
"loss": 0.982,
"step": 32300
},
{
"epoch": 2.488270133066687,
"grad_norm": 0.8909692168235779,
"learning_rate": 3.463626210944602e-05,
"loss": 0.9953,
"step": 32350
},
{
"epoch": 2.4921159910776094,
"grad_norm": 1.806539535522461,
"learning_rate": 3.4376542087629536e-05,
"loss": 1.0069,
"step": 32400
},
{
"epoch": 2.495961849088532,
"grad_norm": 1.4509518146514893,
"learning_rate": 3.4116822065813056e-05,
"loss": 0.9642,
"step": 32450
},
{
"epoch": 2.4998077070994538,
"grad_norm": 1.0312175750732422,
"learning_rate": 3.3857102043996575e-05,
"loss": 0.9921,
"step": 32500
},
{
"epoch": 2.4998077070994538,
"eval_loss": 0.9993879199028015,
"eval_runtime": 17.3389,
"eval_samples_per_second": 57.674,
"eval_steps_per_second": 14.418,
"step": 32500
},
{
"epoch": 2.503653565110376,
"grad_norm": 0.867513120174408,
"learning_rate": 3.359738202218009e-05,
"loss": 0.994,
"step": 32550
},
{
"epoch": 2.507499423121298,
"grad_norm": 1.7425885200500488,
"learning_rate": 3.3337662000363615e-05,
"loss": 1.0068,
"step": 32600
},
{
"epoch": 2.5113452811322206,
"grad_norm": 0.9053608775138855,
"learning_rate": 3.307794197854713e-05,
"loss": 0.9704,
"step": 32650
},
{
"epoch": 2.515191139143143,
"grad_norm": 1.0533051490783691,
"learning_rate": 3.281822195673064e-05,
"loss": 0.9506,
"step": 32700
},
{
"epoch": 2.519036997154065,
"grad_norm": 1.2495230436325073,
"learning_rate": 3.255850193491417e-05,
"loss": 0.9936,
"step": 32750
},
{
"epoch": 2.519036997154065,
"eval_loss": 0.9990929961204529,
"eval_runtime": 17.239,
"eval_samples_per_second": 58.008,
"eval_steps_per_second": 14.502,
"step": 32750
},
{
"epoch": 2.5228828551649873,
"grad_norm": 0.8339760303497314,
"learning_rate": 3.229878191309768e-05,
"loss": 1.0008,
"step": 32800
},
{
"epoch": 2.5267287131759097,
"grad_norm": 1.2839399576187134,
"learning_rate": 3.20390618912812e-05,
"loss": 1.0109,
"step": 32850
},
{
"epoch": 2.5305745711868317,
"grad_norm": 1.2146100997924805,
"learning_rate": 3.177934186946472e-05,
"loss": 0.9834,
"step": 32900
},
{
"epoch": 2.534420429197754,
"grad_norm": 0.7952923774719238,
"learning_rate": 3.151962184764823e-05,
"loss": 1.0013,
"step": 32950
},
{
"epoch": 2.5382662872086765,
"grad_norm": 1.67001211643219,
"learning_rate": 3.125990182583176e-05,
"loss": 0.9797,
"step": 33000
},
{
"epoch": 2.5382662872086765,
"eval_loss": 1.001037359237671,
"eval_runtime": 17.347,
"eval_samples_per_second": 57.647,
"eval_steps_per_second": 14.412,
"step": 33000
},
{
"epoch": 2.5421121452195985,
"grad_norm": 0.7536977529525757,
"learning_rate": 3.100018180401527e-05,
"loss": 0.9818,
"step": 33050
},
{
"epoch": 2.5459580032305205,
"grad_norm": 1.3777302503585815,
"learning_rate": 3.074046178219879e-05,
"loss": 0.953,
"step": 33100
},
{
"epoch": 2.549803861241443,
"grad_norm": 0.9286239743232727,
"learning_rate": 3.048074176038231e-05,
"loss": 1.028,
"step": 33150
},
{
"epoch": 2.5536497192523653,
"grad_norm": 0.6741893291473389,
"learning_rate": 3.022102173856583e-05,
"loss": 0.9693,
"step": 33200
},
{
"epoch": 2.5574955772632872,
"grad_norm": 1.3020586967468262,
"learning_rate": 2.9961301716749346e-05,
"loss": 0.9673,
"step": 33250
},
{
"epoch": 2.5574955772632872,
"eval_loss": 1.0041394233703613,
"eval_runtime": 17.32,
"eval_samples_per_second": 57.737,
"eval_steps_per_second": 14.434,
"step": 33250
},
{
"epoch": 2.5613414352742097,
"grad_norm": 1.4816234111785889,
"learning_rate": 2.9701581694932862e-05,
"loss": 0.9964,
"step": 33300
},
{
"epoch": 2.565187293285132,
"grad_norm": 0.8813285827636719,
"learning_rate": 2.9441861673116382e-05,
"loss": 0.966,
"step": 33350
},
{
"epoch": 2.569033151296054,
"grad_norm": 1.0042293071746826,
"learning_rate": 2.9182141651299898e-05,
"loss": 0.9254,
"step": 33400
},
{
"epoch": 2.5728790093069764,
"grad_norm": 1.5775707960128784,
"learning_rate": 2.892242162948342e-05,
"loss": 1.0215,
"step": 33450
},
{
"epoch": 2.576724867317899,
"grad_norm": 1.0266311168670654,
"learning_rate": 2.8662701607666937e-05,
"loss": 0.9323,
"step": 33500
},
{
"epoch": 2.576724867317899,
"eval_loss": 1.0002070665359497,
"eval_runtime": 17.2317,
"eval_samples_per_second": 58.033,
"eval_steps_per_second": 14.508,
"step": 33500
},
{
"epoch": 2.580570725328821,
"grad_norm": 0.5430648922920227,
"learning_rate": 2.8402981585850453e-05,
"loss": 0.9658,
"step": 33550
},
{
"epoch": 2.5844165833397432,
"grad_norm": 0.6678454279899597,
"learning_rate": 2.8143261564033973e-05,
"loss": 1.0292,
"step": 33600
},
{
"epoch": 2.588262441350665,
"grad_norm": 0.7208724021911621,
"learning_rate": 2.788354154221749e-05,
"loss": 0.9505,
"step": 33650
},
{
"epoch": 2.5921082993615876,
"grad_norm": 1.2248526811599731,
"learning_rate": 2.7623821520401012e-05,
"loss": 0.9728,
"step": 33700
},
{
"epoch": 2.5959541573725096,
"grad_norm": 1.0026588439941406,
"learning_rate": 2.7364101498584525e-05,
"loss": 0.9783,
"step": 33750
},
{
"epoch": 2.5959541573725096,
"eval_loss": 0.9881900548934937,
"eval_runtime": 17.2742,
"eval_samples_per_second": 57.89,
"eval_steps_per_second": 14.472,
"step": 33750
},
{
"epoch": 2.599800015383432,
"grad_norm": 0.9579987525939941,
"learning_rate": 2.710438147676804e-05,
"loss": 0.9882,
"step": 33800
},
{
"epoch": 2.6036458733943544,
"grad_norm": 1.0152076482772827,
"learning_rate": 2.6844661454951564e-05,
"loss": 1.041,
"step": 33850
},
{
"epoch": 2.6074917314052763,
"grad_norm": 1.1370351314544678,
"learning_rate": 2.658494143313508e-05,
"loss": 0.9518,
"step": 33900
},
{
"epoch": 2.6113375894161988,
"grad_norm": 0.9851937890052795,
"learning_rate": 2.63252214113186e-05,
"loss": 0.9125,
"step": 33950
},
{
"epoch": 2.615183447427121,
"grad_norm": 0.8480270504951477,
"learning_rate": 2.6065501389502116e-05,
"loss": 0.9736,
"step": 34000
},
{
"epoch": 2.615183447427121,
"eval_loss": 0.987713634967804,
"eval_runtime": 17.2436,
"eval_samples_per_second": 57.993,
"eval_steps_per_second": 14.498,
"step": 34000
},
{
"epoch": 2.619029305438043,
"grad_norm": 0.5307362079620361,
"learning_rate": 2.580578136768564e-05,
"loss": 1.0031,
"step": 34050
},
{
"epoch": 2.6228751634489655,
"grad_norm": 1.1112557649612427,
"learning_rate": 2.5546061345869156e-05,
"loss": 0.9928,
"step": 34100
},
{
"epoch": 2.626721021459888,
"grad_norm": 0.646759569644928,
"learning_rate": 2.5286341324052672e-05,
"loss": 1.0173,
"step": 34150
},
{
"epoch": 2.63056687947081,
"grad_norm": 1.0114878416061401,
"learning_rate": 2.502662130223619e-05,
"loss": 0.9765,
"step": 34200
},
{
"epoch": 2.6344127374817323,
"grad_norm": 0.8782021403312683,
"learning_rate": 2.4766901280419708e-05,
"loss": 0.9926,
"step": 34250
},
{
"epoch": 2.6344127374817323,
"eval_loss": 0.9932020306587219,
"eval_runtime": 17.1921,
"eval_samples_per_second": 58.166,
"eval_steps_per_second": 14.542,
"step": 34250
},
{
"epoch": 2.6382585954926543,
"grad_norm": 1.0792268514633179,
"learning_rate": 2.4507181258603227e-05,
"loss": 1.0541,
"step": 34300
},
{
"epoch": 2.6421044535035767,
"grad_norm": 0.9647793173789978,
"learning_rate": 2.4247461236786744e-05,
"loss": 0.9779,
"step": 34350
},
{
"epoch": 2.6459503115144987,
"grad_norm": 1.0052498579025269,
"learning_rate": 2.3987741214970263e-05,
"loss": 0.9798,
"step": 34400
},
{
"epoch": 2.649796169525421,
"grad_norm": 0.8122023344039917,
"learning_rate": 2.3728021193153783e-05,
"loss": 1.0205,
"step": 34450
},
{
"epoch": 2.6536420275363435,
"grad_norm": 1.452087163925171,
"learning_rate": 2.3468301171337302e-05,
"loss": 0.9745,
"step": 34500
},
{
"epoch": 2.6536420275363435,
"eval_loss": 0.9959968328475952,
"eval_runtime": 17.2025,
"eval_samples_per_second": 58.131,
"eval_steps_per_second": 14.533,
"step": 34500
},
{
"epoch": 2.6574878855472654,
"grad_norm": 1.3329054117202759,
"learning_rate": 2.3208581149520815e-05,
"loss": 1.0035,
"step": 34550
},
{
"epoch": 2.661333743558188,
"grad_norm": 0.8142715096473694,
"learning_rate": 2.2948861127704335e-05,
"loss": 0.9783,
"step": 34600
},
{
"epoch": 2.6651796015691103,
"grad_norm": 0.5424798130989075,
"learning_rate": 2.2689141105887855e-05,
"loss": 0.9754,
"step": 34650
},
{
"epoch": 2.6690254595800322,
"grad_norm": 0.8890462517738342,
"learning_rate": 2.2429421084071374e-05,
"loss": 0.9707,
"step": 34700
},
{
"epoch": 2.6728713175909546,
"grad_norm": 1.0329838991165161,
"learning_rate": 2.216970106225489e-05,
"loss": 0.9845,
"step": 34750
},
{
"epoch": 2.6728713175909546,
"eval_loss": 0.9899721741676331,
"eval_runtime": 17.3008,
"eval_samples_per_second": 57.801,
"eval_steps_per_second": 14.45,
"step": 34750
},
{
"epoch": 2.676717175601877,
"grad_norm": 1.0697598457336426,
"learning_rate": 2.190998104043841e-05,
"loss": 1.0167,
"step": 34800
},
{
"epoch": 2.680563033612799,
"grad_norm": 0.8134399056434631,
"learning_rate": 2.1650261018621926e-05,
"loss": 0.9001,
"step": 34850
},
{
"epoch": 2.684408891623721,
"grad_norm": 1.0405962467193604,
"learning_rate": 2.1390540996805446e-05,
"loss": 1.0184,
"step": 34900
},
{
"epoch": 2.6882547496346434,
"grad_norm": 1.3229318857192993,
"learning_rate": 2.1130820974988962e-05,
"loss": 0.9556,
"step": 34950
},
{
"epoch": 2.692100607645566,
"grad_norm": 1.2907413244247437,
"learning_rate": 2.087110095317248e-05,
"loss": 0.9355,
"step": 35000
},
{
"epoch": 2.692100607645566,
"eval_loss": 0.9918172359466553,
"eval_runtime": 17.2562,
"eval_samples_per_second": 57.95,
"eval_steps_per_second": 14.488,
"step": 35000
},
{
"epoch": 2.6959464656564878,
"grad_norm": 1.3044216632843018,
"learning_rate": 2.0611380931356e-05,
"loss": 0.9787,
"step": 35050
},
{
"epoch": 2.69979232366741,
"grad_norm": 1.4273097515106201,
"learning_rate": 2.0351660909539518e-05,
"loss": 0.9775,
"step": 35100
},
{
"epoch": 2.7036381816783326,
"grad_norm": 1.6097638607025146,
"learning_rate": 2.0091940887723034e-05,
"loss": 0.9926,
"step": 35150
},
{
"epoch": 2.7074840396892546,
"grad_norm": 1.4893895387649536,
"learning_rate": 1.9832220865906553e-05,
"loss": 1.0435,
"step": 35200
},
{
"epoch": 2.711329897700177,
"grad_norm": 1.1620233058929443,
"learning_rate": 1.9572500844090073e-05,
"loss": 0.9525,
"step": 35250
},
{
"epoch": 2.711329897700177,
"eval_loss": 0.9926208257675171,
"eval_runtime": 17.3977,
"eval_samples_per_second": 57.479,
"eval_steps_per_second": 14.37,
"step": 35250
},
{
"epoch": 2.7151757557110994,
"grad_norm": 1.2087517976760864,
"learning_rate": 1.931278082227359e-05,
"loss": 1.0005,
"step": 35300
},
{
"epoch": 2.7190216137220213,
"grad_norm": 0.8885460495948792,
"learning_rate": 1.9058255200893437e-05,
"loss": 0.9911,
"step": 35350
},
{
"epoch": 2.7228674717329437,
"grad_norm": 0.9543077945709229,
"learning_rate": 1.8798535179076957e-05,
"loss": 1.0437,
"step": 35400
},
{
"epoch": 2.7267133297438657,
"grad_norm": 1.2362306118011475,
"learning_rate": 1.8538815157260476e-05,
"loss": 0.9766,
"step": 35450
},
{
"epoch": 2.730559187754788,
"grad_norm": 1.1946227550506592,
"learning_rate": 1.8279095135443993e-05,
"loss": 1.0333,
"step": 35500
},
{
"epoch": 2.730559187754788,
"eval_loss": 0.9953948855400085,
"eval_runtime": 17.1888,
"eval_samples_per_second": 58.178,
"eval_steps_per_second": 14.544,
"step": 35500
},
{
"epoch": 2.73440504576571,
"grad_norm": 0.66939777135849,
"learning_rate": 1.801937511362751e-05,
"loss": 0.8892,
"step": 35550
},
{
"epoch": 2.7382509037766325,
"grad_norm": 1.0852998495101929,
"learning_rate": 1.775965509181103e-05,
"loss": 0.9502,
"step": 35600
},
{
"epoch": 2.742096761787555,
"grad_norm": 0.7603423595428467,
"learning_rate": 1.7499935069994548e-05,
"loss": 1.0034,
"step": 35650
},
{
"epoch": 2.745942619798477,
"grad_norm": 0.4625702202320099,
"learning_rate": 1.7240215048178064e-05,
"loss": 1.0123,
"step": 35700
},
{
"epoch": 2.7497884778093993,
"grad_norm": 1.4387953281402588,
"learning_rate": 1.6980495026361584e-05,
"loss": 0.9704,
"step": 35750
},
{
"epoch": 2.7497884778093993,
"eval_loss": 0.9952225685119629,
"eval_runtime": 17.4458,
"eval_samples_per_second": 57.32,
"eval_steps_per_second": 14.33,
"step": 35750
},
{
"epoch": 2.7536343358203217,
"grad_norm": 0.6899126172065735,
"learning_rate": 1.67207750045451e-05,
"loss": 0.9627,
"step": 35800
},
{
"epoch": 2.7574801938312437,
"grad_norm": 1.0329424142837524,
"learning_rate": 1.646105498272862e-05,
"loss": 0.9207,
"step": 35850
},
{
"epoch": 2.761326051842166,
"grad_norm": 1.1055504083633423,
"learning_rate": 1.6201334960912136e-05,
"loss": 0.9834,
"step": 35900
},
{
"epoch": 2.7651719098530885,
"grad_norm": 0.7458188533782959,
"learning_rate": 1.5941614939095655e-05,
"loss": 1.003,
"step": 35950
},
{
"epoch": 2.7690177678640104,
"grad_norm": 1.112021803855896,
"learning_rate": 1.5681894917279175e-05,
"loss": 1.0001,
"step": 36000
},
{
"epoch": 2.7690177678640104,
"eval_loss": 0.9910063147544861,
"eval_runtime": 17.2718,
"eval_samples_per_second": 57.898,
"eval_steps_per_second": 14.474,
"step": 36000
},
{
"epoch": 2.772863625874933,
"grad_norm": 0.5958703756332397,
"learning_rate": 1.5422174895462695e-05,
"loss": 0.9503,
"step": 36050
},
{
"epoch": 2.776709483885855,
"grad_norm": 1.1087392568588257,
"learning_rate": 1.516245487364621e-05,
"loss": 0.9725,
"step": 36100
},
{
"epoch": 2.780555341896777,
"grad_norm": 1.2012007236480713,
"learning_rate": 1.4902734851829727e-05,
"loss": 1.0098,
"step": 36150
},
{
"epoch": 2.784401199907699,
"grad_norm": 0.9333285093307495,
"learning_rate": 1.4643014830013247e-05,
"loss": 0.9764,
"step": 36200
},
{
"epoch": 2.7882470579186216,
"grad_norm": 0.9407594203948975,
"learning_rate": 1.4383294808196765e-05,
"loss": 0.9404,
"step": 36250
},
{
"epoch": 2.7882470579186216,
"eval_loss": 0.9894633889198303,
"eval_runtime": 17.3239,
"eval_samples_per_second": 57.724,
"eval_steps_per_second": 14.431,
"step": 36250
},
{
"epoch": 2.792092915929544,
"grad_norm": 0.9794307947158813,
"learning_rate": 1.4123574786380284e-05,
"loss": 0.9445,
"step": 36300
},
{
"epoch": 2.795938773940466,
"grad_norm": 1.1447358131408691,
"learning_rate": 1.3863854764563799e-05,
"loss": 1.0056,
"step": 36350
},
{
"epoch": 2.7997846319513884,
"grad_norm": 0.8741857409477234,
"learning_rate": 1.3604134742747318e-05,
"loss": 0.9511,
"step": 36400
},
{
"epoch": 2.803630489962311,
"grad_norm": 1.0769715309143066,
"learning_rate": 1.3344414720930836e-05,
"loss": 0.9789,
"step": 36450
},
{
"epoch": 2.8074763479732328,
"grad_norm": 1.1045129299163818,
"learning_rate": 1.3084694699114356e-05,
"loss": 0.9958,
"step": 36500
},
{
"epoch": 2.8074763479732328,
"eval_loss": 0.9914500117301941,
"eval_runtime": 17.3793,
"eval_samples_per_second": 57.54,
"eval_steps_per_second": 14.385,
"step": 36500
},
{
"epoch": 2.811322205984155,
"grad_norm": 0.9856983423233032,
"learning_rate": 1.2824974677297874e-05,
"loss": 0.9923,
"step": 36550
},
{
"epoch": 2.8151680639950776,
"grad_norm": 1.2112038135528564,
"learning_rate": 1.2565254655481392e-05,
"loss": 0.9388,
"step": 36600
},
{
"epoch": 2.8190139220059995,
"grad_norm": 1.1824342012405396,
"learning_rate": 1.230553463366491e-05,
"loss": 0.984,
"step": 36650
},
{
"epoch": 2.8228597800169215,
"grad_norm": 1.3278725147247314,
"learning_rate": 1.204581461184843e-05,
"loss": 0.9835,
"step": 36700
},
{
"epoch": 2.826705638027844,
"grad_norm": 1.4297553300857544,
"learning_rate": 1.1786094590031946e-05,
"loss": 0.9999,
"step": 36750
},
{
"epoch": 2.826705638027844,
"eval_loss": 0.9847651720046997,
"eval_runtime": 17.5097,
"eval_samples_per_second": 57.111,
"eval_steps_per_second": 14.278,
"step": 36750
},
{
"epoch": 2.8305514960387663,
"grad_norm": 1.1944117546081543,
"learning_rate": 1.1526374568215465e-05,
"loss": 0.9758,
"step": 36800
},
{
"epoch": 2.8343973540496883,
"grad_norm": 1.1629287004470825,
"learning_rate": 1.1266654546398983e-05,
"loss": 0.988,
"step": 36850
},
{
"epoch": 2.8382432120606107,
"grad_norm": 1.598382592201233,
"learning_rate": 1.1006934524582501e-05,
"loss": 0.9679,
"step": 36900
},
{
"epoch": 2.842089070071533,
"grad_norm": 0.9534172415733337,
"learning_rate": 1.0747214502766019e-05,
"loss": 0.9921,
"step": 36950
},
{
"epoch": 2.845934928082455,
"grad_norm": 1.5716655254364014,
"learning_rate": 1.0487494480949537e-05,
"loss": 0.9898,
"step": 37000
},
{
"epoch": 2.845934928082455,
"eval_loss": 0.9878412485122681,
"eval_runtime": 17.2808,
"eval_samples_per_second": 57.868,
"eval_steps_per_second": 14.467,
"step": 37000
},
{
"epoch": 2.8497807860933775,
"grad_norm": 1.3013421297073364,
"learning_rate": 1.0227774459133055e-05,
"loss": 0.9717,
"step": 37050
},
{
"epoch": 2.8536266441043,
"grad_norm": 1.263071060180664,
"learning_rate": 9.968054437316573e-06,
"loss": 1.0259,
"step": 37100
},
{
"epoch": 2.857472502115222,
"grad_norm": 1.1513851881027222,
"learning_rate": 9.70833441550009e-06,
"loss": 1.0015,
"step": 37150
},
{
"epoch": 2.8613183601261443,
"grad_norm": 0.7431422472000122,
"learning_rate": 9.448614393683609e-06,
"loss": 0.999,
"step": 37200
},
{
"epoch": 2.8651642181370662,
"grad_norm": 0.6744217872619629,
"learning_rate": 9.188894371867128e-06,
"loss": 0.9285,
"step": 37250
},
{
"epoch": 2.8651642181370662,
"eval_loss": 0.994976282119751,
"eval_runtime": 17.3294,
"eval_samples_per_second": 57.705,
"eval_steps_per_second": 14.426,
"step": 37250
},
{
"epoch": 2.8690100761479886,
"grad_norm": 1.2962367534637451,
"learning_rate": 8.929174350050646e-06,
"loss": 0.9433,
"step": 37300
},
{
"epoch": 2.8728559341589106,
"grad_norm": 0.9955423474311829,
"learning_rate": 8.669454328234164e-06,
"loss": 1.0447,
"step": 37350
},
{
"epoch": 2.876701792169833,
"grad_norm": 0.5840064287185669,
"learning_rate": 8.409734306417682e-06,
"loss": 0.9504,
"step": 37400
},
{
"epoch": 2.8805476501807554,
"grad_norm": 1.0777620077133179,
"learning_rate": 8.1500142846012e-06,
"loss": 0.9635,
"step": 37450
},
{
"epoch": 2.8843935081916774,
"grad_norm": 0.9312844276428223,
"learning_rate": 7.890294262784718e-06,
"loss": 1.005,
"step": 37500
},
{
"epoch": 2.8843935081916774,
"eval_loss": 0.9864250421524048,
"eval_runtime": 17.2238,
"eval_samples_per_second": 58.059,
"eval_steps_per_second": 14.515,
"step": 37500
},
{
"epoch": 2.8882393662026,
"grad_norm": 0.6800574660301208,
"learning_rate": 7.630574240968237e-06,
"loss": 0.983,
"step": 37550
},
{
"epoch": 2.892085224213522,
"grad_norm": 0.9417561292648315,
"learning_rate": 7.3708542191517545e-06,
"loss": 0.973,
"step": 37600
},
{
"epoch": 2.895931082224444,
"grad_norm": 0.6454310417175293,
"learning_rate": 7.111134197335273e-06,
"loss": 1.0493,
"step": 37650
},
{
"epoch": 2.8997769402353666,
"grad_norm": 1.0287562608718872,
"learning_rate": 6.851414175518792e-06,
"loss": 0.9821,
"step": 37700
},
{
"epoch": 2.903622798246289,
"grad_norm": 1.6874816417694092,
"learning_rate": 6.591694153702309e-06,
"loss": 0.9761,
"step": 37750
},
{
"epoch": 2.903622798246289,
"eval_loss": 0.9886119961738586,
"eval_runtime": 17.2542,
"eval_samples_per_second": 57.957,
"eval_steps_per_second": 14.489,
"step": 37750
},
{
"epoch": 2.907468656257211,
"grad_norm": 0.7677489519119263,
"learning_rate": 6.331974131885828e-06,
"loss": 1.0421,
"step": 37800
},
{
"epoch": 2.9113145142681334,
"grad_norm": 0.9784395098686218,
"learning_rate": 6.072254110069346e-06,
"loss": 0.9288,
"step": 37850
},
{
"epoch": 2.9151603722790553,
"grad_norm": 0.6236763000488281,
"learning_rate": 5.812534088252864e-06,
"loss": 0.9777,
"step": 37900
},
{
"epoch": 2.9190062302899777,
"grad_norm": 1.0675079822540283,
"learning_rate": 5.552814066436382e-06,
"loss": 0.9659,
"step": 37950
},
{
"epoch": 2.9228520883008997,
"grad_norm": 0.9310262203216553,
"learning_rate": 5.2930940446198996e-06,
"loss": 0.9334,
"step": 38000
},
{
"epoch": 2.9228520883008997,
"eval_loss": 0.98431396484375,
"eval_runtime": 17.3967,
"eval_samples_per_second": 57.482,
"eval_steps_per_second": 14.371,
"step": 38000
},
{
"epoch": 2.926697946311822,
"grad_norm": 0.9615415930747986,
"learning_rate": 5.038568423239748e-06,
"loss": 0.9715,
"step": 38050
},
{
"epoch": 2.9305438043227445,
"grad_norm": 0.5959100127220154,
"learning_rate": 4.778848401423266e-06,
"loss": 0.936,
"step": 38100
},
{
"epoch": 2.9343896623336665,
"grad_norm": 1.0181951522827148,
"learning_rate": 4.519128379606784e-06,
"loss": 0.969,
"step": 38150
},
{
"epoch": 2.938235520344589,
"grad_norm": 0.8716458082199097,
"learning_rate": 4.259408357790302e-06,
"loss": 1.002,
"step": 38200
},
{
"epoch": 2.9420813783555113,
"grad_norm": 1.6100435256958008,
"learning_rate": 3.99968833597382e-06,
"loss": 0.9425,
"step": 38250
},
{
"epoch": 2.9420813783555113,
"eval_loss": 0.9931854009628296,
"eval_runtime": 17.3406,
"eval_samples_per_second": 57.668,
"eval_steps_per_second": 14.417,
"step": 38250
},
{
"epoch": 2.9459272363664333,
"grad_norm": 1.1552485227584839,
"learning_rate": 3.7399683141573383e-06,
"loss": 1.0424,
"step": 38300
},
{
"epoch": 2.9497730943773557,
"grad_norm": 1.441148042678833,
"learning_rate": 3.480248292340857e-06,
"loss": 1.0688,
"step": 38350
},
{
"epoch": 2.953618952388278,
"grad_norm": 0.8127447366714478,
"learning_rate": 3.220528270524375e-06,
"loss": 0.9497,
"step": 38400
},
{
"epoch": 2.9574648103992,
"grad_norm": 0.9566346406936646,
"learning_rate": 2.9608082487078933e-06,
"loss": 0.9628,
"step": 38450
},
{
"epoch": 2.961310668410122,
"grad_norm": 0.7955614924430847,
"learning_rate": 2.7010882268914113e-06,
"loss": 1.016,
"step": 38500
},
{
"epoch": 2.961310668410122,
"eval_loss": 0.9919273257255554,
"eval_runtime": 17.3829,
"eval_samples_per_second": 57.528,
"eval_steps_per_second": 14.382,
"step": 38500
},
{
"epoch": 2.9651565264210444,
"grad_norm": 0.8041125535964966,
"learning_rate": 2.4413682050749296e-06,
"loss": 0.9631,
"step": 38550
},
{
"epoch": 2.969002384431967,
"grad_norm": 1.0755919218063354,
"learning_rate": 2.1816481832584475e-06,
"loss": 0.9689,
"step": 38600
},
{
"epoch": 2.972848242442889,
"grad_norm": 0.8630362153053284,
"learning_rate": 1.9219281614419654e-06,
"loss": 0.9822,
"step": 38650
},
{
"epoch": 2.976694100453811,
"grad_norm": 0.6994553804397583,
"learning_rate": 1.662208139625484e-06,
"loss": 0.9706,
"step": 38700
},
{
"epoch": 2.9805399584647336,
"grad_norm": 1.1746189594268799,
"learning_rate": 1.4024881178090021e-06,
"loss": 0.9695,
"step": 38750
},
{
"epoch": 2.9805399584647336,
"eval_loss": 0.9868382215499878,
"eval_runtime": 17.3783,
"eval_samples_per_second": 57.543,
"eval_steps_per_second": 14.386,
"step": 38750
},
{
"epoch": 2.9843858164756556,
"grad_norm": 1.0636792182922363,
"learning_rate": 1.14276809599252e-06,
"loss": 1.0233,
"step": 38800
},
{
"epoch": 2.988231674486578,
"grad_norm": 0.8173992037773132,
"learning_rate": 8.830480741760383e-07,
"loss": 0.963,
"step": 38850
},
{
"epoch": 2.9920775324975004,
"grad_norm": 1.591539978981018,
"learning_rate": 6.233280523595564e-07,
"loss": 0.949,
"step": 38900
},
{
"epoch": 2.9959233905084224,
"grad_norm": 0.6467046737670898,
"learning_rate": 3.636080305430746e-07,
"loss": 0.9649,
"step": 38950
},
{
"epoch": 2.999769248519345,
"grad_norm": 0.9910799860954285,
"learning_rate": 1.0388800872659275e-07,
"loss": 1.0502,
"step": 39000
},
{
"epoch": 2.999769248519345,
"eval_loss": 0.9894677400588989,
"eval_runtime": 17.4751,
"eval_samples_per_second": 57.224,
"eval_steps_per_second": 14.306,
"step": 39000
}
],
"logging_steps": 50,
"max_steps": 39003,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}