Qwen2.5-1.5B-Open-R1-Distill / trainer_state.json
DigitalLifeYZQiu's picture
Model save
fac8379 verified
raw
history blame
299 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 8564,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0005838393274170948,
"grad_norm": 2.8469952515683437,
"learning_rate": 5.827505827505828e-07,
"loss": 1.1342,
"step": 5
},
{
"epoch": 0.0011676786548341896,
"grad_norm": 2.6897624350510996,
"learning_rate": 1.1655011655011657e-06,
"loss": 1.1192,
"step": 10
},
{
"epoch": 0.0017515179822512844,
"grad_norm": 2.489357614904495,
"learning_rate": 1.7482517482517485e-06,
"loss": 1.1026,
"step": 15
},
{
"epoch": 0.002335357309668379,
"grad_norm": 1.9452947723577663,
"learning_rate": 2.3310023310023313e-06,
"loss": 1.0839,
"step": 20
},
{
"epoch": 0.002919196637085474,
"grad_norm": 1.3334898364058092,
"learning_rate": 2.9137529137529138e-06,
"loss": 1.0333,
"step": 25
},
{
"epoch": 0.003503035964502569,
"grad_norm": 1.3156533821299237,
"learning_rate": 3.496503496503497e-06,
"loss": 1.0142,
"step": 30
},
{
"epoch": 0.004086875291919664,
"grad_norm": 1.0319662501793385,
"learning_rate": 4.079254079254079e-06,
"loss": 0.9595,
"step": 35
},
{
"epoch": 0.004670714619336758,
"grad_norm": 0.9794564419052394,
"learning_rate": 4.662004662004663e-06,
"loss": 0.9241,
"step": 40
},
{
"epoch": 0.005254553946753854,
"grad_norm": 0.8636223012514679,
"learning_rate": 5.244755244755245e-06,
"loss": 0.9067,
"step": 45
},
{
"epoch": 0.005838393274170948,
"grad_norm": 0.8696789275736428,
"learning_rate": 5.8275058275058275e-06,
"loss": 0.8791,
"step": 50
},
{
"epoch": 0.006422232601588043,
"grad_norm": 0.8407794179207635,
"learning_rate": 6.41025641025641e-06,
"loss": 0.8934,
"step": 55
},
{
"epoch": 0.007006071929005138,
"grad_norm": 0.9407370909076542,
"learning_rate": 6.993006993006994e-06,
"loss": 0.8934,
"step": 60
},
{
"epoch": 0.007589911256422233,
"grad_norm": 0.8092787662698692,
"learning_rate": 7.5757575757575764e-06,
"loss": 0.863,
"step": 65
},
{
"epoch": 0.008173750583839328,
"grad_norm": 0.8224130681631353,
"learning_rate": 8.158508158508159e-06,
"loss": 0.8676,
"step": 70
},
{
"epoch": 0.008757589911256422,
"grad_norm": 0.8183320365501537,
"learning_rate": 8.741258741258741e-06,
"loss": 0.8736,
"step": 75
},
{
"epoch": 0.009341429238673517,
"grad_norm": 0.7982114345023084,
"learning_rate": 9.324009324009325e-06,
"loss": 0.8483,
"step": 80
},
{
"epoch": 0.009925268566090611,
"grad_norm": 0.8268601708543242,
"learning_rate": 9.906759906759908e-06,
"loss": 0.8463,
"step": 85
},
{
"epoch": 0.010509107893507707,
"grad_norm": 0.8699340248486512,
"learning_rate": 1.048951048951049e-05,
"loss": 0.8447,
"step": 90
},
{
"epoch": 0.011092947220924802,
"grad_norm": 0.8309891860577984,
"learning_rate": 1.1072261072261073e-05,
"loss": 0.8506,
"step": 95
},
{
"epoch": 0.011676786548341896,
"grad_norm": 0.8767528986933936,
"learning_rate": 1.1655011655011655e-05,
"loss": 0.8229,
"step": 100
},
{
"epoch": 0.01226062587575899,
"grad_norm": 0.8834150944718365,
"learning_rate": 1.2237762237762239e-05,
"loss": 0.8347,
"step": 105
},
{
"epoch": 0.012844465203176086,
"grad_norm": 0.828246433297155,
"learning_rate": 1.282051282051282e-05,
"loss": 0.8258,
"step": 110
},
{
"epoch": 0.01342830453059318,
"grad_norm": 0.8178105785040897,
"learning_rate": 1.3403263403263406e-05,
"loss": 0.817,
"step": 115
},
{
"epoch": 0.014012143858010275,
"grad_norm": 0.8424478993141643,
"learning_rate": 1.3986013986013988e-05,
"loss": 0.825,
"step": 120
},
{
"epoch": 0.014595983185427371,
"grad_norm": 0.8394877026175257,
"learning_rate": 1.456876456876457e-05,
"loss": 0.8147,
"step": 125
},
{
"epoch": 0.015179822512844466,
"grad_norm": 0.89097070697157,
"learning_rate": 1.5151515151515153e-05,
"loss": 0.8146,
"step": 130
},
{
"epoch": 0.01576366184026156,
"grad_norm": 0.9586903801030577,
"learning_rate": 1.5734265734265734e-05,
"loss": 0.8226,
"step": 135
},
{
"epoch": 0.016347501167678656,
"grad_norm": 0.9000551836338954,
"learning_rate": 1.6317016317016318e-05,
"loss": 0.8001,
"step": 140
},
{
"epoch": 0.01693134049509575,
"grad_norm": 0.8067021698055977,
"learning_rate": 1.68997668997669e-05,
"loss": 0.8064,
"step": 145
},
{
"epoch": 0.017515179822512845,
"grad_norm": 0.9644732978901126,
"learning_rate": 1.7482517482517483e-05,
"loss": 0.7948,
"step": 150
},
{
"epoch": 0.01809901914992994,
"grad_norm": 0.8898965604096551,
"learning_rate": 1.8065268065268067e-05,
"loss": 0.8044,
"step": 155
},
{
"epoch": 0.018682858477347034,
"grad_norm": 0.9024973580531401,
"learning_rate": 1.864801864801865e-05,
"loss": 0.8113,
"step": 160
},
{
"epoch": 0.01926669780476413,
"grad_norm": 0.8902533739833289,
"learning_rate": 1.923076923076923e-05,
"loss": 0.812,
"step": 165
},
{
"epoch": 0.019850537132181222,
"grad_norm": 0.8461212901690308,
"learning_rate": 1.9813519813519816e-05,
"loss": 0.7835,
"step": 170
},
{
"epoch": 0.02043437645959832,
"grad_norm": 0.9593997324120669,
"learning_rate": 2.0396270396270396e-05,
"loss": 0.8341,
"step": 175
},
{
"epoch": 0.021018215787015414,
"grad_norm": 0.9319872831334555,
"learning_rate": 2.097902097902098e-05,
"loss": 0.8127,
"step": 180
},
{
"epoch": 0.021602055114432507,
"grad_norm": 0.9164230729247066,
"learning_rate": 2.156177156177156e-05,
"loss": 0.8018,
"step": 185
},
{
"epoch": 0.022185894441849603,
"grad_norm": 0.8722898766994293,
"learning_rate": 2.2144522144522145e-05,
"loss": 0.8126,
"step": 190
},
{
"epoch": 0.0227697337692667,
"grad_norm": 0.9328215048668408,
"learning_rate": 2.272727272727273e-05,
"loss": 0.7799,
"step": 195
},
{
"epoch": 0.023353573096683792,
"grad_norm": 0.9109551187088815,
"learning_rate": 2.331002331002331e-05,
"loss": 0.7911,
"step": 200
},
{
"epoch": 0.023937412424100888,
"grad_norm": 0.8692976775471373,
"learning_rate": 2.3892773892773894e-05,
"loss": 0.7999,
"step": 205
},
{
"epoch": 0.02452125175151798,
"grad_norm": 0.9102711535679596,
"learning_rate": 2.4475524475524478e-05,
"loss": 0.8224,
"step": 210
},
{
"epoch": 0.025105091078935077,
"grad_norm": 1.0021495752651475,
"learning_rate": 2.505827505827506e-05,
"loss": 0.7727,
"step": 215
},
{
"epoch": 0.025688930406352173,
"grad_norm": 0.8838902337075365,
"learning_rate": 2.564102564102564e-05,
"loss": 0.7968,
"step": 220
},
{
"epoch": 0.026272769733769265,
"grad_norm": 0.8062784051996996,
"learning_rate": 2.6223776223776224e-05,
"loss": 0.7917,
"step": 225
},
{
"epoch": 0.02685660906118636,
"grad_norm": 0.8311276241312253,
"learning_rate": 2.680652680652681e-05,
"loss": 0.8031,
"step": 230
},
{
"epoch": 0.027440448388603458,
"grad_norm": 0.895141981016196,
"learning_rate": 2.738927738927739e-05,
"loss": 0.7901,
"step": 235
},
{
"epoch": 0.02802428771602055,
"grad_norm": 0.9022303058168814,
"learning_rate": 2.7972027972027976e-05,
"loss": 0.7652,
"step": 240
},
{
"epoch": 0.028608127043437646,
"grad_norm": 0.913718234352083,
"learning_rate": 2.8554778554778557e-05,
"loss": 0.793,
"step": 245
},
{
"epoch": 0.029191966370854742,
"grad_norm": 0.9464414694961122,
"learning_rate": 2.913752913752914e-05,
"loss": 0.8194,
"step": 250
},
{
"epoch": 0.029775805698271835,
"grad_norm": 0.8798527135755817,
"learning_rate": 2.972027972027972e-05,
"loss": 0.8142,
"step": 255
},
{
"epoch": 0.03035964502568893,
"grad_norm": 0.861473905943677,
"learning_rate": 3.0303030303030306e-05,
"loss": 0.8026,
"step": 260
},
{
"epoch": 0.030943484353106024,
"grad_norm": 0.8890275205284014,
"learning_rate": 3.088578088578088e-05,
"loss": 0.7981,
"step": 265
},
{
"epoch": 0.03152732368052312,
"grad_norm": 0.9184733369917072,
"learning_rate": 3.146853146853147e-05,
"loss": 0.7806,
"step": 270
},
{
"epoch": 0.032111163007940216,
"grad_norm": 0.9624043666630351,
"learning_rate": 3.205128205128206e-05,
"loss": 0.7875,
"step": 275
},
{
"epoch": 0.03269500233535731,
"grad_norm": 0.9589380101558912,
"learning_rate": 3.2634032634032635e-05,
"loss": 0.7884,
"step": 280
},
{
"epoch": 0.0332788416627744,
"grad_norm": 0.9734450757595803,
"learning_rate": 3.321678321678322e-05,
"loss": 0.8107,
"step": 285
},
{
"epoch": 0.0338626809901915,
"grad_norm": 1.0976968736829902,
"learning_rate": 3.37995337995338e-05,
"loss": 0.7985,
"step": 290
},
{
"epoch": 0.034446520317608594,
"grad_norm": 0.8956644445095314,
"learning_rate": 3.438228438228439e-05,
"loss": 0.7864,
"step": 295
},
{
"epoch": 0.03503035964502569,
"grad_norm": 0.8961849644495538,
"learning_rate": 3.4965034965034965e-05,
"loss": 0.7802,
"step": 300
},
{
"epoch": 0.035614198972442786,
"grad_norm": 0.9155647051953708,
"learning_rate": 3.554778554778555e-05,
"loss": 0.7894,
"step": 305
},
{
"epoch": 0.03619803829985988,
"grad_norm": 0.9359545515475477,
"learning_rate": 3.613053613053613e-05,
"loss": 0.8022,
"step": 310
},
{
"epoch": 0.03678187762727697,
"grad_norm": 0.9073994864544219,
"learning_rate": 3.671328671328672e-05,
"loss": 0.7826,
"step": 315
},
{
"epoch": 0.03736571695469407,
"grad_norm": 0.9540970201934222,
"learning_rate": 3.72960372960373e-05,
"loss": 0.7875,
"step": 320
},
{
"epoch": 0.03794955628211116,
"grad_norm": 0.9896974399906573,
"learning_rate": 3.787878787878788e-05,
"loss": 0.7815,
"step": 325
},
{
"epoch": 0.03853339560952826,
"grad_norm": 1.0098673751943326,
"learning_rate": 3.846153846153846e-05,
"loss": 0.7928,
"step": 330
},
{
"epoch": 0.039117234936945355,
"grad_norm": 0.9568961216747747,
"learning_rate": 3.904428904428905e-05,
"loss": 0.8014,
"step": 335
},
{
"epoch": 0.039701074264362445,
"grad_norm": 0.9070273181335483,
"learning_rate": 3.962703962703963e-05,
"loss": 0.7871,
"step": 340
},
{
"epoch": 0.04028491359177954,
"grad_norm": 0.8743288189560451,
"learning_rate": 4.020979020979021e-05,
"loss": 0.774,
"step": 345
},
{
"epoch": 0.04086875291919664,
"grad_norm": 1.0477138979263285,
"learning_rate": 4.079254079254079e-05,
"loss": 0.7866,
"step": 350
},
{
"epoch": 0.04145259224661373,
"grad_norm": 0.845985324382712,
"learning_rate": 4.1375291375291377e-05,
"loss": 0.7945,
"step": 355
},
{
"epoch": 0.04203643157403083,
"grad_norm": 1.067533288565115,
"learning_rate": 4.195804195804196e-05,
"loss": 0.7953,
"step": 360
},
{
"epoch": 0.04262027090144792,
"grad_norm": 0.9916448982734962,
"learning_rate": 4.254079254079254e-05,
"loss": 0.7938,
"step": 365
},
{
"epoch": 0.043204110228865014,
"grad_norm": 0.9867852858846966,
"learning_rate": 4.312354312354312e-05,
"loss": 0.7866,
"step": 370
},
{
"epoch": 0.04378794955628211,
"grad_norm": 0.8871740073671367,
"learning_rate": 4.370629370629371e-05,
"loss": 0.7574,
"step": 375
},
{
"epoch": 0.044371788883699206,
"grad_norm": 0.915178688048331,
"learning_rate": 4.428904428904429e-05,
"loss": 0.7745,
"step": 380
},
{
"epoch": 0.0449556282111163,
"grad_norm": 0.9878189566672229,
"learning_rate": 4.4871794871794874e-05,
"loss": 0.7964,
"step": 385
},
{
"epoch": 0.0455394675385334,
"grad_norm": 1.010681809776873,
"learning_rate": 4.545454545454546e-05,
"loss": 0.7819,
"step": 390
},
{
"epoch": 0.04612330686595049,
"grad_norm": 1.1897933196265749,
"learning_rate": 4.603729603729604e-05,
"loss": 0.7788,
"step": 395
},
{
"epoch": 0.046707146193367584,
"grad_norm": 1.061158392111532,
"learning_rate": 4.662004662004662e-05,
"loss": 0.8035,
"step": 400
},
{
"epoch": 0.04729098552078468,
"grad_norm": 0.8408235558466027,
"learning_rate": 4.7202797202797204e-05,
"loss": 0.7879,
"step": 405
},
{
"epoch": 0.047874824848201776,
"grad_norm": 0.9163616279323427,
"learning_rate": 4.778554778554779e-05,
"loss": 0.8038,
"step": 410
},
{
"epoch": 0.04845866417561887,
"grad_norm": 0.8231726702767893,
"learning_rate": 4.836829836829837e-05,
"loss": 0.7776,
"step": 415
},
{
"epoch": 0.04904250350303596,
"grad_norm": 0.8838721217063888,
"learning_rate": 4.8951048951048956e-05,
"loss": 0.7699,
"step": 420
},
{
"epoch": 0.04962634283045306,
"grad_norm": 0.9845033294686794,
"learning_rate": 4.9533799533799534e-05,
"loss": 0.8037,
"step": 425
},
{
"epoch": 0.050210182157870153,
"grad_norm": 0.8749354730883913,
"learning_rate": 4.999999832221176e-05,
"loss": 0.7775,
"step": 430
},
{
"epoch": 0.05079402148528725,
"grad_norm": 0.9234120904036062,
"learning_rate": 4.999993959964937e-05,
"loss": 0.777,
"step": 435
},
{
"epoch": 0.051377860812704346,
"grad_norm": 0.8214791658110356,
"learning_rate": 4.999979698792484e-05,
"loss": 0.7936,
"step": 440
},
{
"epoch": 0.05196170014012144,
"grad_norm": 1.067662635381822,
"learning_rate": 4.999957048756989e-05,
"loss": 0.7924,
"step": 445
},
{
"epoch": 0.05254553946753853,
"grad_norm": 0.8536251459541943,
"learning_rate": 4.999926009942899e-05,
"loss": 0.7857,
"step": 450
},
{
"epoch": 0.05312937879495563,
"grad_norm": 0.8789484855680025,
"learning_rate": 4.999886582465941e-05,
"loss": 0.8028,
"step": 455
},
{
"epoch": 0.05371321812237272,
"grad_norm": 0.8645381455149383,
"learning_rate": 4.999838766473116e-05,
"loss": 0.7835,
"step": 460
},
{
"epoch": 0.05429705744978982,
"grad_norm": 0.8193748527865334,
"learning_rate": 4.999782562142702e-05,
"loss": 0.7616,
"step": 465
},
{
"epoch": 0.054880896777206915,
"grad_norm": 0.9327485152610473,
"learning_rate": 4.999717969684254e-05,
"loss": 0.7927,
"step": 470
},
{
"epoch": 0.055464736104624004,
"grad_norm": 0.8976109159630459,
"learning_rate": 4.999644989338598e-05,
"loss": 0.8048,
"step": 475
},
{
"epoch": 0.0560485754320411,
"grad_norm": 0.9029757753061991,
"learning_rate": 4.9995636213778354e-05,
"loss": 0.7697,
"step": 480
},
{
"epoch": 0.0566324147594582,
"grad_norm": 0.8037219872770315,
"learning_rate": 4.99947386610534e-05,
"loss": 0.7739,
"step": 485
},
{
"epoch": 0.05721625408687529,
"grad_norm": 0.8876170287138181,
"learning_rate": 4.9993757238557564e-05,
"loss": 0.789,
"step": 490
},
{
"epoch": 0.05780009341429239,
"grad_norm": 0.8049736423744179,
"learning_rate": 4.999269194995001e-05,
"loss": 0.7548,
"step": 495
},
{
"epoch": 0.058383932741709485,
"grad_norm": 0.8398014231160289,
"learning_rate": 4.9991542799202574e-05,
"loss": 0.7658,
"step": 500
},
{
"epoch": 0.058967772069126574,
"grad_norm": 0.7942320658266705,
"learning_rate": 4.999030979059977e-05,
"loss": 0.7873,
"step": 505
},
{
"epoch": 0.05955161139654367,
"grad_norm": 0.8289805635789632,
"learning_rate": 4.998899292873876e-05,
"loss": 0.7909,
"step": 510
},
{
"epoch": 0.060135450723960766,
"grad_norm": 0.7893877662667962,
"learning_rate": 4.9987592218529364e-05,
"loss": 0.7776,
"step": 515
},
{
"epoch": 0.06071929005137786,
"grad_norm": 0.7965955365053803,
"learning_rate": 4.998610766519401e-05,
"loss": 0.7732,
"step": 520
},
{
"epoch": 0.06130312937879496,
"grad_norm": 0.8628562549084506,
"learning_rate": 4.9984539274267726e-05,
"loss": 0.7658,
"step": 525
},
{
"epoch": 0.06188696870621205,
"grad_norm": 0.8318274287434425,
"learning_rate": 4.998288705159815e-05,
"loss": 0.7958,
"step": 530
},
{
"epoch": 0.062470808033629144,
"grad_norm": 0.86123088200466,
"learning_rate": 4.9981151003345436e-05,
"loss": 0.7698,
"step": 535
},
{
"epoch": 0.06305464736104624,
"grad_norm": 0.929794979823692,
"learning_rate": 4.99793311359823e-05,
"loss": 0.7935,
"step": 540
},
{
"epoch": 0.06363848668846334,
"grad_norm": 0.7886084726419026,
"learning_rate": 4.997742745629397e-05,
"loss": 0.7884,
"step": 545
},
{
"epoch": 0.06422232601588043,
"grad_norm": 0.7674475851531968,
"learning_rate": 4.997543997137816e-05,
"loss": 0.7709,
"step": 550
},
{
"epoch": 0.06480616534329753,
"grad_norm": 0.8546622570923904,
"learning_rate": 4.9973368688645034e-05,
"loss": 0.7875,
"step": 555
},
{
"epoch": 0.06539000467071462,
"grad_norm": 0.7843602866632492,
"learning_rate": 4.997121361581721e-05,
"loss": 0.7671,
"step": 560
},
{
"epoch": 0.06597384399813172,
"grad_norm": 0.8189556120793431,
"learning_rate": 4.9968974760929694e-05,
"loss": 0.7586,
"step": 565
},
{
"epoch": 0.0665576833255488,
"grad_norm": 0.7898759730617912,
"learning_rate": 4.996665213232987e-05,
"loss": 0.7377,
"step": 570
},
{
"epoch": 0.0671415226529659,
"grad_norm": 0.8493700109069738,
"learning_rate": 4.9964245738677465e-05,
"loss": 0.7697,
"step": 575
},
{
"epoch": 0.067725361980383,
"grad_norm": 0.8126736004420235,
"learning_rate": 4.996175558894452e-05,
"loss": 0.7624,
"step": 580
},
{
"epoch": 0.06830920130780009,
"grad_norm": 0.8743281435570972,
"learning_rate": 4.9959181692415345e-05,
"loss": 0.7862,
"step": 585
},
{
"epoch": 0.06889304063521719,
"grad_norm": 0.8965919464326476,
"learning_rate": 4.995652405868652e-05,
"loss": 0.7698,
"step": 590
},
{
"epoch": 0.06947687996263428,
"grad_norm": 0.778928402409659,
"learning_rate": 4.99537826976668e-05,
"loss": 0.7685,
"step": 595
},
{
"epoch": 0.07006071929005138,
"grad_norm": 0.8891716181302853,
"learning_rate": 4.9950957619577115e-05,
"loss": 0.7906,
"step": 600
},
{
"epoch": 0.07064455861746848,
"grad_norm": 0.8042288230813074,
"learning_rate": 4.9948048834950546e-05,
"loss": 0.7856,
"step": 605
},
{
"epoch": 0.07122839794488557,
"grad_norm": 0.8829956497108745,
"learning_rate": 4.9945056354632255e-05,
"loss": 0.779,
"step": 610
},
{
"epoch": 0.07181223727230267,
"grad_norm": 0.8942892647415688,
"learning_rate": 4.994198018977945e-05,
"loss": 0.7769,
"step": 615
},
{
"epoch": 0.07239607659971976,
"grad_norm": 0.7921633771179634,
"learning_rate": 4.993882035186136e-05,
"loss": 0.7725,
"step": 620
},
{
"epoch": 0.07297991592713685,
"grad_norm": 0.8160400012903543,
"learning_rate": 4.9935576852659175e-05,
"loss": 0.7628,
"step": 625
},
{
"epoch": 0.07356375525455394,
"grad_norm": 0.7456640206417225,
"learning_rate": 4.993224970426603e-05,
"loss": 0.7516,
"step": 630
},
{
"epoch": 0.07414759458197104,
"grad_norm": 0.8299750150456215,
"learning_rate": 4.99288389190869e-05,
"loss": 0.7678,
"step": 635
},
{
"epoch": 0.07473143390938813,
"grad_norm": 0.7558532259531837,
"learning_rate": 4.992534450983864e-05,
"loss": 0.7815,
"step": 640
},
{
"epoch": 0.07531527323680523,
"grad_norm": 0.8252676659283117,
"learning_rate": 4.9921766489549835e-05,
"loss": 0.7799,
"step": 645
},
{
"epoch": 0.07589911256422233,
"grad_norm": 0.7776976006967014,
"learning_rate": 4.991810487156087e-05,
"loss": 0.7615,
"step": 650
},
{
"epoch": 0.07648295189163942,
"grad_norm": 0.8405891548119552,
"learning_rate": 4.991435966952376e-05,
"loss": 0.7837,
"step": 655
},
{
"epoch": 0.07706679121905652,
"grad_norm": 0.8756113196019248,
"learning_rate": 4.991053089740219e-05,
"loss": 0.7699,
"step": 660
},
{
"epoch": 0.07765063054647361,
"grad_norm": 1.1428051944412874,
"learning_rate": 4.990661856947142e-05,
"loss": 0.7718,
"step": 665
},
{
"epoch": 0.07823446987389071,
"grad_norm": 0.8924432538613407,
"learning_rate": 4.990262270031824e-05,
"loss": 0.7461,
"step": 670
},
{
"epoch": 0.0788183092013078,
"grad_norm": 0.9183291954602351,
"learning_rate": 4.989854330484092e-05,
"loss": 0.789,
"step": 675
},
{
"epoch": 0.07940214852872489,
"grad_norm": 0.8651568066240668,
"learning_rate": 4.9894380398249135e-05,
"loss": 0.7818,
"step": 680
},
{
"epoch": 0.07998598785614199,
"grad_norm": 0.8371330103012771,
"learning_rate": 4.989013399606396e-05,
"loss": 0.7554,
"step": 685
},
{
"epoch": 0.08056982718355908,
"grad_norm": 0.867443980137123,
"learning_rate": 4.988580411411774e-05,
"loss": 0.7547,
"step": 690
},
{
"epoch": 0.08115366651097618,
"grad_norm": 0.7919451608747614,
"learning_rate": 4.988139076855408e-05,
"loss": 0.7782,
"step": 695
},
{
"epoch": 0.08173750583839327,
"grad_norm": 0.9449950228678913,
"learning_rate": 4.9876893975827774e-05,
"loss": 0.7665,
"step": 700
},
{
"epoch": 0.08232134516581037,
"grad_norm": 0.9593745392334234,
"learning_rate": 4.987231375270475e-05,
"loss": 0.7842,
"step": 705
},
{
"epoch": 0.08290518449322747,
"grad_norm": 0.8539716370032612,
"learning_rate": 4.9867650116261994e-05,
"loss": 0.7494,
"step": 710
},
{
"epoch": 0.08348902382064456,
"grad_norm": 0.7557280578139879,
"learning_rate": 4.986290308388747e-05,
"loss": 0.7654,
"step": 715
},
{
"epoch": 0.08407286314806166,
"grad_norm": 0.7185305603275817,
"learning_rate": 4.98580726732801e-05,
"loss": 0.7903,
"step": 720
},
{
"epoch": 0.08465670247547875,
"grad_norm": 0.7803979092440014,
"learning_rate": 4.985315890244969e-05,
"loss": 0.7452,
"step": 725
},
{
"epoch": 0.08524054180289584,
"grad_norm": 0.7966286896957142,
"learning_rate": 4.9848161789716804e-05,
"loss": 0.7871,
"step": 730
},
{
"epoch": 0.08582438113031293,
"grad_norm": 0.8392258470940968,
"learning_rate": 4.9843081353712765e-05,
"loss": 0.7805,
"step": 735
},
{
"epoch": 0.08640822045773003,
"grad_norm": 0.784943072918037,
"learning_rate": 4.983791761337958e-05,
"loss": 0.7501,
"step": 740
},
{
"epoch": 0.08699205978514712,
"grad_norm": 0.9497402810181729,
"learning_rate": 4.9832670587969804e-05,
"loss": 0.7686,
"step": 745
},
{
"epoch": 0.08757589911256422,
"grad_norm": 0.8211803369957865,
"learning_rate": 4.9827340297046546e-05,
"loss": 0.7434,
"step": 750
},
{
"epoch": 0.08815973843998132,
"grad_norm": 0.7188834313405097,
"learning_rate": 4.9821926760483354e-05,
"loss": 0.7693,
"step": 755
},
{
"epoch": 0.08874357776739841,
"grad_norm": 0.7991395439280983,
"learning_rate": 4.9816429998464155e-05,
"loss": 0.7644,
"step": 760
},
{
"epoch": 0.08932741709481551,
"grad_norm": 0.7629383896397622,
"learning_rate": 4.9810850031483155e-05,
"loss": 0.7472,
"step": 765
},
{
"epoch": 0.0899112564222326,
"grad_norm": 0.950532875668034,
"learning_rate": 4.9805186880344826e-05,
"loss": 0.7624,
"step": 770
},
{
"epoch": 0.0904950957496497,
"grad_norm": 0.7831295772019135,
"learning_rate": 4.9799440566163726e-05,
"loss": 0.7601,
"step": 775
},
{
"epoch": 0.0910789350770668,
"grad_norm": 0.7768389598632662,
"learning_rate": 4.979361111036454e-05,
"loss": 0.7763,
"step": 780
},
{
"epoch": 0.09166277440448388,
"grad_norm": 0.7500437315213245,
"learning_rate": 4.9787698534681896e-05,
"loss": 0.781,
"step": 785
},
{
"epoch": 0.09224661373190098,
"grad_norm": 0.725679634631429,
"learning_rate": 4.978170286116035e-05,
"loss": 0.763,
"step": 790
},
{
"epoch": 0.09283045305931807,
"grad_norm": 0.6820916470053155,
"learning_rate": 4.9775624112154275e-05,
"loss": 0.7403,
"step": 795
},
{
"epoch": 0.09341429238673517,
"grad_norm": 0.7021882835127938,
"learning_rate": 4.976946231032777e-05,
"loss": 0.7341,
"step": 800
},
{
"epoch": 0.09399813171415226,
"grad_norm": 0.8227621427323781,
"learning_rate": 4.976321747865462e-05,
"loss": 0.7648,
"step": 805
},
{
"epoch": 0.09458197104156936,
"grad_norm": 0.804890789043882,
"learning_rate": 4.975688964041816e-05,
"loss": 0.7669,
"step": 810
},
{
"epoch": 0.09516581036898646,
"grad_norm": 0.7950162974406255,
"learning_rate": 4.975047881921119e-05,
"loss": 0.7348,
"step": 815
},
{
"epoch": 0.09574964969640355,
"grad_norm": 0.8239463595384365,
"learning_rate": 4.974398503893596e-05,
"loss": 0.7665,
"step": 820
},
{
"epoch": 0.09633348902382065,
"grad_norm": 0.753758358380225,
"learning_rate": 4.973740832380397e-05,
"loss": 0.7751,
"step": 825
},
{
"epoch": 0.09691732835123774,
"grad_norm": 0.8628734451803414,
"learning_rate": 4.9730748698335954e-05,
"loss": 0.7868,
"step": 830
},
{
"epoch": 0.09750116767865484,
"grad_norm": 0.8027759284538418,
"learning_rate": 4.9724006187361794e-05,
"loss": 0.7796,
"step": 835
},
{
"epoch": 0.09808500700607192,
"grad_norm": 0.8509806930142826,
"learning_rate": 4.971718081602037e-05,
"loss": 0.7936,
"step": 840
},
{
"epoch": 0.09866884633348902,
"grad_norm": 0.9698203041083456,
"learning_rate": 4.971027260975952e-05,
"loss": 0.7604,
"step": 845
},
{
"epoch": 0.09925268566090611,
"grad_norm": 0.9130733546941406,
"learning_rate": 4.9703281594335904e-05,
"loss": 0.7564,
"step": 850
},
{
"epoch": 0.09983652498832321,
"grad_norm": 0.7322423075538033,
"learning_rate": 4.969620779581497e-05,
"loss": 0.7601,
"step": 855
},
{
"epoch": 0.10042036431574031,
"grad_norm": 0.964862665560022,
"learning_rate": 4.968905124057077e-05,
"loss": 0.7466,
"step": 860
},
{
"epoch": 0.1010042036431574,
"grad_norm": 0.9994098601217942,
"learning_rate": 4.968181195528594e-05,
"loss": 0.7458,
"step": 865
},
{
"epoch": 0.1015880429705745,
"grad_norm": 0.8404606717951096,
"learning_rate": 4.9674489966951545e-05,
"loss": 0.762,
"step": 870
},
{
"epoch": 0.1021718822979916,
"grad_norm": 0.7250386939916199,
"learning_rate": 4.9667085302867015e-05,
"loss": 0.7461,
"step": 875
},
{
"epoch": 0.10275572162540869,
"grad_norm": 0.7529424335871671,
"learning_rate": 4.9659597990640045e-05,
"loss": 0.7489,
"step": 880
},
{
"epoch": 0.10333956095282579,
"grad_norm": 0.7954471318661211,
"learning_rate": 4.9652028058186435e-05,
"loss": 0.761,
"step": 885
},
{
"epoch": 0.10392340028024288,
"grad_norm": 0.780620168906275,
"learning_rate": 4.9644375533730056e-05,
"loss": 0.7824,
"step": 890
},
{
"epoch": 0.10450723960765997,
"grad_norm": 0.8376075291664477,
"learning_rate": 4.963664044580272e-05,
"loss": 0.7605,
"step": 895
},
{
"epoch": 0.10509107893507706,
"grad_norm": 0.7349182138721907,
"learning_rate": 4.9628822823244056e-05,
"loss": 0.7421,
"step": 900
},
{
"epoch": 0.10567491826249416,
"grad_norm": 0.8037517128405073,
"learning_rate": 4.962092269520143e-05,
"loss": 0.7406,
"step": 905
},
{
"epoch": 0.10625875758991125,
"grad_norm": 0.8672270903867407,
"learning_rate": 4.96129400911298e-05,
"loss": 0.7714,
"step": 910
},
{
"epoch": 0.10684259691732835,
"grad_norm": 0.8132034707233711,
"learning_rate": 4.960487504079166e-05,
"loss": 0.7441,
"step": 915
},
{
"epoch": 0.10742643624474545,
"grad_norm": 0.9227968444282286,
"learning_rate": 4.959672757425688e-05,
"loss": 0.7651,
"step": 920
},
{
"epoch": 0.10801027557216254,
"grad_norm": 0.7885589322496032,
"learning_rate": 4.958849772190261e-05,
"loss": 0.7603,
"step": 925
},
{
"epoch": 0.10859411489957964,
"grad_norm": 0.728877555249776,
"learning_rate": 4.958018551441317e-05,
"loss": 0.7399,
"step": 930
},
{
"epoch": 0.10917795422699673,
"grad_norm": 0.8131296512535063,
"learning_rate": 4.957179098277994e-05,
"loss": 0.7651,
"step": 935
},
{
"epoch": 0.10976179355441383,
"grad_norm": 0.7050317838185277,
"learning_rate": 4.956331415830125e-05,
"loss": 0.7337,
"step": 940
},
{
"epoch": 0.11034563288183093,
"grad_norm": 0.7609060790958637,
"learning_rate": 4.955475507258222e-05,
"loss": 0.7683,
"step": 945
},
{
"epoch": 0.11092947220924801,
"grad_norm": 0.7472253996529664,
"learning_rate": 4.95461137575347e-05,
"loss": 0.7383,
"step": 950
},
{
"epoch": 0.1115133115366651,
"grad_norm": 0.8644167522162455,
"learning_rate": 4.953739024537712e-05,
"loss": 0.7213,
"step": 955
},
{
"epoch": 0.1120971508640822,
"grad_norm": 0.7143865358435283,
"learning_rate": 4.952858456863437e-05,
"loss": 0.775,
"step": 960
},
{
"epoch": 0.1126809901914993,
"grad_norm": 0.828635883067693,
"learning_rate": 4.951969676013768e-05,
"loss": 0.7495,
"step": 965
},
{
"epoch": 0.1132648295189164,
"grad_norm": 0.7571887396783333,
"learning_rate": 4.951072685302452e-05,
"loss": 0.7474,
"step": 970
},
{
"epoch": 0.11384866884633349,
"grad_norm": 0.7502741470779379,
"learning_rate": 4.950167488073844e-05,
"loss": 0.7392,
"step": 975
},
{
"epoch": 0.11443250817375059,
"grad_norm": 0.8496185620948489,
"learning_rate": 4.949254087702896e-05,
"loss": 0.7481,
"step": 980
},
{
"epoch": 0.11501634750116768,
"grad_norm": 0.7272674140304245,
"learning_rate": 4.948332487595148e-05,
"loss": 0.7543,
"step": 985
},
{
"epoch": 0.11560018682858478,
"grad_norm": 0.7136992316712841,
"learning_rate": 4.9474026911867084e-05,
"loss": 0.757,
"step": 990
},
{
"epoch": 0.11618402615600187,
"grad_norm": 0.7174339829157743,
"learning_rate": 4.9464647019442465e-05,
"loss": 0.7239,
"step": 995
},
{
"epoch": 0.11676786548341897,
"grad_norm": 0.8377816293409226,
"learning_rate": 4.945518523364976e-05,
"loss": 0.7548,
"step": 1000
},
{
"epoch": 0.11735170481083605,
"grad_norm": 0.7146458332659746,
"learning_rate": 4.944564158976647e-05,
"loss": 0.7584,
"step": 1005
},
{
"epoch": 0.11793554413825315,
"grad_norm": 0.8123437914876361,
"learning_rate": 4.943601612337528e-05,
"loss": 0.7505,
"step": 1010
},
{
"epoch": 0.11851938346567024,
"grad_norm": 0.938013186912681,
"learning_rate": 4.9426308870363934e-05,
"loss": 0.7513,
"step": 1015
},
{
"epoch": 0.11910322279308734,
"grad_norm": 0.7403213416682412,
"learning_rate": 4.941651986692514e-05,
"loss": 0.7619,
"step": 1020
},
{
"epoch": 0.11968706212050444,
"grad_norm": 0.7221282797840786,
"learning_rate": 4.940664914955637e-05,
"loss": 0.7535,
"step": 1025
},
{
"epoch": 0.12027090144792153,
"grad_norm": 0.6850603164988163,
"learning_rate": 4.939669675505978e-05,
"loss": 0.7648,
"step": 1030
},
{
"epoch": 0.12085474077533863,
"grad_norm": 0.7511737605932157,
"learning_rate": 4.938666272054205e-05,
"loss": 0.7347,
"step": 1035
},
{
"epoch": 0.12143858010275572,
"grad_norm": 0.7184559628537028,
"learning_rate": 4.937654708341425e-05,
"loss": 0.7654,
"step": 1040
},
{
"epoch": 0.12202241943017282,
"grad_norm": 0.7433996377054052,
"learning_rate": 4.93663498813917e-05,
"loss": 0.7529,
"step": 1045
},
{
"epoch": 0.12260625875758992,
"grad_norm": 0.771594757102507,
"learning_rate": 4.9356071152493815e-05,
"loss": 0.7438,
"step": 1050
},
{
"epoch": 0.123190098085007,
"grad_norm": 0.7099006582807407,
"learning_rate": 4.934571093504398e-05,
"loss": 0.7413,
"step": 1055
},
{
"epoch": 0.1237739374124241,
"grad_norm": 0.7753993011006178,
"learning_rate": 4.933526926766943e-05,
"loss": 0.7724,
"step": 1060
},
{
"epoch": 0.12435777673984119,
"grad_norm": 0.6788452808676482,
"learning_rate": 4.9324746189301027e-05,
"loss": 0.76,
"step": 1065
},
{
"epoch": 0.12494161606725829,
"grad_norm": 0.775189516086989,
"learning_rate": 4.9314141739173223e-05,
"loss": 0.7557,
"step": 1070
},
{
"epoch": 0.12552545539467538,
"grad_norm": 0.7464622012093612,
"learning_rate": 4.9303455956823816e-05,
"loss": 0.7675,
"step": 1075
},
{
"epoch": 0.12610929472209248,
"grad_norm": 0.7214821452340838,
"learning_rate": 4.929268888209388e-05,
"loss": 0.7351,
"step": 1080
},
{
"epoch": 0.12669313404950958,
"grad_norm": 0.8130480605834056,
"learning_rate": 4.928184055512754e-05,
"loss": 0.7488,
"step": 1085
},
{
"epoch": 0.12727697337692667,
"grad_norm": 0.6912803562502243,
"learning_rate": 4.927091101637189e-05,
"loss": 0.74,
"step": 1090
},
{
"epoch": 0.12786081270434377,
"grad_norm": 0.7043260852620832,
"learning_rate": 4.9259900306576825e-05,
"loss": 0.7269,
"step": 1095
},
{
"epoch": 0.12844465203176086,
"grad_norm": 0.6686198741044436,
"learning_rate": 4.924880846679485e-05,
"loss": 0.745,
"step": 1100
},
{
"epoch": 0.12902849135917796,
"grad_norm": 0.7390481439750773,
"learning_rate": 4.923763553838098e-05,
"loss": 0.7511,
"step": 1105
},
{
"epoch": 0.12961233068659506,
"grad_norm": 0.6882006426504611,
"learning_rate": 4.9226381562992546e-05,
"loss": 0.7569,
"step": 1110
},
{
"epoch": 0.13019617001401215,
"grad_norm": 0.6876582667989148,
"learning_rate": 4.9215046582589066e-05,
"loss": 0.7484,
"step": 1115
},
{
"epoch": 0.13078000934142925,
"grad_norm": 0.7696618028115917,
"learning_rate": 4.9203630639432083e-05,
"loss": 0.7407,
"step": 1120
},
{
"epoch": 0.13136384866884634,
"grad_norm": 0.9041022334920666,
"learning_rate": 4.919213377608499e-05,
"loss": 0.7577,
"step": 1125
},
{
"epoch": 0.13194768799626344,
"grad_norm": 0.6981879096369867,
"learning_rate": 4.9180556035412876e-05,
"loss": 0.7411,
"step": 1130
},
{
"epoch": 0.13253152732368054,
"grad_norm": 0.8322803511194127,
"learning_rate": 4.916889746058242e-05,
"loss": 0.7646,
"step": 1135
},
{
"epoch": 0.1331153666510976,
"grad_norm": 0.8419357871549217,
"learning_rate": 4.9157158095061636e-05,
"loss": 0.7408,
"step": 1140
},
{
"epoch": 0.1336992059785147,
"grad_norm": 0.7278929815006521,
"learning_rate": 4.914533798261977e-05,
"loss": 0.7358,
"step": 1145
},
{
"epoch": 0.1342830453059318,
"grad_norm": 0.8230990854901412,
"learning_rate": 4.913343716732713e-05,
"loss": 0.7558,
"step": 1150
},
{
"epoch": 0.1348668846333489,
"grad_norm": 0.7572916262517394,
"learning_rate": 4.912145569355495e-05,
"loss": 0.7428,
"step": 1155
},
{
"epoch": 0.135450723960766,
"grad_norm": 0.7491860421832062,
"learning_rate": 4.910939360597514e-05,
"loss": 0.7296,
"step": 1160
},
{
"epoch": 0.13603456328818309,
"grad_norm": 0.7690234675019328,
"learning_rate": 4.909725094956019e-05,
"loss": 0.7538,
"step": 1165
},
{
"epoch": 0.13661840261560018,
"grad_norm": 0.7299285218059413,
"learning_rate": 4.908502776958301e-05,
"loss": 0.7384,
"step": 1170
},
{
"epoch": 0.13720224194301728,
"grad_norm": 0.9300595802171107,
"learning_rate": 4.907272411161668e-05,
"loss": 0.7383,
"step": 1175
},
{
"epoch": 0.13778608127043437,
"grad_norm": 0.7337159957589131,
"learning_rate": 4.9060340021534415e-05,
"loss": 0.74,
"step": 1180
},
{
"epoch": 0.13836992059785147,
"grad_norm": 0.7259120613159923,
"learning_rate": 4.9047875545509235e-05,
"loss": 0.7341,
"step": 1185
},
{
"epoch": 0.13895375992526857,
"grad_norm": 0.7044955504234521,
"learning_rate": 4.9035330730013926e-05,
"loss": 0.7352,
"step": 1190
},
{
"epoch": 0.13953759925268566,
"grad_norm": 0.729708864538447,
"learning_rate": 4.9022705621820786e-05,
"loss": 0.7214,
"step": 1195
},
{
"epoch": 0.14012143858010276,
"grad_norm": 0.7149710232809768,
"learning_rate": 4.901000026800148e-05,
"loss": 0.7492,
"step": 1200
},
{
"epoch": 0.14070527790751985,
"grad_norm": 0.7537344737207899,
"learning_rate": 4.899721471592688e-05,
"loss": 0.7294,
"step": 1205
},
{
"epoch": 0.14128911723493695,
"grad_norm": 0.742958645757194,
"learning_rate": 4.898434901326685e-05,
"loss": 0.723,
"step": 1210
},
{
"epoch": 0.14187295656235405,
"grad_norm": 0.7286315223401272,
"learning_rate": 4.897140320799011e-05,
"loss": 0.7301,
"step": 1215
},
{
"epoch": 0.14245679588977114,
"grad_norm": 0.7421491652969855,
"learning_rate": 4.8958377348364e-05,
"loss": 0.7498,
"step": 1220
},
{
"epoch": 0.14304063521718824,
"grad_norm": 0.7414371484478484,
"learning_rate": 4.894527148295438e-05,
"loss": 0.7477,
"step": 1225
},
{
"epoch": 0.14362447454460534,
"grad_norm": 0.9459047173148238,
"learning_rate": 4.8932085660625374e-05,
"loss": 0.751,
"step": 1230
},
{
"epoch": 0.14420831387202243,
"grad_norm": 0.7886617214535956,
"learning_rate": 4.8918819930539244e-05,
"loss": 0.7365,
"step": 1235
},
{
"epoch": 0.14479215319943953,
"grad_norm": 0.846908444975893,
"learning_rate": 4.8905474342156144e-05,
"loss": 0.7453,
"step": 1240
},
{
"epoch": 0.1453759925268566,
"grad_norm": 0.7469618310264896,
"learning_rate": 4.889204894523401e-05,
"loss": 0.7457,
"step": 1245
},
{
"epoch": 0.1459598318542737,
"grad_norm": 0.7542688436084458,
"learning_rate": 4.8878543789828314e-05,
"loss": 0.7377,
"step": 1250
},
{
"epoch": 0.1465436711816908,
"grad_norm": 0.7035185258217505,
"learning_rate": 4.886495892629191e-05,
"loss": 0.7524,
"step": 1255
},
{
"epoch": 0.14712751050910788,
"grad_norm": 0.7088045343755188,
"learning_rate": 4.8851294405274855e-05,
"loss": 0.722,
"step": 1260
},
{
"epoch": 0.14771134983652498,
"grad_norm": 0.7155027999309841,
"learning_rate": 4.8837550277724165e-05,
"loss": 0.7308,
"step": 1265
},
{
"epoch": 0.14829518916394208,
"grad_norm": 0.7781935060677897,
"learning_rate": 4.8823726594883696e-05,
"loss": 0.7436,
"step": 1270
},
{
"epoch": 0.14887902849135917,
"grad_norm": 0.7515943007662474,
"learning_rate": 4.8809823408293887e-05,
"loss": 0.7443,
"step": 1275
},
{
"epoch": 0.14946286781877627,
"grad_norm": 0.7058532715642434,
"learning_rate": 4.8795840769791634e-05,
"loss": 0.7487,
"step": 1280
},
{
"epoch": 0.15004670714619336,
"grad_norm": 0.7950171862665167,
"learning_rate": 4.878177873151004e-05,
"loss": 0.7434,
"step": 1285
},
{
"epoch": 0.15063054647361046,
"grad_norm": 0.7274427585504183,
"learning_rate": 4.876763734587825e-05,
"loss": 0.7345,
"step": 1290
},
{
"epoch": 0.15121438580102756,
"grad_norm": 0.6601643434457186,
"learning_rate": 4.8753416665621255e-05,
"loss": 0.7186,
"step": 1295
},
{
"epoch": 0.15179822512844465,
"grad_norm": 0.6860883461904687,
"learning_rate": 4.873911674375968e-05,
"loss": 0.7501,
"step": 1300
},
{
"epoch": 0.15238206445586175,
"grad_norm": 0.8052239552003218,
"learning_rate": 4.87247376336096e-05,
"loss": 0.745,
"step": 1305
},
{
"epoch": 0.15296590378327884,
"grad_norm": 0.7778289155585636,
"learning_rate": 4.8710279388782345e-05,
"loss": 0.7369,
"step": 1310
},
{
"epoch": 0.15354974311069594,
"grad_norm": 0.7450131152812045,
"learning_rate": 4.869574206318427e-05,
"loss": 0.7281,
"step": 1315
},
{
"epoch": 0.15413358243811304,
"grad_norm": 0.6824661780591835,
"learning_rate": 4.868112571101659e-05,
"loss": 0.7115,
"step": 1320
},
{
"epoch": 0.15471742176553013,
"grad_norm": 0.7908470051891671,
"learning_rate": 4.866643038677519e-05,
"loss": 0.7415,
"step": 1325
},
{
"epoch": 0.15530126109294723,
"grad_norm": 0.8076628638288811,
"learning_rate": 4.865165614525033e-05,
"loss": 0.7425,
"step": 1330
},
{
"epoch": 0.15588510042036433,
"grad_norm": 0.7177283048497456,
"learning_rate": 4.863680304152657e-05,
"loss": 0.7457,
"step": 1335
},
{
"epoch": 0.15646893974778142,
"grad_norm": 0.7665358490767893,
"learning_rate": 4.862187113098249e-05,
"loss": 0.7458,
"step": 1340
},
{
"epoch": 0.15705277907519852,
"grad_norm": 0.807745344355968,
"learning_rate": 4.8606860469290454e-05,
"loss": 0.733,
"step": 1345
},
{
"epoch": 0.1576366184026156,
"grad_norm": 0.7713265125912636,
"learning_rate": 4.859177111241649e-05,
"loss": 0.7175,
"step": 1350
},
{
"epoch": 0.15822045773003268,
"grad_norm": 0.7831841494722501,
"learning_rate": 4.8576603116620004e-05,
"loss": 0.7318,
"step": 1355
},
{
"epoch": 0.15880429705744978,
"grad_norm": 0.8627481452369917,
"learning_rate": 4.8561356538453625e-05,
"loss": 0.721,
"step": 1360
},
{
"epoch": 0.15938813638486687,
"grad_norm": 0.7424567289172562,
"learning_rate": 4.8546031434762954e-05,
"loss": 0.7548,
"step": 1365
},
{
"epoch": 0.15997197571228397,
"grad_norm": 0.7337905743450628,
"learning_rate": 4.853062786268636e-05,
"loss": 0.7497,
"step": 1370
},
{
"epoch": 0.16055581503970107,
"grad_norm": 0.7476013018793656,
"learning_rate": 4.85151458796548e-05,
"loss": 0.7163,
"step": 1375
},
{
"epoch": 0.16113965436711816,
"grad_norm": 0.8043212279278502,
"learning_rate": 4.849958554339156e-05,
"loss": 0.7531,
"step": 1380
},
{
"epoch": 0.16172349369453526,
"grad_norm": 0.8977875103848776,
"learning_rate": 4.8483946911912064e-05,
"loss": 0.7444,
"step": 1385
},
{
"epoch": 0.16230733302195235,
"grad_norm": 0.7817709982244301,
"learning_rate": 4.846823004352366e-05,
"loss": 0.7258,
"step": 1390
},
{
"epoch": 0.16289117234936945,
"grad_norm": 0.8098880241300683,
"learning_rate": 4.845243499682539e-05,
"loss": 0.7509,
"step": 1395
},
{
"epoch": 0.16347501167678655,
"grad_norm": 0.770709782942639,
"learning_rate": 4.8436561830707786e-05,
"loss": 0.7327,
"step": 1400
},
{
"epoch": 0.16405885100420364,
"grad_norm": 0.7668990726836052,
"learning_rate": 4.842061060435261e-05,
"loss": 0.729,
"step": 1405
},
{
"epoch": 0.16464269033162074,
"grad_norm": 0.7164525231757792,
"learning_rate": 4.840458137723271e-05,
"loss": 0.7014,
"step": 1410
},
{
"epoch": 0.16522652965903784,
"grad_norm": 0.8840366074204892,
"learning_rate": 4.838847420911172e-05,
"loss": 0.7221,
"step": 1415
},
{
"epoch": 0.16581036898645493,
"grad_norm": 0.7844725444750456,
"learning_rate": 4.8372289160043895e-05,
"loss": 0.7388,
"step": 1420
},
{
"epoch": 0.16639420831387203,
"grad_norm": 0.7986461499605559,
"learning_rate": 4.835602629037384e-05,
"loss": 0.7298,
"step": 1425
},
{
"epoch": 0.16697804764128912,
"grad_norm": 0.6968449481556948,
"learning_rate": 4.8339685660736324e-05,
"loss": 0.723,
"step": 1430
},
{
"epoch": 0.16756188696870622,
"grad_norm": 0.7822286977526196,
"learning_rate": 4.8323267332056026e-05,
"loss": 0.7334,
"step": 1435
},
{
"epoch": 0.16814572629612332,
"grad_norm": 0.6954977452551738,
"learning_rate": 4.830677136554733e-05,
"loss": 0.7311,
"step": 1440
},
{
"epoch": 0.1687295656235404,
"grad_norm": 0.8436171213392395,
"learning_rate": 4.829019782271408e-05,
"loss": 0.7519,
"step": 1445
},
{
"epoch": 0.1693134049509575,
"grad_norm": 0.7046391770080698,
"learning_rate": 4.827354676534937e-05,
"loss": 0.7459,
"step": 1450
},
{
"epoch": 0.1698972442783746,
"grad_norm": 0.719980747686207,
"learning_rate": 4.825681825553527e-05,
"loss": 0.739,
"step": 1455
},
{
"epoch": 0.17048108360579167,
"grad_norm": 0.7103564668833614,
"learning_rate": 4.824001235564265e-05,
"loss": 0.7458,
"step": 1460
},
{
"epoch": 0.17106492293320877,
"grad_norm": 0.6864678038650494,
"learning_rate": 4.822312912833092e-05,
"loss": 0.7297,
"step": 1465
},
{
"epoch": 0.17164876226062586,
"grad_norm": 0.7328592434686132,
"learning_rate": 4.82061686365478e-05,
"loss": 0.7327,
"step": 1470
},
{
"epoch": 0.17223260158804296,
"grad_norm": 0.7571325709685475,
"learning_rate": 4.818913094352907e-05,
"loss": 0.7302,
"step": 1475
},
{
"epoch": 0.17281644091546006,
"grad_norm": 0.7179359161509148,
"learning_rate": 4.8172016112798364e-05,
"loss": 0.7444,
"step": 1480
},
{
"epoch": 0.17340028024287715,
"grad_norm": 0.7722153034297365,
"learning_rate": 4.8154824208166906e-05,
"loss": 0.7595,
"step": 1485
},
{
"epoch": 0.17398411957029425,
"grad_norm": 0.7363478679562057,
"learning_rate": 4.8137555293733294e-05,
"loss": 0.7263,
"step": 1490
},
{
"epoch": 0.17456795889771135,
"grad_norm": 0.8124460587287587,
"learning_rate": 4.812020943388324e-05,
"loss": 0.7124,
"step": 1495
},
{
"epoch": 0.17515179822512844,
"grad_norm": 0.7148964393172884,
"learning_rate": 4.810278669328935e-05,
"loss": 0.7429,
"step": 1500
},
{
"epoch": 0.17573563755254554,
"grad_norm": 0.8468862512438853,
"learning_rate": 4.808528713691087e-05,
"loss": 0.7166,
"step": 1505
},
{
"epoch": 0.17631947687996263,
"grad_norm": 0.7834963307553772,
"learning_rate": 4.806771082999346e-05,
"loss": 0.7488,
"step": 1510
},
{
"epoch": 0.17690331620737973,
"grad_norm": 0.7261513735919553,
"learning_rate": 4.8050057838068904e-05,
"loss": 0.7421,
"step": 1515
},
{
"epoch": 0.17748715553479683,
"grad_norm": 0.7551923969507343,
"learning_rate": 4.803232822695493e-05,
"loss": 0.7337,
"step": 1520
},
{
"epoch": 0.17807099486221392,
"grad_norm": 0.8296978285255314,
"learning_rate": 4.801452206275493e-05,
"loss": 0.7281,
"step": 1525
},
{
"epoch": 0.17865483418963102,
"grad_norm": 0.7081211489282102,
"learning_rate": 4.79966394118577e-05,
"loss": 0.7216,
"step": 1530
},
{
"epoch": 0.1792386735170481,
"grad_norm": 0.803686277740153,
"learning_rate": 4.797868034093724e-05,
"loss": 0.7476,
"step": 1535
},
{
"epoch": 0.1798225128444652,
"grad_norm": 0.7401591404567419,
"learning_rate": 4.7960644916952444e-05,
"loss": 0.7417,
"step": 1540
},
{
"epoch": 0.1804063521718823,
"grad_norm": 0.7893838464964483,
"learning_rate": 4.7942533207146916e-05,
"loss": 0.7498,
"step": 1545
},
{
"epoch": 0.1809901914992994,
"grad_norm": 0.8188234554767124,
"learning_rate": 4.792434527904864e-05,
"loss": 0.7422,
"step": 1550
},
{
"epoch": 0.1815740308267165,
"grad_norm": 0.6461357446193222,
"learning_rate": 4.7906081200469835e-05,
"loss": 0.7303,
"step": 1555
},
{
"epoch": 0.1821578701541336,
"grad_norm": 0.6843427149385607,
"learning_rate": 4.788774103950657e-05,
"loss": 0.726,
"step": 1560
},
{
"epoch": 0.1827417094815507,
"grad_norm": 0.9580042770236428,
"learning_rate": 4.7869324864538636e-05,
"loss": 0.732,
"step": 1565
},
{
"epoch": 0.18332554880896776,
"grad_norm": 0.7146642982640992,
"learning_rate": 4.7850832744229216e-05,
"loss": 0.7397,
"step": 1570
},
{
"epoch": 0.18390938813638485,
"grad_norm": 0.8365768389202328,
"learning_rate": 4.783226474752465e-05,
"loss": 0.7383,
"step": 1575
},
{
"epoch": 0.18449322746380195,
"grad_norm": 0.7990966260259673,
"learning_rate": 4.781362094365417e-05,
"loss": 0.7196,
"step": 1580
},
{
"epoch": 0.18507706679121905,
"grad_norm": 0.8345679211677212,
"learning_rate": 4.779490140212966e-05,
"loss": 0.7182,
"step": 1585
},
{
"epoch": 0.18566090611863614,
"grad_norm": 0.7251252240357575,
"learning_rate": 4.777610619274539e-05,
"loss": 0.735,
"step": 1590
},
{
"epoch": 0.18624474544605324,
"grad_norm": 0.7016107904965909,
"learning_rate": 4.775723538557772e-05,
"loss": 0.7144,
"step": 1595
},
{
"epoch": 0.18682858477347034,
"grad_norm": 0.7741910944117159,
"learning_rate": 4.7738289050984905e-05,
"loss": 0.7393,
"step": 1600
},
{
"epoch": 0.18741242410088743,
"grad_norm": 0.826306902362388,
"learning_rate": 4.7719267259606795e-05,
"loss": 0.7278,
"step": 1605
},
{
"epoch": 0.18799626342830453,
"grad_norm": 0.7428959175092528,
"learning_rate": 4.770017008236455e-05,
"loss": 0.7404,
"step": 1610
},
{
"epoch": 0.18858010275572162,
"grad_norm": 0.7152156671162925,
"learning_rate": 4.768099759046042e-05,
"loss": 0.711,
"step": 1615
},
{
"epoch": 0.18916394208313872,
"grad_norm": 0.7243354015604727,
"learning_rate": 4.766174985537745e-05,
"loss": 0.7043,
"step": 1620
},
{
"epoch": 0.18974778141055582,
"grad_norm": 0.7070692111855642,
"learning_rate": 4.7642426948879234e-05,
"loss": 0.7115,
"step": 1625
},
{
"epoch": 0.1903316207379729,
"grad_norm": 0.7526365457947733,
"learning_rate": 4.762302894300962e-05,
"loss": 0.7443,
"step": 1630
},
{
"epoch": 0.19091546006539,
"grad_norm": 0.7400669421223262,
"learning_rate": 4.760355591009247e-05,
"loss": 0.7397,
"step": 1635
},
{
"epoch": 0.1914992993928071,
"grad_norm": 0.6973523839093395,
"learning_rate": 4.7584007922731383e-05,
"loss": 0.7357,
"step": 1640
},
{
"epoch": 0.1920831387202242,
"grad_norm": 0.7222361309671663,
"learning_rate": 4.75643850538094e-05,
"loss": 0.7169,
"step": 1645
},
{
"epoch": 0.1926669780476413,
"grad_norm": 0.765478532510268,
"learning_rate": 4.754468737648878e-05,
"loss": 0.7252,
"step": 1650
},
{
"epoch": 0.1932508173750584,
"grad_norm": 0.7492897945719802,
"learning_rate": 4.752491496421066e-05,
"loss": 0.7308,
"step": 1655
},
{
"epoch": 0.1938346567024755,
"grad_norm": 0.6713564084526321,
"learning_rate": 4.750506789069486e-05,
"loss": 0.7257,
"step": 1660
},
{
"epoch": 0.19441849602989258,
"grad_norm": 0.7445920572569636,
"learning_rate": 4.7485146229939545e-05,
"loss": 0.7235,
"step": 1665
},
{
"epoch": 0.19500233535730968,
"grad_norm": 0.7538942068389883,
"learning_rate": 4.746515005622097e-05,
"loss": 0.7179,
"step": 1670
},
{
"epoch": 0.19558617468472678,
"grad_norm": 0.7816896420662496,
"learning_rate": 4.744507944409322e-05,
"loss": 0.7437,
"step": 1675
},
{
"epoch": 0.19617001401214385,
"grad_norm": 0.6863940434432573,
"learning_rate": 4.742493446838791e-05,
"loss": 0.7104,
"step": 1680
},
{
"epoch": 0.19675385333956094,
"grad_norm": 0.7694997012291778,
"learning_rate": 4.740471520421392e-05,
"loss": 0.7256,
"step": 1685
},
{
"epoch": 0.19733769266697804,
"grad_norm": 0.7486931532940556,
"learning_rate": 4.73844217269571e-05,
"loss": 0.7238,
"step": 1690
},
{
"epoch": 0.19792153199439513,
"grad_norm": 0.6904775873596434,
"learning_rate": 4.736405411228e-05,
"loss": 0.7143,
"step": 1695
},
{
"epoch": 0.19850537132181223,
"grad_norm": 0.7482659898128301,
"learning_rate": 4.7343612436121575e-05,
"loss": 0.7623,
"step": 1700
},
{
"epoch": 0.19908921064922933,
"grad_norm": 0.7314789656035418,
"learning_rate": 4.732309677469693e-05,
"loss": 0.7412,
"step": 1705
},
{
"epoch": 0.19967304997664642,
"grad_norm": 0.6899940552720958,
"learning_rate": 4.7302507204497026e-05,
"loss": 0.719,
"step": 1710
},
{
"epoch": 0.20025688930406352,
"grad_norm": 0.7282691554214673,
"learning_rate": 4.728184380228834e-05,
"loss": 0.7374,
"step": 1715
},
{
"epoch": 0.20084072863148061,
"grad_norm": 0.7013315312116346,
"learning_rate": 4.7261106645112677e-05,
"loss": 0.7274,
"step": 1720
},
{
"epoch": 0.2014245679588977,
"grad_norm": 0.7819768242805057,
"learning_rate": 4.72402958102868e-05,
"loss": 0.7461,
"step": 1725
},
{
"epoch": 0.2020084072863148,
"grad_norm": 0.7309301514396227,
"learning_rate": 4.72194113754022e-05,
"loss": 0.7296,
"step": 1730
},
{
"epoch": 0.2025922466137319,
"grad_norm": 0.7487524023177833,
"learning_rate": 4.719845341832475e-05,
"loss": 0.6885,
"step": 1735
},
{
"epoch": 0.203176085941149,
"grad_norm": 0.9788588285692689,
"learning_rate": 4.7177422017194464e-05,
"loss": 0.7199,
"step": 1740
},
{
"epoch": 0.2037599252685661,
"grad_norm": 0.7918374375300556,
"learning_rate": 4.715631725042517e-05,
"loss": 0.7422,
"step": 1745
},
{
"epoch": 0.2043437645959832,
"grad_norm": 0.8055570505724186,
"learning_rate": 4.7135139196704254e-05,
"loss": 0.7235,
"step": 1750
},
{
"epoch": 0.2049276039234003,
"grad_norm": 0.7196921493988918,
"learning_rate": 4.711388793499233e-05,
"loss": 0.7204,
"step": 1755
},
{
"epoch": 0.20551144325081738,
"grad_norm": 0.8221920631131578,
"learning_rate": 4.709256354452298e-05,
"loss": 0.7296,
"step": 1760
},
{
"epoch": 0.20609528257823448,
"grad_norm": 0.8127617904984341,
"learning_rate": 4.7071166104802415e-05,
"loss": 0.7245,
"step": 1765
},
{
"epoch": 0.20667912190565157,
"grad_norm": 0.7043304691367893,
"learning_rate": 4.7049695695609224e-05,
"loss": 0.7103,
"step": 1770
},
{
"epoch": 0.20726296123306867,
"grad_norm": 0.649014116022384,
"learning_rate": 4.702815239699405e-05,
"loss": 0.7146,
"step": 1775
},
{
"epoch": 0.20784680056048577,
"grad_norm": 0.7140594319480008,
"learning_rate": 4.7006536289279285e-05,
"loss": 0.7175,
"step": 1780
},
{
"epoch": 0.20843063988790284,
"grad_norm": 0.7241345764959499,
"learning_rate": 4.698484745305882e-05,
"loss": 0.7007,
"step": 1785
},
{
"epoch": 0.20901447921531993,
"grad_norm": 0.7813784340987022,
"learning_rate": 4.696308596919767e-05,
"loss": 0.7457,
"step": 1790
},
{
"epoch": 0.20959831854273703,
"grad_norm": 0.7070774934064795,
"learning_rate": 4.694125191883174e-05,
"loss": 0.7188,
"step": 1795
},
{
"epoch": 0.21018215787015412,
"grad_norm": 0.6804544243910975,
"learning_rate": 4.691934538336746e-05,
"loss": 0.725,
"step": 1800
},
{
"epoch": 0.21076599719757122,
"grad_norm": 0.6728574938989879,
"learning_rate": 4.6897366444481545e-05,
"loss": 0.7181,
"step": 1805
},
{
"epoch": 0.21134983652498832,
"grad_norm": 0.7631516107921738,
"learning_rate": 4.687531518412065e-05,
"loss": 0.73,
"step": 1810
},
{
"epoch": 0.2119336758524054,
"grad_norm": 0.7241736870057862,
"learning_rate": 4.685319168450107e-05,
"loss": 0.6938,
"step": 1815
},
{
"epoch": 0.2125175151798225,
"grad_norm": 0.7095684376204812,
"learning_rate": 4.683099602810845e-05,
"loss": 0.7445,
"step": 1820
},
{
"epoch": 0.2131013545072396,
"grad_norm": 0.6789241533004805,
"learning_rate": 4.680872829769745e-05,
"loss": 0.7139,
"step": 1825
},
{
"epoch": 0.2136851938346567,
"grad_norm": 0.6959715900470123,
"learning_rate": 4.6786388576291446e-05,
"loss": 0.738,
"step": 1830
},
{
"epoch": 0.2142690331620738,
"grad_norm": 0.7209917479235074,
"learning_rate": 4.6763976947182256e-05,
"loss": 0.7143,
"step": 1835
},
{
"epoch": 0.2148528724894909,
"grad_norm": 0.7846008482256338,
"learning_rate": 4.6741493493929794e-05,
"loss": 0.7193,
"step": 1840
},
{
"epoch": 0.215436711816908,
"grad_norm": 0.8188656249048588,
"learning_rate": 4.671893830036174e-05,
"loss": 0.7274,
"step": 1845
},
{
"epoch": 0.21602055114432508,
"grad_norm": 0.8194046437238091,
"learning_rate": 4.6696311450573266e-05,
"loss": 0.7446,
"step": 1850
},
{
"epoch": 0.21660439047174218,
"grad_norm": 0.8430873361879598,
"learning_rate": 4.667361302892671e-05,
"loss": 0.7034,
"step": 1855
},
{
"epoch": 0.21718822979915928,
"grad_norm": 0.8546100923096231,
"learning_rate": 4.665084312005126e-05,
"loss": 0.6924,
"step": 1860
},
{
"epoch": 0.21777206912657637,
"grad_norm": 0.710893431040641,
"learning_rate": 4.662800180884263e-05,
"loss": 0.7386,
"step": 1865
},
{
"epoch": 0.21835590845399347,
"grad_norm": 0.7830156422797829,
"learning_rate": 4.660508918046277e-05,
"loss": 0.7225,
"step": 1870
},
{
"epoch": 0.21893974778141057,
"grad_norm": 0.691958056855305,
"learning_rate": 4.658210532033951e-05,
"loss": 0.7287,
"step": 1875
},
{
"epoch": 0.21952358710882766,
"grad_norm": 0.711708610990222,
"learning_rate": 4.6559050314166264e-05,
"loss": 0.7167,
"step": 1880
},
{
"epoch": 0.22010742643624476,
"grad_norm": 0.7122500688909733,
"learning_rate": 4.653592424790172e-05,
"loss": 0.731,
"step": 1885
},
{
"epoch": 0.22069126576366185,
"grad_norm": 0.743470720917725,
"learning_rate": 4.6512727207769504e-05,
"loss": 0.7335,
"step": 1890
},
{
"epoch": 0.22127510509107892,
"grad_norm": 0.7437561107243308,
"learning_rate": 4.6489459280257856e-05,
"loss": 0.716,
"step": 1895
},
{
"epoch": 0.22185894441849602,
"grad_norm": 0.7146414813002067,
"learning_rate": 4.646612055211933e-05,
"loss": 0.7393,
"step": 1900
},
{
"epoch": 0.22244278374591311,
"grad_norm": 0.709665611856685,
"learning_rate": 4.6442711110370424e-05,
"loss": 0.697,
"step": 1905
},
{
"epoch": 0.2230266230733302,
"grad_norm": 0.9843066670816354,
"learning_rate": 4.64192310422913e-05,
"loss": 0.7088,
"step": 1910
},
{
"epoch": 0.2236104624007473,
"grad_norm": 0.6457620908315918,
"learning_rate": 4.639568043542548e-05,
"loss": 0.7035,
"step": 1915
},
{
"epoch": 0.2241943017281644,
"grad_norm": 0.708708239642946,
"learning_rate": 4.6372059377579414e-05,
"loss": 0.7217,
"step": 1920
},
{
"epoch": 0.2247781410555815,
"grad_norm": 0.7500113479201387,
"learning_rate": 4.634836795682228e-05,
"loss": 0.7147,
"step": 1925
},
{
"epoch": 0.2253619803829986,
"grad_norm": 0.7119356065347904,
"learning_rate": 4.632460626148558e-05,
"loss": 0.7259,
"step": 1930
},
{
"epoch": 0.2259458197104157,
"grad_norm": 0.6943499671147205,
"learning_rate": 4.6300774380162825e-05,
"loss": 0.7299,
"step": 1935
},
{
"epoch": 0.2265296590378328,
"grad_norm": 0.7890957040844898,
"learning_rate": 4.627687240170921e-05,
"loss": 0.72,
"step": 1940
},
{
"epoch": 0.22711349836524988,
"grad_norm": 0.7431464489795075,
"learning_rate": 4.625290041524128e-05,
"loss": 0.7227,
"step": 1945
},
{
"epoch": 0.22769733769266698,
"grad_norm": 0.6790888225168289,
"learning_rate": 4.6228858510136616e-05,
"loss": 0.7137,
"step": 1950
},
{
"epoch": 0.22828117702008408,
"grad_norm": 0.7157071181310045,
"learning_rate": 4.620474677603345e-05,
"loss": 0.7235,
"step": 1955
},
{
"epoch": 0.22886501634750117,
"grad_norm": 0.7285656420756189,
"learning_rate": 4.61805653028304e-05,
"loss": 0.7188,
"step": 1960
},
{
"epoch": 0.22944885567491827,
"grad_norm": 0.798435835459099,
"learning_rate": 4.615631418068609e-05,
"loss": 0.736,
"step": 1965
},
{
"epoch": 0.23003269500233536,
"grad_norm": 0.7297626012745467,
"learning_rate": 4.613199350001881e-05,
"loss": 0.7306,
"step": 1970
},
{
"epoch": 0.23061653432975246,
"grad_norm": 0.7425217549848289,
"learning_rate": 4.6107603351506205e-05,
"loss": 0.7235,
"step": 1975
},
{
"epoch": 0.23120037365716956,
"grad_norm": 0.7220004291903487,
"learning_rate": 4.608314382608493e-05,
"loss": 0.7173,
"step": 1980
},
{
"epoch": 0.23178421298458665,
"grad_norm": 0.7173901963009753,
"learning_rate": 4.6058615014950315e-05,
"loss": 0.7338,
"step": 1985
},
{
"epoch": 0.23236805231200375,
"grad_norm": 0.6790506753795957,
"learning_rate": 4.6034017009555975e-05,
"loss": 0.7272,
"step": 1990
},
{
"epoch": 0.23295189163942084,
"grad_norm": 0.7047642196241657,
"learning_rate": 4.600934990161355e-05,
"loss": 0.7139,
"step": 1995
},
{
"epoch": 0.23353573096683794,
"grad_norm": 0.758097998754232,
"learning_rate": 4.598461378309231e-05,
"loss": 0.7262,
"step": 2000
},
{
"epoch": 0.234119570294255,
"grad_norm": 0.6628045373040543,
"learning_rate": 4.5959808746218823e-05,
"loss": 0.7118,
"step": 2005
},
{
"epoch": 0.2347034096216721,
"grad_norm": 0.7054626068525848,
"learning_rate": 4.593493488347662e-05,
"loss": 0.7231,
"step": 2010
},
{
"epoch": 0.2352872489490892,
"grad_norm": 0.7292550895736654,
"learning_rate": 4.590999228760583e-05,
"loss": 0.7192,
"step": 2015
},
{
"epoch": 0.2358710882765063,
"grad_norm": 0.7267049938742051,
"learning_rate": 4.5884981051602873e-05,
"loss": 0.7015,
"step": 2020
},
{
"epoch": 0.2364549276039234,
"grad_norm": 0.6455506674222412,
"learning_rate": 4.585990126872006e-05,
"loss": 0.7084,
"step": 2025
},
{
"epoch": 0.2370387669313405,
"grad_norm": 0.659962308727248,
"learning_rate": 4.583475303246527e-05,
"loss": 0.7055,
"step": 2030
},
{
"epoch": 0.23762260625875758,
"grad_norm": 0.6480329607760619,
"learning_rate": 4.580953643660165e-05,
"loss": 0.7121,
"step": 2035
},
{
"epoch": 0.23820644558617468,
"grad_norm": 0.7606715846725551,
"learning_rate": 4.5784251575147176e-05,
"loss": 0.7292,
"step": 2040
},
{
"epoch": 0.23879028491359178,
"grad_norm": 0.715112048397896,
"learning_rate": 4.5758898542374354e-05,
"loss": 0.7255,
"step": 2045
},
{
"epoch": 0.23937412424100887,
"grad_norm": 0.6867461810383618,
"learning_rate": 4.5733477432809884e-05,
"loss": 0.7022,
"step": 2050
},
{
"epoch": 0.23995796356842597,
"grad_norm": 0.643764497714989,
"learning_rate": 4.570798834123425e-05,
"loss": 0.7079,
"step": 2055
},
{
"epoch": 0.24054180289584307,
"grad_norm": 0.6735689952721187,
"learning_rate": 4.5682431362681435e-05,
"loss": 0.7087,
"step": 2060
},
{
"epoch": 0.24112564222326016,
"grad_norm": 0.6466696791306183,
"learning_rate": 4.565680659243851e-05,
"loss": 0.6947,
"step": 2065
},
{
"epoch": 0.24170948155067726,
"grad_norm": 0.6809481402735733,
"learning_rate": 4.5631114126045315e-05,
"loss": 0.7152,
"step": 2070
},
{
"epoch": 0.24229332087809435,
"grad_norm": 0.6348521978462173,
"learning_rate": 4.560535405929408e-05,
"loss": 0.7037,
"step": 2075
},
{
"epoch": 0.24287716020551145,
"grad_norm": 0.6796805499806474,
"learning_rate": 4.557952648822908e-05,
"loss": 0.7162,
"step": 2080
},
{
"epoch": 0.24346099953292855,
"grad_norm": 0.7185561925920844,
"learning_rate": 4.555363150914628e-05,
"loss": 0.7287,
"step": 2085
},
{
"epoch": 0.24404483886034564,
"grad_norm": 0.6818164108488127,
"learning_rate": 4.552766921859297e-05,
"loss": 0.7048,
"step": 2090
},
{
"epoch": 0.24462867818776274,
"grad_norm": 0.7900603785829974,
"learning_rate": 4.5501639713367386e-05,
"loss": 0.6947,
"step": 2095
},
{
"epoch": 0.24521251751517983,
"grad_norm": 0.8122940747156205,
"learning_rate": 4.547554309051839e-05,
"loss": 0.7334,
"step": 2100
},
{
"epoch": 0.24579635684259693,
"grad_norm": 0.7539481589316036,
"learning_rate": 4.5449379447345084e-05,
"loss": 0.7192,
"step": 2105
},
{
"epoch": 0.246380196170014,
"grad_norm": 0.7714231523906511,
"learning_rate": 4.5423148881396444e-05,
"loss": 0.7252,
"step": 2110
},
{
"epoch": 0.2469640354974311,
"grad_norm": 0.7271119619546056,
"learning_rate": 4.539685149047097e-05,
"loss": 0.7151,
"step": 2115
},
{
"epoch": 0.2475478748248482,
"grad_norm": 0.7408265661316124,
"learning_rate": 4.5370487372616285e-05,
"loss": 0.7064,
"step": 2120
},
{
"epoch": 0.2481317141522653,
"grad_norm": 0.6622899868579666,
"learning_rate": 4.5344056626128847e-05,
"loss": 0.7248,
"step": 2125
},
{
"epoch": 0.24871555347968238,
"grad_norm": 0.7402058026741981,
"learning_rate": 4.53175593495535e-05,
"loss": 0.7198,
"step": 2130
},
{
"epoch": 0.24929939280709948,
"grad_norm": 0.6980377237367063,
"learning_rate": 4.529099564168312e-05,
"loss": 0.7281,
"step": 2135
},
{
"epoch": 0.24988323213451658,
"grad_norm": 0.6874041759692474,
"learning_rate": 4.526436560155833e-05,
"loss": 0.7128,
"step": 2140
},
{
"epoch": 0.2504670714619337,
"grad_norm": 0.7071543054401014,
"learning_rate": 4.5237669328467e-05,
"loss": 0.7013,
"step": 2145
},
{
"epoch": 0.25105091078935077,
"grad_norm": 0.8532610188441806,
"learning_rate": 4.5210906921944e-05,
"loss": 0.7175,
"step": 2150
},
{
"epoch": 0.2516347501167679,
"grad_norm": 0.7039095592542086,
"learning_rate": 4.518407848177073e-05,
"loss": 0.7254,
"step": 2155
},
{
"epoch": 0.25221858944418496,
"grad_norm": 0.714656007800895,
"learning_rate": 4.515718410797481e-05,
"loss": 0.7381,
"step": 2160
},
{
"epoch": 0.25280242877160203,
"grad_norm": 0.8251368166360454,
"learning_rate": 4.513022390082969e-05,
"loss": 0.7019,
"step": 2165
},
{
"epoch": 0.25338626809901915,
"grad_norm": 0.6503458888327505,
"learning_rate": 4.510319796085428e-05,
"loss": 0.7318,
"step": 2170
},
{
"epoch": 0.2539701074264362,
"grad_norm": 0.70200635610956,
"learning_rate": 4.5076106388812534e-05,
"loss": 0.7242,
"step": 2175
},
{
"epoch": 0.25455394675385334,
"grad_norm": 0.6976305202866769,
"learning_rate": 4.504894928571315e-05,
"loss": 0.7154,
"step": 2180
},
{
"epoch": 0.2551377860812704,
"grad_norm": 0.6885582444665725,
"learning_rate": 4.502172675280915e-05,
"loss": 0.7078,
"step": 2185
},
{
"epoch": 0.25572162540868754,
"grad_norm": 0.7713787736354188,
"learning_rate": 4.4994438891597486e-05,
"loss": 0.7604,
"step": 2190
},
{
"epoch": 0.2563054647361046,
"grad_norm": 0.6839857411452482,
"learning_rate": 4.496708580381868e-05,
"loss": 0.7309,
"step": 2195
},
{
"epoch": 0.25688930406352173,
"grad_norm": 0.7530727282616239,
"learning_rate": 4.4939667591456465e-05,
"loss": 0.7093,
"step": 2200
},
{
"epoch": 0.2574731433909388,
"grad_norm": 0.7522898332263996,
"learning_rate": 4.491218435673737e-05,
"loss": 0.7188,
"step": 2205
},
{
"epoch": 0.2580569827183559,
"grad_norm": 0.6687177720045462,
"learning_rate": 4.4884636202130365e-05,
"loss": 0.7091,
"step": 2210
},
{
"epoch": 0.258640822045773,
"grad_norm": 0.7463298115893197,
"learning_rate": 4.485702323034647e-05,
"loss": 0.7179,
"step": 2215
},
{
"epoch": 0.2592246613731901,
"grad_norm": 0.7302081712396381,
"learning_rate": 4.4829345544338355e-05,
"loss": 0.7241,
"step": 2220
},
{
"epoch": 0.2598085007006072,
"grad_norm": 0.7606959932317213,
"learning_rate": 4.480160324729998e-05,
"loss": 0.7067,
"step": 2225
},
{
"epoch": 0.2603923400280243,
"grad_norm": 0.7189168959720732,
"learning_rate": 4.477379644266621e-05,
"loss": 0.7105,
"step": 2230
},
{
"epoch": 0.2609761793554414,
"grad_norm": 0.6519886164895211,
"learning_rate": 4.47459252341124e-05,
"loss": 0.7463,
"step": 2235
},
{
"epoch": 0.2615600186828585,
"grad_norm": 0.6945678381172148,
"learning_rate": 4.471798972555407e-05,
"loss": 0.7127,
"step": 2240
},
{
"epoch": 0.26214385801027557,
"grad_norm": 0.7717732570913138,
"learning_rate": 4.468999002114642e-05,
"loss": 0.7293,
"step": 2245
},
{
"epoch": 0.2627276973376927,
"grad_norm": 0.6834180271932923,
"learning_rate": 4.4661926225284057e-05,
"loss": 0.7104,
"step": 2250
},
{
"epoch": 0.26331153666510976,
"grad_norm": 0.6953443040568222,
"learning_rate": 4.463379844260051e-05,
"loss": 0.7229,
"step": 2255
},
{
"epoch": 0.2638953759925269,
"grad_norm": 0.7559329906618139,
"learning_rate": 4.460560677796788e-05,
"loss": 0.7036,
"step": 2260
},
{
"epoch": 0.26447921531994395,
"grad_norm": 0.7214193536775777,
"learning_rate": 4.4577351336496466e-05,
"loss": 0.717,
"step": 2265
},
{
"epoch": 0.2650630546473611,
"grad_norm": 0.7534161123387066,
"learning_rate": 4.454903222353433e-05,
"loss": 0.7013,
"step": 2270
},
{
"epoch": 0.26564689397477814,
"grad_norm": 0.6774263838856932,
"learning_rate": 4.4520649544666955e-05,
"loss": 0.713,
"step": 2275
},
{
"epoch": 0.2662307333021952,
"grad_norm": 0.6707315471580606,
"learning_rate": 4.4492203405716804e-05,
"loss": 0.7066,
"step": 2280
},
{
"epoch": 0.26681457262961233,
"grad_norm": 0.7262172413571504,
"learning_rate": 4.4463693912742944e-05,
"loss": 0.707,
"step": 2285
},
{
"epoch": 0.2673984119570294,
"grad_norm": 0.7280740771734855,
"learning_rate": 4.4435121172040674e-05,
"loss": 0.7073,
"step": 2290
},
{
"epoch": 0.2679822512844465,
"grad_norm": 0.6934167945142579,
"learning_rate": 4.4406485290141075e-05,
"loss": 0.6996,
"step": 2295
},
{
"epoch": 0.2685660906118636,
"grad_norm": 0.7041971735017232,
"learning_rate": 4.437778637381068e-05,
"loss": 0.7007,
"step": 2300
},
{
"epoch": 0.2691499299392807,
"grad_norm": 0.7866703196821602,
"learning_rate": 4.434902453005101e-05,
"loss": 0.7369,
"step": 2305
},
{
"epoch": 0.2697337692666978,
"grad_norm": 0.7061581650441343,
"learning_rate": 4.4320199866098216e-05,
"loss": 0.7065,
"step": 2310
},
{
"epoch": 0.2703176085941149,
"grad_norm": 0.711781682573645,
"learning_rate": 4.4291312489422684e-05,
"loss": 0.7282,
"step": 2315
},
{
"epoch": 0.270901447921532,
"grad_norm": 0.6745562342287523,
"learning_rate": 4.426236250772859e-05,
"loss": 0.708,
"step": 2320
},
{
"epoch": 0.2714852872489491,
"grad_norm": 0.6740154414616895,
"learning_rate": 4.423335002895358e-05,
"loss": 0.7042,
"step": 2325
},
{
"epoch": 0.27206912657636617,
"grad_norm": 0.7056547935721885,
"learning_rate": 4.420427516126822e-05,
"loss": 0.7317,
"step": 2330
},
{
"epoch": 0.2726529659037833,
"grad_norm": 0.6657623911020857,
"learning_rate": 4.4175138013075804e-05,
"loss": 0.7111,
"step": 2335
},
{
"epoch": 0.27323680523120036,
"grad_norm": 0.7553581713777339,
"learning_rate": 4.4145938693011747e-05,
"loss": 0.7108,
"step": 2340
},
{
"epoch": 0.2738206445586175,
"grad_norm": 0.8733645372868298,
"learning_rate": 4.4116677309943295e-05,
"loss": 0.7156,
"step": 2345
},
{
"epoch": 0.27440448388603456,
"grad_norm": 0.6923216566350364,
"learning_rate": 4.40873539729691e-05,
"loss": 0.7247,
"step": 2350
},
{
"epoch": 0.2749883232134517,
"grad_norm": 0.7032793290907281,
"learning_rate": 4.405796879141881e-05,
"loss": 0.7038,
"step": 2355
},
{
"epoch": 0.27557216254086875,
"grad_norm": 0.7366632971967906,
"learning_rate": 4.402852187485262e-05,
"loss": 0.7201,
"step": 2360
},
{
"epoch": 0.27615600186828587,
"grad_norm": 0.7421535169907494,
"learning_rate": 4.3999013333060936e-05,
"loss": 0.7227,
"step": 2365
},
{
"epoch": 0.27673984119570294,
"grad_norm": 0.7054520030021889,
"learning_rate": 4.396944327606389e-05,
"loss": 0.7031,
"step": 2370
},
{
"epoch": 0.27732368052312006,
"grad_norm": 0.6757686926397057,
"learning_rate": 4.393981181411102e-05,
"loss": 0.7241,
"step": 2375
},
{
"epoch": 0.27790751985053713,
"grad_norm": 0.6949088234243647,
"learning_rate": 4.3910119057680765e-05,
"loss": 0.6881,
"step": 2380
},
{
"epoch": 0.2784913591779542,
"grad_norm": 0.7569213644796755,
"learning_rate": 4.3880365117480114e-05,
"loss": 0.7225,
"step": 2385
},
{
"epoch": 0.2790751985053713,
"grad_norm": 0.7530951446510832,
"learning_rate": 4.385055010444416e-05,
"loss": 0.7028,
"step": 2390
},
{
"epoch": 0.2796590378327884,
"grad_norm": 0.7409067166871964,
"learning_rate": 4.382067412973573e-05,
"loss": 0.7003,
"step": 2395
},
{
"epoch": 0.2802428771602055,
"grad_norm": 0.681951341188302,
"learning_rate": 4.3790737304744906e-05,
"loss": 0.7189,
"step": 2400
},
{
"epoch": 0.2808267164876226,
"grad_norm": 0.6465059909172851,
"learning_rate": 4.376073974108866e-05,
"loss": 0.7225,
"step": 2405
},
{
"epoch": 0.2814105558150397,
"grad_norm": 0.6656442992103169,
"learning_rate": 4.373068155061043e-05,
"loss": 0.7069,
"step": 2410
},
{
"epoch": 0.2819943951424568,
"grad_norm": 0.6534552264402784,
"learning_rate": 4.37005628453797e-05,
"loss": 0.7,
"step": 2415
},
{
"epoch": 0.2825782344698739,
"grad_norm": 0.6952651195555617,
"learning_rate": 4.367038373769155e-05,
"loss": 0.7299,
"step": 2420
},
{
"epoch": 0.28316207379729097,
"grad_norm": 0.6842809970843287,
"learning_rate": 4.36401443400663e-05,
"loss": 0.7258,
"step": 2425
},
{
"epoch": 0.2837459131247081,
"grad_norm": 0.7331445788168038,
"learning_rate": 4.3609844765249034e-05,
"loss": 0.6946,
"step": 2430
},
{
"epoch": 0.28432975245212516,
"grad_norm": 0.7347255587932956,
"learning_rate": 4.357948512620922e-05,
"loss": 0.7048,
"step": 2435
},
{
"epoch": 0.2849135917795423,
"grad_norm": 0.8121169217743148,
"learning_rate": 4.354906553614024e-05,
"loss": 0.7125,
"step": 2440
},
{
"epoch": 0.28549743110695935,
"grad_norm": 0.6800833661450737,
"learning_rate": 4.3518586108459034e-05,
"loss": 0.6926,
"step": 2445
},
{
"epoch": 0.2860812704343765,
"grad_norm": 0.736383144444763,
"learning_rate": 4.34880469568056e-05,
"loss": 0.6916,
"step": 2450
},
{
"epoch": 0.28666510976179355,
"grad_norm": 0.7294023217733528,
"learning_rate": 4.345744819504266e-05,
"loss": 0.7112,
"step": 2455
},
{
"epoch": 0.28724894908921067,
"grad_norm": 0.6890872175614355,
"learning_rate": 4.342678993725517e-05,
"loss": 0.7039,
"step": 2460
},
{
"epoch": 0.28783278841662774,
"grad_norm": 0.6649888788261715,
"learning_rate": 4.339607229774989e-05,
"loss": 0.7108,
"step": 2465
},
{
"epoch": 0.28841662774404486,
"grad_norm": 0.6522467369641055,
"learning_rate": 4.3365295391054996e-05,
"loss": 0.7044,
"step": 2470
},
{
"epoch": 0.28900046707146193,
"grad_norm": 0.7181727438474074,
"learning_rate": 4.333445933191964e-05,
"loss": 0.7175,
"step": 2475
},
{
"epoch": 0.28958430639887905,
"grad_norm": 0.6795701001412194,
"learning_rate": 4.330356423531352e-05,
"loss": 0.7112,
"step": 2480
},
{
"epoch": 0.2901681457262961,
"grad_norm": 0.7039262534649463,
"learning_rate": 4.327261021642644e-05,
"loss": 0.7192,
"step": 2485
},
{
"epoch": 0.2907519850537132,
"grad_norm": 0.7104481909459923,
"learning_rate": 4.32415973906679e-05,
"loss": 0.7072,
"step": 2490
},
{
"epoch": 0.2913358243811303,
"grad_norm": 0.6730367434130707,
"learning_rate": 4.3210525873666656e-05,
"loss": 0.7062,
"step": 2495
},
{
"epoch": 0.2919196637085474,
"grad_norm": 0.8316755995846691,
"learning_rate": 4.317939578127029e-05,
"loss": 0.7101,
"step": 2500
},
{
"epoch": 0.2925035030359645,
"grad_norm": 0.7776101899040683,
"learning_rate": 4.314820722954476e-05,
"loss": 0.7102,
"step": 2505
},
{
"epoch": 0.2930873423633816,
"grad_norm": 0.7759230288668068,
"learning_rate": 4.3116960334774e-05,
"loss": 0.7088,
"step": 2510
},
{
"epoch": 0.2936711816907987,
"grad_norm": 0.7300987708649985,
"learning_rate": 4.308565521345949e-05,
"loss": 0.7174,
"step": 2515
},
{
"epoch": 0.29425502101821577,
"grad_norm": 0.691563983565198,
"learning_rate": 4.305429198231977e-05,
"loss": 0.7094,
"step": 2520
},
{
"epoch": 0.2948388603456329,
"grad_norm": 0.6857850906341554,
"learning_rate": 4.302287075829005e-05,
"loss": 0.7043,
"step": 2525
},
{
"epoch": 0.29542269967304996,
"grad_norm": 0.6168868887138788,
"learning_rate": 4.2991391658521765e-05,
"loss": 0.6994,
"step": 2530
},
{
"epoch": 0.2960065390004671,
"grad_norm": 0.7084081647037271,
"learning_rate": 4.2959854800382136e-05,
"loss": 0.7047,
"step": 2535
},
{
"epoch": 0.29659037832788415,
"grad_norm": 0.6454455832599233,
"learning_rate": 4.292826030145372e-05,
"loss": 0.7107,
"step": 2540
},
{
"epoch": 0.2971742176553013,
"grad_norm": 0.7537875064830446,
"learning_rate": 4.289660827953399e-05,
"loss": 0.6881,
"step": 2545
},
{
"epoch": 0.29775805698271834,
"grad_norm": 0.665124089965744,
"learning_rate": 4.28648988526349e-05,
"loss": 0.7252,
"step": 2550
},
{
"epoch": 0.29834189631013547,
"grad_norm": 10.191918716713266,
"learning_rate": 4.2833132138982415e-05,
"loss": 0.717,
"step": 2555
},
{
"epoch": 0.29892573563755254,
"grad_norm": 0.659368592659605,
"learning_rate": 4.280130825701609e-05,
"loss": 0.711,
"step": 2560
},
{
"epoch": 0.29950957496496966,
"grad_norm": 0.7185064005962647,
"learning_rate": 4.276942732538866e-05,
"loss": 0.7008,
"step": 2565
},
{
"epoch": 0.30009341429238673,
"grad_norm": 0.6847656968163185,
"learning_rate": 4.273748946296552e-05,
"loss": 0.7344,
"step": 2570
},
{
"epoch": 0.30067725361980385,
"grad_norm": 0.7003587004137025,
"learning_rate": 4.2705494788824345e-05,
"loss": 0.7121,
"step": 2575
},
{
"epoch": 0.3012610929472209,
"grad_norm": 0.7621456361852244,
"learning_rate": 4.267344342225463e-05,
"loss": 0.7231,
"step": 2580
},
{
"epoch": 0.30184493227463804,
"grad_norm": 0.6976812884927045,
"learning_rate": 4.264133548275725e-05,
"loss": 0.7009,
"step": 2585
},
{
"epoch": 0.3024287716020551,
"grad_norm": 0.6443472776826159,
"learning_rate": 4.2609171090044e-05,
"loss": 0.6905,
"step": 2590
},
{
"epoch": 0.30301261092947224,
"grad_norm": 0.6585015539397339,
"learning_rate": 4.257695036403714e-05,
"loss": 0.7033,
"step": 2595
},
{
"epoch": 0.3035964502568893,
"grad_norm": 0.6939183770223464,
"learning_rate": 4.2544673424868994e-05,
"loss": 0.7006,
"step": 2600
},
{
"epoch": 0.3041802895843064,
"grad_norm": 0.6470850459309977,
"learning_rate": 4.251234039288145e-05,
"loss": 0.7064,
"step": 2605
},
{
"epoch": 0.3047641289117235,
"grad_norm": 0.7018952484634018,
"learning_rate": 4.2479951388625546e-05,
"loss": 0.7287,
"step": 2610
},
{
"epoch": 0.30534796823914057,
"grad_norm": 0.6707270948528987,
"learning_rate": 4.2447506532861e-05,
"loss": 0.7247,
"step": 2615
},
{
"epoch": 0.3059318075665577,
"grad_norm": 0.6873978437611662,
"learning_rate": 4.241500594655577e-05,
"loss": 0.7049,
"step": 2620
},
{
"epoch": 0.30651564689397476,
"grad_norm": 0.6587849385828438,
"learning_rate": 4.2382449750885604e-05,
"loss": 0.6914,
"step": 2625
},
{
"epoch": 0.3070994862213919,
"grad_norm": 0.7235185916363532,
"learning_rate": 4.2349838067233575e-05,
"loss": 0.6889,
"step": 2630
},
{
"epoch": 0.30768332554880895,
"grad_norm": 0.7438779983004897,
"learning_rate": 4.231717101718967e-05,
"loss": 0.6854,
"step": 2635
},
{
"epoch": 0.3082671648762261,
"grad_norm": 0.6358770824948266,
"learning_rate": 4.228444872255025e-05,
"loss": 0.6982,
"step": 2640
},
{
"epoch": 0.30885100420364314,
"grad_norm": 0.7557882249486343,
"learning_rate": 4.2251671305317696e-05,
"loss": 0.7095,
"step": 2645
},
{
"epoch": 0.30943484353106027,
"grad_norm": 0.6559135507236117,
"learning_rate": 4.2218838887699894e-05,
"loss": 0.7143,
"step": 2650
},
{
"epoch": 0.31001868285847733,
"grad_norm": 0.8256383447956446,
"learning_rate": 4.2185951592109794e-05,
"loss": 0.6974,
"step": 2655
},
{
"epoch": 0.31060252218589446,
"grad_norm": 0.7970632629634542,
"learning_rate": 4.2153009541164965e-05,
"loss": 0.7143,
"step": 2660
},
{
"epoch": 0.3111863615133115,
"grad_norm": 0.7317917328687641,
"learning_rate": 4.21200128576871e-05,
"loss": 0.7248,
"step": 2665
},
{
"epoch": 0.31177020084072865,
"grad_norm": 0.6721444475469468,
"learning_rate": 4.208696166470161e-05,
"loss": 0.718,
"step": 2670
},
{
"epoch": 0.3123540401681457,
"grad_norm": 0.7858254702932888,
"learning_rate": 4.2053856085437124e-05,
"loss": 0.7106,
"step": 2675
},
{
"epoch": 0.31293787949556284,
"grad_norm": 0.6846940486554121,
"learning_rate": 4.202069624332507e-05,
"loss": 0.7071,
"step": 2680
},
{
"epoch": 0.3135217188229799,
"grad_norm": 0.7014521181871499,
"learning_rate": 4.1987482261999164e-05,
"loss": 0.7235,
"step": 2685
},
{
"epoch": 0.31410555815039704,
"grad_norm": 0.6825589502569331,
"learning_rate": 4.1954214265294985e-05,
"loss": 0.6855,
"step": 2690
},
{
"epoch": 0.3146893974778141,
"grad_norm": 0.5691899107187525,
"learning_rate": 4.192089237724951e-05,
"loss": 0.6909,
"step": 2695
},
{
"epoch": 0.3152732368052312,
"grad_norm": 0.6814010212005029,
"learning_rate": 4.188751672210063e-05,
"loss": 0.71,
"step": 2700
},
{
"epoch": 0.3158570761326483,
"grad_norm": 0.689336907740428,
"learning_rate": 4.1854087424286725e-05,
"loss": 0.6859,
"step": 2705
},
{
"epoch": 0.31644091546006536,
"grad_norm": 0.7473246366953188,
"learning_rate": 4.182060460844615e-05,
"loss": 0.6962,
"step": 2710
},
{
"epoch": 0.3170247547874825,
"grad_norm": 1.0456906948881868,
"learning_rate": 4.1787068399416825e-05,
"loss": 0.6992,
"step": 2715
},
{
"epoch": 0.31760859411489956,
"grad_norm": 0.6930676304656502,
"learning_rate": 4.175347892223572e-05,
"loss": 0.6783,
"step": 2720
},
{
"epoch": 0.3181924334423167,
"grad_norm": 0.760108322896286,
"learning_rate": 4.1719836302138426e-05,
"loss": 0.6989,
"step": 2725
},
{
"epoch": 0.31877627276973375,
"grad_norm": 0.7328309953315881,
"learning_rate": 4.168614066455867e-05,
"loss": 0.7007,
"step": 2730
},
{
"epoch": 0.31936011209715087,
"grad_norm": 0.7344695169247055,
"learning_rate": 4.165239213512784e-05,
"loss": 0.6931,
"step": 2735
},
{
"epoch": 0.31994395142456794,
"grad_norm": 0.696926721127825,
"learning_rate": 4.161859083967454e-05,
"loss": 0.7007,
"step": 2740
},
{
"epoch": 0.32052779075198506,
"grad_norm": 0.6590558010589381,
"learning_rate": 4.158473690422409e-05,
"loss": 0.6918,
"step": 2745
},
{
"epoch": 0.32111163007940213,
"grad_norm": 0.6476753037089318,
"learning_rate": 4.15508304549981e-05,
"loss": 0.6747,
"step": 2750
},
{
"epoch": 0.32169546940681926,
"grad_norm": 0.7031519295271327,
"learning_rate": 4.1516871618413947e-05,
"loss": 0.7064,
"step": 2755
},
{
"epoch": 0.3222793087342363,
"grad_norm": 0.6439952158257722,
"learning_rate": 4.148286052108436e-05,
"loss": 0.7097,
"step": 2760
},
{
"epoch": 0.32286314806165345,
"grad_norm": 0.6991435194699337,
"learning_rate": 4.144879728981688e-05,
"loss": 0.7048,
"step": 2765
},
{
"epoch": 0.3234469873890705,
"grad_norm": 0.6677321854539165,
"learning_rate": 4.141468205161345e-05,
"loss": 0.6984,
"step": 2770
},
{
"epoch": 0.32403082671648764,
"grad_norm": 0.6595171804385687,
"learning_rate": 4.1380514933669916e-05,
"loss": 0.6784,
"step": 2775
},
{
"epoch": 0.3246146660439047,
"grad_norm": 1.015021139511113,
"learning_rate": 4.134629606337555e-05,
"loss": 0.6925,
"step": 2780
},
{
"epoch": 0.32519850537132183,
"grad_norm": 0.7214804008392196,
"learning_rate": 4.131202556831257e-05,
"loss": 0.7085,
"step": 2785
},
{
"epoch": 0.3257823446987389,
"grad_norm": 0.731348240194823,
"learning_rate": 4.1277703576255685e-05,
"loss": 0.6931,
"step": 2790
},
{
"epoch": 0.326366184026156,
"grad_norm": 0.8049304402215585,
"learning_rate": 4.12433302151716e-05,
"loss": 0.6947,
"step": 2795
},
{
"epoch": 0.3269500233535731,
"grad_norm": 0.6361610980352036,
"learning_rate": 4.1208905613218547e-05,
"loss": 0.7007,
"step": 2800
},
{
"epoch": 0.3275338626809902,
"grad_norm": 0.6580608817341838,
"learning_rate": 4.1174429898745795e-05,
"loss": 0.6851,
"step": 2805
},
{
"epoch": 0.3281177020084073,
"grad_norm": 0.657110312928303,
"learning_rate": 4.113990320029321e-05,
"loss": 0.6806,
"step": 2810
},
{
"epoch": 0.32870154133582435,
"grad_norm": 0.6569678973108026,
"learning_rate": 4.1105325646590714e-05,
"loss": 0.7137,
"step": 2815
},
{
"epoch": 0.3292853806632415,
"grad_norm": 0.6555287805338297,
"learning_rate": 4.1070697366557856e-05,
"loss": 0.6947,
"step": 2820
},
{
"epoch": 0.32986921999065855,
"grad_norm": 0.7188920865527592,
"learning_rate": 4.103601848930332e-05,
"loss": 0.6875,
"step": 2825
},
{
"epoch": 0.33045305931807567,
"grad_norm": 0.7899413062439027,
"learning_rate": 4.100128914412442e-05,
"loss": 0.7319,
"step": 2830
},
{
"epoch": 0.33103689864549274,
"grad_norm": 0.733267680492785,
"learning_rate": 4.096650946050664e-05,
"loss": 0.688,
"step": 2835
},
{
"epoch": 0.33162073797290986,
"grad_norm": 0.7098708668395597,
"learning_rate": 4.0931679568123174e-05,
"loss": 0.7132,
"step": 2840
},
{
"epoch": 0.33220457730032693,
"grad_norm": 0.6758172761287962,
"learning_rate": 4.0896799596834365e-05,
"loss": 0.6856,
"step": 2845
},
{
"epoch": 0.33278841662774405,
"grad_norm": 0.7459994296367798,
"learning_rate": 4.086186967668731e-05,
"loss": 0.7159,
"step": 2850
},
{
"epoch": 0.3333722559551611,
"grad_norm": 0.7288354533071518,
"learning_rate": 4.082688993791533e-05,
"loss": 0.6734,
"step": 2855
},
{
"epoch": 0.33395609528257825,
"grad_norm": 0.9075510914012299,
"learning_rate": 4.079186051093747e-05,
"loss": 0.6948,
"step": 2860
},
{
"epoch": 0.3345399346099953,
"grad_norm": 0.7536999910281159,
"learning_rate": 4.075678152635807e-05,
"loss": 0.7141,
"step": 2865
},
{
"epoch": 0.33512377393741244,
"grad_norm": 0.6776220892035995,
"learning_rate": 4.07216531149662e-05,
"loss": 0.7186,
"step": 2870
},
{
"epoch": 0.3357076132648295,
"grad_norm": 0.6698565336124284,
"learning_rate": 4.068647540773524e-05,
"loss": 0.7061,
"step": 2875
},
{
"epoch": 0.33629145259224663,
"grad_norm": 0.7102555949426216,
"learning_rate": 4.065124853582237e-05,
"loss": 0.7114,
"step": 2880
},
{
"epoch": 0.3368752919196637,
"grad_norm": 0.6493150216278054,
"learning_rate": 4.0615972630568055e-05,
"loss": 0.703,
"step": 2885
},
{
"epoch": 0.3374591312470808,
"grad_norm": 0.6775510172923045,
"learning_rate": 4.0580647823495587e-05,
"loss": 0.7045,
"step": 2890
},
{
"epoch": 0.3380429705744979,
"grad_norm": 0.6692491951480167,
"learning_rate": 4.054527424631059e-05,
"loss": 0.7315,
"step": 2895
},
{
"epoch": 0.338626809901915,
"grad_norm": 0.6738566077502742,
"learning_rate": 4.0509852030900506e-05,
"loss": 0.6768,
"step": 2900
},
{
"epoch": 0.3392106492293321,
"grad_norm": 0.6409417570471981,
"learning_rate": 4.047438130933415e-05,
"loss": 0.703,
"step": 2905
},
{
"epoch": 0.3397944885567492,
"grad_norm": 0.7101594035977382,
"learning_rate": 4.043886221386117e-05,
"loss": 0.6896,
"step": 2910
},
{
"epoch": 0.3403783278841663,
"grad_norm": 0.6346742841875289,
"learning_rate": 4.040329487691155e-05,
"loss": 0.6973,
"step": 2915
},
{
"epoch": 0.34096216721158334,
"grad_norm": 0.8103244660091896,
"learning_rate": 4.036767943109519e-05,
"loss": 0.7054,
"step": 2920
},
{
"epoch": 0.34154600653900047,
"grad_norm": 0.688994173113095,
"learning_rate": 4.0332016009201315e-05,
"loss": 0.7048,
"step": 2925
},
{
"epoch": 0.34212984586641754,
"grad_norm": 0.6578485129978351,
"learning_rate": 4.0296304744198045e-05,
"loss": 0.6956,
"step": 2930
},
{
"epoch": 0.34271368519383466,
"grad_norm": 0.6377412125496434,
"learning_rate": 4.0260545769231875e-05,
"loss": 0.6908,
"step": 2935
},
{
"epoch": 0.34329752452125173,
"grad_norm": 0.670481994026621,
"learning_rate": 4.022473921762719e-05,
"loss": 0.7069,
"step": 2940
},
{
"epoch": 0.34388136384866885,
"grad_norm": 0.7121697851728561,
"learning_rate": 4.018888522288574e-05,
"loss": 0.7129,
"step": 2945
},
{
"epoch": 0.3444652031760859,
"grad_norm": 0.6887026519736897,
"learning_rate": 4.0152983918686175e-05,
"loss": 0.6939,
"step": 2950
},
{
"epoch": 0.34504904250350305,
"grad_norm": 0.7017636868082506,
"learning_rate": 4.0117035438883553e-05,
"loss": 0.6812,
"step": 2955
},
{
"epoch": 0.3456328818309201,
"grad_norm": 0.7655292468384295,
"learning_rate": 4.008103991750879e-05,
"loss": 0.715,
"step": 2960
},
{
"epoch": 0.34621672115833724,
"grad_norm": 0.6287342780769691,
"learning_rate": 4.004499748876821e-05,
"loss": 0.694,
"step": 2965
},
{
"epoch": 0.3468005604857543,
"grad_norm": 0.6396219403886257,
"learning_rate": 4.000890828704304e-05,
"loss": 0.681,
"step": 2970
},
{
"epoch": 0.34738439981317143,
"grad_norm": 0.6866315857071885,
"learning_rate": 3.997277244688886e-05,
"loss": 0.7081,
"step": 2975
},
{
"epoch": 0.3479682391405885,
"grad_norm": 0.7105639621395489,
"learning_rate": 3.993659010303517e-05,
"loss": 0.7073,
"step": 2980
},
{
"epoch": 0.3485520784680056,
"grad_norm": 0.6483059974789481,
"learning_rate": 3.9900361390384836e-05,
"loss": 0.6721,
"step": 2985
},
{
"epoch": 0.3491359177954227,
"grad_norm": 0.6579177689380913,
"learning_rate": 3.986408644401362e-05,
"loss": 0.7076,
"step": 2990
},
{
"epoch": 0.3497197571228398,
"grad_norm": 0.6842426698159001,
"learning_rate": 3.982776539916966e-05,
"loss": 0.6946,
"step": 2995
},
{
"epoch": 0.3503035964502569,
"grad_norm": 0.6214608870632571,
"learning_rate": 3.979139839127296e-05,
"loss": 0.679,
"step": 3000
},
{
"epoch": 0.350887435777674,
"grad_norm": 0.7832466335636389,
"learning_rate": 3.975498555591489e-05,
"loss": 0.694,
"step": 3005
},
{
"epoch": 0.3514712751050911,
"grad_norm": 0.6827242334386869,
"learning_rate": 3.971852702885772e-05,
"loss": 0.686,
"step": 3010
},
{
"epoch": 0.3520551144325082,
"grad_norm": 0.6580285082416581,
"learning_rate": 3.9682022946034006e-05,
"loss": 0.6915,
"step": 3015
},
{
"epoch": 0.35263895375992527,
"grad_norm": 0.6644485295248183,
"learning_rate": 3.964547344354624e-05,
"loss": 0.6955,
"step": 3020
},
{
"epoch": 0.3532227930873424,
"grad_norm": 0.6797616001519512,
"learning_rate": 3.9608878657666195e-05,
"loss": 0.7005,
"step": 3025
},
{
"epoch": 0.35380663241475946,
"grad_norm": 0.6882723067231326,
"learning_rate": 3.9572238724834503e-05,
"loss": 0.7009,
"step": 3030
},
{
"epoch": 0.3543904717421765,
"grad_norm": 0.7025514914124491,
"learning_rate": 3.953555378166012e-05,
"loss": 0.7015,
"step": 3035
},
{
"epoch": 0.35497431106959365,
"grad_norm": 0.7682225416014826,
"learning_rate": 3.9498823964919827e-05,
"loss": 0.7039,
"step": 3040
},
{
"epoch": 0.3555581503970107,
"grad_norm": 0.7107066995180754,
"learning_rate": 3.94620494115577e-05,
"loss": 0.688,
"step": 3045
},
{
"epoch": 0.35614198972442784,
"grad_norm": 0.603746881271006,
"learning_rate": 3.942523025868461e-05,
"loss": 0.6915,
"step": 3050
},
{
"epoch": 0.3567258290518449,
"grad_norm": 0.6977580760284987,
"learning_rate": 3.9388366643577745e-05,
"loss": 0.689,
"step": 3055
},
{
"epoch": 0.35730966837926204,
"grad_norm": 0.6254718988195369,
"learning_rate": 3.9351458703680017e-05,
"loss": 0.7093,
"step": 3060
},
{
"epoch": 0.3578935077066791,
"grad_norm": 0.6794200041896331,
"learning_rate": 3.931450657659963e-05,
"loss": 0.6954,
"step": 3065
},
{
"epoch": 0.3584773470340962,
"grad_norm": 0.7000832034897168,
"learning_rate": 3.927751040010954e-05,
"loss": 0.6751,
"step": 3070
},
{
"epoch": 0.3590611863615133,
"grad_norm": 0.7011267366601257,
"learning_rate": 3.924047031214691e-05,
"loss": 0.708,
"step": 3075
},
{
"epoch": 0.3596450256889304,
"grad_norm": 0.7552720574103187,
"learning_rate": 3.920338645081266e-05,
"loss": 0.6939,
"step": 3080
},
{
"epoch": 0.3602288650163475,
"grad_norm": 0.709414134987964,
"learning_rate": 3.916625895437089e-05,
"loss": 0.6684,
"step": 3085
},
{
"epoch": 0.3608127043437646,
"grad_norm": 0.7038732195784704,
"learning_rate": 3.912908796124839e-05,
"loss": 0.6953,
"step": 3090
},
{
"epoch": 0.3613965436711817,
"grad_norm": 0.7064185255920825,
"learning_rate": 3.909187361003414e-05,
"loss": 0.6847,
"step": 3095
},
{
"epoch": 0.3619803829985988,
"grad_norm": 0.7314898303717124,
"learning_rate": 3.905461603947878e-05,
"loss": 0.6987,
"step": 3100
},
{
"epoch": 0.3625642223260159,
"grad_norm": 0.6889637879433151,
"learning_rate": 3.9017315388494044e-05,
"loss": 0.6853,
"step": 3105
},
{
"epoch": 0.363148061653433,
"grad_norm": 0.8312571876184839,
"learning_rate": 3.8979971796152346e-05,
"loss": 0.6914,
"step": 3110
},
{
"epoch": 0.36373190098085006,
"grad_norm": 0.582698659718714,
"learning_rate": 3.894258540168618e-05,
"loss": 0.6728,
"step": 3115
},
{
"epoch": 0.3643157403082672,
"grad_norm": 0.6619313870899733,
"learning_rate": 3.89051563444876e-05,
"loss": 0.6785,
"step": 3120
},
{
"epoch": 0.36489957963568426,
"grad_norm": 0.6298363673198333,
"learning_rate": 3.886768476410777e-05,
"loss": 0.6833,
"step": 3125
},
{
"epoch": 0.3654834189631014,
"grad_norm": 0.8219887753285791,
"learning_rate": 3.883017080025638e-05,
"loss": 0.6837,
"step": 3130
},
{
"epoch": 0.36606725829051845,
"grad_norm": 0.7880218471799671,
"learning_rate": 3.879261459280111e-05,
"loss": 0.6868,
"step": 3135
},
{
"epoch": 0.3666510976179355,
"grad_norm": 0.7345112746587662,
"learning_rate": 3.875501628176719e-05,
"loss": 0.6934,
"step": 3140
},
{
"epoch": 0.36723493694535264,
"grad_norm": 0.686416317008603,
"learning_rate": 3.8717376007336814e-05,
"loss": 0.6877,
"step": 3145
},
{
"epoch": 0.3678187762727697,
"grad_norm": 0.7334866029192383,
"learning_rate": 3.867969390984862e-05,
"loss": 0.7131,
"step": 3150
},
{
"epoch": 0.36840261560018683,
"grad_norm": 0.6732681172574688,
"learning_rate": 3.864197012979719e-05,
"loss": 0.7155,
"step": 3155
},
{
"epoch": 0.3689864549276039,
"grad_norm": 0.6887026642062427,
"learning_rate": 3.8604204807832516e-05,
"loss": 0.7267,
"step": 3160
},
{
"epoch": 0.369570294255021,
"grad_norm": 0.6714593414758345,
"learning_rate": 3.856639808475947e-05,
"loss": 0.6795,
"step": 3165
},
{
"epoch": 0.3701541335824381,
"grad_norm": 0.6649220433540397,
"learning_rate": 3.85285501015373e-05,
"loss": 0.6889,
"step": 3170
},
{
"epoch": 0.3707379729098552,
"grad_norm": 0.6235959590521644,
"learning_rate": 3.8490660999279074e-05,
"loss": 0.6862,
"step": 3175
},
{
"epoch": 0.3713218122372723,
"grad_norm": 0.626992375665342,
"learning_rate": 3.8452730919251174e-05,
"loss": 0.6964,
"step": 3180
},
{
"epoch": 0.3719056515646894,
"grad_norm": 0.6155958131892619,
"learning_rate": 3.841476000287275e-05,
"loss": 0.7053,
"step": 3185
},
{
"epoch": 0.3724894908921065,
"grad_norm": 0.6851410102440836,
"learning_rate": 3.837674839171524e-05,
"loss": 0.6864,
"step": 3190
},
{
"epoch": 0.3730733302195236,
"grad_norm": 0.6357427022808335,
"learning_rate": 3.833869622750177e-05,
"loss": 0.7065,
"step": 3195
},
{
"epoch": 0.37365716954694067,
"grad_norm": 0.8472517069635026,
"learning_rate": 3.8300603652106704e-05,
"loss": 0.695,
"step": 3200
},
{
"epoch": 0.3742410088743578,
"grad_norm": 0.756250084595756,
"learning_rate": 3.8262470807555045e-05,
"loss": 0.6802,
"step": 3205
},
{
"epoch": 0.37482484820177486,
"grad_norm": 0.6839568898754494,
"learning_rate": 3.822429783602195e-05,
"loss": 0.6991,
"step": 3210
},
{
"epoch": 0.375408687529192,
"grad_norm": 0.6868351371148217,
"learning_rate": 3.818608487983218e-05,
"loss": 0.6769,
"step": 3215
},
{
"epoch": 0.37599252685660906,
"grad_norm": 0.6203850879359878,
"learning_rate": 3.8147832081459574e-05,
"loss": 0.6861,
"step": 3220
},
{
"epoch": 0.3765763661840262,
"grad_norm": 0.7422862863096844,
"learning_rate": 3.810953958352653e-05,
"loss": 0.7102,
"step": 3225
},
{
"epoch": 0.37716020551144325,
"grad_norm": 0.6192695326949685,
"learning_rate": 3.807120752880346e-05,
"loss": 0.6889,
"step": 3230
},
{
"epoch": 0.37774404483886037,
"grad_norm": 0.6414835917217437,
"learning_rate": 3.8032836060208265e-05,
"loss": 0.6769,
"step": 3235
},
{
"epoch": 0.37832788416627744,
"grad_norm": 0.6636929047568134,
"learning_rate": 3.799442532080577e-05,
"loss": 0.6792,
"step": 3240
},
{
"epoch": 0.3789117234936945,
"grad_norm": 0.7334579643106445,
"learning_rate": 3.795597545380724e-05,
"loss": 0.6921,
"step": 3245
},
{
"epoch": 0.37949556282111163,
"grad_norm": 0.7242201259541307,
"learning_rate": 3.791748660256983e-05,
"loss": 0.6894,
"step": 3250
},
{
"epoch": 0.3800794021485287,
"grad_norm": 0.6800906858596535,
"learning_rate": 3.787895891059603e-05,
"loss": 0.6732,
"step": 3255
},
{
"epoch": 0.3806632414759458,
"grad_norm": 0.6694030660200606,
"learning_rate": 3.784039252153315e-05,
"loss": 0.6833,
"step": 3260
},
{
"epoch": 0.3812470808033629,
"grad_norm": 0.7837093304371092,
"learning_rate": 3.780178757917278e-05,
"loss": 0.7044,
"step": 3265
},
{
"epoch": 0.38183092013078,
"grad_norm": 0.6212998846441561,
"learning_rate": 3.776314422745026e-05,
"loss": 0.6743,
"step": 3270
},
{
"epoch": 0.3824147594581971,
"grad_norm": 0.7428058913367551,
"learning_rate": 3.772446261044411e-05,
"loss": 0.6851,
"step": 3275
},
{
"epoch": 0.3829985987856142,
"grad_norm": 0.6845689857219531,
"learning_rate": 3.768574287237555e-05,
"loss": 0.6824,
"step": 3280
},
{
"epoch": 0.3835824381130313,
"grad_norm": 0.7390118487736407,
"learning_rate": 3.7646985157607915e-05,
"loss": 0.6899,
"step": 3285
},
{
"epoch": 0.3841662774404484,
"grad_norm": 0.7772545878003212,
"learning_rate": 3.760818961064614e-05,
"loss": 0.6831,
"step": 3290
},
{
"epoch": 0.38475011676786547,
"grad_norm": 0.7965196041346534,
"learning_rate": 3.75693563761362e-05,
"loss": 0.6892,
"step": 3295
},
{
"epoch": 0.3853339560952826,
"grad_norm": 0.6577827435561913,
"learning_rate": 3.75304855988646e-05,
"loss": 0.7007,
"step": 3300
},
{
"epoch": 0.38591779542269966,
"grad_norm": 0.7685136213719611,
"learning_rate": 3.749157742375782e-05,
"loss": 0.697,
"step": 3305
},
{
"epoch": 0.3865016347501168,
"grad_norm": 0.7295181909577287,
"learning_rate": 3.745263199588176e-05,
"loss": 0.6931,
"step": 3310
},
{
"epoch": 0.38708547407753385,
"grad_norm": 0.6878827489363338,
"learning_rate": 3.741364946044123e-05,
"loss": 0.6886,
"step": 3315
},
{
"epoch": 0.387669313404951,
"grad_norm": 0.6687780142846725,
"learning_rate": 3.737462996277939e-05,
"loss": 0.686,
"step": 3320
},
{
"epoch": 0.38825315273236805,
"grad_norm": 0.6822226346049627,
"learning_rate": 3.73355736483772e-05,
"loss": 0.7238,
"step": 3325
},
{
"epoch": 0.38883699205978517,
"grad_norm": 0.7793024322443424,
"learning_rate": 3.72964806628529e-05,
"loss": 0.6793,
"step": 3330
},
{
"epoch": 0.38942083138720224,
"grad_norm": 0.7332530440310608,
"learning_rate": 3.725735115196145e-05,
"loss": 0.7077,
"step": 3335
},
{
"epoch": 0.39000467071461936,
"grad_norm": 0.7291717740569443,
"learning_rate": 3.7218185261593984e-05,
"loss": 0.6939,
"step": 3340
},
{
"epoch": 0.39058851004203643,
"grad_norm": 0.678089956357006,
"learning_rate": 3.717898313777729e-05,
"loss": 0.6971,
"step": 3345
},
{
"epoch": 0.39117234936945355,
"grad_norm": 0.7147921035532178,
"learning_rate": 3.713974492667324e-05,
"loss": 0.6957,
"step": 3350
},
{
"epoch": 0.3917561886968706,
"grad_norm": 0.6522228386343375,
"learning_rate": 3.710047077457826e-05,
"loss": 0.6891,
"step": 3355
},
{
"epoch": 0.3923400280242877,
"grad_norm": 0.6710362987910684,
"learning_rate": 3.706116082792276e-05,
"loss": 0.6815,
"step": 3360
},
{
"epoch": 0.3929238673517048,
"grad_norm": 0.6291887390159168,
"learning_rate": 3.702181523327064e-05,
"loss": 0.7092,
"step": 3365
},
{
"epoch": 0.3935077066791219,
"grad_norm": 0.6995020859725057,
"learning_rate": 3.698243413731867e-05,
"loss": 0.6861,
"step": 3370
},
{
"epoch": 0.394091546006539,
"grad_norm": 0.7633653352248395,
"learning_rate": 3.694301768689603e-05,
"loss": 0.6917,
"step": 3375
},
{
"epoch": 0.3946753853339561,
"grad_norm": 0.7314601602571308,
"learning_rate": 3.690356602896368e-05,
"loss": 0.6982,
"step": 3380
},
{
"epoch": 0.3952592246613732,
"grad_norm": 0.8403954593326415,
"learning_rate": 3.686407931061386e-05,
"loss": 0.6769,
"step": 3385
},
{
"epoch": 0.39584306398879027,
"grad_norm": 0.747466874618807,
"learning_rate": 3.682455767906954e-05,
"loss": 0.6928,
"step": 3390
},
{
"epoch": 0.3964269033162074,
"grad_norm": 0.6767936645773085,
"learning_rate": 3.678500128168384e-05,
"loss": 0.6744,
"step": 3395
},
{
"epoch": 0.39701074264362446,
"grad_norm": 0.7332913208913894,
"learning_rate": 3.674541026593952e-05,
"loss": 0.6872,
"step": 3400
},
{
"epoch": 0.3975945819710416,
"grad_norm": 0.6812167590557383,
"learning_rate": 3.6705784779448405e-05,
"loss": 0.6771,
"step": 3405
},
{
"epoch": 0.39817842129845865,
"grad_norm": 0.6245513451979265,
"learning_rate": 3.6666124969950835e-05,
"loss": 0.6834,
"step": 3410
},
{
"epoch": 0.3987622606258758,
"grad_norm": 0.6561995817045527,
"learning_rate": 3.662643098531513e-05,
"loss": 0.6942,
"step": 3415
},
{
"epoch": 0.39934609995329284,
"grad_norm": 0.7088078943293247,
"learning_rate": 3.6586702973537025e-05,
"loss": 0.6921,
"step": 3420
},
{
"epoch": 0.39992993928070997,
"grad_norm": 0.6426536532575015,
"learning_rate": 3.654694108273912e-05,
"loss": 0.6696,
"step": 3425
},
{
"epoch": 0.40051377860812704,
"grad_norm": 0.6457654101479425,
"learning_rate": 3.6507145461170345e-05,
"loss": 0.6657,
"step": 3430
},
{
"epoch": 0.40109761793554416,
"grad_norm": 0.7405170378891968,
"learning_rate": 3.646731625720537e-05,
"loss": 0.6714,
"step": 3435
},
{
"epoch": 0.40168145726296123,
"grad_norm": 0.722411587375372,
"learning_rate": 3.642745361934408e-05,
"loss": 0.6947,
"step": 3440
},
{
"epoch": 0.40226529659037835,
"grad_norm": 0.7570639839348096,
"learning_rate": 3.638755769621104e-05,
"loss": 0.6937,
"step": 3445
},
{
"epoch": 0.4028491359177954,
"grad_norm": 0.6831672177275794,
"learning_rate": 3.634762863655487e-05,
"loss": 0.682,
"step": 3450
},
{
"epoch": 0.40343297524521254,
"grad_norm": 0.7543450989841435,
"learning_rate": 3.630766658924779e-05,
"loss": 0.6977,
"step": 3455
},
{
"epoch": 0.4040168145726296,
"grad_norm": 0.6552784947688012,
"learning_rate": 3.6267671703284963e-05,
"loss": 0.6912,
"step": 3460
},
{
"epoch": 0.4046006539000467,
"grad_norm": 0.8909899789769518,
"learning_rate": 3.6227644127784026e-05,
"loss": 0.6987,
"step": 3465
},
{
"epoch": 0.4051844932274638,
"grad_norm": 0.7456834932185622,
"learning_rate": 3.618758401198447e-05,
"loss": 0.6829,
"step": 3470
},
{
"epoch": 0.4057683325548809,
"grad_norm": 0.8023255073012434,
"learning_rate": 3.6147491505247124e-05,
"loss": 0.6904,
"step": 3475
},
{
"epoch": 0.406352171882298,
"grad_norm": 0.6740423733656838,
"learning_rate": 3.6107366757053586e-05,
"loss": 0.6916,
"step": 3480
},
{
"epoch": 0.40693601120971507,
"grad_norm": 0.7245632825508872,
"learning_rate": 3.606720991700565e-05,
"loss": 0.6972,
"step": 3485
},
{
"epoch": 0.4075198505371322,
"grad_norm": 0.6922520102615611,
"learning_rate": 3.602702113482477e-05,
"loss": 0.692,
"step": 3490
},
{
"epoch": 0.40810368986454926,
"grad_norm": 0.7075741705174543,
"learning_rate": 3.59868005603515e-05,
"loss": 0.6982,
"step": 3495
},
{
"epoch": 0.4086875291919664,
"grad_norm": 0.6844993230920785,
"learning_rate": 3.5946548343544925e-05,
"loss": 0.6761,
"step": 3500
},
{
"epoch": 0.40927136851938345,
"grad_norm": 0.663357862984397,
"learning_rate": 3.5906264634482084e-05,
"loss": 0.6762,
"step": 3505
},
{
"epoch": 0.4098552078468006,
"grad_norm": 0.6834924687016519,
"learning_rate": 3.586594958335747e-05,
"loss": 0.68,
"step": 3510
},
{
"epoch": 0.41043904717421764,
"grad_norm": 0.6593085285066986,
"learning_rate": 3.582560334048241e-05,
"loss": 0.6754,
"step": 3515
},
{
"epoch": 0.41102288650163477,
"grad_norm": 0.6908960562089073,
"learning_rate": 3.578522605628453e-05,
"loss": 0.6654,
"step": 3520
},
{
"epoch": 0.41160672582905183,
"grad_norm": 0.6399414636774441,
"learning_rate": 3.5744817881307184e-05,
"loss": 0.6801,
"step": 3525
},
{
"epoch": 0.41219056515646896,
"grad_norm": 0.6935459398136206,
"learning_rate": 3.570437896620891e-05,
"loss": 0.6823,
"step": 3530
},
{
"epoch": 0.412774404483886,
"grad_norm": 0.6936385265984067,
"learning_rate": 3.566390946176286e-05,
"loss": 0.6966,
"step": 3535
},
{
"epoch": 0.41335824381130315,
"grad_norm": 0.733685837451808,
"learning_rate": 3.5623409518856225e-05,
"loss": 0.7024,
"step": 3540
},
{
"epoch": 0.4139420831387202,
"grad_norm": 0.6476767995892281,
"learning_rate": 3.55828792884897e-05,
"loss": 0.6945,
"step": 3545
},
{
"epoch": 0.41452592246613734,
"grad_norm": 0.7048321026918248,
"learning_rate": 3.5542318921776886e-05,
"loss": 0.6824,
"step": 3550
},
{
"epoch": 0.4151097617935544,
"grad_norm": 0.626231298342904,
"learning_rate": 3.5501728569943746e-05,
"loss": 0.6798,
"step": 3555
},
{
"epoch": 0.41569360112097153,
"grad_norm": 0.7000118781897608,
"learning_rate": 3.546110838432806e-05,
"loss": 0.6767,
"step": 3560
},
{
"epoch": 0.4162774404483886,
"grad_norm": 0.7861580029223385,
"learning_rate": 3.542045851637883e-05,
"loss": 0.6767,
"step": 3565
},
{
"epoch": 0.41686127977580567,
"grad_norm": 0.8074448945180608,
"learning_rate": 3.53797791176557e-05,
"loss": 0.7018,
"step": 3570
},
{
"epoch": 0.4174451191032228,
"grad_norm": 0.7146281056667712,
"learning_rate": 3.5339070339828466e-05,
"loss": 0.6865,
"step": 3575
},
{
"epoch": 0.41802895843063986,
"grad_norm": 0.7441724024271111,
"learning_rate": 3.529833233467642e-05,
"loss": 0.6617,
"step": 3580
},
{
"epoch": 0.418612797758057,
"grad_norm": 0.6449017553102977,
"learning_rate": 3.525756525408785e-05,
"loss": 0.6938,
"step": 3585
},
{
"epoch": 0.41919663708547406,
"grad_norm": 0.7257859088018427,
"learning_rate": 3.521676925005945e-05,
"loss": 0.6899,
"step": 3590
},
{
"epoch": 0.4197804764128912,
"grad_norm": 0.7124825002659122,
"learning_rate": 3.517594447469572e-05,
"loss": 0.6744,
"step": 3595
},
{
"epoch": 0.42036431574030825,
"grad_norm": 0.634046926253278,
"learning_rate": 3.513509108020846e-05,
"loss": 0.6927,
"step": 3600
},
{
"epoch": 0.42094815506772537,
"grad_norm": 0.6432772267862358,
"learning_rate": 3.5094209218916185e-05,
"loss": 0.6775,
"step": 3605
},
{
"epoch": 0.42153199439514244,
"grad_norm": 0.6879752638775296,
"learning_rate": 3.505329904324351e-05,
"loss": 0.6728,
"step": 3610
},
{
"epoch": 0.42211583372255956,
"grad_norm": 0.7595999082326345,
"learning_rate": 3.501236070572066e-05,
"loss": 0.6768,
"step": 3615
},
{
"epoch": 0.42269967304997663,
"grad_norm": 0.6505503285843467,
"learning_rate": 3.497139435898283e-05,
"loss": 0.6849,
"step": 3620
},
{
"epoch": 0.42328351237739376,
"grad_norm": 0.6499660319606886,
"learning_rate": 3.4930400155769644e-05,
"loss": 0.6754,
"step": 3625
},
{
"epoch": 0.4238673517048108,
"grad_norm": 0.6251892497211252,
"learning_rate": 3.488937824892461e-05,
"loss": 0.697,
"step": 3630
},
{
"epoch": 0.42445119103222795,
"grad_norm": 0.6581284790376016,
"learning_rate": 3.48483287913945e-05,
"loss": 0.6965,
"step": 3635
},
{
"epoch": 0.425035030359645,
"grad_norm": 0.6744230932975706,
"learning_rate": 3.480725193622881e-05,
"loss": 0.6751,
"step": 3640
},
{
"epoch": 0.42561886968706214,
"grad_norm": 0.6793166788586336,
"learning_rate": 3.476614783657922e-05,
"loss": 0.7013,
"step": 3645
},
{
"epoch": 0.4262027090144792,
"grad_norm": 0.6669494613932773,
"learning_rate": 3.472501664569894e-05,
"loss": 0.6886,
"step": 3650
},
{
"epoch": 0.42678654834189633,
"grad_norm": 0.6252882985306817,
"learning_rate": 3.468385851694222e-05,
"loss": 0.6863,
"step": 3655
},
{
"epoch": 0.4273703876693134,
"grad_norm": 0.7019249203412884,
"learning_rate": 3.464267360376373e-05,
"loss": 0.6798,
"step": 3660
},
{
"epoch": 0.4279542269967305,
"grad_norm": 0.6806791026761212,
"learning_rate": 3.460146205971802e-05,
"loss": 0.6948,
"step": 3665
},
{
"epoch": 0.4285380663241476,
"grad_norm": 0.6241325042876864,
"learning_rate": 3.456022403845891e-05,
"loss": 0.6864,
"step": 3670
},
{
"epoch": 0.4291219056515647,
"grad_norm": 0.7118272522580762,
"learning_rate": 3.4518959693738944e-05,
"loss": 0.6782,
"step": 3675
},
{
"epoch": 0.4297057449789818,
"grad_norm": 0.6972804885973292,
"learning_rate": 3.4477669179408834e-05,
"loss": 0.6826,
"step": 3680
},
{
"epoch": 0.43028958430639885,
"grad_norm": 0.734923021450161,
"learning_rate": 3.443635264941682e-05,
"loss": 0.6834,
"step": 3685
},
{
"epoch": 0.430873423633816,
"grad_norm": 0.6523407996281446,
"learning_rate": 3.4395010257808185e-05,
"loss": 0.6871,
"step": 3690
},
{
"epoch": 0.43145726296123305,
"grad_norm": 0.6589822288230873,
"learning_rate": 3.43536421587246e-05,
"loss": 0.6964,
"step": 3695
},
{
"epoch": 0.43204110228865017,
"grad_norm": 0.6714736680537675,
"learning_rate": 3.431224850640361e-05,
"loss": 0.6807,
"step": 3700
},
{
"epoch": 0.43262494161606724,
"grad_norm": 0.6798905897218975,
"learning_rate": 3.427082945517801e-05,
"loss": 0.6751,
"step": 3705
},
{
"epoch": 0.43320878094348436,
"grad_norm": 0.6499474511854968,
"learning_rate": 3.422938515947531e-05,
"loss": 0.67,
"step": 3710
},
{
"epoch": 0.43379262027090143,
"grad_norm": 0.621575208209989,
"learning_rate": 3.418791577381713e-05,
"loss": 0.6763,
"step": 3715
},
{
"epoch": 0.43437645959831855,
"grad_norm": 0.7243770387236024,
"learning_rate": 3.4146421452818657e-05,
"loss": 0.6989,
"step": 3720
},
{
"epoch": 0.4349602989257356,
"grad_norm": 0.6610910256831685,
"learning_rate": 3.4104902351188e-05,
"loss": 0.679,
"step": 3725
},
{
"epoch": 0.43554413825315275,
"grad_norm": 0.6531424740316809,
"learning_rate": 3.406335862372573e-05,
"loss": 0.6683,
"step": 3730
},
{
"epoch": 0.4361279775805698,
"grad_norm": 0.7388404354779999,
"learning_rate": 3.402179042532417e-05,
"loss": 0.6895,
"step": 3735
},
{
"epoch": 0.43671181690798694,
"grad_norm": 0.6570601621253606,
"learning_rate": 3.3980197910966915e-05,
"loss": 0.6855,
"step": 3740
},
{
"epoch": 0.437295656235404,
"grad_norm": 0.6882979008253807,
"learning_rate": 3.3938581235728214e-05,
"loss": 0.6727,
"step": 3745
},
{
"epoch": 0.43787949556282113,
"grad_norm": 0.6360394378243861,
"learning_rate": 3.389694055477238e-05,
"loss": 0.6942,
"step": 3750
},
{
"epoch": 0.4384633348902382,
"grad_norm": 0.6630358815549592,
"learning_rate": 3.385527602335327e-05,
"loss": 0.6459,
"step": 3755
},
{
"epoch": 0.4390471742176553,
"grad_norm": 0.6293041537091805,
"learning_rate": 3.381358779681362e-05,
"loss": 0.6707,
"step": 3760
},
{
"epoch": 0.4396310135450724,
"grad_norm": 0.6518689521648657,
"learning_rate": 3.377187603058454e-05,
"loss": 0.6858,
"step": 3765
},
{
"epoch": 0.4402148528724895,
"grad_norm": 0.721640342120163,
"learning_rate": 3.373014088018489e-05,
"loss": 0.702,
"step": 3770
},
{
"epoch": 0.4407986921999066,
"grad_norm": 0.6236090703227648,
"learning_rate": 3.3688382501220727e-05,
"loss": 0.6691,
"step": 3775
},
{
"epoch": 0.4413825315273237,
"grad_norm": 0.7348340644041736,
"learning_rate": 3.364660104938472e-05,
"loss": 0.6894,
"step": 3780
},
{
"epoch": 0.4419663708547408,
"grad_norm": 0.7284618161561375,
"learning_rate": 3.3604796680455546e-05,
"loss": 0.7028,
"step": 3785
},
{
"epoch": 0.44255021018215784,
"grad_norm": 0.6364473797128645,
"learning_rate": 3.356296955029733e-05,
"loss": 0.6783,
"step": 3790
},
{
"epoch": 0.44313404950957497,
"grad_norm": 0.6779869540793965,
"learning_rate": 3.3521119814859063e-05,
"loss": 0.6838,
"step": 3795
},
{
"epoch": 0.44371788883699204,
"grad_norm": 0.6074968287383505,
"learning_rate": 3.347924763017403e-05,
"loss": 0.6713,
"step": 3800
},
{
"epoch": 0.44430172816440916,
"grad_norm": 0.6605903491808311,
"learning_rate": 3.3437353152359195e-05,
"loss": 0.6764,
"step": 3805
},
{
"epoch": 0.44488556749182623,
"grad_norm": 0.6650848208031424,
"learning_rate": 3.339543653761466e-05,
"loss": 0.6849,
"step": 3810
},
{
"epoch": 0.44546940681924335,
"grad_norm": 0.696409487160532,
"learning_rate": 3.335349794222304e-05,
"loss": 0.6732,
"step": 3815
},
{
"epoch": 0.4460532461466604,
"grad_norm": 0.6826813844096611,
"learning_rate": 3.331153752254893e-05,
"loss": 0.6954,
"step": 3820
},
{
"epoch": 0.44663708547407754,
"grad_norm": 0.7001544241959847,
"learning_rate": 3.326955543503827e-05,
"loss": 0.7036,
"step": 3825
},
{
"epoch": 0.4472209248014946,
"grad_norm": 0.708253678105146,
"learning_rate": 3.322755183621779e-05,
"loss": 0.6756,
"step": 3830
},
{
"epoch": 0.44780476412891174,
"grad_norm": 0.7002269675727447,
"learning_rate": 3.318552688269446e-05,
"loss": 0.6694,
"step": 3835
},
{
"epoch": 0.4483886034563288,
"grad_norm": 0.71437997612647,
"learning_rate": 3.314348073115481e-05,
"loss": 0.6839,
"step": 3840
},
{
"epoch": 0.44897244278374593,
"grad_norm": 0.6324613379256642,
"learning_rate": 3.310141353836446e-05,
"loss": 0.6881,
"step": 3845
},
{
"epoch": 0.449556282111163,
"grad_norm": 0.722441314751814,
"learning_rate": 3.305932546116743e-05,
"loss": 0.6787,
"step": 3850
},
{
"epoch": 0.4501401214385801,
"grad_norm": 0.6536803096515039,
"learning_rate": 3.301721665648566e-05,
"loss": 0.6967,
"step": 3855
},
{
"epoch": 0.4507239607659972,
"grad_norm": 0.6797978945571883,
"learning_rate": 3.297508728131832e-05,
"loss": 0.6738,
"step": 3860
},
{
"epoch": 0.4513078000934143,
"grad_norm": 0.7092983307165541,
"learning_rate": 3.29329374927413e-05,
"loss": 0.684,
"step": 3865
},
{
"epoch": 0.4518916394208314,
"grad_norm": 0.6438253823064166,
"learning_rate": 3.2890767447906615e-05,
"loss": 0.6702,
"step": 3870
},
{
"epoch": 0.4524754787482485,
"grad_norm": 0.6771145985577135,
"learning_rate": 3.284857730404176e-05,
"loss": 0.6741,
"step": 3875
},
{
"epoch": 0.4530593180756656,
"grad_norm": 0.7049524098486649,
"learning_rate": 3.2806367218449216e-05,
"loss": 0.68,
"step": 3880
},
{
"epoch": 0.4536431574030827,
"grad_norm": 0.7689565685544832,
"learning_rate": 3.2764137348505785e-05,
"loss": 0.6771,
"step": 3885
},
{
"epoch": 0.45422699673049977,
"grad_norm": 0.7184216342968225,
"learning_rate": 3.2721887851662044e-05,
"loss": 0.6691,
"step": 3890
},
{
"epoch": 0.45481083605791683,
"grad_norm": 1.0037489244304667,
"learning_rate": 3.267961888544173e-05,
"loss": 0.6869,
"step": 3895
},
{
"epoch": 0.45539467538533396,
"grad_norm": 0.6392092275759652,
"learning_rate": 3.263733060744121e-05,
"loss": 0.6768,
"step": 3900
},
{
"epoch": 0.455978514712751,
"grad_norm": 0.7279101918913876,
"learning_rate": 3.25950231753288e-05,
"loss": 0.6694,
"step": 3905
},
{
"epoch": 0.45656235404016815,
"grad_norm": 0.6380599876186313,
"learning_rate": 3.255269674684427e-05,
"loss": 0.6965,
"step": 3910
},
{
"epoch": 0.4571461933675852,
"grad_norm": 1.1241632593541853,
"learning_rate": 3.2510351479798214e-05,
"loss": 0.6528,
"step": 3915
},
{
"epoch": 0.45773003269500234,
"grad_norm": 0.7377068542020085,
"learning_rate": 3.2467987532071436e-05,
"loss": 0.6758,
"step": 3920
},
{
"epoch": 0.4583138720224194,
"grad_norm": 0.7636434619576733,
"learning_rate": 3.242560506161442e-05,
"loss": 0.6755,
"step": 3925
},
{
"epoch": 0.45889771134983653,
"grad_norm": 0.7539143479823989,
"learning_rate": 3.2383204226446706e-05,
"loss": 0.6873,
"step": 3930
},
{
"epoch": 0.4594815506772536,
"grad_norm": 0.7306998933828304,
"learning_rate": 3.234078518465628e-05,
"loss": 0.6913,
"step": 3935
},
{
"epoch": 0.4600653900046707,
"grad_norm": 0.6947217093339101,
"learning_rate": 3.229834809439904e-05,
"loss": 0.6798,
"step": 3940
},
{
"epoch": 0.4606492293320878,
"grad_norm": 0.7803653857617466,
"learning_rate": 3.225589311389816e-05,
"loss": 0.6761,
"step": 3945
},
{
"epoch": 0.4612330686595049,
"grad_norm": 0.6901357785716413,
"learning_rate": 3.221342040144352e-05,
"loss": 0.685,
"step": 3950
},
{
"epoch": 0.461816907986922,
"grad_norm": 0.6654140453249453,
"learning_rate": 3.217093011539111e-05,
"loss": 0.6895,
"step": 3955
},
{
"epoch": 0.4624007473143391,
"grad_norm": 0.7594359234519504,
"learning_rate": 3.2128422414162454e-05,
"loss": 0.681,
"step": 3960
},
{
"epoch": 0.4629845866417562,
"grad_norm": 0.6334597443860617,
"learning_rate": 3.2085897456243986e-05,
"loss": 0.679,
"step": 3965
},
{
"epoch": 0.4635684259691733,
"grad_norm": 0.734979838291988,
"learning_rate": 3.204335540018649e-05,
"loss": 0.6755,
"step": 3970
},
{
"epoch": 0.46415226529659037,
"grad_norm": 0.6136264666954013,
"learning_rate": 3.200079640460451e-05,
"loss": 0.68,
"step": 3975
},
{
"epoch": 0.4647361046240075,
"grad_norm": 0.6453670043694373,
"learning_rate": 3.195822062817573e-05,
"loss": 0.6635,
"step": 3980
},
{
"epoch": 0.46531994395142456,
"grad_norm": 0.6362575487106,
"learning_rate": 3.191562822964041e-05,
"loss": 0.6996,
"step": 3985
},
{
"epoch": 0.4659037832788417,
"grad_norm": 0.6409760600871875,
"learning_rate": 3.187301936780079e-05,
"loss": 0.6537,
"step": 3990
},
{
"epoch": 0.46648762260625876,
"grad_norm": 0.6559646351695331,
"learning_rate": 3.183039420152047e-05,
"loss": 0.6832,
"step": 3995
},
{
"epoch": 0.4670714619336759,
"grad_norm": 0.6018024828748533,
"learning_rate": 3.178775288972386e-05,
"loss": 0.681,
"step": 4000
},
{
"epoch": 0.46765530126109295,
"grad_norm": 0.6377120529644873,
"learning_rate": 3.174509559139556e-05,
"loss": 0.685,
"step": 4005
},
{
"epoch": 0.46823914058851,
"grad_norm": 0.6400426176086523,
"learning_rate": 3.170242246557978e-05,
"loss": 0.6784,
"step": 4010
},
{
"epoch": 0.46882297991592714,
"grad_norm": 0.6640252613798415,
"learning_rate": 3.1659733671379735e-05,
"loss": 0.6856,
"step": 4015
},
{
"epoch": 0.4694068192433442,
"grad_norm": 0.646288274877907,
"learning_rate": 3.1617029367957053e-05,
"loss": 0.6579,
"step": 4020
},
{
"epoch": 0.46999065857076133,
"grad_norm": 0.6953884425787434,
"learning_rate": 3.1574309714531195e-05,
"loss": 0.6812,
"step": 4025
},
{
"epoch": 0.4705744978981784,
"grad_norm": 0.6389292862822579,
"learning_rate": 3.153157487037887e-05,
"loss": 0.658,
"step": 4030
},
{
"epoch": 0.4711583372255955,
"grad_norm": 0.6192373540978041,
"learning_rate": 3.1488824994833395e-05,
"loss": 0.6719,
"step": 4035
},
{
"epoch": 0.4717421765530126,
"grad_norm": 0.677062843729221,
"learning_rate": 3.1446060247284134e-05,
"loss": 0.6748,
"step": 4040
},
{
"epoch": 0.4723260158804297,
"grad_norm": 0.7002790776981689,
"learning_rate": 3.140328078717591e-05,
"loss": 0.6835,
"step": 4045
},
{
"epoch": 0.4729098552078468,
"grad_norm": 0.684084141492704,
"learning_rate": 3.1360486774008415e-05,
"loss": 0.6788,
"step": 4050
},
{
"epoch": 0.4734936945352639,
"grad_norm": 0.6127438339585026,
"learning_rate": 3.131767836733556e-05,
"loss": 0.6837,
"step": 4055
},
{
"epoch": 0.474077533862681,
"grad_norm": 0.6533785884118601,
"learning_rate": 3.127485572676496e-05,
"loss": 0.6717,
"step": 4060
},
{
"epoch": 0.4746613731900981,
"grad_norm": 0.7442708321520219,
"learning_rate": 3.1232019011957294e-05,
"loss": 0.6678,
"step": 4065
},
{
"epoch": 0.47524521251751517,
"grad_norm": 0.6987027444038157,
"learning_rate": 3.118916838262568e-05,
"loss": 0.6608,
"step": 4070
},
{
"epoch": 0.4758290518449323,
"grad_norm": 0.6241023108030676,
"learning_rate": 3.114630399853517e-05,
"loss": 0.6651,
"step": 4075
},
{
"epoch": 0.47641289117234936,
"grad_norm": 0.6189067007303835,
"learning_rate": 3.1103426019502055e-05,
"loss": 0.6556,
"step": 4080
},
{
"epoch": 0.4769967304997665,
"grad_norm": 0.712436710640221,
"learning_rate": 3.1060534605393345e-05,
"loss": 0.6905,
"step": 4085
},
{
"epoch": 0.47758056982718355,
"grad_norm": 0.641440082239808,
"learning_rate": 3.101762991612611e-05,
"loss": 0.6741,
"step": 4090
},
{
"epoch": 0.4781644091546007,
"grad_norm": 0.8062934249889537,
"learning_rate": 3.0974712111666935e-05,
"loss": 0.6852,
"step": 4095
},
{
"epoch": 0.47874824848201775,
"grad_norm": 0.6820374065368244,
"learning_rate": 3.09317813520313e-05,
"loss": 0.6934,
"step": 4100
},
{
"epoch": 0.47933208780943487,
"grad_norm": 0.6752506606689288,
"learning_rate": 3.0888837797283005e-05,
"loss": 0.6825,
"step": 4105
},
{
"epoch": 0.47991592713685194,
"grad_norm": 0.699965080274038,
"learning_rate": 3.0845881607533524e-05,
"loss": 0.6767,
"step": 4110
},
{
"epoch": 0.480499766464269,
"grad_norm": 0.654159080249834,
"learning_rate": 3.0802912942941453e-05,
"loss": 0.683,
"step": 4115
},
{
"epoch": 0.48108360579168613,
"grad_norm": 0.6442413472135359,
"learning_rate": 3.0759931963711913e-05,
"loss": 0.678,
"step": 4120
},
{
"epoch": 0.4816674451191032,
"grad_norm": 0.7381324903332581,
"learning_rate": 3.071693883009591e-05,
"loss": 0.683,
"step": 4125
},
{
"epoch": 0.4822512844465203,
"grad_norm": 0.6457233216074335,
"learning_rate": 3.06739337023898e-05,
"loss": 0.6736,
"step": 4130
},
{
"epoch": 0.4828351237739374,
"grad_norm": 0.7359729311119515,
"learning_rate": 3.0630916740934626e-05,
"loss": 0.6757,
"step": 4135
},
{
"epoch": 0.4834189631013545,
"grad_norm": 0.6233470516611314,
"learning_rate": 3.058788810611558e-05,
"loss": 0.67,
"step": 4140
},
{
"epoch": 0.4840028024287716,
"grad_norm": 1.0294485166376433,
"learning_rate": 3.054484795836136e-05,
"loss": 0.6768,
"step": 4145
},
{
"epoch": 0.4845866417561887,
"grad_norm": 0.680402589218942,
"learning_rate": 3.0501796458143593e-05,
"loss": 0.674,
"step": 4150
},
{
"epoch": 0.4851704810836058,
"grad_norm": 0.6623427781284357,
"learning_rate": 3.045873376597624e-05,
"loss": 0.6673,
"step": 4155
},
{
"epoch": 0.4857543204110229,
"grad_norm": 0.767286105688504,
"learning_rate": 3.041566004241498e-05,
"loss": 0.6651,
"step": 4160
},
{
"epoch": 0.48633815973843997,
"grad_norm": 0.6849899087779138,
"learning_rate": 3.037257544805661e-05,
"loss": 0.6694,
"step": 4165
},
{
"epoch": 0.4869219990658571,
"grad_norm": 0.6409872957590486,
"learning_rate": 3.0329480143538498e-05,
"loss": 0.6811,
"step": 4170
},
{
"epoch": 0.48750583839327416,
"grad_norm": 0.7215248798470533,
"learning_rate": 3.0286374289537912e-05,
"loss": 0.6435,
"step": 4175
},
{
"epoch": 0.4880896777206913,
"grad_norm": 0.6092177315893011,
"learning_rate": 3.0243258046771446e-05,
"loss": 0.6668,
"step": 4180
},
{
"epoch": 0.48867351704810835,
"grad_norm": 0.7112159612723071,
"learning_rate": 3.0200131575994456e-05,
"loss": 0.6849,
"step": 4185
},
{
"epoch": 0.4892573563755255,
"grad_norm": 0.650423693744085,
"learning_rate": 3.0156995038000418e-05,
"loss": 0.6553,
"step": 4190
},
{
"epoch": 0.48984119570294254,
"grad_norm": 0.6800090154074253,
"learning_rate": 3.011384859362034e-05,
"loss": 0.6785,
"step": 4195
},
{
"epoch": 0.49042503503035967,
"grad_norm": 0.7750493391037397,
"learning_rate": 3.0070692403722162e-05,
"loss": 0.6877,
"step": 4200
},
{
"epoch": 0.49100887435777674,
"grad_norm": 0.652952148206805,
"learning_rate": 3.002752662921018e-05,
"loss": 0.6917,
"step": 4205
},
{
"epoch": 0.49159271368519386,
"grad_norm": 0.702734786226677,
"learning_rate": 2.9984351431024394e-05,
"loss": 0.67,
"step": 4210
},
{
"epoch": 0.49217655301261093,
"grad_norm": 0.616359579386879,
"learning_rate": 2.9941166970139968e-05,
"loss": 0.6465,
"step": 4215
},
{
"epoch": 0.492760392340028,
"grad_norm": 0.6803956080030293,
"learning_rate": 2.9897973407566583e-05,
"loss": 0.6648,
"step": 4220
},
{
"epoch": 0.4933442316674451,
"grad_norm": 0.7410062068580452,
"learning_rate": 2.985477090434786e-05,
"loss": 0.6878,
"step": 4225
},
{
"epoch": 0.4939280709948622,
"grad_norm": 0.6690192640617082,
"learning_rate": 2.9811559621560765e-05,
"loss": 0.679,
"step": 4230
},
{
"epoch": 0.4945119103222793,
"grad_norm": 0.7463974638227667,
"learning_rate": 2.976833972031498e-05,
"loss": 0.6542,
"step": 4235
},
{
"epoch": 0.4950957496496964,
"grad_norm": 0.636239524473288,
"learning_rate": 2.9725111361752333e-05,
"loss": 0.6946,
"step": 4240
},
{
"epoch": 0.4956795889771135,
"grad_norm": 0.643160861690113,
"learning_rate": 2.968187470704618e-05,
"loss": 0.6665,
"step": 4245
},
{
"epoch": 0.4962634283045306,
"grad_norm": 0.6501656010029333,
"learning_rate": 2.9638629917400806e-05,
"loss": 0.6911,
"step": 4250
},
{
"epoch": 0.4968472676319477,
"grad_norm": 0.6799041381274764,
"learning_rate": 2.9595377154050836e-05,
"loss": 0.6907,
"step": 4255
},
{
"epoch": 0.49743110695936477,
"grad_norm": 0.6406866621038129,
"learning_rate": 2.955211657826061e-05,
"loss": 0.6823,
"step": 4260
},
{
"epoch": 0.4980149462867819,
"grad_norm": 0.6161212623768382,
"learning_rate": 2.9508848351323597e-05,
"loss": 0.6551,
"step": 4265
},
{
"epoch": 0.49859878561419896,
"grad_norm": 0.6815666894559284,
"learning_rate": 2.9465572634561815e-05,
"loss": 0.6844,
"step": 4270
},
{
"epoch": 0.4991826249416161,
"grad_norm": 0.6442254841664492,
"learning_rate": 2.9422289589325187e-05,
"loss": 0.6706,
"step": 4275
},
{
"epoch": 0.49976646426903315,
"grad_norm": 0.8310936243096908,
"learning_rate": 2.9378999376990958e-05,
"loss": 0.6448,
"step": 4280
},
{
"epoch": 0.5003503035964503,
"grad_norm": 0.6760003318405201,
"learning_rate": 2.9335702158963107e-05,
"loss": 0.6605,
"step": 4285
},
{
"epoch": 0.5009341429238674,
"grad_norm": 0.7025461997484286,
"learning_rate": 2.929239809667172e-05,
"loss": 0.642,
"step": 4290
},
{
"epoch": 0.5015179822512844,
"grad_norm": 0.664276542414707,
"learning_rate": 2.9249087351572414e-05,
"loss": 0.6882,
"step": 4295
},
{
"epoch": 0.5021018215787015,
"grad_norm": 0.6645716985768038,
"learning_rate": 2.9205770085145716e-05,
"loss": 0.6899,
"step": 4300
},
{
"epoch": 0.5026856609061187,
"grad_norm": 0.6880641459644484,
"learning_rate": 2.916244645889647e-05,
"loss": 0.6748,
"step": 4305
},
{
"epoch": 0.5032695002335358,
"grad_norm": 1.1945610733947678,
"learning_rate": 2.911911663435322e-05,
"loss": 0.661,
"step": 4310
},
{
"epoch": 0.5038533395609528,
"grad_norm": 0.706990728557796,
"learning_rate": 2.9075780773067644e-05,
"loss": 0.6545,
"step": 4315
},
{
"epoch": 0.5044371788883699,
"grad_norm": 0.7181341291853678,
"learning_rate": 2.9032439036613907e-05,
"loss": 0.692,
"step": 4320
},
{
"epoch": 0.505021018215787,
"grad_norm": 0.6447564749525979,
"learning_rate": 2.8989091586588085e-05,
"loss": 0.6708,
"step": 4325
},
{
"epoch": 0.5056048575432041,
"grad_norm": 0.5977179669644185,
"learning_rate": 2.894573858460755e-05,
"loss": 0.6742,
"step": 4330
},
{
"epoch": 0.5061886968706212,
"grad_norm": 0.7458347028703832,
"learning_rate": 2.8902380192310392e-05,
"loss": 0.6634,
"step": 4335
},
{
"epoch": 0.5067725361980383,
"grad_norm": 0.7177340143136662,
"learning_rate": 2.8859016571354778e-05,
"loss": 0.6672,
"step": 4340
},
{
"epoch": 0.5073563755254554,
"grad_norm": 0.682261699673471,
"learning_rate": 2.881564788341839e-05,
"loss": 0.6853,
"step": 4345
},
{
"epoch": 0.5079402148528724,
"grad_norm": 0.6182174980938593,
"learning_rate": 2.877227429019778e-05,
"loss": 0.6674,
"step": 4350
},
{
"epoch": 0.5085240541802896,
"grad_norm": 0.6532096511243414,
"learning_rate": 2.872889595340781e-05,
"loss": 0.67,
"step": 4355
},
{
"epoch": 0.5091078935077067,
"grad_norm": 0.6181422508422711,
"learning_rate": 2.8685513034781003e-05,
"loss": 0.6615,
"step": 4360
},
{
"epoch": 0.5096917328351238,
"grad_norm": 0.6266040555229238,
"learning_rate": 2.864212569606699e-05,
"loss": 0.6608,
"step": 4365
},
{
"epoch": 0.5102755721625408,
"grad_norm": 0.6352722725804308,
"learning_rate": 2.8598734099031878e-05,
"loss": 0.6561,
"step": 4370
},
{
"epoch": 0.510859411489958,
"grad_norm": 0.6793855189516165,
"learning_rate": 2.8555338405457628e-05,
"loss": 0.6854,
"step": 4375
},
{
"epoch": 0.5114432508173751,
"grad_norm": 0.6768950418408346,
"learning_rate": 2.85119387771415e-05,
"loss": 0.6688,
"step": 4380
},
{
"epoch": 0.5120270901447922,
"grad_norm": 0.6580694433254123,
"learning_rate": 2.8468535375895417e-05,
"loss": 0.6421,
"step": 4385
},
{
"epoch": 0.5126109294722092,
"grad_norm": 0.6538999168432518,
"learning_rate": 2.8425128363545362e-05,
"loss": 0.6869,
"step": 4390
},
{
"epoch": 0.5131947687996263,
"grad_norm": 0.6158714003319267,
"learning_rate": 2.8381717901930792e-05,
"loss": 0.6528,
"step": 4395
},
{
"epoch": 0.5137786081270435,
"grad_norm": 0.6279279783568744,
"learning_rate": 2.8338304152904016e-05,
"loss": 0.6825,
"step": 4400
},
{
"epoch": 0.5143624474544606,
"grad_norm": 0.8185637479398602,
"learning_rate": 2.8294887278329606e-05,
"loss": 0.6748,
"step": 4405
},
{
"epoch": 0.5149462867818776,
"grad_norm": 0.6541467739492082,
"learning_rate": 2.825146744008378e-05,
"loss": 0.6552,
"step": 4410
},
{
"epoch": 0.5155301261092947,
"grad_norm": 0.6843236154552048,
"learning_rate": 2.8208044800053822e-05,
"loss": 0.7005,
"step": 4415
},
{
"epoch": 0.5161139654367118,
"grad_norm": 0.5775959172366985,
"learning_rate": 2.8164619520137437e-05,
"loss": 0.6677,
"step": 4420
},
{
"epoch": 0.516697804764129,
"grad_norm": 0.6428121987084588,
"learning_rate": 2.8121191762242188e-05,
"loss": 0.6664,
"step": 4425
},
{
"epoch": 0.517281644091546,
"grad_norm": 0.6205198887681351,
"learning_rate": 2.8077761688284886e-05,
"loss": 0.65,
"step": 4430
},
{
"epoch": 0.5178654834189631,
"grad_norm": 0.645133941769941,
"learning_rate": 2.803432946019095e-05,
"loss": 0.6677,
"step": 4435
},
{
"epoch": 0.5184493227463802,
"grad_norm": 0.694500378663462,
"learning_rate": 2.7990895239893866e-05,
"loss": 0.6627,
"step": 4440
},
{
"epoch": 0.5190331620737972,
"grad_norm": 0.6404349225526806,
"learning_rate": 2.7947459189334514e-05,
"loss": 0.6451,
"step": 4445
},
{
"epoch": 0.5196170014012144,
"grad_norm": 0.6664652978552327,
"learning_rate": 2.790402147046062e-05,
"loss": 0.6718,
"step": 4450
},
{
"epoch": 0.5202008407286315,
"grad_norm": 0.6931947237337528,
"learning_rate": 2.7860582245226114e-05,
"loss": 0.6566,
"step": 4455
},
{
"epoch": 0.5207846800560486,
"grad_norm": 0.6405779834688036,
"learning_rate": 2.781714167559056e-05,
"loss": 0.6637,
"step": 4460
},
{
"epoch": 0.5213685193834656,
"grad_norm": 0.782585563978562,
"learning_rate": 2.7773699923518527e-05,
"loss": 0.6783,
"step": 4465
},
{
"epoch": 0.5219523587108827,
"grad_norm": 0.6165660688313139,
"learning_rate": 2.7730257150978985e-05,
"loss": 0.6786,
"step": 4470
},
{
"epoch": 0.5225361980382999,
"grad_norm": 0.6107830822770479,
"learning_rate": 2.7686813519944716e-05,
"loss": 0.6735,
"step": 4475
},
{
"epoch": 0.523120037365717,
"grad_norm": 0.6248137517649573,
"learning_rate": 2.7643369192391705e-05,
"loss": 0.6603,
"step": 4480
},
{
"epoch": 0.523703876693134,
"grad_norm": 0.679269210927074,
"learning_rate": 2.759992433029852e-05,
"loss": 0.6714,
"step": 4485
},
{
"epoch": 0.5242877160205511,
"grad_norm": 0.65480206508057,
"learning_rate": 2.7556479095645753e-05,
"loss": 0.6572,
"step": 4490
},
{
"epoch": 0.5248715553479683,
"grad_norm": 0.6333322424967925,
"learning_rate": 2.7513033650415352e-05,
"loss": 0.6814,
"step": 4495
},
{
"epoch": 0.5254553946753854,
"grad_norm": 0.6752748268520505,
"learning_rate": 2.7469588156590065e-05,
"loss": 0.6529,
"step": 4500
},
{
"epoch": 0.5260392340028024,
"grad_norm": 0.6352970270986175,
"learning_rate": 2.742614277615282e-05,
"loss": 0.6769,
"step": 4505
},
{
"epoch": 0.5266230733302195,
"grad_norm": 0.5979789727114955,
"learning_rate": 2.7382697671086115e-05,
"loss": 0.6755,
"step": 4510
},
{
"epoch": 0.5272069126576366,
"grad_norm": 0.6076973729335945,
"learning_rate": 2.7339253003371434e-05,
"loss": 0.6753,
"step": 4515
},
{
"epoch": 0.5277907519850538,
"grad_norm": 0.6802597909982014,
"learning_rate": 2.729580893498862e-05,
"loss": 0.6814,
"step": 4520
},
{
"epoch": 0.5283745913124708,
"grad_norm": 0.6326424153261881,
"learning_rate": 2.725236562791529e-05,
"loss": 0.6576,
"step": 4525
},
{
"epoch": 0.5289584306398879,
"grad_norm": 0.5895411638479591,
"learning_rate": 2.7208923244126218e-05,
"loss": 0.6681,
"step": 4530
},
{
"epoch": 0.529542269967305,
"grad_norm": 0.6902118471733315,
"learning_rate": 2.716548194559273e-05,
"loss": 0.6884,
"step": 4535
},
{
"epoch": 0.5301261092947221,
"grad_norm": 0.6336203964319361,
"learning_rate": 2.7122041894282113e-05,
"loss": 0.6796,
"step": 4540
},
{
"epoch": 0.5307099486221392,
"grad_norm": 0.646528570657251,
"learning_rate": 2.707860325215701e-05,
"loss": 0.687,
"step": 4545
},
{
"epoch": 0.5312937879495563,
"grad_norm": 0.6563953698422303,
"learning_rate": 2.7035166181174786e-05,
"loss": 0.6717,
"step": 4550
},
{
"epoch": 0.5318776272769734,
"grad_norm": 0.6263578982587643,
"learning_rate": 2.6991730843286985e-05,
"loss": 0.6684,
"step": 4555
},
{
"epoch": 0.5324614666043904,
"grad_norm": 0.7538703227143297,
"learning_rate": 2.6948297400438654e-05,
"loss": 0.6919,
"step": 4560
},
{
"epoch": 0.5330453059318075,
"grad_norm": 0.6709526573941172,
"learning_rate": 2.6904866014567792e-05,
"loss": 0.6812,
"step": 4565
},
{
"epoch": 0.5336291452592247,
"grad_norm": 0.7425754597552296,
"learning_rate": 2.686143684760473e-05,
"loss": 0.679,
"step": 4570
},
{
"epoch": 0.5342129845866418,
"grad_norm": 0.6318974476260896,
"learning_rate": 2.6818010061471516e-05,
"loss": 0.6531,
"step": 4575
},
{
"epoch": 0.5347968239140588,
"grad_norm": 0.6462920552490208,
"learning_rate": 2.6774585818081332e-05,
"loss": 0.6544,
"step": 4580
},
{
"epoch": 0.5353806632414759,
"grad_norm": 0.6583779577031856,
"learning_rate": 2.6731164279337867e-05,
"loss": 0.6855,
"step": 4585
},
{
"epoch": 0.535964502568893,
"grad_norm": 0.66094901945792,
"learning_rate": 2.668774560713474e-05,
"loss": 0.6828,
"step": 4590
},
{
"epoch": 0.5365483418963102,
"grad_norm": 0.6454191499272093,
"learning_rate": 2.6644329963354882e-05,
"loss": 0.6768,
"step": 4595
},
{
"epoch": 0.5371321812237272,
"grad_norm": 0.6423029422108348,
"learning_rate": 2.6600917509869912e-05,
"loss": 0.6634,
"step": 4600
},
{
"epoch": 0.5377160205511443,
"grad_norm": 0.6120074184419095,
"learning_rate": 2.655750840853958e-05,
"loss": 0.6514,
"step": 4605
},
{
"epoch": 0.5382998598785614,
"grad_norm": 0.6618483496621712,
"learning_rate": 2.6514102821211117e-05,
"loss": 0.6827,
"step": 4610
},
{
"epoch": 0.5388836992059786,
"grad_norm": 0.6393827695784342,
"learning_rate": 2.647070090971867e-05,
"loss": 0.6705,
"step": 4615
},
{
"epoch": 0.5394675385333956,
"grad_norm": 0.6889961180224594,
"learning_rate": 2.6427302835882672e-05,
"loss": 0.6517,
"step": 4620
},
{
"epoch": 0.5400513778608127,
"grad_norm": 0.6716609039557061,
"learning_rate": 2.6383908761509252e-05,
"loss": 0.6571,
"step": 4625
},
{
"epoch": 0.5406352171882298,
"grad_norm": 0.6487842275728426,
"learning_rate": 2.634051884838961e-05,
"loss": 0.6715,
"step": 4630
},
{
"epoch": 0.541219056515647,
"grad_norm": 0.6320037420726218,
"learning_rate": 2.629713325829946e-05,
"loss": 0.6855,
"step": 4635
},
{
"epoch": 0.541802895843064,
"grad_norm": 0.654475506494646,
"learning_rate": 2.625375215299838e-05,
"loss": 0.6727,
"step": 4640
},
{
"epoch": 0.5423867351704811,
"grad_norm": 0.6258679210045515,
"learning_rate": 2.6210375694229227e-05,
"loss": 0.6861,
"step": 4645
},
{
"epoch": 0.5429705744978982,
"grad_norm": 0.6658148645412844,
"learning_rate": 2.6167004043717535e-05,
"loss": 0.6855,
"step": 4650
},
{
"epoch": 0.5435544138253152,
"grad_norm": 0.7529018085571549,
"learning_rate": 2.6123637363170912e-05,
"loss": 0.6347,
"step": 4655
},
{
"epoch": 0.5441382531527323,
"grad_norm": 0.6507054576235803,
"learning_rate": 2.6080275814278444e-05,
"loss": 0.6666,
"step": 4660
},
{
"epoch": 0.5447220924801495,
"grad_norm": 0.6127728778558662,
"learning_rate": 2.6036919558710064e-05,
"loss": 0.6783,
"step": 4665
},
{
"epoch": 0.5453059318075666,
"grad_norm": 0.6683246687451804,
"learning_rate": 2.599356875811599e-05,
"loss": 0.6727,
"step": 4670
},
{
"epoch": 0.5458897711349836,
"grad_norm": 0.6397245959830973,
"learning_rate": 2.595022357412609e-05,
"loss": 0.6412,
"step": 4675
},
{
"epoch": 0.5464736104624007,
"grad_norm": 0.7821742087008638,
"learning_rate": 2.59068841683493e-05,
"loss": 0.6644,
"step": 4680
},
{
"epoch": 0.5470574497898179,
"grad_norm": 0.7160839696281952,
"learning_rate": 2.586355070237301e-05,
"loss": 0.6712,
"step": 4685
},
{
"epoch": 0.547641289117235,
"grad_norm": 0.6546505519277086,
"learning_rate": 2.5820223337762438e-05,
"loss": 0.6484,
"step": 4690
},
{
"epoch": 0.548225128444652,
"grad_norm": 0.6836017954976104,
"learning_rate": 2.5776902236060096e-05,
"loss": 0.6635,
"step": 4695
},
{
"epoch": 0.5488089677720691,
"grad_norm": 0.6193223946781837,
"learning_rate": 2.5733587558785126e-05,
"loss": 0.6706,
"step": 4700
},
{
"epoch": 0.5493928070994862,
"grad_norm": 0.6870210502329172,
"learning_rate": 2.569027946743271e-05,
"loss": 0.6725,
"step": 4705
},
{
"epoch": 0.5499766464269034,
"grad_norm": 0.7597886046577483,
"learning_rate": 2.5646978123473477e-05,
"loss": 0.6594,
"step": 4710
},
{
"epoch": 0.5505604857543204,
"grad_norm": 0.6579081912335516,
"learning_rate": 2.560368368835291e-05,
"loss": 0.6626,
"step": 4715
},
{
"epoch": 0.5511443250817375,
"grad_norm": 0.6748205385114101,
"learning_rate": 2.5560396323490725e-05,
"loss": 0.6691,
"step": 4720
},
{
"epoch": 0.5517281644091546,
"grad_norm": 0.6372677695854689,
"learning_rate": 2.5517116190280284e-05,
"loss": 0.686,
"step": 4725
},
{
"epoch": 0.5523120037365717,
"grad_norm": 0.5944092246162501,
"learning_rate": 2.547384345008797e-05,
"loss": 0.6758,
"step": 4730
},
{
"epoch": 0.5528958430639888,
"grad_norm": 0.6061343724493072,
"learning_rate": 2.5430578264252612e-05,
"loss": 0.6704,
"step": 4735
},
{
"epoch": 0.5534796823914059,
"grad_norm": 0.6168180686552578,
"learning_rate": 2.538732079408489e-05,
"loss": 0.6666,
"step": 4740
},
{
"epoch": 0.554063521718823,
"grad_norm": 0.6591886118078801,
"learning_rate": 2.534407120086668e-05,
"loss": 0.6761,
"step": 4745
},
{
"epoch": 0.5546473610462401,
"grad_norm": 0.6285893259645317,
"learning_rate": 2.5300829645850533e-05,
"loss": 0.6494,
"step": 4750
},
{
"epoch": 0.5552312003736571,
"grad_norm": 0.6613463510402504,
"learning_rate": 2.5257596290258983e-05,
"loss": 0.6421,
"step": 4755
},
{
"epoch": 0.5558150397010743,
"grad_norm": 0.6509470734947815,
"learning_rate": 2.5214371295284028e-05,
"loss": 0.6791,
"step": 4760
},
{
"epoch": 0.5563988790284914,
"grad_norm": 0.6559471033653244,
"learning_rate": 2.517115482208649e-05,
"loss": 0.6499,
"step": 4765
},
{
"epoch": 0.5569827183559084,
"grad_norm": 0.6262128494929659,
"learning_rate": 2.5127947031795397e-05,
"loss": 0.6598,
"step": 4770
},
{
"epoch": 0.5575665576833255,
"grad_norm": 0.6470174794084477,
"learning_rate": 2.5084748085507432e-05,
"loss": 0.6661,
"step": 4775
},
{
"epoch": 0.5581503970107426,
"grad_norm": 0.7264040252574807,
"learning_rate": 2.5041558144286282e-05,
"loss": 0.6742,
"step": 4780
},
{
"epoch": 0.5587342363381598,
"grad_norm": 0.7333937929350072,
"learning_rate": 2.499837736916207e-05,
"loss": 0.6572,
"step": 4785
},
{
"epoch": 0.5593180756655768,
"grad_norm": 0.6754263101679833,
"learning_rate": 2.495520592113074e-05,
"loss": 0.6456,
"step": 4790
},
{
"epoch": 0.5599019149929939,
"grad_norm": 0.7090304084632133,
"learning_rate": 2.4912043961153468e-05,
"loss": 0.6655,
"step": 4795
},
{
"epoch": 0.560485754320411,
"grad_norm": 0.7061385294457425,
"learning_rate": 2.486889165015604e-05,
"loss": 0.6637,
"step": 4800
},
{
"epoch": 0.5610695936478282,
"grad_norm": 0.614318719517407,
"learning_rate": 2.4825749149028277e-05,
"loss": 0.6586,
"step": 4805
},
{
"epoch": 0.5616534329752452,
"grad_norm": 0.6780896077775943,
"learning_rate": 2.4782616618623428e-05,
"loss": 0.6597,
"step": 4810
},
{
"epoch": 0.5622372723026623,
"grad_norm": 0.682690271463293,
"learning_rate": 2.4739494219757554e-05,
"loss": 0.673,
"step": 4815
},
{
"epoch": 0.5628211116300794,
"grad_norm": 0.6042977851704044,
"learning_rate": 2.4696382113208956e-05,
"loss": 0.6629,
"step": 4820
},
{
"epoch": 0.5634049509574965,
"grad_norm": 0.6523864414349894,
"learning_rate": 2.465328045971755e-05,
"loss": 0.6627,
"step": 4825
},
{
"epoch": 0.5639887902849136,
"grad_norm": 0.6704225986429544,
"learning_rate": 2.4610189419984285e-05,
"loss": 0.6718,
"step": 4830
},
{
"epoch": 0.5645726296123307,
"grad_norm": 0.7144111554427477,
"learning_rate": 2.4567109154670542e-05,
"loss": 0.6583,
"step": 4835
},
{
"epoch": 0.5651564689397478,
"grad_norm": 0.5940084029786278,
"learning_rate": 2.452403982439751e-05,
"loss": 0.6861,
"step": 4840
},
{
"epoch": 0.5657403082671649,
"grad_norm": 0.6389581252882928,
"learning_rate": 2.4480981589745632e-05,
"loss": 0.6443,
"step": 4845
},
{
"epoch": 0.5663241475945819,
"grad_norm": 0.6261683415598047,
"learning_rate": 2.4437934611253972e-05,
"loss": 0.666,
"step": 4850
},
{
"epoch": 0.5669079869219991,
"grad_norm": 0.6293838161623035,
"learning_rate": 2.4394899049419612e-05,
"loss": 0.6528,
"step": 4855
},
{
"epoch": 0.5674918262494162,
"grad_norm": 0.603682637870287,
"learning_rate": 2.4351875064697093e-05,
"loss": 0.6587,
"step": 4860
},
{
"epoch": 0.5680756655768333,
"grad_norm": 0.6017431879927584,
"learning_rate": 2.4308862817497786e-05,
"loss": 0.6536,
"step": 4865
},
{
"epoch": 0.5686595049042503,
"grad_norm": 0.6379759520988977,
"learning_rate": 2.4265862468189283e-05,
"loss": 0.6665,
"step": 4870
},
{
"epoch": 0.5692433442316674,
"grad_norm": 0.6159240207916252,
"learning_rate": 2.4222874177094823e-05,
"loss": 0.6654,
"step": 4875
},
{
"epoch": 0.5698271835590846,
"grad_norm": 0.593380408089785,
"learning_rate": 2.4179898104492705e-05,
"loss": 0.6483,
"step": 4880
},
{
"epoch": 0.5704110228865016,
"grad_norm": 0.6322387205454781,
"learning_rate": 2.4136934410615646e-05,
"loss": 0.6616,
"step": 4885
},
{
"epoch": 0.5709948622139187,
"grad_norm": 0.6292676873494932,
"learning_rate": 2.4093983255650227e-05,
"loss": 0.6547,
"step": 4890
},
{
"epoch": 0.5715787015413358,
"grad_norm": 0.5837583585711659,
"learning_rate": 2.405104479973628e-05,
"loss": 0.6798,
"step": 4895
},
{
"epoch": 0.572162540868753,
"grad_norm": 0.6041134607397431,
"learning_rate": 2.400811920296627e-05,
"loss": 0.6531,
"step": 4900
},
{
"epoch": 0.57274638019617,
"grad_norm": 0.6464625146228011,
"learning_rate": 2.396520662538474e-05,
"loss": 0.6346,
"step": 4905
},
{
"epoch": 0.5733302195235871,
"grad_norm": 0.6378859032761719,
"learning_rate": 2.3922307226987678e-05,
"loss": 0.6554,
"step": 4910
},
{
"epoch": 0.5739140588510042,
"grad_norm": 0.6668773983877149,
"learning_rate": 2.3879421167721944e-05,
"loss": 0.6775,
"step": 4915
},
{
"epoch": 0.5744978981784213,
"grad_norm": 0.6389338845894582,
"learning_rate": 2.383654860748466e-05,
"loss": 0.676,
"step": 4920
},
{
"epoch": 0.5750817375058384,
"grad_norm": 0.6268842116797327,
"learning_rate": 2.379368970612261e-05,
"loss": 0.6466,
"step": 4925
},
{
"epoch": 0.5756655768332555,
"grad_norm": 0.626165018277812,
"learning_rate": 2.375084462343167e-05,
"loss": 0.6444,
"step": 4930
},
{
"epoch": 0.5762494161606726,
"grad_norm": 0.6373562595855843,
"learning_rate": 2.370801351915617e-05,
"loss": 0.6473,
"step": 4935
},
{
"epoch": 0.5768332554880897,
"grad_norm": 0.6351844260626782,
"learning_rate": 2.3665196552988357e-05,
"loss": 0.6668,
"step": 4940
},
{
"epoch": 0.5774170948155067,
"grad_norm": 0.6341326881996405,
"learning_rate": 2.362239388456773e-05,
"loss": 0.6732,
"step": 4945
},
{
"epoch": 0.5780009341429239,
"grad_norm": 0.6011659795188354,
"learning_rate": 2.357960567348049e-05,
"loss": 0.6603,
"step": 4950
},
{
"epoch": 0.578584773470341,
"grad_norm": 0.6062672094765592,
"learning_rate": 2.3536832079258952e-05,
"loss": 0.6605,
"step": 4955
},
{
"epoch": 0.5791686127977581,
"grad_norm": 0.6120030700584078,
"learning_rate": 2.3494073261380915e-05,
"loss": 0.6745,
"step": 4960
},
{
"epoch": 0.5797524521251751,
"grad_norm": 0.5924738657071491,
"learning_rate": 2.34513293792691e-05,
"loss": 0.6622,
"step": 4965
},
{
"epoch": 0.5803362914525922,
"grad_norm": 0.651965227102999,
"learning_rate": 2.340860059229052e-05,
"loss": 0.6418,
"step": 4970
},
{
"epoch": 0.5809201307800094,
"grad_norm": 0.6560128763326548,
"learning_rate": 2.3365887059755925e-05,
"loss": 0.6758,
"step": 4975
},
{
"epoch": 0.5815039701074264,
"grad_norm": 0.6506643684559978,
"learning_rate": 2.3323188940919188e-05,
"loss": 0.6533,
"step": 4980
},
{
"epoch": 0.5820878094348435,
"grad_norm": 0.6835264310080673,
"learning_rate": 2.328050639497671e-05,
"loss": 0.6647,
"step": 4985
},
{
"epoch": 0.5826716487622606,
"grad_norm": 0.5988300708192568,
"learning_rate": 2.3237839581066828e-05,
"loss": 0.6537,
"step": 4990
},
{
"epoch": 0.5832554880896778,
"grad_norm": 0.6750579550887138,
"learning_rate": 2.3195188658269224e-05,
"loss": 0.6486,
"step": 4995
},
{
"epoch": 0.5838393274170948,
"grad_norm": 0.6440448351107685,
"learning_rate": 2.3152553785604336e-05,
"loss": 0.6592,
"step": 5000
},
{
"epoch": 0.5844231667445119,
"grad_norm": 0.731759601804783,
"learning_rate": 2.3109935122032754e-05,
"loss": 0.6466,
"step": 5005
},
{
"epoch": 0.585007006071929,
"grad_norm": 0.6959499030612664,
"learning_rate": 2.3067332826454647e-05,
"loss": 0.6636,
"step": 5010
},
{
"epoch": 0.5855908453993461,
"grad_norm": 0.6113554156310252,
"learning_rate": 2.3024747057709132e-05,
"loss": 0.6763,
"step": 5015
},
{
"epoch": 0.5861746847267632,
"grad_norm": 0.6909357785524548,
"learning_rate": 2.2982177974573733e-05,
"loss": 0.6683,
"step": 5020
},
{
"epoch": 0.5867585240541803,
"grad_norm": 0.6298425983174408,
"learning_rate": 2.2939625735763743e-05,
"loss": 0.6535,
"step": 5025
},
{
"epoch": 0.5873423633815974,
"grad_norm": 0.7109872764357104,
"learning_rate": 2.2897090499931674e-05,
"loss": 0.6424,
"step": 5030
},
{
"epoch": 0.5879262027090145,
"grad_norm": 0.7580889572674745,
"learning_rate": 2.285457242566662e-05,
"loss": 0.6374,
"step": 5035
},
{
"epoch": 0.5885100420364315,
"grad_norm": 0.7017847671946715,
"learning_rate": 2.2812071671493713e-05,
"loss": 0.6529,
"step": 5040
},
{
"epoch": 0.5890938813638487,
"grad_norm": 0.6090406210084096,
"learning_rate": 2.2769588395873482e-05,
"loss": 0.6449,
"step": 5045
},
{
"epoch": 0.5896777206912658,
"grad_norm": 0.6403885858292363,
"learning_rate": 2.272712275720132e-05,
"loss": 0.6661,
"step": 5050
},
{
"epoch": 0.5902615600186829,
"grad_norm": 0.6578980666805756,
"learning_rate": 2.268467491380683e-05,
"loss": 0.6579,
"step": 5055
},
{
"epoch": 0.5908453993460999,
"grad_norm": 0.5999401661047145,
"learning_rate": 2.264224502395329e-05,
"loss": 0.6603,
"step": 5060
},
{
"epoch": 0.591429238673517,
"grad_norm": 0.6338741234363249,
"learning_rate": 2.2599833245837032e-05,
"loss": 0.6787,
"step": 5065
},
{
"epoch": 0.5920130780009342,
"grad_norm": 0.74795467731577,
"learning_rate": 2.2557439737586856e-05,
"loss": 0.6722,
"step": 5070
},
{
"epoch": 0.5925969173283513,
"grad_norm": 0.7217752543711582,
"learning_rate": 2.2515064657263447e-05,
"loss": 0.6498,
"step": 5075
},
{
"epoch": 0.5931807566557683,
"grad_norm": 0.7161765238355562,
"learning_rate": 2.2472708162858792e-05,
"loss": 0.6576,
"step": 5080
},
{
"epoch": 0.5937645959831854,
"grad_norm": 0.6535890562908294,
"learning_rate": 2.2430370412295566e-05,
"loss": 0.6714,
"step": 5085
},
{
"epoch": 0.5943484353106026,
"grad_norm": 0.6710772968914349,
"learning_rate": 2.2388051563426577e-05,
"loss": 0.6899,
"step": 5090
},
{
"epoch": 0.5949322746380196,
"grad_norm": 0.6681349138065166,
"learning_rate": 2.2345751774034135e-05,
"loss": 0.6558,
"step": 5095
},
{
"epoch": 0.5955161139654367,
"grad_norm": 0.637682685332637,
"learning_rate": 2.230347120182951e-05,
"loss": 0.6689,
"step": 5100
},
{
"epoch": 0.5960999532928538,
"grad_norm": 0.6245322541135276,
"learning_rate": 2.226121000445232e-05,
"loss": 0.6533,
"step": 5105
},
{
"epoch": 0.5966837926202709,
"grad_norm": 0.6587924015023942,
"learning_rate": 2.2218968339469932e-05,
"loss": 0.6437,
"step": 5110
},
{
"epoch": 0.597267631947688,
"grad_norm": 0.6379201786043399,
"learning_rate": 2.2176746364376904e-05,
"loss": 0.6567,
"step": 5115
},
{
"epoch": 0.5978514712751051,
"grad_norm": 0.7123489421996893,
"learning_rate": 2.2134544236594374e-05,
"loss": 0.6382,
"step": 5120
},
{
"epoch": 0.5984353106025222,
"grad_norm": 0.7171028751940878,
"learning_rate": 2.2092362113469474e-05,
"loss": 0.6628,
"step": 5125
},
{
"epoch": 0.5990191499299393,
"grad_norm": 0.6450226856301006,
"learning_rate": 2.2050200152274763e-05,
"loss": 0.6756,
"step": 5130
},
{
"epoch": 0.5996029892573563,
"grad_norm": 0.6323784574000388,
"learning_rate": 2.2008058510207635e-05,
"loss": 0.6743,
"step": 5135
},
{
"epoch": 0.6001868285847735,
"grad_norm": 0.5909602763814854,
"learning_rate": 2.1965937344389692e-05,
"loss": 0.6513,
"step": 5140
},
{
"epoch": 0.6007706679121906,
"grad_norm": 0.6395839042983865,
"learning_rate": 2.1923836811866227e-05,
"loss": 0.6534,
"step": 5145
},
{
"epoch": 0.6013545072396077,
"grad_norm": 0.6353994749336142,
"learning_rate": 2.188175706960559e-05,
"loss": 0.66,
"step": 5150
},
{
"epoch": 0.6019383465670247,
"grad_norm": 0.5922251888014807,
"learning_rate": 2.1839698274498616e-05,
"loss": 0.6808,
"step": 5155
},
{
"epoch": 0.6025221858944418,
"grad_norm": 0.6299887546972825,
"learning_rate": 2.1797660583358032e-05,
"loss": 0.658,
"step": 5160
},
{
"epoch": 0.603106025221859,
"grad_norm": 0.6531078824421405,
"learning_rate": 2.1755644152917903e-05,
"loss": 0.6481,
"step": 5165
},
{
"epoch": 0.6036898645492761,
"grad_norm": 0.6461278007928728,
"learning_rate": 2.1713649139833e-05,
"loss": 0.6719,
"step": 5170
},
{
"epoch": 0.6042737038766931,
"grad_norm": 0.6223291663451023,
"learning_rate": 2.1671675700678257e-05,
"loss": 0.6601,
"step": 5175
},
{
"epoch": 0.6048575432041102,
"grad_norm": 0.6357576147483377,
"learning_rate": 2.1629723991948176e-05,
"loss": 0.6618,
"step": 5180
},
{
"epoch": 0.6054413825315274,
"grad_norm": 0.6059467007562491,
"learning_rate": 2.1587794170056213e-05,
"loss": 0.6642,
"step": 5185
},
{
"epoch": 0.6060252218589445,
"grad_norm": 0.6095500305300035,
"learning_rate": 2.154588639133425e-05,
"loss": 0.6722,
"step": 5190
},
{
"epoch": 0.6066090611863615,
"grad_norm": 0.6726290914535656,
"learning_rate": 2.1504000812031966e-05,
"loss": 0.6659,
"step": 5195
},
{
"epoch": 0.6071929005137786,
"grad_norm": 0.5841342154320484,
"learning_rate": 2.1462137588316268e-05,
"loss": 0.6404,
"step": 5200
},
{
"epoch": 0.6077767398411957,
"grad_norm": 0.7123498891125715,
"learning_rate": 2.142029687627074e-05,
"loss": 0.6489,
"step": 5205
},
{
"epoch": 0.6083605791686127,
"grad_norm": 0.6913002613908555,
"learning_rate": 2.1378478831895e-05,
"loss": 0.6436,
"step": 5210
},
{
"epoch": 0.6089444184960299,
"grad_norm": 0.6997647280613319,
"learning_rate": 2.133668361110417e-05,
"loss": 0.6667,
"step": 5215
},
{
"epoch": 0.609528257823447,
"grad_norm": 0.6158618256900755,
"learning_rate": 2.129491136972826e-05,
"loss": 0.6378,
"step": 5220
},
{
"epoch": 0.6101120971508641,
"grad_norm": 0.6547526608685518,
"learning_rate": 2.125316226351163e-05,
"loss": 0.6703,
"step": 5225
},
{
"epoch": 0.6106959364782811,
"grad_norm": 0.6391213550824983,
"learning_rate": 2.1211436448112356e-05,
"loss": 0.6578,
"step": 5230
},
{
"epoch": 0.6112797758056983,
"grad_norm": 0.6757603390258218,
"learning_rate": 2.1169734079101684e-05,
"loss": 0.6635,
"step": 5235
},
{
"epoch": 0.6118636151331154,
"grad_norm": 0.6940231794284659,
"learning_rate": 2.1128055311963453e-05,
"loss": 0.6587,
"step": 5240
},
{
"epoch": 0.6124474544605325,
"grad_norm": 0.6946228292627638,
"learning_rate": 2.1086400302093483e-05,
"loss": 0.6635,
"step": 5245
},
{
"epoch": 0.6130312937879495,
"grad_norm": 0.6161808239430036,
"learning_rate": 2.104476920479905e-05,
"loss": 0.651,
"step": 5250
},
{
"epoch": 0.6136151331153666,
"grad_norm": 0.6427474786201127,
"learning_rate": 2.1003162175298234e-05,
"loss": 0.642,
"step": 5255
},
{
"epoch": 0.6141989724427838,
"grad_norm": 0.6308918177017456,
"learning_rate": 2.0961579368719407e-05,
"loss": 0.6604,
"step": 5260
},
{
"epoch": 0.6147828117702009,
"grad_norm": 0.6229236551665738,
"learning_rate": 2.0920020940100626e-05,
"loss": 0.6675,
"step": 5265
},
{
"epoch": 0.6153666510976179,
"grad_norm": 0.6268464665703455,
"learning_rate": 2.087848704438905e-05,
"loss": 0.6606,
"step": 5270
},
{
"epoch": 0.615950490425035,
"grad_norm": 0.6422041257326545,
"learning_rate": 2.0836977836440364e-05,
"loss": 0.6415,
"step": 5275
},
{
"epoch": 0.6165343297524521,
"grad_norm": 0.6479209627639781,
"learning_rate": 2.0795493471018222e-05,
"loss": 0.6704,
"step": 5280
},
{
"epoch": 0.6171181690798693,
"grad_norm": 0.6163247141093213,
"learning_rate": 2.075403410279364e-05,
"loss": 0.6419,
"step": 5285
},
{
"epoch": 0.6177020084072863,
"grad_norm": 0.6241956989559475,
"learning_rate": 2.0712599886344447e-05,
"loss": 0.6468,
"step": 5290
},
{
"epoch": 0.6182858477347034,
"grad_norm": 0.7656906484840937,
"learning_rate": 2.067119097615468e-05,
"loss": 0.6651,
"step": 5295
},
{
"epoch": 0.6188696870621205,
"grad_norm": 0.6965465049861614,
"learning_rate": 2.0629807526614037e-05,
"loss": 0.67,
"step": 5300
},
{
"epoch": 0.6194535263895375,
"grad_norm": 0.6023945687060536,
"learning_rate": 2.0588449692017287e-05,
"loss": 0.6507,
"step": 5305
},
{
"epoch": 0.6200373657169547,
"grad_norm": 0.6326497518909883,
"learning_rate": 2.054711762656369e-05,
"loss": 0.6695,
"step": 5310
},
{
"epoch": 0.6206212050443718,
"grad_norm": 0.6310767137058264,
"learning_rate": 2.0505811484356424e-05,
"loss": 0.6797,
"step": 5315
},
{
"epoch": 0.6212050443717889,
"grad_norm": 0.6906028300828547,
"learning_rate": 2.0464531419402026e-05,
"loss": 0.651,
"step": 5320
},
{
"epoch": 0.6217888836992059,
"grad_norm": 0.6044802850120902,
"learning_rate": 2.0423277585609806e-05,
"loss": 0.6576,
"step": 5325
},
{
"epoch": 0.622372723026623,
"grad_norm": 0.6339515721943146,
"learning_rate": 2.038205013679127e-05,
"loss": 0.6633,
"step": 5330
},
{
"epoch": 0.6229565623540402,
"grad_norm": 0.6359485528317174,
"learning_rate": 2.034084922665953e-05,
"loss": 0.645,
"step": 5335
},
{
"epoch": 0.6235404016814573,
"grad_norm": 0.6073123010252469,
"learning_rate": 2.0299675008828783e-05,
"loss": 0.6331,
"step": 5340
},
{
"epoch": 0.6241242410088743,
"grad_norm": 0.6176061294859415,
"learning_rate": 2.025852763681369e-05,
"loss": 0.6403,
"step": 5345
},
{
"epoch": 0.6247080803362914,
"grad_norm": 0.664866798044163,
"learning_rate": 2.021740726402882e-05,
"loss": 0.655,
"step": 5350
},
{
"epoch": 0.6252919196637086,
"grad_norm": 0.6086215680093997,
"learning_rate": 2.0176314043788077e-05,
"loss": 0.6611,
"step": 5355
},
{
"epoch": 0.6258757589911257,
"grad_norm": 0.7004052612104904,
"learning_rate": 2.0135248129304124e-05,
"loss": 0.6465,
"step": 5360
},
{
"epoch": 0.6264595983185427,
"grad_norm": 0.6076350077833853,
"learning_rate": 2.009420967368784e-05,
"loss": 0.6382,
"step": 5365
},
{
"epoch": 0.6270434376459598,
"grad_norm": 0.5881892530834505,
"learning_rate": 2.00531988299477e-05,
"loss": 0.6648,
"step": 5370
},
{
"epoch": 0.627627276973377,
"grad_norm": 0.6333360202885088,
"learning_rate": 2.0012215750989242e-05,
"loss": 0.6516,
"step": 5375
},
{
"epoch": 0.6282111163007941,
"grad_norm": 0.62973394403979,
"learning_rate": 1.997126058961448e-05,
"loss": 0.6567,
"step": 5380
},
{
"epoch": 0.6287949556282111,
"grad_norm": 0.6063915894035368,
"learning_rate": 1.9930333498521354e-05,
"loss": 0.6428,
"step": 5385
},
{
"epoch": 0.6293787949556282,
"grad_norm": 0.6408961647702774,
"learning_rate": 1.9889434630303118e-05,
"loss": 0.6582,
"step": 5390
},
{
"epoch": 0.6299626342830453,
"grad_norm": 0.6036349823080569,
"learning_rate": 1.9848564137447823e-05,
"loss": 0.6557,
"step": 5395
},
{
"epoch": 0.6305464736104625,
"grad_norm": 0.5919332703394341,
"learning_rate": 1.9807722172337724e-05,
"loss": 0.6487,
"step": 5400
},
{
"epoch": 0.6311303129378795,
"grad_norm": 0.6036747835371132,
"learning_rate": 1.9766908887248697e-05,
"loss": 0.6521,
"step": 5405
},
{
"epoch": 0.6317141522652966,
"grad_norm": 0.6212050444795796,
"learning_rate": 1.9726124434349706e-05,
"loss": 0.6475,
"step": 5410
},
{
"epoch": 0.6322979915927137,
"grad_norm": 0.6262816679505311,
"learning_rate": 1.9685368965702204e-05,
"loss": 0.6697,
"step": 5415
},
{
"epoch": 0.6328818309201307,
"grad_norm": 0.6124621585414037,
"learning_rate": 1.9644642633259575e-05,
"loss": 0.6302,
"step": 5420
},
{
"epoch": 0.6334656702475479,
"grad_norm": 0.6770338916705769,
"learning_rate": 1.960394558886659e-05,
"loss": 0.6435,
"step": 5425
},
{
"epoch": 0.634049509574965,
"grad_norm": 0.6501085864358357,
"learning_rate": 1.95632779842588e-05,
"loss": 0.6522,
"step": 5430
},
{
"epoch": 0.6346333489023821,
"grad_norm": 0.6130740281906146,
"learning_rate": 1.9522639971062008e-05,
"loss": 0.6689,
"step": 5435
},
{
"epoch": 0.6352171882297991,
"grad_norm": 0.5804300738072693,
"learning_rate": 1.948203170079168e-05,
"loss": 0.6609,
"step": 5440
},
{
"epoch": 0.6358010275572162,
"grad_norm": 0.5788490909752783,
"learning_rate": 1.9441453324852387e-05,
"loss": 0.6353,
"step": 5445
},
{
"epoch": 0.6363848668846334,
"grad_norm": 0.6615239820449326,
"learning_rate": 1.9400904994537257e-05,
"loss": 0.6512,
"step": 5450
},
{
"epoch": 0.6369687062120505,
"grad_norm": 0.60774343414679,
"learning_rate": 1.936038686102736e-05,
"loss": 0.6441,
"step": 5455
},
{
"epoch": 0.6375525455394675,
"grad_norm": 0.602791772845411,
"learning_rate": 1.931989907539123e-05,
"loss": 0.6603,
"step": 5460
},
{
"epoch": 0.6381363848668846,
"grad_norm": 0.6326090876764604,
"learning_rate": 1.92794417885842e-05,
"loss": 0.6654,
"step": 5465
},
{
"epoch": 0.6387202241943017,
"grad_norm": 0.643694166081535,
"learning_rate": 1.9239015151447927e-05,
"loss": 0.65,
"step": 5470
},
{
"epoch": 0.6393040635217189,
"grad_norm": 0.6786655863886062,
"learning_rate": 1.919861931470978e-05,
"loss": 0.6576,
"step": 5475
},
{
"epoch": 0.6398879028491359,
"grad_norm": 0.6106055689740919,
"learning_rate": 1.9158254428982293e-05,
"loss": 0.6552,
"step": 5480
},
{
"epoch": 0.640471742176553,
"grad_norm": 0.6007153329631955,
"learning_rate": 1.9117920644762594e-05,
"loss": 0.6712,
"step": 5485
},
{
"epoch": 0.6410555815039701,
"grad_norm": 0.6524003118058777,
"learning_rate": 1.907761811243186e-05,
"loss": 0.6408,
"step": 5490
},
{
"epoch": 0.6416394208313873,
"grad_norm": 0.6415758998316752,
"learning_rate": 1.9037346982254755e-05,
"loss": 0.6514,
"step": 5495
},
{
"epoch": 0.6422232601588043,
"grad_norm": 0.6295681316621072,
"learning_rate": 1.8997107404378846e-05,
"loss": 0.6555,
"step": 5500
},
{
"epoch": 0.6428070994862214,
"grad_norm": 0.6100423766821667,
"learning_rate": 1.8956899528834065e-05,
"loss": 0.6378,
"step": 5505
},
{
"epoch": 0.6433909388136385,
"grad_norm": 0.6450917653118526,
"learning_rate": 1.8916723505532157e-05,
"loss": 0.6454,
"step": 5510
},
{
"epoch": 0.6439747781410555,
"grad_norm": 0.6558351860092898,
"learning_rate": 1.8876579484266094e-05,
"loss": 0.64,
"step": 5515
},
{
"epoch": 0.6445586174684726,
"grad_norm": 0.631323811726284,
"learning_rate": 1.8836467614709535e-05,
"loss": 0.659,
"step": 5520
},
{
"epoch": 0.6451424567958898,
"grad_norm": 0.6153787346866092,
"learning_rate": 1.8796388046416253e-05,
"loss": 0.6619,
"step": 5525
},
{
"epoch": 0.6457262961233069,
"grad_norm": 0.6176755770472118,
"learning_rate": 1.875634092881963e-05,
"loss": 0.6552,
"step": 5530
},
{
"epoch": 0.6463101354507239,
"grad_norm": 0.6517515108908513,
"learning_rate": 1.8716326411232016e-05,
"loss": 0.6529,
"step": 5535
},
{
"epoch": 0.646893974778141,
"grad_norm": 0.6320389903656185,
"learning_rate": 1.8676344642844217e-05,
"loss": 0.6419,
"step": 5540
},
{
"epoch": 0.6474778141055582,
"grad_norm": 0.6577567743805603,
"learning_rate": 1.8636395772724952e-05,
"loss": 0.6517,
"step": 5545
},
{
"epoch": 0.6480616534329753,
"grad_norm": 0.5779409429803866,
"learning_rate": 1.8596479949820273e-05,
"loss": 0.6516,
"step": 5550
},
{
"epoch": 0.6486454927603923,
"grad_norm": 0.5982564290583605,
"learning_rate": 1.8556597322953035e-05,
"loss": 0.6628,
"step": 5555
},
{
"epoch": 0.6492293320878094,
"grad_norm": 0.6033275354473104,
"learning_rate": 1.8516748040822295e-05,
"loss": 0.6462,
"step": 5560
},
{
"epoch": 0.6498131714152265,
"grad_norm": 0.7478615750937032,
"learning_rate": 1.847693225200281e-05,
"loss": 0.652,
"step": 5565
},
{
"epoch": 0.6503970107426437,
"grad_norm": 0.6028133177899443,
"learning_rate": 1.843715010494445e-05,
"loss": 0.6554,
"step": 5570
},
{
"epoch": 0.6509808500700607,
"grad_norm": 0.6335265470082463,
"learning_rate": 1.839740174797166e-05,
"loss": 0.6683,
"step": 5575
},
{
"epoch": 0.6515646893974778,
"grad_norm": 0.6384594597330329,
"learning_rate": 1.8357687329282896e-05,
"loss": 0.6583,
"step": 5580
},
{
"epoch": 0.6521485287248949,
"grad_norm": 0.6634517144325908,
"learning_rate": 1.831800699695008e-05,
"loss": 0.6518,
"step": 5585
},
{
"epoch": 0.652732368052312,
"grad_norm": 0.6324345938898167,
"learning_rate": 1.827836089891805e-05,
"loss": 0.6595,
"step": 5590
},
{
"epoch": 0.6533162073797291,
"grad_norm": 0.61649113535426,
"learning_rate": 1.823874918300399e-05,
"loss": 0.6515,
"step": 5595
},
{
"epoch": 0.6539000467071462,
"grad_norm": 0.5897384254074928,
"learning_rate": 1.8199171996896912e-05,
"loss": 0.6517,
"step": 5600
},
{
"epoch": 0.6544838860345633,
"grad_norm": 0.6677989875296673,
"learning_rate": 1.8159629488157082e-05,
"loss": 0.6698,
"step": 5605
},
{
"epoch": 0.6550677253619804,
"grad_norm": 0.6455168459371002,
"learning_rate": 1.8120121804215466e-05,
"loss": 0.6537,
"step": 5610
},
{
"epoch": 0.6556515646893974,
"grad_norm": 0.6397653850550866,
"learning_rate": 1.8080649092373187e-05,
"loss": 0.6584,
"step": 5615
},
{
"epoch": 0.6562354040168146,
"grad_norm": 0.626530275084573,
"learning_rate": 1.8041211499800992e-05,
"loss": 0.6483,
"step": 5620
},
{
"epoch": 0.6568192433442317,
"grad_norm": 0.6227195752834476,
"learning_rate": 1.8001809173538676e-05,
"loss": 0.6449,
"step": 5625
},
{
"epoch": 0.6574030826716487,
"grad_norm": 0.6115177962874717,
"learning_rate": 1.796244226049455e-05,
"loss": 0.6541,
"step": 5630
},
{
"epoch": 0.6579869219990658,
"grad_norm": 0.6110717594233479,
"learning_rate": 1.792311090744489e-05,
"loss": 0.6234,
"step": 5635
},
{
"epoch": 0.658570761326483,
"grad_norm": 0.6345530467904902,
"learning_rate": 1.7883815261033393e-05,
"loss": 0.6517,
"step": 5640
},
{
"epoch": 0.6591546006539001,
"grad_norm": 0.6059373392098923,
"learning_rate": 1.7844555467770624e-05,
"loss": 0.6655,
"step": 5645
},
{
"epoch": 0.6597384399813171,
"grad_norm": 0.5919917566253382,
"learning_rate": 1.7805331674033466e-05,
"loss": 0.6533,
"step": 5650
},
{
"epoch": 0.6603222793087342,
"grad_norm": 0.6500691097454692,
"learning_rate": 1.776614402606459e-05,
"loss": 0.6516,
"step": 5655
},
{
"epoch": 0.6609061186361513,
"grad_norm": 0.6554113582331242,
"learning_rate": 1.7726992669971904e-05,
"loss": 0.6513,
"step": 5660
},
{
"epoch": 0.6614899579635685,
"grad_norm": 0.587579022092992,
"learning_rate": 1.768787775172799e-05,
"loss": 0.6541,
"step": 5665
},
{
"epoch": 0.6620737972909855,
"grad_norm": 0.6461208021423955,
"learning_rate": 1.7648799417169588e-05,
"loss": 0.6536,
"step": 5670
},
{
"epoch": 0.6626576366184026,
"grad_norm": 0.6446174965711384,
"learning_rate": 1.7609757811997023e-05,
"loss": 0.6501,
"step": 5675
},
{
"epoch": 0.6632414759458197,
"grad_norm": 0.6563309768757624,
"learning_rate": 1.75707530817737e-05,
"loss": 0.6492,
"step": 5680
},
{
"epoch": 0.6638253152732368,
"grad_norm": 0.6207469875080636,
"learning_rate": 1.753178537192551e-05,
"loss": 0.6533,
"step": 5685
},
{
"epoch": 0.6644091546006539,
"grad_norm": 0.6369501224014037,
"learning_rate": 1.7492854827740353e-05,
"loss": 0.6596,
"step": 5690
},
{
"epoch": 0.664992993928071,
"grad_norm": 0.6136259565703345,
"learning_rate": 1.7453961594367528e-05,
"loss": 0.6519,
"step": 5695
},
{
"epoch": 0.6655768332554881,
"grad_norm": 0.6684512394811364,
"learning_rate": 1.741510581681724e-05,
"loss": 0.6478,
"step": 5700
},
{
"epoch": 0.6661606725829052,
"grad_norm": 0.6889730982477917,
"learning_rate": 1.737628763996005e-05,
"loss": 0.6663,
"step": 5705
},
{
"epoch": 0.6667445119103222,
"grad_norm": 0.6993773876386039,
"learning_rate": 1.7337507208526295e-05,
"loss": 0.6537,
"step": 5710
},
{
"epoch": 0.6673283512377394,
"grad_norm": 0.5948998317909232,
"learning_rate": 1.729876466710561e-05,
"loss": 0.6525,
"step": 5715
},
{
"epoch": 0.6679121905651565,
"grad_norm": 0.6349196689747608,
"learning_rate": 1.726006016014637e-05,
"loss": 0.6697,
"step": 5720
},
{
"epoch": 0.6684960298925736,
"grad_norm": 0.6247396476759942,
"learning_rate": 1.7221393831955102e-05,
"loss": 0.6517,
"step": 5725
},
{
"epoch": 0.6690798692199906,
"grad_norm": 0.7102052156590236,
"learning_rate": 1.718276582669602e-05,
"loss": 0.6408,
"step": 5730
},
{
"epoch": 0.6696637085474078,
"grad_norm": 0.6098663239917782,
"learning_rate": 1.7144176288390448e-05,
"loss": 0.6411,
"step": 5735
},
{
"epoch": 0.6702475478748249,
"grad_norm": 0.6236060967217535,
"learning_rate": 1.7105625360916276e-05,
"loss": 0.6614,
"step": 5740
},
{
"epoch": 0.6708313872022419,
"grad_norm": 0.6514443423723102,
"learning_rate": 1.7067113188007457e-05,
"loss": 0.6391,
"step": 5745
},
{
"epoch": 0.671415226529659,
"grad_norm": 0.6145992521310542,
"learning_rate": 1.7028639913253426e-05,
"loss": 0.6526,
"step": 5750
},
{
"epoch": 0.6719990658570761,
"grad_norm": 0.6392156062042794,
"learning_rate": 1.6990205680098612e-05,
"loss": 0.6613,
"step": 5755
},
{
"epoch": 0.6725829051844933,
"grad_norm": 0.6093115717729776,
"learning_rate": 1.695181063184187e-05,
"loss": 0.6597,
"step": 5760
},
{
"epoch": 0.6731667445119103,
"grad_norm": 0.6158268949318265,
"learning_rate": 1.6913454911635954e-05,
"loss": 0.6454,
"step": 5765
},
{
"epoch": 0.6737505838393274,
"grad_norm": 0.6295476392889442,
"learning_rate": 1.6875138662486997e-05,
"loss": 0.6408,
"step": 5770
},
{
"epoch": 0.6743344231667445,
"grad_norm": 0.6754149318261963,
"learning_rate": 1.6836862027253963e-05,
"loss": 0.6345,
"step": 5775
},
{
"epoch": 0.6749182624941616,
"grad_norm": 0.7471938496073851,
"learning_rate": 1.6798625148648113e-05,
"loss": 0.6557,
"step": 5780
},
{
"epoch": 0.6755021018215787,
"grad_norm": 0.6355574202574713,
"learning_rate": 1.6760428169232483e-05,
"loss": 0.6365,
"step": 5785
},
{
"epoch": 0.6760859411489958,
"grad_norm": 0.5835350863392492,
"learning_rate": 1.672227123142136e-05,
"loss": 0.6575,
"step": 5790
},
{
"epoch": 0.6766697804764129,
"grad_norm": 0.5942137459790245,
"learning_rate": 1.668415447747971e-05,
"loss": 0.6326,
"step": 5795
},
{
"epoch": 0.67725361980383,
"grad_norm": 0.6282690457346377,
"learning_rate": 1.6646078049522706e-05,
"loss": 0.6383,
"step": 5800
},
{
"epoch": 0.677837459131247,
"grad_norm": 0.6219412563102987,
"learning_rate": 1.660804208951516e-05,
"loss": 0.6519,
"step": 5805
},
{
"epoch": 0.6784212984586642,
"grad_norm": 0.5701155156677662,
"learning_rate": 1.6570046739270988e-05,
"loss": 0.6519,
"step": 5810
},
{
"epoch": 0.6790051377860813,
"grad_norm": 0.5878066897687751,
"learning_rate": 1.6532092140452725e-05,
"loss": 0.654,
"step": 5815
},
{
"epoch": 0.6795889771134984,
"grad_norm": 0.5908028121861336,
"learning_rate": 1.649417843457094e-05,
"loss": 0.6373,
"step": 5820
},
{
"epoch": 0.6801728164409154,
"grad_norm": 0.639959083744787,
"learning_rate": 1.6456305762983742e-05,
"loss": 0.6553,
"step": 5825
},
{
"epoch": 0.6807566557683326,
"grad_norm": 0.5986487080483318,
"learning_rate": 1.6418474266896257e-05,
"loss": 0.6427,
"step": 5830
},
{
"epoch": 0.6813404950957497,
"grad_norm": 0.6435561857144427,
"learning_rate": 1.6380684087360088e-05,
"loss": 0.6319,
"step": 5835
},
{
"epoch": 0.6819243344231667,
"grad_norm": 0.6281643119655214,
"learning_rate": 1.6342935365272785e-05,
"loss": 0.6595,
"step": 5840
},
{
"epoch": 0.6825081737505838,
"grad_norm": 0.62328606989937,
"learning_rate": 1.6305228241377347e-05,
"loss": 0.6337,
"step": 5845
},
{
"epoch": 0.6830920130780009,
"grad_norm": 0.6131419563564767,
"learning_rate": 1.6267562856261638e-05,
"loss": 0.6455,
"step": 5850
},
{
"epoch": 0.6836758524054181,
"grad_norm": 0.6126007307298542,
"learning_rate": 1.6229939350357952e-05,
"loss": 0.6423,
"step": 5855
},
{
"epoch": 0.6842596917328351,
"grad_norm": 0.6039540770428432,
"learning_rate": 1.6192357863942415e-05,
"loss": 0.6459,
"step": 5860
},
{
"epoch": 0.6848435310602522,
"grad_norm": 0.6406333234122924,
"learning_rate": 1.615481853713448e-05,
"loss": 0.6331,
"step": 5865
},
{
"epoch": 0.6854273703876693,
"grad_norm": 0.6561121020637503,
"learning_rate": 1.6117321509896422e-05,
"loss": 0.6431,
"step": 5870
},
{
"epoch": 0.6860112097150864,
"grad_norm": 0.6034665695981549,
"learning_rate": 1.60798669220328e-05,
"loss": 0.6404,
"step": 5875
},
{
"epoch": 0.6865950490425035,
"grad_norm": 0.6602912607998124,
"learning_rate": 1.6042454913189946e-05,
"loss": 0.6735,
"step": 5880
},
{
"epoch": 0.6871788883699206,
"grad_norm": 0.6492996294271381,
"learning_rate": 1.600508562285544e-05,
"loss": 0.6708,
"step": 5885
},
{
"epoch": 0.6877627276973377,
"grad_norm": 0.6018342905291776,
"learning_rate": 1.5967759190357585e-05,
"loss": 0.6361,
"step": 5890
},
{
"epoch": 0.6883465670247548,
"grad_norm": 0.6329922232888995,
"learning_rate": 1.5930475754864898e-05,
"loss": 0.6468,
"step": 5895
},
{
"epoch": 0.6889304063521718,
"grad_norm": 0.6525189222677755,
"learning_rate": 1.5893235455385575e-05,
"loss": 0.6601,
"step": 5900
},
{
"epoch": 0.689514245679589,
"grad_norm": 0.5877002885557084,
"learning_rate": 1.5856038430766994e-05,
"loss": 0.6493,
"step": 5905
},
{
"epoch": 0.6900980850070061,
"grad_norm": 0.6163486286664249,
"learning_rate": 1.5818884819695184e-05,
"loss": 0.6414,
"step": 5910
},
{
"epoch": 0.6906819243344232,
"grad_norm": 0.6686532667952824,
"learning_rate": 1.5781774760694304e-05,
"loss": 0.6307,
"step": 5915
},
{
"epoch": 0.6912657636618402,
"grad_norm": 0.6422509284417167,
"learning_rate": 1.5744708392126138e-05,
"loss": 0.6431,
"step": 5920
},
{
"epoch": 0.6918496029892574,
"grad_norm": 0.6414099958899228,
"learning_rate": 1.5707685852189573e-05,
"loss": 0.6535,
"step": 5925
},
{
"epoch": 0.6924334423166745,
"grad_norm": 0.6286824558971477,
"learning_rate": 1.5670707278920084e-05,
"loss": 0.652,
"step": 5930
},
{
"epoch": 0.6930172816440916,
"grad_norm": 0.617985919524731,
"learning_rate": 1.563377281018922e-05,
"loss": 0.6433,
"step": 5935
},
{
"epoch": 0.6936011209715086,
"grad_norm": 0.6398079036319524,
"learning_rate": 1.5596882583704092e-05,
"loss": 0.6691,
"step": 5940
},
{
"epoch": 0.6941849602989257,
"grad_norm": 0.587760118379386,
"learning_rate": 1.5560036737006856e-05,
"loss": 0.646,
"step": 5945
},
{
"epoch": 0.6947687996263429,
"grad_norm": 0.6269915871686259,
"learning_rate": 1.5523235407474195e-05,
"loss": 0.6506,
"step": 5950
},
{
"epoch": 0.6953526389537599,
"grad_norm": 0.6653984631772404,
"learning_rate": 1.5486478732316827e-05,
"loss": 0.668,
"step": 5955
},
{
"epoch": 0.695936478281177,
"grad_norm": 0.606714091384374,
"learning_rate": 1.5449766848578968e-05,
"loss": 0.6463,
"step": 5960
},
{
"epoch": 0.6965203176085941,
"grad_norm": 0.6520045148882592,
"learning_rate": 1.541309989313784e-05,
"loss": 0.6675,
"step": 5965
},
{
"epoch": 0.6971041569360112,
"grad_norm": 0.6604691030846023,
"learning_rate": 1.5376478002703154e-05,
"loss": 0.6539,
"step": 5970
},
{
"epoch": 0.6976879962634283,
"grad_norm": 0.6256562528791287,
"learning_rate": 1.5339901313816584e-05,
"loss": 0.6391,
"step": 5975
},
{
"epoch": 0.6982718355908454,
"grad_norm": 0.6131593198270057,
"learning_rate": 1.5303369962851298e-05,
"loss": 0.6588,
"step": 5980
},
{
"epoch": 0.6988556749182625,
"grad_norm": 0.5802257376801692,
"learning_rate": 1.5266884086011406e-05,
"loss": 0.6542,
"step": 5985
},
{
"epoch": 0.6994395142456796,
"grad_norm": 0.5816165522921771,
"learning_rate": 1.5230443819331492e-05,
"loss": 0.6431,
"step": 5990
},
{
"epoch": 0.7000233535730966,
"grad_norm": 0.612328598266228,
"learning_rate": 1.5194049298676061e-05,
"loss": 0.6281,
"step": 5995
},
{
"epoch": 0.7006071929005138,
"grad_norm": 0.6653597780756229,
"learning_rate": 1.515770065973907e-05,
"loss": 0.6431,
"step": 6000
},
{
"epoch": 0.7011910322279309,
"grad_norm": 0.6376102143697321,
"learning_rate": 1.5121398038043421e-05,
"loss": 0.6573,
"step": 6005
},
{
"epoch": 0.701774871555348,
"grad_norm": 0.636891782596444,
"learning_rate": 1.5085141568940419e-05,
"loss": 0.6403,
"step": 6010
},
{
"epoch": 0.702358710882765,
"grad_norm": 0.6134843917603902,
"learning_rate": 1.5048931387609321e-05,
"loss": 0.6372,
"step": 6015
},
{
"epoch": 0.7029425502101821,
"grad_norm": 0.6111357219189175,
"learning_rate": 1.501276762905679e-05,
"loss": 0.6522,
"step": 6020
},
{
"epoch": 0.7035263895375993,
"grad_norm": 0.7624771516374586,
"learning_rate": 1.4976650428116401e-05,
"loss": 0.6591,
"step": 6025
},
{
"epoch": 0.7041102288650164,
"grad_norm": 0.6082248979619495,
"learning_rate": 1.4940579919448147e-05,
"loss": 0.6437,
"step": 6030
},
{
"epoch": 0.7046940681924334,
"grad_norm": 0.6060066653601676,
"learning_rate": 1.4904556237537936e-05,
"loss": 0.6816,
"step": 6035
},
{
"epoch": 0.7052779075198505,
"grad_norm": 0.6428653416363093,
"learning_rate": 1.4868579516697079e-05,
"loss": 0.6505,
"step": 6040
},
{
"epoch": 0.7058617468472677,
"grad_norm": 0.6196446538185576,
"learning_rate": 1.4832649891061811e-05,
"loss": 0.6413,
"step": 6045
},
{
"epoch": 0.7064455861746848,
"grad_norm": 0.6370871328272835,
"learning_rate": 1.4796767494592757e-05,
"loss": 0.6458,
"step": 6050
},
{
"epoch": 0.7070294255021018,
"grad_norm": 0.5896264915560131,
"learning_rate": 1.4760932461074467e-05,
"loss": 0.6409,
"step": 6055
},
{
"epoch": 0.7076132648295189,
"grad_norm": 0.6123918490122225,
"learning_rate": 1.4725144924114891e-05,
"loss": 0.6356,
"step": 6060
},
{
"epoch": 0.708197104156936,
"grad_norm": 0.6159099756196177,
"learning_rate": 1.4689405017144908e-05,
"loss": 0.6396,
"step": 6065
},
{
"epoch": 0.708780943484353,
"grad_norm": 0.7028868922003575,
"learning_rate": 1.4653712873417796e-05,
"loss": 0.6572,
"step": 6070
},
{
"epoch": 0.7093647828117702,
"grad_norm": 0.5920818715762003,
"learning_rate": 1.4618068626008755e-05,
"loss": 0.6319,
"step": 6075
},
{
"epoch": 0.7099486221391873,
"grad_norm": 0.5995994181370015,
"learning_rate": 1.4582472407814419e-05,
"loss": 0.643,
"step": 6080
},
{
"epoch": 0.7105324614666044,
"grad_norm": 0.5860343193735567,
"learning_rate": 1.4546924351552333e-05,
"loss": 0.6545,
"step": 6085
},
{
"epoch": 0.7111163007940214,
"grad_norm": 0.6128351666184559,
"learning_rate": 1.4511424589760486e-05,
"loss": 0.6349,
"step": 6090
},
{
"epoch": 0.7117001401214386,
"grad_norm": 0.6252244905566473,
"learning_rate": 1.4475973254796799e-05,
"loss": 0.6648,
"step": 6095
},
{
"epoch": 0.7122839794488557,
"grad_norm": 0.6046517958913106,
"learning_rate": 1.4440570478838645e-05,
"loss": 0.6541,
"step": 6100
},
{
"epoch": 0.7128678187762728,
"grad_norm": 0.6371397956094005,
"learning_rate": 1.440521639388233e-05,
"loss": 0.6442,
"step": 6105
},
{
"epoch": 0.7134516581036898,
"grad_norm": 0.6334205971437344,
"learning_rate": 1.436991113174265e-05,
"loss": 0.6568,
"step": 6110
},
{
"epoch": 0.714035497431107,
"grad_norm": 0.6136702164991507,
"learning_rate": 1.4334654824052351e-05,
"loss": 0.6356,
"step": 6115
},
{
"epoch": 0.7146193367585241,
"grad_norm": 0.5740255870328927,
"learning_rate": 1.429944760226164e-05,
"loss": 0.6554,
"step": 6120
},
{
"epoch": 0.7152031760859412,
"grad_norm": 0.647412850303488,
"learning_rate": 1.4264289597637741e-05,
"loss": 0.6594,
"step": 6125
},
{
"epoch": 0.7157870154133582,
"grad_norm": 0.6023144532026925,
"learning_rate": 1.4229180941264364e-05,
"loss": 0.6346,
"step": 6130
},
{
"epoch": 0.7163708547407753,
"grad_norm": 0.6566751465495958,
"learning_rate": 1.4194121764041224e-05,
"loss": 0.6396,
"step": 6135
},
{
"epoch": 0.7169546940681925,
"grad_norm": 0.6297042271423003,
"learning_rate": 1.4159112196683564e-05,
"loss": 0.6377,
"step": 6140
},
{
"epoch": 0.7175385333956096,
"grad_norm": 0.6734484611856958,
"learning_rate": 1.4124152369721655e-05,
"loss": 0.6537,
"step": 6145
},
{
"epoch": 0.7181223727230266,
"grad_norm": 0.6317487967098724,
"learning_rate": 1.408924241350032e-05,
"loss": 0.6209,
"step": 6150
},
{
"epoch": 0.7187062120504437,
"grad_norm": 0.5866856203616632,
"learning_rate": 1.4054382458178439e-05,
"loss": 0.6421,
"step": 6155
},
{
"epoch": 0.7192900513778608,
"grad_norm": 0.68949122971027,
"learning_rate": 1.4019572633728473e-05,
"loss": 0.6525,
"step": 6160
},
{
"epoch": 0.7198738907052779,
"grad_norm": 0.6472560827371897,
"learning_rate": 1.3984813069935967e-05,
"loss": 0.6473,
"step": 6165
},
{
"epoch": 0.720457730032695,
"grad_norm": 0.6070589006805013,
"learning_rate": 1.395010389639908e-05,
"loss": 0.6574,
"step": 6170
},
{
"epoch": 0.7210415693601121,
"grad_norm": 0.5650663505862649,
"learning_rate": 1.391544524252808e-05,
"loss": 0.6402,
"step": 6175
},
{
"epoch": 0.7216254086875292,
"grad_norm": 0.6390461835721468,
"learning_rate": 1.388083723754491e-05,
"loss": 0.6494,
"step": 6180
},
{
"epoch": 0.7222092480149462,
"grad_norm": 0.6103399495927149,
"learning_rate": 1.384628001048264e-05,
"loss": 0.6418,
"step": 6185
},
{
"epoch": 0.7227930873423634,
"grad_norm": 0.6954938172005755,
"learning_rate": 1.381177369018503e-05,
"loss": 0.6406,
"step": 6190
},
{
"epoch": 0.7233769266697805,
"grad_norm": 0.6218535372380833,
"learning_rate": 1.377731840530604e-05,
"loss": 0.6432,
"step": 6195
},
{
"epoch": 0.7239607659971976,
"grad_norm": 0.6089100037812815,
"learning_rate": 1.374291428430935e-05,
"loss": 0.6515,
"step": 6200
},
{
"epoch": 0.7245446053246146,
"grad_norm": 0.6572788470504004,
"learning_rate": 1.3708561455467872e-05,
"loss": 0.646,
"step": 6205
},
{
"epoch": 0.7251284446520317,
"grad_norm": 0.5960825154712381,
"learning_rate": 1.3674260046863285e-05,
"loss": 0.6362,
"step": 6210
},
{
"epoch": 0.7257122839794489,
"grad_norm": 0.5993510671824555,
"learning_rate": 1.3640010186385552e-05,
"loss": 0.6382,
"step": 6215
},
{
"epoch": 0.726296123306866,
"grad_norm": 0.6337346841685323,
"learning_rate": 1.3605812001732444e-05,
"loss": 0.6531,
"step": 6220
},
{
"epoch": 0.726879962634283,
"grad_norm": 0.6349038847860046,
"learning_rate": 1.3571665620409064e-05,
"loss": 0.6488,
"step": 6225
},
{
"epoch": 0.7274638019617001,
"grad_norm": 0.6294965655367071,
"learning_rate": 1.3537571169727359e-05,
"loss": 0.6365,
"step": 6230
},
{
"epoch": 0.7280476412891173,
"grad_norm": 0.6201172499902672,
"learning_rate": 1.3503528776805676e-05,
"loss": 0.6295,
"step": 6235
},
{
"epoch": 0.7286314806165344,
"grad_norm": 0.6110861537928787,
"learning_rate": 1.3469538568568255e-05,
"loss": 0.6284,
"step": 6240
},
{
"epoch": 0.7292153199439514,
"grad_norm": 0.6215210859610367,
"learning_rate": 1.3435600671744768e-05,
"loss": 0.6274,
"step": 6245
},
{
"epoch": 0.7297991592713685,
"grad_norm": 0.5847702315380873,
"learning_rate": 1.3401715212869864e-05,
"loss": 0.6208,
"step": 6250
},
{
"epoch": 0.7303829985987856,
"grad_norm": 0.5828875582378416,
"learning_rate": 1.3367882318282666e-05,
"loss": 0.6475,
"step": 6255
},
{
"epoch": 0.7309668379262028,
"grad_norm": 0.6127808516375031,
"learning_rate": 1.3334102114126314e-05,
"loss": 0.6504,
"step": 6260
},
{
"epoch": 0.7315506772536198,
"grad_norm": 0.6671099859763688,
"learning_rate": 1.330037472634752e-05,
"loss": 0.6512,
"step": 6265
},
{
"epoch": 0.7321345165810369,
"grad_norm": 0.6142444220893487,
"learning_rate": 1.3266700280696042e-05,
"loss": 0.6318,
"step": 6270
},
{
"epoch": 0.732718355908454,
"grad_norm": 0.5938810178983895,
"learning_rate": 1.3233078902724266e-05,
"loss": 0.6568,
"step": 6275
},
{
"epoch": 0.733302195235871,
"grad_norm": 0.70988806135586,
"learning_rate": 1.3199510717786714e-05,
"loss": 0.6521,
"step": 6280
},
{
"epoch": 0.7338860345632882,
"grad_norm": 0.6512938097399694,
"learning_rate": 1.3165995851039591e-05,
"loss": 0.6713,
"step": 6285
},
{
"epoch": 0.7344698738907053,
"grad_norm": 0.6601449949865233,
"learning_rate": 1.3132534427440301e-05,
"loss": 0.6417,
"step": 6290
},
{
"epoch": 0.7350537132181224,
"grad_norm": 0.6701809451404614,
"learning_rate": 1.309912657174699e-05,
"loss": 0.6483,
"step": 6295
},
{
"epoch": 0.7356375525455394,
"grad_norm": 0.6158604399069405,
"learning_rate": 1.3065772408518085e-05,
"loss": 0.6363,
"step": 6300
},
{
"epoch": 0.7362213918729565,
"grad_norm": 0.6476859953305574,
"learning_rate": 1.3032472062111823e-05,
"loss": 0.6379,
"step": 6305
},
{
"epoch": 0.7368052312003737,
"grad_norm": 0.6658408876225623,
"learning_rate": 1.2999225656685781e-05,
"loss": 0.6441,
"step": 6310
},
{
"epoch": 0.7373890705277908,
"grad_norm": 0.6288192770121708,
"learning_rate": 1.2966033316196435e-05,
"loss": 0.6526,
"step": 6315
},
{
"epoch": 0.7379729098552078,
"grad_norm": 0.6007763475196312,
"learning_rate": 1.2932895164398684e-05,
"loss": 0.6462,
"step": 6320
},
{
"epoch": 0.7385567491826249,
"grad_norm": 0.5998562845562518,
"learning_rate": 1.2899811324845373e-05,
"loss": 0.6447,
"step": 6325
},
{
"epoch": 0.739140588510042,
"grad_norm": 0.6066992473051736,
"learning_rate": 1.2866781920886873e-05,
"loss": 0.6599,
"step": 6330
},
{
"epoch": 0.7397244278374592,
"grad_norm": 0.5610174159935682,
"learning_rate": 1.2833807075670564e-05,
"loss": 0.6404,
"step": 6335
},
{
"epoch": 0.7403082671648762,
"grad_norm": 0.6733209472989998,
"learning_rate": 1.2800886912140433e-05,
"loss": 0.6426,
"step": 6340
},
{
"epoch": 0.7408921064922933,
"grad_norm": 0.6339702252792792,
"learning_rate": 1.2768021553036596e-05,
"loss": 0.6278,
"step": 6345
},
{
"epoch": 0.7414759458197104,
"grad_norm": 0.6258093250705813,
"learning_rate": 1.2735211120894813e-05,
"loss": 0.647,
"step": 6350
},
{
"epoch": 0.7420597851471276,
"grad_norm": 0.5868680496778399,
"learning_rate": 1.2702455738046068e-05,
"loss": 0.6468,
"step": 6355
},
{
"epoch": 0.7426436244745446,
"grad_norm": 0.7012725486888974,
"learning_rate": 1.2669755526616093e-05,
"loss": 0.6573,
"step": 6360
},
{
"epoch": 0.7432274638019617,
"grad_norm": 0.6943371402211006,
"learning_rate": 1.2637110608524916e-05,
"loss": 0.6213,
"step": 6365
},
{
"epoch": 0.7438113031293788,
"grad_norm": 0.6618196344032842,
"learning_rate": 1.2604521105486417e-05,
"loss": 0.6397,
"step": 6370
},
{
"epoch": 0.7443951424567959,
"grad_norm": 0.5761032360591999,
"learning_rate": 1.2571987139007856e-05,
"loss": 0.6417,
"step": 6375
},
{
"epoch": 0.744978981784213,
"grad_norm": 0.6296066629510068,
"learning_rate": 1.253950883038944e-05,
"loss": 0.6506,
"step": 6380
},
{
"epoch": 0.7455628211116301,
"grad_norm": 0.6262153811941109,
"learning_rate": 1.2507086300723846e-05,
"loss": 0.6532,
"step": 6385
},
{
"epoch": 0.7461466604390472,
"grad_norm": 0.6605485302237062,
"learning_rate": 1.2474719670895796e-05,
"loss": 0.6458,
"step": 6390
},
{
"epoch": 0.7467304997664642,
"grad_norm": 0.6302511565958615,
"learning_rate": 1.2442409061581587e-05,
"loss": 0.6262,
"step": 6395
},
{
"epoch": 0.7473143390938813,
"grad_norm": 0.6287357429748764,
"learning_rate": 1.2410154593248657e-05,
"loss": 0.6248,
"step": 6400
},
{
"epoch": 0.7478981784212985,
"grad_norm": 0.5949685378563663,
"learning_rate": 1.2377956386155114e-05,
"loss": 0.6403,
"step": 6405
},
{
"epoch": 0.7484820177487156,
"grad_norm": 0.6462039582391939,
"learning_rate": 1.2345814560349316e-05,
"loss": 0.6557,
"step": 6410
},
{
"epoch": 0.7490658570761326,
"grad_norm": 0.7239013066491163,
"learning_rate": 1.231372923566939e-05,
"loss": 0.6426,
"step": 6415
},
{
"epoch": 0.7496496964035497,
"grad_norm": 0.6795140306032631,
"learning_rate": 1.2281700531742818e-05,
"loss": 0.6493,
"step": 6420
},
{
"epoch": 0.7502335357309668,
"grad_norm": 0.6152699940480184,
"learning_rate": 1.2249728567985966e-05,
"loss": 0.6507,
"step": 6425
},
{
"epoch": 0.750817375058384,
"grad_norm": 0.6604449508028087,
"learning_rate": 1.2217813463603664e-05,
"loss": 0.6467,
"step": 6430
},
{
"epoch": 0.751401214385801,
"grad_norm": 0.5787590414167918,
"learning_rate": 1.2185955337588727e-05,
"loss": 0.6323,
"step": 6435
},
{
"epoch": 0.7519850537132181,
"grad_norm": 0.6562834320938398,
"learning_rate": 1.2154154308721546e-05,
"loss": 0.6406,
"step": 6440
},
{
"epoch": 0.7525688930406352,
"grad_norm": 0.5899101687555925,
"learning_rate": 1.2122410495569623e-05,
"loss": 0.6386,
"step": 6445
},
{
"epoch": 0.7531527323680524,
"grad_norm": 0.6351186664661537,
"learning_rate": 1.2090724016487137e-05,
"loss": 0.66,
"step": 6450
},
{
"epoch": 0.7537365716954694,
"grad_norm": 0.61481417247934,
"learning_rate": 1.2059094989614503e-05,
"loss": 0.639,
"step": 6455
},
{
"epoch": 0.7543204110228865,
"grad_norm": 0.6730266072773823,
"learning_rate": 1.2027523532877928e-05,
"loss": 0.6327,
"step": 6460
},
{
"epoch": 0.7549042503503036,
"grad_norm": 0.6479664121197497,
"learning_rate": 1.1996009763988974e-05,
"loss": 0.6297,
"step": 6465
},
{
"epoch": 0.7554880896777207,
"grad_norm": 0.6349758804122987,
"learning_rate": 1.1964553800444123e-05,
"loss": 0.6459,
"step": 6470
},
{
"epoch": 0.7560719290051378,
"grad_norm": 0.605193746860756,
"learning_rate": 1.1933155759524332e-05,
"loss": 0.6374,
"step": 6475
},
{
"epoch": 0.7566557683325549,
"grad_norm": 0.6774832969822996,
"learning_rate": 1.1901815758294589e-05,
"loss": 0.6334,
"step": 6480
},
{
"epoch": 0.757239607659972,
"grad_norm": 0.672744571091384,
"learning_rate": 1.18705339136035e-05,
"loss": 0.6567,
"step": 6485
},
{
"epoch": 0.757823446987389,
"grad_norm": 0.629538825683775,
"learning_rate": 1.1839310342082835e-05,
"loss": 0.6678,
"step": 6490
},
{
"epoch": 0.7584072863148061,
"grad_norm": 0.613775052214992,
"learning_rate": 1.1808145160147092e-05,
"loss": 0.6166,
"step": 6495
},
{
"epoch": 0.7589911256422233,
"grad_norm": 0.5753134401277333,
"learning_rate": 1.1777038483993066e-05,
"loss": 0.6349,
"step": 6500
},
{
"epoch": 0.7595749649696404,
"grad_norm": 0.6226123133457339,
"learning_rate": 1.1745990429599439e-05,
"loss": 0.6567,
"step": 6505
},
{
"epoch": 0.7601588042970574,
"grad_norm": 0.6287406461309847,
"learning_rate": 1.1715001112726304e-05,
"loss": 0.6086,
"step": 6510
},
{
"epoch": 0.7607426436244745,
"grad_norm": 0.6065648011677031,
"learning_rate": 1.1684070648914763e-05,
"loss": 0.65,
"step": 6515
},
{
"epoch": 0.7613264829518916,
"grad_norm": 0.5720273238666315,
"learning_rate": 1.1653199153486488e-05,
"loss": 0.6336,
"step": 6520
},
{
"epoch": 0.7619103222793088,
"grad_norm": 0.6588805705393938,
"learning_rate": 1.1622386741543295e-05,
"loss": 0.6333,
"step": 6525
},
{
"epoch": 0.7624941616067258,
"grad_norm": 0.6299712032540549,
"learning_rate": 1.1591633527966713e-05,
"loss": 0.6455,
"step": 6530
},
{
"epoch": 0.7630780009341429,
"grad_norm": 0.6220052731578186,
"learning_rate": 1.1560939627417555e-05,
"loss": 0.6504,
"step": 6535
},
{
"epoch": 0.76366184026156,
"grad_norm": 0.6343636477004547,
"learning_rate": 1.1530305154335482e-05,
"loss": 0.6319,
"step": 6540
},
{
"epoch": 0.7642456795889772,
"grad_norm": 0.6165498750966297,
"learning_rate": 1.1499730222938595e-05,
"loss": 0.6418,
"step": 6545
},
{
"epoch": 0.7648295189163942,
"grad_norm": 0.598828343635059,
"learning_rate": 1.1469214947222993e-05,
"loss": 0.6348,
"step": 6550
},
{
"epoch": 0.7654133582438113,
"grad_norm": 0.6161720451502942,
"learning_rate": 1.1438759440962353e-05,
"loss": 0.6231,
"step": 6555
},
{
"epoch": 0.7659971975712284,
"grad_norm": 0.5912844516303455,
"learning_rate": 1.1408363817707523e-05,
"loss": 0.6439,
"step": 6560
},
{
"epoch": 0.7665810368986455,
"grad_norm": 0.686397942467021,
"learning_rate": 1.1378028190786045e-05,
"loss": 0.632,
"step": 6565
},
{
"epoch": 0.7671648762260626,
"grad_norm": 0.598527869533545,
"learning_rate": 1.134775267330181e-05,
"loss": 0.621,
"step": 6570
},
{
"epoch": 0.7677487155534797,
"grad_norm": 0.6459305188029775,
"learning_rate": 1.1317537378134568e-05,
"loss": 0.6497,
"step": 6575
},
{
"epoch": 0.7683325548808968,
"grad_norm": 0.5997827134823882,
"learning_rate": 1.1287382417939555e-05,
"loss": 0.6491,
"step": 6580
},
{
"epoch": 0.7689163942083139,
"grad_norm": 0.63181338212674,
"learning_rate": 1.1257287905147035e-05,
"loss": 0.6316,
"step": 6585
},
{
"epoch": 0.7695002335357309,
"grad_norm": 0.5691139530472041,
"learning_rate": 1.1227253951961911e-05,
"loss": 0.6194,
"step": 6590
},
{
"epoch": 0.7700840728631481,
"grad_norm": 0.571434056270116,
"learning_rate": 1.1197280670363297e-05,
"loss": 0.6367,
"step": 6595
},
{
"epoch": 0.7706679121905652,
"grad_norm": 0.5876708801749873,
"learning_rate": 1.1167368172104084e-05,
"loss": 0.6393,
"step": 6600
},
{
"epoch": 0.7712517515179822,
"grad_norm": 0.6010005050901622,
"learning_rate": 1.1137516568710548e-05,
"loss": 0.6448,
"step": 6605
},
{
"epoch": 0.7718355908453993,
"grad_norm": 0.580336564305556,
"learning_rate": 1.1107725971481923e-05,
"loss": 0.638,
"step": 6610
},
{
"epoch": 0.7724194301728164,
"grad_norm": 0.5690526195490969,
"learning_rate": 1.107799649148998e-05,
"loss": 0.6273,
"step": 6615
},
{
"epoch": 0.7730032695002336,
"grad_norm": 0.621848022577463,
"learning_rate": 1.1048328239578631e-05,
"loss": 0.6396,
"step": 6620
},
{
"epoch": 0.7735871088276506,
"grad_norm": 0.590747044398145,
"learning_rate": 1.1018721326363493e-05,
"loss": 0.6278,
"step": 6625
},
{
"epoch": 0.7741709481550677,
"grad_norm": 0.632179461583809,
"learning_rate": 1.0989175862231488e-05,
"loss": 0.6501,
"step": 6630
},
{
"epoch": 0.7747547874824848,
"grad_norm": 0.5798177631629294,
"learning_rate": 1.095969195734044e-05,
"loss": 0.6519,
"step": 6635
},
{
"epoch": 0.775338626809902,
"grad_norm": 0.5641917730026553,
"learning_rate": 1.0930269721618641e-05,
"loss": 0.6107,
"step": 6640
},
{
"epoch": 0.775922466137319,
"grad_norm": 0.593753176024571,
"learning_rate": 1.0900909264764463e-05,
"loss": 0.6262,
"step": 6645
},
{
"epoch": 0.7765063054647361,
"grad_norm": 0.6176877432973589,
"learning_rate": 1.0871610696245941e-05,
"loss": 0.6575,
"step": 6650
},
{
"epoch": 0.7770901447921532,
"grad_norm": 0.6220895199951411,
"learning_rate": 1.0842374125300364e-05,
"loss": 0.647,
"step": 6655
},
{
"epoch": 0.7776739841195703,
"grad_norm": 0.6514844978329911,
"learning_rate": 1.081319966093386e-05,
"loss": 0.6653,
"step": 6660
},
{
"epoch": 0.7782578234469874,
"grad_norm": 0.6229635390512143,
"learning_rate": 1.0784087411921e-05,
"loss": 0.6207,
"step": 6665
},
{
"epoch": 0.7788416627744045,
"grad_norm": 0.5906786528297869,
"learning_rate": 1.0755037486804411e-05,
"loss": 0.645,
"step": 6670
},
{
"epoch": 0.7794255021018216,
"grad_norm": 0.5790238518612034,
"learning_rate": 1.0726049993894324e-05,
"loss": 0.6283,
"step": 6675
},
{
"epoch": 0.7800093414292387,
"grad_norm": 0.6137457769476979,
"learning_rate": 1.0697125041268207e-05,
"loss": 0.6278,
"step": 6680
},
{
"epoch": 0.7805931807566557,
"grad_norm": 0.6070225802535901,
"learning_rate": 1.0668262736770356e-05,
"loss": 0.6306,
"step": 6685
},
{
"epoch": 0.7811770200840729,
"grad_norm": 0.594752473424898,
"learning_rate": 1.0639463188011476e-05,
"loss": 0.654,
"step": 6690
},
{
"epoch": 0.78176085941149,
"grad_norm": 0.6022443205432915,
"learning_rate": 1.0610726502368303e-05,
"loss": 0.6491,
"step": 6695
},
{
"epoch": 0.7823446987389071,
"grad_norm": 0.6273411095037073,
"learning_rate": 1.0582052786983194e-05,
"loss": 0.6408,
"step": 6700
},
{
"epoch": 0.7829285380663241,
"grad_norm": 0.599000533291735,
"learning_rate": 1.0553442148763725e-05,
"loss": 0.6626,
"step": 6705
},
{
"epoch": 0.7835123773937412,
"grad_norm": 0.6393108507284879,
"learning_rate": 1.0524894694382284e-05,
"loss": 0.6339,
"step": 6710
},
{
"epoch": 0.7840962167211584,
"grad_norm": 0.6143401990401224,
"learning_rate": 1.0496410530275694e-05,
"loss": 0.638,
"step": 6715
},
{
"epoch": 0.7846800560485754,
"grad_norm": 0.6058792735537258,
"learning_rate": 1.0467989762644803e-05,
"loss": 0.6271,
"step": 6720
},
{
"epoch": 0.7852638953759925,
"grad_norm": 0.6178694094145576,
"learning_rate": 1.0439632497454093e-05,
"loss": 0.6419,
"step": 6725
},
{
"epoch": 0.7858477347034096,
"grad_norm": 0.5953596115650422,
"learning_rate": 1.0411338840431278e-05,
"loss": 0.6402,
"step": 6730
},
{
"epoch": 0.7864315740308268,
"grad_norm": 0.5971529371806052,
"learning_rate": 1.0383108897066915e-05,
"loss": 0.6399,
"step": 6735
},
{
"epoch": 0.7870154133582438,
"grad_norm": 0.5995025251220168,
"learning_rate": 1.035494277261401e-05,
"loss": 0.6103,
"step": 6740
},
{
"epoch": 0.7875992526856609,
"grad_norm": 0.6046302699367826,
"learning_rate": 1.0326840572087633e-05,
"loss": 0.6261,
"step": 6745
},
{
"epoch": 0.788183092013078,
"grad_norm": 0.5778379971801313,
"learning_rate": 1.0298802400264502e-05,
"loss": 0.6292,
"step": 6750
},
{
"epoch": 0.7887669313404951,
"grad_norm": 0.6248841480809375,
"learning_rate": 1.0270828361682628e-05,
"loss": 0.6377,
"step": 6755
},
{
"epoch": 0.7893507706679121,
"grad_norm": 0.6371232840447588,
"learning_rate": 1.0242918560640893e-05,
"loss": 0.6161,
"step": 6760
},
{
"epoch": 0.7899346099953293,
"grad_norm": 0.6497075961211516,
"learning_rate": 1.0215073101198683e-05,
"loss": 0.6357,
"step": 6765
},
{
"epoch": 0.7905184493227464,
"grad_norm": 0.5909439271188935,
"learning_rate": 1.0187292087175485e-05,
"loss": 0.6236,
"step": 6770
},
{
"epoch": 0.7911022886501635,
"grad_norm": 0.5976434941734019,
"learning_rate": 1.0159575622150513e-05,
"loss": 0.6398,
"step": 6775
},
{
"epoch": 0.7916861279775805,
"grad_norm": 0.6201574893269625,
"learning_rate": 1.0131923809462313e-05,
"loss": 0.6474,
"step": 6780
},
{
"epoch": 0.7922699673049977,
"grad_norm": 0.6428020951178163,
"learning_rate": 1.0104336752208374e-05,
"loss": 0.6187,
"step": 6785
},
{
"epoch": 0.7928538066324148,
"grad_norm": 0.634116819275743,
"learning_rate": 1.0076814553244762e-05,
"loss": 0.6451,
"step": 6790
},
{
"epoch": 0.7934376459598319,
"grad_norm": 0.5773391894733362,
"learning_rate": 1.0049357315185711e-05,
"loss": 0.6355,
"step": 6795
},
{
"epoch": 0.7940214852872489,
"grad_norm": 0.5920164601744748,
"learning_rate": 1.0021965140403267e-05,
"loss": 0.6108,
"step": 6800
},
{
"epoch": 0.794605324614666,
"grad_norm": 0.6404362819261163,
"learning_rate": 9.99463813102688e-06,
"loss": 0.6211,
"step": 6805
},
{
"epoch": 0.7951891639420832,
"grad_norm": 0.6327289926377916,
"learning_rate": 9.967376388943042e-06,
"loss": 0.641,
"step": 6810
},
{
"epoch": 0.7957730032695002,
"grad_norm": 0.597921078351381,
"learning_rate": 9.940180015794908e-06,
"loss": 0.6566,
"step": 6815
},
{
"epoch": 0.7963568425969173,
"grad_norm": 0.6021040015646795,
"learning_rate": 9.913049112981897e-06,
"loss": 0.6487,
"step": 6820
},
{
"epoch": 0.7969406819243344,
"grad_norm": 0.5973580692426544,
"learning_rate": 9.885983781659332e-06,
"loss": 0.6548,
"step": 6825
},
{
"epoch": 0.7975245212517516,
"grad_norm": 0.6334946405203701,
"learning_rate": 9.858984122738072e-06,
"loss": 0.6262,
"step": 6830
},
{
"epoch": 0.7981083605791686,
"grad_norm": 0.6491494108765251,
"learning_rate": 9.832050236884102e-06,
"loss": 0.6485,
"step": 6835
},
{
"epoch": 0.7986921999065857,
"grad_norm": 0.6023997514406285,
"learning_rate": 9.805182224518186e-06,
"loss": 0.6404,
"step": 6840
},
{
"epoch": 0.7992760392340028,
"grad_norm": 0.572125578630417,
"learning_rate": 9.778380185815486e-06,
"loss": 0.6369,
"step": 6845
},
{
"epoch": 0.7998598785614199,
"grad_norm": 0.6308711498259555,
"learning_rate": 9.751644220705187e-06,
"loss": 0.5927,
"step": 6850
},
{
"epoch": 0.800443717888837,
"grad_norm": 0.6173581391316637,
"learning_rate": 9.72497442887012e-06,
"loss": 0.6579,
"step": 6855
},
{
"epoch": 0.8010275572162541,
"grad_norm": 0.6263463329181845,
"learning_rate": 9.698370909746387e-06,
"loss": 0.6302,
"step": 6860
},
{
"epoch": 0.8016113965436712,
"grad_norm": 0.5951782783655484,
"learning_rate": 9.671833762523016e-06,
"loss": 0.6413,
"step": 6865
},
{
"epoch": 0.8021952358710883,
"grad_norm": 0.630199983460248,
"learning_rate": 9.645363086141561e-06,
"loss": 0.6676,
"step": 6870
},
{
"epoch": 0.8027790751985053,
"grad_norm": 0.6054299149277153,
"learning_rate": 9.618958979295747e-06,
"loss": 0.6487,
"step": 6875
},
{
"epoch": 0.8033629145259225,
"grad_norm": 0.5994022645181672,
"learning_rate": 9.592621540431101e-06,
"loss": 0.6318,
"step": 6880
},
{
"epoch": 0.8039467538533396,
"grad_norm": 0.5837832389369707,
"learning_rate": 9.566350867744584e-06,
"loss": 0.6481,
"step": 6885
},
{
"epoch": 0.8045305931807567,
"grad_norm": 0.5698387763091648,
"learning_rate": 9.540147059184226e-06,
"loss": 0.6293,
"step": 6890
},
{
"epoch": 0.8051144325081737,
"grad_norm": 0.6984312056126706,
"learning_rate": 9.514010212448751e-06,
"loss": 0.6319,
"step": 6895
},
{
"epoch": 0.8056982718355908,
"grad_norm": 0.5740328225103342,
"learning_rate": 9.487940424987235e-06,
"loss": 0.6401,
"step": 6900
},
{
"epoch": 0.806282111163008,
"grad_norm": 0.5775510495309711,
"learning_rate": 9.461937793998723e-06,
"loss": 0.6547,
"step": 6905
},
{
"epoch": 0.8068659504904251,
"grad_norm": 0.6278805432704445,
"learning_rate": 9.436002416431868e-06,
"loss": 0.6298,
"step": 6910
},
{
"epoch": 0.8074497898178421,
"grad_norm": 0.5999967276015522,
"learning_rate": 9.41013438898458e-06,
"loss": 0.631,
"step": 6915
},
{
"epoch": 0.8080336291452592,
"grad_norm": 0.6286509812121734,
"learning_rate": 9.384333808103656e-06,
"loss": 0.6309,
"step": 6920
},
{
"epoch": 0.8086174684726763,
"grad_norm": 0.6488958189176001,
"learning_rate": 9.358600769984432e-06,
"loss": 0.6347,
"step": 6925
},
{
"epoch": 0.8092013078000934,
"grad_norm": 0.5968006420157496,
"learning_rate": 9.332935370570402e-06,
"loss": 0.6491,
"step": 6930
},
{
"epoch": 0.8097851471275105,
"grad_norm": 0.6014926523303887,
"learning_rate": 9.30733770555289e-06,
"loss": 0.6388,
"step": 6935
},
{
"epoch": 0.8103689864549276,
"grad_norm": 0.6304655837034348,
"learning_rate": 9.281807870370666e-06,
"loss": 0.6436,
"step": 6940
},
{
"epoch": 0.8109528257823447,
"grad_norm": 0.6149766049986034,
"learning_rate": 9.256345960209608e-06,
"loss": 0.6251,
"step": 6945
},
{
"epoch": 0.8115366651097617,
"grad_norm": 0.6681829744274775,
"learning_rate": 9.23095207000234e-06,
"loss": 0.6732,
"step": 6950
},
{
"epoch": 0.8121205044371789,
"grad_norm": 0.6403388685978876,
"learning_rate": 9.205626294427885e-06,
"loss": 0.6287,
"step": 6955
},
{
"epoch": 0.812704343764596,
"grad_norm": 0.6164760499868315,
"learning_rate": 9.18036872791129e-06,
"loss": 0.6485,
"step": 6960
},
{
"epoch": 0.8132881830920131,
"grad_norm": 0.5654692225018587,
"learning_rate": 9.155179464623312e-06,
"loss": 0.6315,
"step": 6965
},
{
"epoch": 0.8138720224194301,
"grad_norm": 0.6235874243866195,
"learning_rate": 9.130058598480027e-06,
"loss": 0.6297,
"step": 6970
},
{
"epoch": 0.8144558617468473,
"grad_norm": 0.611311993620239,
"learning_rate": 9.105006223142507e-06,
"loss": 0.6285,
"step": 6975
},
{
"epoch": 0.8150397010742644,
"grad_norm": 0.6060839143598995,
"learning_rate": 9.080022432016457e-06,
"loss": 0.6264,
"step": 6980
},
{
"epoch": 0.8156235404016815,
"grad_norm": 0.5749876598916189,
"learning_rate": 9.05510731825188e-06,
"loss": 0.6444,
"step": 6985
},
{
"epoch": 0.8162073797290985,
"grad_norm": 0.6281041451457502,
"learning_rate": 9.030260974742701e-06,
"loss": 0.6455,
"step": 6990
},
{
"epoch": 0.8167912190565156,
"grad_norm": 0.6134476446305431,
"learning_rate": 9.005483494126474e-06,
"loss": 0.6373,
"step": 6995
},
{
"epoch": 0.8173750583839328,
"grad_norm": 0.5916584563540126,
"learning_rate": 8.980774968783978e-06,
"loss": 0.6483,
"step": 7000
},
{
"epoch": 0.8179588977113499,
"grad_norm": 0.6259270613286148,
"learning_rate": 8.9561354908389e-06,
"loss": 0.647,
"step": 7005
},
{
"epoch": 0.8185427370387669,
"grad_norm": 0.6689331791172888,
"learning_rate": 8.931565152157492e-06,
"loss": 0.6499,
"step": 7010
},
{
"epoch": 0.819126576366184,
"grad_norm": 0.6681257239383737,
"learning_rate": 8.907064044348232e-06,
"loss": 0.6447,
"step": 7015
},
{
"epoch": 0.8197104156936011,
"grad_norm": 0.6250826494469736,
"learning_rate": 8.88263225876147e-06,
"loss": 0.6034,
"step": 7020
},
{
"epoch": 0.8202942550210183,
"grad_norm": 0.5711401947607472,
"learning_rate": 8.858269886489099e-06,
"loss": 0.6114,
"step": 7025
},
{
"epoch": 0.8208780943484353,
"grad_norm": 0.5985615408353293,
"learning_rate": 8.8339770183642e-06,
"loss": 0.64,
"step": 7030
},
{
"epoch": 0.8214619336758524,
"grad_norm": 0.6006406679048145,
"learning_rate": 8.809753744960733e-06,
"loss": 0.646,
"step": 7035
},
{
"epoch": 0.8220457730032695,
"grad_norm": 0.5811155925766809,
"learning_rate": 8.785600156593157e-06,
"loss": 0.6308,
"step": 7040
},
{
"epoch": 0.8226296123306865,
"grad_norm": 0.5833779089569627,
"learning_rate": 8.761516343316131e-06,
"loss": 0.6428,
"step": 7045
},
{
"epoch": 0.8232134516581037,
"grad_norm": 0.5919532280968446,
"learning_rate": 8.737502394924158e-06,
"loss": 0.6363,
"step": 7050
},
{
"epoch": 0.8237972909855208,
"grad_norm": 0.6034532255231587,
"learning_rate": 8.713558400951254e-06,
"loss": 0.6306,
"step": 7055
},
{
"epoch": 0.8243811303129379,
"grad_norm": 0.6278732337136065,
"learning_rate": 8.689684450670627e-06,
"loss": 0.6437,
"step": 7060
},
{
"epoch": 0.8249649696403549,
"grad_norm": 0.6265141928072414,
"learning_rate": 8.665880633094314e-06,
"loss": 0.6132,
"step": 7065
},
{
"epoch": 0.825548808967772,
"grad_norm": 0.5510714562301102,
"learning_rate": 8.642147036972887e-06,
"loss": 0.6322,
"step": 7070
},
{
"epoch": 0.8261326482951892,
"grad_norm": 0.5990688750904606,
"learning_rate": 8.618483750795087e-06,
"loss": 0.6426,
"step": 7075
},
{
"epoch": 0.8267164876226063,
"grad_norm": 0.58078220118751,
"learning_rate": 8.594890862787518e-06,
"loss": 0.622,
"step": 7080
},
{
"epoch": 0.8273003269500233,
"grad_norm": 0.6009559251228864,
"learning_rate": 8.571368460914316e-06,
"loss": 0.6304,
"step": 7085
},
{
"epoch": 0.8278841662774404,
"grad_norm": 0.5966997292141136,
"learning_rate": 8.547916632876806e-06,
"loss": 0.6416,
"step": 7090
},
{
"epoch": 0.8284680056048576,
"grad_norm": 0.5985492978905353,
"learning_rate": 8.524535466113185e-06,
"loss": 0.6278,
"step": 7095
},
{
"epoch": 0.8290518449322747,
"grad_norm": 0.6119607849123736,
"learning_rate": 8.5012250477982e-06,
"loss": 0.6324,
"step": 7100
},
{
"epoch": 0.8296356842596917,
"grad_norm": 0.582057910730287,
"learning_rate": 8.477985464842816e-06,
"loss": 0.6347,
"step": 7105
},
{
"epoch": 0.8302195235871088,
"grad_norm": 0.6407160513095409,
"learning_rate": 8.454816803893893e-06,
"loss": 0.6361,
"step": 7110
},
{
"epoch": 0.830803362914526,
"grad_norm": 0.6273654016841408,
"learning_rate": 8.431719151333864e-06,
"loss": 0.6234,
"step": 7115
},
{
"epoch": 0.8313872022419431,
"grad_norm": 0.603508156859576,
"learning_rate": 8.40869259328042e-06,
"loss": 0.643,
"step": 7120
},
{
"epoch": 0.8319710415693601,
"grad_norm": 0.6185148314476354,
"learning_rate": 8.385737215586171e-06,
"loss": 0.6447,
"step": 7125
},
{
"epoch": 0.8325548808967772,
"grad_norm": 0.5771583068818322,
"learning_rate": 8.362853103838344e-06,
"loss": 0.6246,
"step": 7130
},
{
"epoch": 0.8331387202241943,
"grad_norm": 0.6249572682049243,
"learning_rate": 8.340040343358455e-06,
"loss": 0.6102,
"step": 7135
},
{
"epoch": 0.8337225595516113,
"grad_norm": 0.6419451327422312,
"learning_rate": 8.317299019201996e-06,
"loss": 0.6372,
"step": 7140
},
{
"epoch": 0.8343063988790285,
"grad_norm": 0.5841961031320476,
"learning_rate": 8.294629216158107e-06,
"loss": 0.6373,
"step": 7145
},
{
"epoch": 0.8348902382064456,
"grad_norm": 0.5995927444477767,
"learning_rate": 8.272031018749272e-06,
"loss": 0.6433,
"step": 7150
},
{
"epoch": 0.8354740775338627,
"grad_norm": 0.582344410588876,
"learning_rate": 8.249504511231005e-06,
"loss": 0.6487,
"step": 7155
},
{
"epoch": 0.8360579168612797,
"grad_norm": 0.5962429914662568,
"learning_rate": 8.227049777591516e-06,
"loss": 0.6268,
"step": 7160
},
{
"epoch": 0.8366417561886969,
"grad_norm": 0.608107604021329,
"learning_rate": 8.204666901551428e-06,
"loss": 0.6457,
"step": 7165
},
{
"epoch": 0.837225595516114,
"grad_norm": 0.6058460527295055,
"learning_rate": 8.182355966563438e-06,
"loss": 0.6296,
"step": 7170
},
{
"epoch": 0.8378094348435311,
"grad_norm": 0.6280047949198018,
"learning_rate": 8.160117055812019e-06,
"loss": 0.6481,
"step": 7175
},
{
"epoch": 0.8383932741709481,
"grad_norm": 0.5930435152936081,
"learning_rate": 8.13795025221311e-06,
"loss": 0.6255,
"step": 7180
},
{
"epoch": 0.8389771134983652,
"grad_norm": 0.5929303430476128,
"learning_rate": 8.115855638413806e-06,
"loss": 0.6207,
"step": 7185
},
{
"epoch": 0.8395609528257824,
"grad_norm": 0.6302797296442603,
"learning_rate": 8.09383329679204e-06,
"loss": 0.626,
"step": 7190
},
{
"epoch": 0.8401447921531995,
"grad_norm": 0.611869681356887,
"learning_rate": 8.071883309456292e-06,
"loss": 0.6258,
"step": 7195
},
{
"epoch": 0.8407286314806165,
"grad_norm": 0.5908565394267222,
"learning_rate": 8.050005758245274e-06,
"loss": 0.6021,
"step": 7200
},
{
"epoch": 0.8413124708080336,
"grad_norm": 0.6213769971363277,
"learning_rate": 8.028200724727623e-06,
"loss": 0.6261,
"step": 7205
},
{
"epoch": 0.8418963101354507,
"grad_norm": 0.6090574028964819,
"learning_rate": 8.006468290201603e-06,
"loss": 0.6245,
"step": 7210
},
{
"epoch": 0.8424801494628679,
"grad_norm": 0.5529484666318186,
"learning_rate": 7.984808535694794e-06,
"loss": 0.633,
"step": 7215
},
{
"epoch": 0.8430639887902849,
"grad_norm": 0.6131465277863353,
"learning_rate": 7.963221541963799e-06,
"loss": 0.6421,
"step": 7220
},
{
"epoch": 0.843647828117702,
"grad_norm": 0.5705723507812558,
"learning_rate": 7.94170738949394e-06,
"loss": 0.6209,
"step": 7225
},
{
"epoch": 0.8442316674451191,
"grad_norm": 0.5870298806748856,
"learning_rate": 7.920266158498948e-06,
"loss": 0.6382,
"step": 7230
},
{
"epoch": 0.8448155067725363,
"grad_norm": 0.6024114225879414,
"learning_rate": 7.898897928920684e-06,
"loss": 0.6194,
"step": 7235
},
{
"epoch": 0.8453993460999533,
"grad_norm": 0.5955204114457981,
"learning_rate": 7.877602780428816e-06,
"loss": 0.6183,
"step": 7240
},
{
"epoch": 0.8459831854273704,
"grad_norm": 0.593227028079345,
"learning_rate": 7.856380792420549e-06,
"loss": 0.63,
"step": 7245
},
{
"epoch": 0.8465670247547875,
"grad_norm": 0.614407216262562,
"learning_rate": 7.835232044020304e-06,
"loss": 0.6365,
"step": 7250
},
{
"epoch": 0.8471508640822045,
"grad_norm": 0.585501938437949,
"learning_rate": 7.81415661407944e-06,
"loss": 0.6231,
"step": 7255
},
{
"epoch": 0.8477347034096216,
"grad_norm": 0.6051940564461002,
"learning_rate": 7.793154581175954e-06,
"loss": 0.6305,
"step": 7260
},
{
"epoch": 0.8483185427370388,
"grad_norm": 0.5559347387737404,
"learning_rate": 7.772226023614185e-06,
"loss": 0.6337,
"step": 7265
},
{
"epoch": 0.8489023820644559,
"grad_norm": 0.5493920577269864,
"learning_rate": 7.751371019424528e-06,
"loss": 0.6287,
"step": 7270
},
{
"epoch": 0.8494862213918729,
"grad_norm": 0.6548373342721107,
"learning_rate": 7.730589646363141e-06,
"loss": 0.6348,
"step": 7275
},
{
"epoch": 0.85007006071929,
"grad_norm": 0.570177127165855,
"learning_rate": 7.709881981911648e-06,
"loss": 0.6257,
"step": 7280
},
{
"epoch": 0.8506539000467072,
"grad_norm": 0.6276195548347564,
"learning_rate": 7.689248103276873e-06,
"loss": 0.6418,
"step": 7285
},
{
"epoch": 0.8512377393741243,
"grad_norm": 0.6332505550767229,
"learning_rate": 7.668688087390509e-06,
"loss": 0.6208,
"step": 7290
},
{
"epoch": 0.8518215787015413,
"grad_norm": 0.5723828888486714,
"learning_rate": 7.648202010908884e-06,
"loss": 0.6488,
"step": 7295
},
{
"epoch": 0.8524054180289584,
"grad_norm": 0.5789633418286964,
"learning_rate": 7.627789950212635e-06,
"loss": 0.6434,
"step": 7300
},
{
"epoch": 0.8529892573563755,
"grad_norm": 0.5841746070572975,
"learning_rate": 7.607451981406441e-06,
"loss": 0.6399,
"step": 7305
},
{
"epoch": 0.8535730966837927,
"grad_norm": 0.6242425057438395,
"learning_rate": 7.587188180318736e-06,
"loss": 0.6325,
"step": 7310
},
{
"epoch": 0.8541569360112097,
"grad_norm": 0.6265552276446478,
"learning_rate": 7.5669986225014215e-06,
"loss": 0.6133,
"step": 7315
},
{
"epoch": 0.8547407753386268,
"grad_norm": 0.5868392976953574,
"learning_rate": 7.546883383229594e-06,
"loss": 0.6293,
"step": 7320
},
{
"epoch": 0.8553246146660439,
"grad_norm": 0.5615113171449998,
"learning_rate": 7.526842537501259e-06,
"loss": 0.6184,
"step": 7325
},
{
"epoch": 0.855908453993461,
"grad_norm": 0.5993576030041161,
"learning_rate": 7.50687616003705e-06,
"loss": 0.6505,
"step": 7330
},
{
"epoch": 0.8564922933208781,
"grad_norm": 0.5867887403151126,
"learning_rate": 7.486984325279956e-06,
"loss": 0.6341,
"step": 7335
},
{
"epoch": 0.8570761326482952,
"grad_norm": 0.6051624859315319,
"learning_rate": 7.467167107395028e-06,
"loss": 0.6504,
"step": 7340
},
{
"epoch": 0.8576599719757123,
"grad_norm": 0.6037147824039196,
"learning_rate": 7.44742458026913e-06,
"loss": 0.6448,
"step": 7345
},
{
"epoch": 0.8582438113031294,
"grad_norm": 0.5946601208566309,
"learning_rate": 7.427756817510634e-06,
"loss": 0.631,
"step": 7350
},
{
"epoch": 0.8588276506305464,
"grad_norm": 0.5741556297197494,
"learning_rate": 7.408163892449172e-06,
"loss": 0.6563,
"step": 7355
},
{
"epoch": 0.8594114899579636,
"grad_norm": 0.6303180031417959,
"learning_rate": 7.388645878135338e-06,
"loss": 0.6247,
"step": 7360
},
{
"epoch": 0.8599953292853807,
"grad_norm": 0.5995656192902652,
"learning_rate": 7.369202847340432e-06,
"loss": 0.6342,
"step": 7365
},
{
"epoch": 0.8605791686127977,
"grad_norm": 0.5906249879434073,
"learning_rate": 7.349834872556187e-06,
"loss": 0.6178,
"step": 7370
},
{
"epoch": 0.8611630079402148,
"grad_norm": 0.65165057184435,
"learning_rate": 7.330542025994495e-06,
"loss": 0.6395,
"step": 7375
},
{
"epoch": 0.861746847267632,
"grad_norm": 0.5643528091874781,
"learning_rate": 7.311324379587136e-06,
"loss": 0.6236,
"step": 7380
},
{
"epoch": 0.8623306865950491,
"grad_norm": 0.5922391769747286,
"learning_rate": 7.292182004985511e-06,
"loss": 0.6236,
"step": 7385
},
{
"epoch": 0.8629145259224661,
"grad_norm": 0.5850760407995166,
"learning_rate": 7.2731149735603825e-06,
"loss": 0.6375,
"step": 7390
},
{
"epoch": 0.8634983652498832,
"grad_norm": 0.5381785704672893,
"learning_rate": 7.254123356401597e-06,
"loss": 0.6435,
"step": 7395
},
{
"epoch": 0.8640822045773003,
"grad_norm": 0.5743079078236182,
"learning_rate": 7.23520722431783e-06,
"loss": 0.6323,
"step": 7400
},
{
"epoch": 0.8646660439047175,
"grad_norm": 0.6261936371916004,
"learning_rate": 7.216366647836306e-06,
"loss": 0.6106,
"step": 7405
},
{
"epoch": 0.8652498832321345,
"grad_norm": 0.634094214371704,
"learning_rate": 7.197601697202565e-06,
"loss": 0.6514,
"step": 7410
},
{
"epoch": 0.8658337225595516,
"grad_norm": 0.6366357404750688,
"learning_rate": 7.1789124423801645e-06,
"loss": 0.6371,
"step": 7415
},
{
"epoch": 0.8664175618869687,
"grad_norm": 0.5957182423994334,
"learning_rate": 7.160298953050448e-06,
"loss": 0.619,
"step": 7420
},
{
"epoch": 0.8670014012143858,
"grad_norm": 0.6214057365825675,
"learning_rate": 7.141761298612267e-06,
"loss": 0.6493,
"step": 7425
},
{
"epoch": 0.8675852405418029,
"grad_norm": 0.6313558445784697,
"learning_rate": 7.123299548181732e-06,
"loss": 0.6465,
"step": 7430
},
{
"epoch": 0.86816907986922,
"grad_norm": 0.6421258222661969,
"learning_rate": 7.104913770591953e-06,
"loss": 0.6317,
"step": 7435
},
{
"epoch": 0.8687529191966371,
"grad_norm": 0.6111310721240305,
"learning_rate": 7.086604034392777e-06,
"loss": 0.6518,
"step": 7440
},
{
"epoch": 0.8693367585240542,
"grad_norm": 0.594650062721218,
"learning_rate": 7.068370407850541e-06,
"loss": 0.6501,
"step": 7445
},
{
"epoch": 0.8699205978514712,
"grad_norm": 0.5899132242091928,
"learning_rate": 7.050212958947813e-06,
"loss": 0.6178,
"step": 7450
},
{
"epoch": 0.8705044371788884,
"grad_norm": 0.6063821474980347,
"learning_rate": 7.032131755383134e-06,
"loss": 0.6121,
"step": 7455
},
{
"epoch": 0.8710882765063055,
"grad_norm": 0.5965686651312564,
"learning_rate": 7.014126864570782e-06,
"loss": 0.6087,
"step": 7460
},
{
"epoch": 0.8716721158337225,
"grad_norm": 0.575294738225672,
"learning_rate": 6.996198353640495e-06,
"loss": 0.632,
"step": 7465
},
{
"epoch": 0.8722559551611396,
"grad_norm": 0.6032994099485566,
"learning_rate": 6.978346289437245e-06,
"loss": 0.6377,
"step": 7470
},
{
"epoch": 0.8728397944885568,
"grad_norm": 0.5966900854298699,
"learning_rate": 6.9605707385209755e-06,
"loss": 0.6314,
"step": 7475
},
{
"epoch": 0.8734236338159739,
"grad_norm": 0.610898441392739,
"learning_rate": 6.942871767166354e-06,
"loss": 0.6166,
"step": 7480
},
{
"epoch": 0.8740074731433909,
"grad_norm": 0.5877123729177149,
"learning_rate": 6.925249441362533e-06,
"loss": 0.6445,
"step": 7485
},
{
"epoch": 0.874591312470808,
"grad_norm": 0.5783211230488218,
"learning_rate": 6.907703826812895e-06,
"loss": 0.6248,
"step": 7490
},
{
"epoch": 0.8751751517982251,
"grad_norm": 0.6179082340790674,
"learning_rate": 6.89023498893481e-06,
"loss": 0.6193,
"step": 7495
},
{
"epoch": 0.8757589911256423,
"grad_norm": 0.5730522084112291,
"learning_rate": 6.872842992859395e-06,
"loss": 0.6415,
"step": 7500
},
{
"epoch": 0.8763428304530593,
"grad_norm": 0.5982698829404914,
"learning_rate": 6.855527903431267e-06,
"loss": 0.6545,
"step": 7505
},
{
"epoch": 0.8769266697804764,
"grad_norm": 0.5961084151743405,
"learning_rate": 6.838289785208303e-06,
"loss": 0.622,
"step": 7510
},
{
"epoch": 0.8775105091078935,
"grad_norm": 0.6117282352748235,
"learning_rate": 6.821128702461401e-06,
"loss": 0.6221,
"step": 7515
},
{
"epoch": 0.8780943484353106,
"grad_norm": 0.6251541376005386,
"learning_rate": 6.804044719174235e-06,
"loss": 0.6387,
"step": 7520
},
{
"epoch": 0.8786781877627277,
"grad_norm": 0.6333406133218736,
"learning_rate": 6.787037899043027e-06,
"loss": 0.6487,
"step": 7525
},
{
"epoch": 0.8792620270901448,
"grad_norm": 0.6202248190034936,
"learning_rate": 6.770108305476293e-06,
"loss": 0.6327,
"step": 7530
},
{
"epoch": 0.8798458664175619,
"grad_norm": 0.5738748315197423,
"learning_rate": 6.753256001594622e-06,
"loss": 0.6261,
"step": 7535
},
{
"epoch": 0.880429705744979,
"grad_norm": 0.6092187481579089,
"learning_rate": 6.736481050230438e-06,
"loss": 0.649,
"step": 7540
},
{
"epoch": 0.881013545072396,
"grad_norm": 0.6620406400149058,
"learning_rate": 6.719783513927755e-06,
"loss": 0.6473,
"step": 7545
},
{
"epoch": 0.8815973843998132,
"grad_norm": 0.5329023889984801,
"learning_rate": 6.703163454941953e-06,
"loss": 0.6226,
"step": 7550
},
{
"epoch": 0.8821812237272303,
"grad_norm": 0.595989753599616,
"learning_rate": 6.686620935239552e-06,
"loss": 0.6293,
"step": 7555
},
{
"epoch": 0.8827650630546474,
"grad_norm": 0.5929933589279678,
"learning_rate": 6.670156016497958e-06,
"loss": 0.6156,
"step": 7560
},
{
"epoch": 0.8833489023820644,
"grad_norm": 0.5871775777565705,
"learning_rate": 6.653768760105268e-06,
"loss": 0.6316,
"step": 7565
},
{
"epoch": 0.8839327417094816,
"grad_norm": 0.5852189786127018,
"learning_rate": 6.637459227160004e-06,
"loss": 0.6264,
"step": 7570
},
{
"epoch": 0.8845165810368987,
"grad_norm": 0.5793680620401365,
"learning_rate": 6.621227478470911e-06,
"loss": 0.6379,
"step": 7575
},
{
"epoch": 0.8851004203643157,
"grad_norm": 0.5972498950686099,
"learning_rate": 6.605073574556721e-06,
"loss": 0.6471,
"step": 7580
},
{
"epoch": 0.8856842596917328,
"grad_norm": 0.5985472050403043,
"learning_rate": 6.588997575645929e-06,
"loss": 0.6137,
"step": 7585
},
{
"epoch": 0.8862680990191499,
"grad_norm": 0.6018567480056775,
"learning_rate": 6.572999541676563e-06,
"loss": 0.6468,
"step": 7590
},
{
"epoch": 0.8868519383465671,
"grad_norm": 0.6019158386265433,
"learning_rate": 6.557079532295968e-06,
"loss": 0.6315,
"step": 7595
},
{
"epoch": 0.8874357776739841,
"grad_norm": 0.5745280786194661,
"learning_rate": 6.541237606860582e-06,
"loss": 0.6242,
"step": 7600
},
{
"epoch": 0.8880196170014012,
"grad_norm": 0.5604599589548389,
"learning_rate": 6.525473824435714e-06,
"loss": 0.6429,
"step": 7605
},
{
"epoch": 0.8886034563288183,
"grad_norm": 0.5719551021185416,
"learning_rate": 6.5097882437953205e-06,
"loss": 0.6215,
"step": 7610
},
{
"epoch": 0.8891872956562354,
"grad_norm": 0.6037011288977054,
"learning_rate": 6.49418092342179e-06,
"loss": 0.644,
"step": 7615
},
{
"epoch": 0.8897711349836525,
"grad_norm": 0.5582874213963103,
"learning_rate": 6.478651921505727e-06,
"loss": 0.6329,
"step": 7620
},
{
"epoch": 0.8903549743110696,
"grad_norm": 0.5869464431191361,
"learning_rate": 6.463201295945727e-06,
"loss": 0.6403,
"step": 7625
},
{
"epoch": 0.8909388136384867,
"grad_norm": 0.5695738691482084,
"learning_rate": 6.447829104348171e-06,
"loss": 0.6341,
"step": 7630
},
{
"epoch": 0.8915226529659038,
"grad_norm": 0.5884434441417679,
"learning_rate": 6.432535404026997e-06,
"loss": 0.631,
"step": 7635
},
{
"epoch": 0.8921064922933208,
"grad_norm": 0.5613896008670317,
"learning_rate": 6.417320252003505e-06,
"loss": 0.6231,
"step": 7640
},
{
"epoch": 0.892690331620738,
"grad_norm": 0.5733433479092367,
"learning_rate": 6.402183705006127e-06,
"loss": 0.6247,
"step": 7645
},
{
"epoch": 0.8932741709481551,
"grad_norm": 0.5789784381483709,
"learning_rate": 6.387125819470231e-06,
"loss": 0.6275,
"step": 7650
},
{
"epoch": 0.8938580102755722,
"grad_norm": 0.5855870723039609,
"learning_rate": 6.372146651537892e-06,
"loss": 0.6334,
"step": 7655
},
{
"epoch": 0.8944418496029892,
"grad_norm": 0.5790357716517759,
"learning_rate": 6.3572462570576985e-06,
"loss": 0.6369,
"step": 7660
},
{
"epoch": 0.8950256889304063,
"grad_norm": 0.5917387925639677,
"learning_rate": 6.3424246915845395e-06,
"loss": 0.65,
"step": 7665
},
{
"epoch": 0.8956095282578235,
"grad_norm": 0.6005126599817849,
"learning_rate": 6.327682010379392e-06,
"loss": 0.6315,
"step": 7670
},
{
"epoch": 0.8961933675852406,
"grad_norm": 0.6156495229109331,
"learning_rate": 6.313018268409122e-06,
"loss": 0.612,
"step": 7675
},
{
"epoch": 0.8967772069126576,
"grad_norm": 0.5874738604549495,
"learning_rate": 6.2984335203462825e-06,
"loss": 0.6362,
"step": 7680
},
{
"epoch": 0.8973610462400747,
"grad_norm": 0.5635575151716498,
"learning_rate": 6.283927820568894e-06,
"loss": 0.6152,
"step": 7685
},
{
"epoch": 0.8979448855674919,
"grad_norm": 0.6131277247350563,
"learning_rate": 6.269501223160259e-06,
"loss": 0.6304,
"step": 7690
},
{
"epoch": 0.8985287248949089,
"grad_norm": 0.5611246976662609,
"learning_rate": 6.255153781908754e-06,
"loss": 0.635,
"step": 7695
},
{
"epoch": 0.899112564222326,
"grad_norm": 0.5763254004910536,
"learning_rate": 6.240885550307624e-06,
"loss": 0.6119,
"step": 7700
},
{
"epoch": 0.8996964035497431,
"grad_norm": 0.5986985930592617,
"learning_rate": 6.2266965815547865e-06,
"loss": 0.6241,
"step": 7705
},
{
"epoch": 0.9002802428771602,
"grad_norm": 0.6011640913293429,
"learning_rate": 6.212586928552641e-06,
"loss": 0.6156,
"step": 7710
},
{
"epoch": 0.9008640822045773,
"grad_norm": 0.5873329366206337,
"learning_rate": 6.19855664390786e-06,
"loss": 0.6262,
"step": 7715
},
{
"epoch": 0.9014479215319944,
"grad_norm": 0.5922643236306615,
"learning_rate": 6.184605779931197e-06,
"loss": 0.6424,
"step": 7720
},
{
"epoch": 0.9020317608594115,
"grad_norm": 0.5965372758227344,
"learning_rate": 6.170734388637294e-06,
"loss": 0.6339,
"step": 7725
},
{
"epoch": 0.9026156001868286,
"grad_norm": 0.5789469748793128,
"learning_rate": 6.156942521744484e-06,
"loss": 0.6186,
"step": 7730
},
{
"epoch": 0.9031994395142456,
"grad_norm": 0.5845062220347781,
"learning_rate": 6.143230230674602e-06,
"loss": 0.6282,
"step": 7735
},
{
"epoch": 0.9037832788416628,
"grad_norm": 0.5871311104639835,
"learning_rate": 6.12959756655279e-06,
"loss": 0.6268,
"step": 7740
},
{
"epoch": 0.9043671181690799,
"grad_norm": 0.6144246132041297,
"learning_rate": 6.11604458020731e-06,
"loss": 0.6361,
"step": 7745
},
{
"epoch": 0.904950957496497,
"grad_norm": 0.5632685742171856,
"learning_rate": 6.102571322169347e-06,
"loss": 0.6351,
"step": 7750
},
{
"epoch": 0.905534796823914,
"grad_norm": 0.6125320826036844,
"learning_rate": 6.089177842672826e-06,
"loss": 0.6469,
"step": 7755
},
{
"epoch": 0.9061186361513311,
"grad_norm": 0.604521634652814,
"learning_rate": 6.075864191654231e-06,
"loss": 0.6269,
"step": 7760
},
{
"epoch": 0.9067024754787483,
"grad_norm": 0.575565398186447,
"learning_rate": 6.062630418752404e-06,
"loss": 0.621,
"step": 7765
},
{
"epoch": 0.9072863148061654,
"grad_norm": 0.60593584998349,
"learning_rate": 6.049476573308375e-06,
"loss": 0.6472,
"step": 7770
},
{
"epoch": 0.9078701541335824,
"grad_norm": 0.5555915365416894,
"learning_rate": 6.036402704365168e-06,
"loss": 0.6239,
"step": 7775
},
{
"epoch": 0.9084539934609995,
"grad_norm": 0.6061859054502865,
"learning_rate": 6.023408860667617e-06,
"loss": 0.6447,
"step": 7780
},
{
"epoch": 0.9090378327884167,
"grad_norm": 0.5939193378003372,
"learning_rate": 6.010495090662197e-06,
"loss": 0.6209,
"step": 7785
},
{
"epoch": 0.9096216721158337,
"grad_norm": 0.5992459593790246,
"learning_rate": 5.9976614424968245e-06,
"loss": 0.6231,
"step": 7790
},
{
"epoch": 0.9102055114432508,
"grad_norm": 0.5926513959690488,
"learning_rate": 5.9849079640207e-06,
"loss": 0.6281,
"step": 7795
},
{
"epoch": 0.9107893507706679,
"grad_norm": 0.6549532492152256,
"learning_rate": 5.972234702784106e-06,
"loss": 0.6392,
"step": 7800
},
{
"epoch": 0.911373190098085,
"grad_norm": 0.5997057009691191,
"learning_rate": 5.9596417060382545e-06,
"loss": 0.6308,
"step": 7805
},
{
"epoch": 0.911957029425502,
"grad_norm": 0.5956499191366402,
"learning_rate": 5.9471290207350925e-06,
"loss": 0.6122,
"step": 7810
},
{
"epoch": 0.9125408687529192,
"grad_norm": 0.6003046930206634,
"learning_rate": 5.934696693527123e-06,
"loss": 0.6326,
"step": 7815
},
{
"epoch": 0.9131247080803363,
"grad_norm": 0.5678487478127593,
"learning_rate": 5.9223447707672564e-06,
"loss": 0.615,
"step": 7820
},
{
"epoch": 0.9137085474077534,
"grad_norm": 0.6110308127531793,
"learning_rate": 5.910073298508609e-06,
"loss": 0.6424,
"step": 7825
},
{
"epoch": 0.9142923867351704,
"grad_norm": 0.5807846331837968,
"learning_rate": 5.8978823225043555e-06,
"loss": 0.6242,
"step": 7830
},
{
"epoch": 0.9148762260625876,
"grad_norm": 0.5947050905634069,
"learning_rate": 5.8857718882075325e-06,
"loss": 0.6456,
"step": 7835
},
{
"epoch": 0.9154600653900047,
"grad_norm": 0.5643509149559328,
"learning_rate": 5.8737420407708985e-06,
"loss": 0.6282,
"step": 7840
},
{
"epoch": 0.9160439047174218,
"grad_norm": 0.5853474055836014,
"learning_rate": 5.861792825046739e-06,
"loss": 0.6247,
"step": 7845
},
{
"epoch": 0.9166277440448388,
"grad_norm": 0.5676070256446623,
"learning_rate": 5.849924285586719e-06,
"loss": 0.6382,
"step": 7850
},
{
"epoch": 0.917211583372256,
"grad_norm": 0.5920050545049914,
"learning_rate": 5.838136466641704e-06,
"loss": 0.6282,
"step": 7855
},
{
"epoch": 0.9177954226996731,
"grad_norm": 0.587809192772673,
"learning_rate": 5.8264294121616e-06,
"loss": 0.6266,
"step": 7860
},
{
"epoch": 0.9183792620270902,
"grad_norm": 0.6148621426813374,
"learning_rate": 5.814803165795194e-06,
"loss": 0.6331,
"step": 7865
},
{
"epoch": 0.9189631013545072,
"grad_norm": 0.5876900466381477,
"learning_rate": 5.803257770889978e-06,
"loss": 0.6398,
"step": 7870
},
{
"epoch": 0.9195469406819243,
"grad_norm": 0.5939774769190836,
"learning_rate": 5.791793270492006e-06,
"loss": 0.6322,
"step": 7875
},
{
"epoch": 0.9201307800093415,
"grad_norm": 0.5789466055435857,
"learning_rate": 5.780409707345714e-06,
"loss": 0.6463,
"step": 7880
},
{
"epoch": 0.9207146193367586,
"grad_norm": 0.5609575621697309,
"learning_rate": 5.769107123893781e-06,
"loss": 0.6173,
"step": 7885
},
{
"epoch": 0.9212984586641756,
"grad_norm": 0.5577225899589808,
"learning_rate": 5.757885562276948e-06,
"loss": 0.6214,
"step": 7890
},
{
"epoch": 0.9218822979915927,
"grad_norm": 0.5898473069884559,
"learning_rate": 5.7467450643338804e-06,
"loss": 0.6285,
"step": 7895
},
{
"epoch": 0.9224661373190098,
"grad_norm": 0.5721961907102839,
"learning_rate": 5.7356856716010014e-06,
"loss": 0.6247,
"step": 7900
},
{
"epoch": 0.9230499766464269,
"grad_norm": 0.6238808048027492,
"learning_rate": 5.724707425312344e-06,
"loss": 0.6246,
"step": 7905
},
{
"epoch": 0.923633815973844,
"grad_norm": 0.5791143715327993,
"learning_rate": 5.7138103663993895e-06,
"loss": 0.6254,
"step": 7910
},
{
"epoch": 0.9242176553012611,
"grad_norm": 0.5893152674667929,
"learning_rate": 5.70299453549092e-06,
"loss": 0.632,
"step": 7915
},
{
"epoch": 0.9248014946286782,
"grad_norm": 0.6070842874428386,
"learning_rate": 5.692259972912865e-06,
"loss": 0.6251,
"step": 7920
},
{
"epoch": 0.9253853339560952,
"grad_norm": 0.5962820894650079,
"learning_rate": 5.681606718688152e-06,
"loss": 0.6275,
"step": 7925
},
{
"epoch": 0.9259691732835124,
"grad_norm": 0.5553459724718108,
"learning_rate": 5.671034812536561e-06,
"loss": 0.5999,
"step": 7930
},
{
"epoch": 0.9265530126109295,
"grad_norm": 0.613232503868812,
"learning_rate": 5.660544293874561e-06,
"loss": 0.6133,
"step": 7935
},
{
"epoch": 0.9271368519383466,
"grad_norm": 0.5801623987095087,
"learning_rate": 5.65013520181519e-06,
"loss": 0.6213,
"step": 7940
},
{
"epoch": 0.9277206912657636,
"grad_norm": 0.5531859501490597,
"learning_rate": 5.639807575167886e-06,
"loss": 0.6217,
"step": 7945
},
{
"epoch": 0.9283045305931807,
"grad_norm": 0.5843472632939827,
"learning_rate": 5.6295614524383436e-06,
"loss": 0.6238,
"step": 7950
},
{
"epoch": 0.9288883699205979,
"grad_norm": 0.5953076232672708,
"learning_rate": 5.619396871828387e-06,
"loss": 0.6281,
"step": 7955
},
{
"epoch": 0.929472209248015,
"grad_norm": 0.6537058483917418,
"learning_rate": 5.6093138712358155e-06,
"loss": 0.6321,
"step": 7960
},
{
"epoch": 0.930056048575432,
"grad_norm": 0.6050956625859224,
"learning_rate": 5.5993124882542584e-06,
"loss": 0.633,
"step": 7965
},
{
"epoch": 0.9306398879028491,
"grad_norm": 0.5938596525123857,
"learning_rate": 5.589392760173047e-06,
"loss": 0.6336,
"step": 7970
},
{
"epoch": 0.9312237272302663,
"grad_norm": 0.5808287306501443,
"learning_rate": 5.579554723977065e-06,
"loss": 0.6205,
"step": 7975
},
{
"epoch": 0.9318075665576834,
"grad_norm": 0.5741959017637736,
"learning_rate": 5.569798416346615e-06,
"loss": 0.6271,
"step": 7980
},
{
"epoch": 0.9323914058851004,
"grad_norm": 0.620961853533032,
"learning_rate": 5.560123873657284e-06,
"loss": 0.6217,
"step": 7985
},
{
"epoch": 0.9329752452125175,
"grad_norm": 0.5735575986272371,
"learning_rate": 5.550531131979804e-06,
"loss": 0.6314,
"step": 7990
},
{
"epoch": 0.9335590845399346,
"grad_norm": 0.57393459936111,
"learning_rate": 5.5410202270799165e-06,
"loss": 0.6327,
"step": 7995
},
{
"epoch": 0.9341429238673518,
"grad_norm": 0.5887604120106468,
"learning_rate": 5.531591194418244e-06,
"loss": 0.6235,
"step": 8000
},
{
"epoch": 0.9347267631947688,
"grad_norm": 0.5625514258585821,
"learning_rate": 5.5222440691501534e-06,
"loss": 0.623,
"step": 8005
},
{
"epoch": 0.9353106025221859,
"grad_norm": 0.579437947711775,
"learning_rate": 5.512978886125628e-06,
"loss": 0.6078,
"step": 8010
},
{
"epoch": 0.935894441849603,
"grad_norm": 0.5746479892458379,
"learning_rate": 5.5037956798891345e-06,
"loss": 0.626,
"step": 8015
},
{
"epoch": 0.93647828117702,
"grad_norm": 0.6248729593430751,
"learning_rate": 5.494694484679501e-06,
"loss": 0.6097,
"step": 8020
},
{
"epoch": 0.9370621205044372,
"grad_norm": 0.5976544998280431,
"learning_rate": 5.485675334429776e-06,
"loss": 0.6042,
"step": 8025
},
{
"epoch": 0.9376459598318543,
"grad_norm": 0.5953741944030403,
"learning_rate": 5.476738262767116e-06,
"loss": 0.6401,
"step": 8030
},
{
"epoch": 0.9382297991592714,
"grad_norm": 0.5820653871322563,
"learning_rate": 5.467883303012653e-06,
"loss": 0.6403,
"step": 8035
},
{
"epoch": 0.9388136384866884,
"grad_norm": 0.548717776517309,
"learning_rate": 5.459110488181373e-06,
"loss": 0.6169,
"step": 8040
},
{
"epoch": 0.9393974778141055,
"grad_norm": 0.6017633488697844,
"learning_rate": 5.450419850981987e-06,
"loss": 0.6487,
"step": 8045
},
{
"epoch": 0.9399813171415227,
"grad_norm": 0.5874919853407551,
"learning_rate": 5.441811423816817e-06,
"loss": 0.6397,
"step": 8050
},
{
"epoch": 0.9405651564689398,
"grad_norm": 0.5855086777496958,
"learning_rate": 5.433285238781674e-06,
"loss": 0.6418,
"step": 8055
},
{
"epoch": 0.9411489957963568,
"grad_norm": 0.5796809419680976,
"learning_rate": 5.424841327665728e-06,
"loss": 0.6353,
"step": 8060
},
{
"epoch": 0.9417328351237739,
"grad_norm": 0.6073576400260668,
"learning_rate": 5.416479721951409e-06,
"loss": 0.6234,
"step": 8065
},
{
"epoch": 0.942316674451191,
"grad_norm": 0.5763590392057469,
"learning_rate": 5.408200452814265e-06,
"loss": 0.6167,
"step": 8070
},
{
"epoch": 0.9429005137786082,
"grad_norm": 0.5592518641443556,
"learning_rate": 5.400003551122871e-06,
"loss": 0.6356,
"step": 8075
},
{
"epoch": 0.9434843531060252,
"grad_norm": 0.6114852642140004,
"learning_rate": 5.391889047438692e-06,
"loss": 0.6231,
"step": 8080
},
{
"epoch": 0.9440681924334423,
"grad_norm": 0.6239977699979702,
"learning_rate": 5.383856972015984e-06,
"loss": 0.6216,
"step": 8085
},
{
"epoch": 0.9446520317608594,
"grad_norm": 0.5934103583221862,
"learning_rate": 5.3759073548016776e-06,
"loss": 0.6394,
"step": 8090
},
{
"epoch": 0.9452358710882766,
"grad_norm": 0.5749305173515513,
"learning_rate": 5.368040225435264e-06,
"loss": 0.6178,
"step": 8095
},
{
"epoch": 0.9458197104156936,
"grad_norm": 0.5450188710666751,
"learning_rate": 5.360255613248679e-06,
"loss": 0.6144,
"step": 8100
},
{
"epoch": 0.9464035497431107,
"grad_norm": 0.5688620542709263,
"learning_rate": 5.352553547266205e-06,
"loss": 0.6338,
"step": 8105
},
{
"epoch": 0.9469873890705278,
"grad_norm": 0.5788749409178702,
"learning_rate": 5.34493405620436e-06,
"loss": 0.6444,
"step": 8110
},
{
"epoch": 0.9475712283979448,
"grad_norm": 0.63750470201553,
"learning_rate": 5.337397168471786e-06,
"loss": 0.6394,
"step": 8115
},
{
"epoch": 0.948155067725362,
"grad_norm": 0.5993406233897292,
"learning_rate": 5.329942912169144e-06,
"loss": 0.6326,
"step": 8120
},
{
"epoch": 0.9487389070527791,
"grad_norm": 0.5758372086950723,
"learning_rate": 5.322571315089009e-06,
"loss": 0.6255,
"step": 8125
},
{
"epoch": 0.9493227463801962,
"grad_norm": 0.5645325322106245,
"learning_rate": 5.315282404715776e-06,
"loss": 0.6258,
"step": 8130
},
{
"epoch": 0.9499065857076132,
"grad_norm": 0.5631286057965411,
"learning_rate": 5.308076208225538e-06,
"loss": 0.6252,
"step": 8135
},
{
"epoch": 0.9504904250350303,
"grad_norm": 0.6324457688892521,
"learning_rate": 5.300952752486006e-06,
"loss": 0.6407,
"step": 8140
},
{
"epoch": 0.9510742643624475,
"grad_norm": 0.5738895175594562,
"learning_rate": 5.293912064056394e-06,
"loss": 0.6218,
"step": 8145
},
{
"epoch": 0.9516581036898646,
"grad_norm": 0.5758169636621945,
"learning_rate": 5.286954169187325e-06,
"loss": 0.6158,
"step": 8150
},
{
"epoch": 0.9522419430172816,
"grad_norm": 0.6504160410551667,
"learning_rate": 5.280079093820737e-06,
"loss": 0.6441,
"step": 8155
},
{
"epoch": 0.9528257823446987,
"grad_norm": 0.5954950695118845,
"learning_rate": 5.273286863589776e-06,
"loss": 0.6405,
"step": 8160
},
{
"epoch": 0.9534096216721158,
"grad_norm": 0.5690006780375534,
"learning_rate": 5.266577503818708e-06,
"loss": 0.6516,
"step": 8165
},
{
"epoch": 0.953993460999533,
"grad_norm": 0.5984737264452394,
"learning_rate": 5.259951039522832e-06,
"loss": 0.6452,
"step": 8170
},
{
"epoch": 0.95457730032695,
"grad_norm": 0.6010086090108752,
"learning_rate": 5.253407495408368e-06,
"loss": 0.623,
"step": 8175
},
{
"epoch": 0.9551611396543671,
"grad_norm": 0.5623765170749638,
"learning_rate": 5.24694689587238e-06,
"loss": 0.6232,
"step": 8180
},
{
"epoch": 0.9557449789817842,
"grad_norm": 0.5660884439507055,
"learning_rate": 5.240569265002673e-06,
"loss": 0.6229,
"step": 8185
},
{
"epoch": 0.9563288183092014,
"grad_norm": 0.5856955251220529,
"learning_rate": 5.234274626577723e-06,
"loss": 0.6416,
"step": 8190
},
{
"epoch": 0.9569126576366184,
"grad_norm": 0.5942822574063878,
"learning_rate": 5.228063004066567e-06,
"loss": 0.6179,
"step": 8195
},
{
"epoch": 0.9574964969640355,
"grad_norm": 0.6209192638572009,
"learning_rate": 5.22193442062872e-06,
"loss": 0.6222,
"step": 8200
},
{
"epoch": 0.9580803362914526,
"grad_norm": 0.6332384096267408,
"learning_rate": 5.2158888991141055e-06,
"loss": 0.6333,
"step": 8205
},
{
"epoch": 0.9586641756188697,
"grad_norm": 0.5824197330774395,
"learning_rate": 5.2099264620629425e-06,
"loss": 0.6343,
"step": 8210
},
{
"epoch": 0.9592480149462868,
"grad_norm": 0.580708798795656,
"learning_rate": 5.204047131705689e-06,
"loss": 0.6194,
"step": 8215
},
{
"epoch": 0.9598318542737039,
"grad_norm": 0.6071990294751999,
"learning_rate": 5.198250929962939e-06,
"loss": 0.6405,
"step": 8220
},
{
"epoch": 0.960415693601121,
"grad_norm": 0.5904035551150635,
"learning_rate": 5.192537878445356e-06,
"loss": 0.6375,
"step": 8225
},
{
"epoch": 0.960999532928538,
"grad_norm": 0.5707654417055844,
"learning_rate": 5.186907998453573e-06,
"loss": 0.6239,
"step": 8230
},
{
"epoch": 0.9615833722559551,
"grad_norm": 0.6110530879520853,
"learning_rate": 5.181361310978133e-06,
"loss": 0.6157,
"step": 8235
},
{
"epoch": 0.9621672115833723,
"grad_norm": 0.5713240749254127,
"learning_rate": 5.175897836699403e-06,
"loss": 0.6427,
"step": 8240
},
{
"epoch": 0.9627510509107894,
"grad_norm": 0.6113466535850735,
"learning_rate": 5.170517595987493e-06,
"loss": 0.633,
"step": 8245
},
{
"epoch": 0.9633348902382064,
"grad_norm": 0.6352232672030907,
"learning_rate": 5.165220608902186e-06,
"loss": 0.6283,
"step": 8250
},
{
"epoch": 0.9639187295656235,
"grad_norm": 0.6248726494982753,
"learning_rate": 5.160006895192858e-06,
"loss": 0.627,
"step": 8255
},
{
"epoch": 0.9645025688930406,
"grad_norm": 0.5977144170498508,
"learning_rate": 5.154876474298412e-06,
"loss": 0.6079,
"step": 8260
},
{
"epoch": 0.9650864082204578,
"grad_norm": 0.6176936440149047,
"learning_rate": 5.149829365347197e-06,
"loss": 0.6287,
"step": 8265
},
{
"epoch": 0.9656702475478748,
"grad_norm": 0.5930270091080995,
"learning_rate": 5.14486558715694e-06,
"loss": 0.6327,
"step": 8270
},
{
"epoch": 0.9662540868752919,
"grad_norm": 0.6038829668240889,
"learning_rate": 5.139985158234677e-06,
"loss": 0.6649,
"step": 8275
},
{
"epoch": 0.966837926202709,
"grad_norm": 0.6054431949079073,
"learning_rate": 5.135188096776682e-06,
"loss": 0.6263,
"step": 8280
},
{
"epoch": 0.9674217655301262,
"grad_norm": 0.5883184566217126,
"learning_rate": 5.130474420668403e-06,
"loss": 0.6393,
"step": 8285
},
{
"epoch": 0.9680056048575432,
"grad_norm": 0.6114296749695102,
"learning_rate": 5.125844147484391e-06,
"loss": 0.6194,
"step": 8290
},
{
"epoch": 0.9685894441849603,
"grad_norm": 0.6480286564406009,
"learning_rate": 5.121297294488237e-06,
"loss": 0.6508,
"step": 8295
},
{
"epoch": 0.9691732835123774,
"grad_norm": 0.5652153579112486,
"learning_rate": 5.1168338786325025e-06,
"loss": 0.6234,
"step": 8300
},
{
"epoch": 0.9697571228397945,
"grad_norm": 0.655551749204273,
"learning_rate": 5.112453916558671e-06,
"loss": 0.6485,
"step": 8305
},
{
"epoch": 0.9703409621672116,
"grad_norm": 0.5723069124877626,
"learning_rate": 5.108157424597062e-06,
"loss": 0.6224,
"step": 8310
},
{
"epoch": 0.9709248014946287,
"grad_norm": 0.5959024092431895,
"learning_rate": 5.103944418766791e-06,
"loss": 0.634,
"step": 8315
},
{
"epoch": 0.9715086408220458,
"grad_norm": 0.5985134268757333,
"learning_rate": 5.099814914775706e-06,
"loss": 0.6226,
"step": 8320
},
{
"epoch": 0.9720924801494629,
"grad_norm": 0.5897338045813747,
"learning_rate": 5.095768928020314e-06,
"loss": 0.639,
"step": 8325
},
{
"epoch": 0.9726763194768799,
"grad_norm": 0.651914632203561,
"learning_rate": 5.09180647358575e-06,
"loss": 0.6239,
"step": 8330
},
{
"epoch": 0.9732601588042971,
"grad_norm": 0.6099486128222116,
"learning_rate": 5.087927566245688e-06,
"loss": 0.6226,
"step": 8335
},
{
"epoch": 0.9738439981317142,
"grad_norm": 0.546666413702602,
"learning_rate": 5.0841322204623205e-06,
"loss": 0.617,
"step": 8340
},
{
"epoch": 0.9744278374591312,
"grad_norm": 0.5704580268593588,
"learning_rate": 5.080420450386274e-06,
"loss": 0.6289,
"step": 8345
},
{
"epoch": 0.9750116767865483,
"grad_norm": 0.6016070522149358,
"learning_rate": 5.076792269856582e-06,
"loss": 0.6254,
"step": 8350
},
{
"epoch": 0.9755955161139654,
"grad_norm": 0.6144686115339516,
"learning_rate": 5.073247692400609e-06,
"loss": 0.6188,
"step": 8355
},
{
"epoch": 0.9761793554413826,
"grad_norm": 0.5912193203791197,
"learning_rate": 5.069786731234025e-06,
"loss": 0.6349,
"step": 8360
},
{
"epoch": 0.9767631947687996,
"grad_norm": 0.5690491120853608,
"learning_rate": 5.066409399260733e-06,
"loss": 0.647,
"step": 8365
},
{
"epoch": 0.9773470340962167,
"grad_norm": 0.5595435206008783,
"learning_rate": 5.063115709072837e-06,
"loss": 0.6356,
"step": 8370
},
{
"epoch": 0.9779308734236338,
"grad_norm": 0.5831436839137033,
"learning_rate": 5.059905672950588e-06,
"loss": 0.6333,
"step": 8375
},
{
"epoch": 0.978514712751051,
"grad_norm": 0.6175484375330281,
"learning_rate": 5.056779302862337e-06,
"loss": 0.6341,
"step": 8380
},
{
"epoch": 0.979098552078468,
"grad_norm": 0.5821033777881047,
"learning_rate": 5.0537366104645e-06,
"loss": 0.6186,
"step": 8385
},
{
"epoch": 0.9796823914058851,
"grad_norm": 0.6004196509282287,
"learning_rate": 5.050777607101506e-06,
"loss": 0.6264,
"step": 8390
},
{
"epoch": 0.9802662307333022,
"grad_norm": 0.615930287431196,
"learning_rate": 5.047902303805746e-06,
"loss": 0.6471,
"step": 8395
},
{
"epoch": 0.9808500700607193,
"grad_norm": 0.5882780472877527,
"learning_rate": 5.045110711297557e-06,
"loss": 0.6249,
"step": 8400
},
{
"epoch": 0.9814339093881363,
"grad_norm": 0.5649174129976797,
"learning_rate": 5.042402839985161e-06,
"loss": 0.6123,
"step": 8405
},
{
"epoch": 0.9820177487155535,
"grad_norm": 0.6170975536166484,
"learning_rate": 5.039778699964626e-06,
"loss": 0.637,
"step": 8410
},
{
"epoch": 0.9826015880429706,
"grad_norm": 0.5811148214585863,
"learning_rate": 5.037238301019845e-06,
"loss": 0.6136,
"step": 8415
},
{
"epoch": 0.9831854273703877,
"grad_norm": 0.5891383001640484,
"learning_rate": 5.034781652622484e-06,
"loss": 0.6163,
"step": 8420
},
{
"epoch": 0.9837692666978047,
"grad_norm": 0.5778442239111855,
"learning_rate": 5.032408763931956e-06,
"loss": 0.6011,
"step": 8425
},
{
"epoch": 0.9843531060252219,
"grad_norm": 0.6213289305639533,
"learning_rate": 5.0301196437953755e-06,
"loss": 0.6219,
"step": 8430
},
{
"epoch": 0.984936945352639,
"grad_norm": 0.5936267104807765,
"learning_rate": 5.0279143007475425e-06,
"loss": 0.6338,
"step": 8435
},
{
"epoch": 0.985520784680056,
"grad_norm": 0.6102251067582517,
"learning_rate": 5.02579274301089e-06,
"loss": 0.6321,
"step": 8440
},
{
"epoch": 0.9861046240074731,
"grad_norm": 0.6211508033536209,
"learning_rate": 5.0237549784954745e-06,
"loss": 0.6076,
"step": 8445
},
{
"epoch": 0.9866884633348902,
"grad_norm": 0.6012756936641813,
"learning_rate": 5.021801014798933e-06,
"loss": 0.6137,
"step": 8450
},
{
"epoch": 0.9872723026623074,
"grad_norm": 0.5574258509117747,
"learning_rate": 5.0199308592064535e-06,
"loss": 0.6175,
"step": 8455
},
{
"epoch": 0.9878561419897244,
"grad_norm": 0.5843933121249048,
"learning_rate": 5.018144518690761e-06,
"loss": 0.6243,
"step": 8460
},
{
"epoch": 0.9884399813171415,
"grad_norm": 0.5567147895943964,
"learning_rate": 5.016441999912074e-06,
"loss": 0.6252,
"step": 8465
},
{
"epoch": 0.9890238206445586,
"grad_norm": 0.6198445874725916,
"learning_rate": 5.014823309218096e-06,
"loss": 0.6216,
"step": 8470
},
{
"epoch": 0.9896076599719758,
"grad_norm": 0.6235986624021734,
"learning_rate": 5.013288452643979e-06,
"loss": 0.6358,
"step": 8475
},
{
"epoch": 0.9901914992993928,
"grad_norm": 0.5920757806320878,
"learning_rate": 5.011837435912308e-06,
"loss": 0.6242,
"step": 8480
},
{
"epoch": 0.9907753386268099,
"grad_norm": 0.5510879318024842,
"learning_rate": 5.010470264433083e-06,
"loss": 0.6321,
"step": 8485
},
{
"epoch": 0.991359177954227,
"grad_norm": 0.553704155240658,
"learning_rate": 5.009186943303684e-06,
"loss": 0.6365,
"step": 8490
},
{
"epoch": 0.9919430172816441,
"grad_norm": 0.5664918391691475,
"learning_rate": 5.0079874773088735e-06,
"loss": 0.6305,
"step": 8495
},
{
"epoch": 0.9925268566090611,
"grad_norm": 0.5874914103537708,
"learning_rate": 5.006871870920757e-06,
"loss": 0.6198,
"step": 8500
},
{
"epoch": 0.9931106959364783,
"grad_norm": 0.5452067196739758,
"learning_rate": 5.005840128298783e-06,
"loss": 0.6219,
"step": 8505
},
{
"epoch": 0.9936945352638954,
"grad_norm": 0.5894932818967217,
"learning_rate": 5.004892253289714e-06,
"loss": 0.631,
"step": 8510
},
{
"epoch": 0.9942783745913125,
"grad_norm": 0.5942366273834488,
"learning_rate": 5.004028249427629e-06,
"loss": 0.6221,
"step": 8515
},
{
"epoch": 0.9948622139187295,
"grad_norm": 0.5576430535004351,
"learning_rate": 5.003248119933894e-06,
"loss": 0.6137,
"step": 8520
},
{
"epoch": 0.9954460532461467,
"grad_norm": 0.5620291144392621,
"learning_rate": 5.002551867717153e-06,
"loss": 0.6152,
"step": 8525
},
{
"epoch": 0.9960298925735638,
"grad_norm": 0.615196076482753,
"learning_rate": 5.00193949537333e-06,
"loss": 0.6451,
"step": 8530
},
{
"epoch": 0.9966137319009809,
"grad_norm": 0.5979491929387287,
"learning_rate": 5.0014110051856e-06,
"loss": 0.653,
"step": 8535
},
{
"epoch": 0.9971975712283979,
"grad_norm": 0.5740297550730898,
"learning_rate": 5.000966399124398e-06,
"loss": 0.629,
"step": 8540
},
{
"epoch": 0.997781410555815,
"grad_norm": 0.621793395309707,
"learning_rate": 5.000605678847399e-06,
"loss": 0.6204,
"step": 8545
},
{
"epoch": 0.9983652498832322,
"grad_norm": 0.5650577119664021,
"learning_rate": 5.000328845699522e-06,
"loss": 0.6251,
"step": 8550
},
{
"epoch": 0.9989490892106492,
"grad_norm": 0.6069800107074691,
"learning_rate": 5.000135900712914e-06,
"loss": 0.6241,
"step": 8555
},
{
"epoch": 0.9995329285380663,
"grad_norm": 0.5758118307684206,
"learning_rate": 5.000026844606953e-06,
"loss": 0.6308,
"step": 8560
},
{
"epoch": 1.0,
"step": 8564,
"total_flos": 487937544290304.0,
"train_loss": 0.6853283835259179,
"train_runtime": 17987.0771,
"train_samples_per_second": 30.472,
"train_steps_per_second": 0.476
}
],
"logging_steps": 5,
"max_steps": 8564,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 487937544290304.0,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}