granite-lora-development / trainer_state.json
mqddd's picture
Upload folder using huggingface_hub
f32a993 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 6.961584633853541,
"eval_steps": 500,
"global_step": 5799,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.012004801920768308,
"grad_norm": 0.02549879066646099,
"learning_rate": 4.0000000000000003e-07,
"loss": 0.9862,
"step": 10
},
{
"epoch": 0.024009603841536616,
"grad_norm": 0.03309444710612297,
"learning_rate": 8.000000000000001e-07,
"loss": 0.9386,
"step": 20
},
{
"epoch": 0.03601440576230492,
"grad_norm": 0.02672477439045906,
"learning_rate": 1.2000000000000002e-06,
"loss": 0.9425,
"step": 30
},
{
"epoch": 0.04801920768307323,
"grad_norm": 0.02686227485537529,
"learning_rate": 1.6000000000000001e-06,
"loss": 0.9175,
"step": 40
},
{
"epoch": 0.060024009603841535,
"grad_norm": 0.025844795629382133,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.9963,
"step": 50
},
{
"epoch": 0.07202881152460984,
"grad_norm": 0.032747991383075714,
"learning_rate": 2.4000000000000003e-06,
"loss": 0.9103,
"step": 60
},
{
"epoch": 0.08403361344537816,
"grad_norm": 0.03013288602232933,
"learning_rate": 2.8000000000000003e-06,
"loss": 0.9381,
"step": 70
},
{
"epoch": 0.09603841536614646,
"grad_norm": 0.03312341496348381,
"learning_rate": 3.2000000000000003e-06,
"loss": 0.9105,
"step": 80
},
{
"epoch": 0.10804321728691477,
"grad_norm": 0.03535911813378334,
"learning_rate": 3.6000000000000003e-06,
"loss": 0.9662,
"step": 90
},
{
"epoch": 0.12004801920768307,
"grad_norm": 0.03163569048047066,
"learning_rate": 4.000000000000001e-06,
"loss": 0.8886,
"step": 100
},
{
"epoch": 0.13205282112845138,
"grad_norm": 0.03126745671033859,
"learning_rate": 4.4e-06,
"loss": 0.9677,
"step": 110
},
{
"epoch": 0.14405762304921968,
"grad_norm": 0.03188377246260643,
"learning_rate": 4.800000000000001e-06,
"loss": 0.9096,
"step": 120
},
{
"epoch": 0.15606242496998798,
"grad_norm": 0.03766326606273651,
"learning_rate": 5.2e-06,
"loss": 0.9868,
"step": 130
},
{
"epoch": 0.16806722689075632,
"grad_norm": 0.03288175165653229,
"learning_rate": 5.600000000000001e-06,
"loss": 0.9614,
"step": 140
},
{
"epoch": 0.18007202881152462,
"grad_norm": 0.03540439158678055,
"learning_rate": 6e-06,
"loss": 0.9471,
"step": 150
},
{
"epoch": 0.19207683073229292,
"grad_norm": 0.0429881289601326,
"learning_rate": 6.4000000000000006e-06,
"loss": 0.8901,
"step": 160
},
{
"epoch": 0.20408163265306123,
"grad_norm": 0.04951930418610573,
"learning_rate": 6.800000000000001e-06,
"loss": 0.9403,
"step": 170
},
{
"epoch": 0.21608643457382953,
"grad_norm": 0.049168169498443604,
"learning_rate": 7.2000000000000005e-06,
"loss": 0.8943,
"step": 180
},
{
"epoch": 0.22809123649459784,
"grad_norm": 0.04803888127207756,
"learning_rate": 7.600000000000001e-06,
"loss": 0.8852,
"step": 190
},
{
"epoch": 0.24009603841536614,
"grad_norm": 0.04201197251677513,
"learning_rate": 8.000000000000001e-06,
"loss": 0.8882,
"step": 200
},
{
"epoch": 0.25210084033613445,
"grad_norm": 0.04449041560292244,
"learning_rate": 8.400000000000001e-06,
"loss": 0.8777,
"step": 210
},
{
"epoch": 0.26410564225690275,
"grad_norm": 0.036004405468702316,
"learning_rate": 8.8e-06,
"loss": 0.8496,
"step": 220
},
{
"epoch": 0.27611044417767105,
"grad_norm": 0.04503042995929718,
"learning_rate": 9.200000000000002e-06,
"loss": 0.8712,
"step": 230
},
{
"epoch": 0.28811524609843936,
"grad_norm": 0.04312640801072121,
"learning_rate": 9.600000000000001e-06,
"loss": 0.8459,
"step": 240
},
{
"epoch": 0.30012004801920766,
"grad_norm": 0.03977839648723602,
"learning_rate": 1e-05,
"loss": 0.8705,
"step": 250
},
{
"epoch": 0.31212484993997597,
"grad_norm": 0.039520714432001114,
"learning_rate": 9.999919867416569e-06,
"loss": 0.8751,
"step": 260
},
{
"epoch": 0.3241296518607443,
"grad_norm": 0.037476055324077606,
"learning_rate": 9.999679472234764e-06,
"loss": 0.8137,
"step": 270
},
{
"epoch": 0.33613445378151263,
"grad_norm": 0.0433887243270874,
"learning_rate": 9.999278822159985e-06,
"loss": 0.8147,
"step": 280
},
{
"epoch": 0.34813925570228094,
"grad_norm": 0.04283862188458443,
"learning_rate": 9.998717930034278e-06,
"loss": 0.8642,
"step": 290
},
{
"epoch": 0.36014405762304924,
"grad_norm": 0.03704555705189705,
"learning_rate": 9.997996813835938e-06,
"loss": 0.8481,
"step": 300
},
{
"epoch": 0.37214885954381755,
"grad_norm": 0.038123104721307755,
"learning_rate": 9.997115496678926e-06,
"loss": 0.8328,
"step": 310
},
{
"epoch": 0.38415366146458585,
"grad_norm": 0.03972741588950157,
"learning_rate": 9.996074006812133e-06,
"loss": 0.8846,
"step": 320
},
{
"epoch": 0.39615846338535415,
"grad_norm": 0.03823426738381386,
"learning_rate": 9.994872377618467e-06,
"loss": 0.8571,
"step": 330
},
{
"epoch": 0.40816326530612246,
"grad_norm": 0.03882453590631485,
"learning_rate": 9.993510647613787e-06,
"loss": 0.8664,
"step": 340
},
{
"epoch": 0.42016806722689076,
"grad_norm": 0.040137797594070435,
"learning_rate": 9.991988860445672e-06,
"loss": 0.8386,
"step": 350
},
{
"epoch": 0.43217286914765907,
"grad_norm": 0.0372915081679821,
"learning_rate": 9.990307064892015e-06,
"loss": 0.8799,
"step": 360
},
{
"epoch": 0.44417767106842737,
"grad_norm": 0.03684937208890915,
"learning_rate": 9.988465314859469e-06,
"loss": 0.8352,
"step": 370
},
{
"epoch": 0.4561824729891957,
"grad_norm": 0.03629535436630249,
"learning_rate": 9.986463669381705e-06,
"loss": 0.8084,
"step": 380
},
{
"epoch": 0.468187274909964,
"grad_norm": 0.034760646522045135,
"learning_rate": 9.984302192617534e-06,
"loss": 0.8363,
"step": 390
},
{
"epoch": 0.4801920768307323,
"grad_norm": 0.033850379288196564,
"learning_rate": 9.981980953848844e-06,
"loss": 0.8209,
"step": 400
},
{
"epoch": 0.4921968787515006,
"grad_norm": 0.031105060130357742,
"learning_rate": 9.979500027478377e-06,
"loss": 0.8589,
"step": 410
},
{
"epoch": 0.5042016806722689,
"grad_norm": 0.034456316381692886,
"learning_rate": 9.97685949302735e-06,
"loss": 0.8599,
"step": 420
},
{
"epoch": 0.5162064825930373,
"grad_norm": 0.03859219700098038,
"learning_rate": 9.974059435132901e-06,
"loss": 0.8862,
"step": 430
},
{
"epoch": 0.5282112845138055,
"grad_norm": 0.031010838225483894,
"learning_rate": 9.97109994354538e-06,
"loss": 0.8565,
"step": 440
},
{
"epoch": 0.5402160864345739,
"grad_norm": 0.03447699174284935,
"learning_rate": 9.967981113125466e-06,
"loss": 0.8691,
"step": 450
},
{
"epoch": 0.5522208883553421,
"grad_norm": 0.033097319304943085,
"learning_rate": 9.964703043841139e-06,
"loss": 0.8003,
"step": 460
},
{
"epoch": 0.5642256902761105,
"grad_norm": 0.03505665436387062,
"learning_rate": 9.961265840764463e-06,
"loss": 0.863,
"step": 470
},
{
"epoch": 0.5762304921968787,
"grad_norm": 0.04121314734220505,
"learning_rate": 9.957669614068222e-06,
"loss": 0.8208,
"step": 480
},
{
"epoch": 0.5882352941176471,
"grad_norm": 0.03139583393931389,
"learning_rate": 9.953914479022388e-06,
"loss": 0.8637,
"step": 490
},
{
"epoch": 0.6002400960384153,
"grad_norm": 0.03294196352362633,
"learning_rate": 9.950000555990434e-06,
"loss": 0.804,
"step": 500
},
{
"epoch": 0.6122448979591837,
"grad_norm": 0.03441837430000305,
"learning_rate": 9.945927970425462e-06,
"loss": 0.842,
"step": 510
},
{
"epoch": 0.6242496998799519,
"grad_norm": 0.03234728053212166,
"learning_rate": 9.941696852866195e-06,
"loss": 0.8194,
"step": 520
},
{
"epoch": 0.6362545018007203,
"grad_norm": 0.03413551673293114,
"learning_rate": 9.937307338932783e-06,
"loss": 0.7902,
"step": 530
},
{
"epoch": 0.6482593037214885,
"grad_norm": 0.03235715255141258,
"learning_rate": 9.932759569322467e-06,
"loss": 0.8163,
"step": 540
},
{
"epoch": 0.6602641056422569,
"grad_norm": 0.03173699602484703,
"learning_rate": 9.928053689805054e-06,
"loss": 0.8364,
"step": 550
},
{
"epoch": 0.6722689075630253,
"grad_norm": 0.035562604665756226,
"learning_rate": 9.923189851218259e-06,
"loss": 0.8336,
"step": 560
},
{
"epoch": 0.6842737094837935,
"grad_norm": 0.03246847167611122,
"learning_rate": 9.918168209462862e-06,
"loss": 0.8354,
"step": 570
},
{
"epoch": 0.6962785114045619,
"grad_norm": 0.03735557198524475,
"learning_rate": 9.912988925497716e-06,
"loss": 0.8705,
"step": 580
},
{
"epoch": 0.7082833133253301,
"grad_norm": 0.04147843271493912,
"learning_rate": 9.907652165334579e-06,
"loss": 0.8296,
"step": 590
},
{
"epoch": 0.7202881152460985,
"grad_norm": 0.03415181115269661,
"learning_rate": 9.902158100032804e-06,
"loss": 0.8191,
"step": 600
},
{
"epoch": 0.7322929171668667,
"grad_norm": 0.03535843268036842,
"learning_rate": 9.89650690569385e-06,
"loss": 0.8365,
"step": 610
},
{
"epoch": 0.7442977190876351,
"grad_norm": 0.036314673721790314,
"learning_rate": 9.890698763455637e-06,
"loss": 0.8575,
"step": 620
},
{
"epoch": 0.7563025210084033,
"grad_norm": 0.033384788781404495,
"learning_rate": 9.884733859486745e-06,
"loss": 0.8736,
"step": 630
},
{
"epoch": 0.7683073229291717,
"grad_norm": 0.036313604563474655,
"learning_rate": 9.878612384980435e-06,
"loss": 0.8104,
"step": 640
},
{
"epoch": 0.78031212484994,
"grad_norm": 0.03539718687534332,
"learning_rate": 9.872334536148538e-06,
"loss": 0.8535,
"step": 650
},
{
"epoch": 0.7923169267707083,
"grad_norm": 0.040795594453811646,
"learning_rate": 9.86590051421515e-06,
"loss": 0.859,
"step": 660
},
{
"epoch": 0.8043217286914766,
"grad_norm": 0.03459113836288452,
"learning_rate": 9.85931052541019e-06,
"loss": 0.8637,
"step": 670
},
{
"epoch": 0.8163265306122449,
"grad_norm": 0.03594481945037842,
"learning_rate": 9.852564780962793e-06,
"loss": 0.8462,
"step": 680
},
{
"epoch": 0.8283313325330132,
"grad_norm": 0.03311854600906372,
"learning_rate": 9.845663497094526e-06,
"loss": 0.8201,
"step": 690
},
{
"epoch": 0.8403361344537815,
"grad_norm": 0.04675761237740517,
"learning_rate": 9.838606895012474e-06,
"loss": 0.8724,
"step": 700
},
{
"epoch": 0.8523409363745498,
"grad_norm": 0.035599809139966965,
"learning_rate": 9.831395200902137e-06,
"loss": 0.7937,
"step": 710
},
{
"epoch": 0.8643457382953181,
"grad_norm": 0.036689773201942444,
"learning_rate": 9.82402864592019e-06,
"loss": 0.8179,
"step": 720
},
{
"epoch": 0.8763505402160864,
"grad_norm": 0.041255176067352295,
"learning_rate": 9.816507466187064e-06,
"loss": 0.8175,
"step": 730
},
{
"epoch": 0.8883553421368547,
"grad_norm": 0.03711531683802605,
"learning_rate": 9.808831902779382e-06,
"loss": 0.8199,
"step": 740
},
{
"epoch": 0.9003601440576231,
"grad_norm": 0.04047127068042755,
"learning_rate": 9.801002201722237e-06,
"loss": 0.8439,
"step": 750
},
{
"epoch": 0.9123649459783914,
"grad_norm": 0.03690866753458977,
"learning_rate": 9.793018613981298e-06,
"loss": 0.8206,
"step": 760
},
{
"epoch": 0.9243697478991597,
"grad_norm": 0.03672650083899498,
"learning_rate": 9.784881395454767e-06,
"loss": 0.8057,
"step": 770
},
{
"epoch": 0.936374549819928,
"grad_norm": 0.0367751307785511,
"learning_rate": 9.776590806965183e-06,
"loss": 0.8021,
"step": 780
},
{
"epoch": 0.9483793517406963,
"grad_norm": 0.03809288516640663,
"learning_rate": 9.768147114251056e-06,
"loss": 0.841,
"step": 790
},
{
"epoch": 0.9603841536614646,
"grad_norm": 0.038219958543777466,
"learning_rate": 9.75955058795835e-06,
"loss": 0.8226,
"step": 800
},
{
"epoch": 0.9723889555822329,
"grad_norm": 0.03921416401863098,
"learning_rate": 9.750801503631809e-06,
"loss": 0.8298,
"step": 810
},
{
"epoch": 0.9843937575030012,
"grad_norm": 0.03702569007873535,
"learning_rate": 9.741900141706122e-06,
"loss": 0.8547,
"step": 820
},
{
"epoch": 0.9963985594237695,
"grad_norm": 0.041355907917022705,
"learning_rate": 9.732846787496943e-06,
"loss": 0.8252,
"step": 830
},
{
"epoch": 1.0084033613445378,
"grad_norm": 0.039719488471746445,
"learning_rate": 9.723641731191737e-06,
"loss": 0.8116,
"step": 840
},
{
"epoch": 1.0204081632653061,
"grad_norm": 0.041531551629304886,
"learning_rate": 9.71428526784048e-06,
"loss": 0.8268,
"step": 850
},
{
"epoch": 1.0324129651860745,
"grad_norm": 0.03957758843898773,
"learning_rate": 9.704777697346204e-06,
"loss": 0.8147,
"step": 860
},
{
"epoch": 1.0444177671068426,
"grad_norm": 0.040645621716976166,
"learning_rate": 9.695119324455383e-06,
"loss": 0.8487,
"step": 870
},
{
"epoch": 1.056422569027611,
"grad_norm": 0.037387099117040634,
"learning_rate": 9.685310458748165e-06,
"loss": 0.8016,
"step": 880
},
{
"epoch": 1.0684273709483794,
"grad_norm": 0.0387730598449707,
"learning_rate": 9.675351414628452e-06,
"loss": 0.8389,
"step": 890
},
{
"epoch": 1.0804321728691477,
"grad_norm": 0.03654830902814865,
"learning_rate": 9.665242511313816e-06,
"loss": 0.8571,
"step": 900
},
{
"epoch": 1.092436974789916,
"grad_norm": 0.03688422963023186,
"learning_rate": 9.654984072825273e-06,
"loss": 0.8148,
"step": 910
},
{
"epoch": 1.1044417767106842,
"grad_norm": 0.04102282598614693,
"learning_rate": 9.644576427976893e-06,
"loss": 0.8252,
"step": 920
},
{
"epoch": 1.1164465786314526,
"grad_norm": 0.04032592847943306,
"learning_rate": 9.634019910365265e-06,
"loss": 0.8362,
"step": 930
},
{
"epoch": 1.128451380552221,
"grad_norm": 0.038995206356048584,
"learning_rate": 9.6233148583588e-06,
"loss": 0.8071,
"step": 940
},
{
"epoch": 1.140456182472989,
"grad_norm": 0.04222116991877556,
"learning_rate": 9.612461615086888e-06,
"loss": 0.8162,
"step": 950
},
{
"epoch": 1.1524609843937574,
"grad_norm": 0.03999168053269386,
"learning_rate": 9.601460528428894e-06,
"loss": 0.8124,
"step": 960
},
{
"epoch": 1.1644657863145258,
"grad_norm": 0.04248590022325516,
"learning_rate": 9.590311951003022e-06,
"loss": 0.809,
"step": 970
},
{
"epoch": 1.1764705882352942,
"grad_norm": 0.041043415665626526,
"learning_rate": 9.57901624015499e-06,
"loss": 0.7999,
"step": 980
},
{
"epoch": 1.1884753901560625,
"grad_norm": 0.037975288927555084,
"learning_rate": 9.5675737579466e-06,
"loss": 0.8839,
"step": 990
},
{
"epoch": 1.2004801920768307,
"grad_norm": 0.038373079150915146,
"learning_rate": 9.555984871144112e-06,
"loss": 0.8297,
"step": 1000
},
{
"epoch": 1.212484993997599,
"grad_norm": 0.03899591043591499,
"learning_rate": 9.544249951206504e-06,
"loss": 0.7856,
"step": 1010
},
{
"epoch": 1.2244897959183674,
"grad_norm": 0.0428166463971138,
"learning_rate": 9.532369374273555e-06,
"loss": 0.8018,
"step": 1020
},
{
"epoch": 1.2364945978391357,
"grad_norm": 0.04245438799262047,
"learning_rate": 9.520343521153796e-06,
"loss": 0.8404,
"step": 1030
},
{
"epoch": 1.2484993997599039,
"grad_norm": 0.04125148430466652,
"learning_rate": 9.508172777312297e-06,
"loss": 0.8264,
"step": 1040
},
{
"epoch": 1.2605042016806722,
"grad_norm": 0.04240383952856064,
"learning_rate": 9.495857532858314e-06,
"loss": 0.8442,
"step": 1050
},
{
"epoch": 1.2725090036014406,
"grad_norm": 0.04034110903739929,
"learning_rate": 9.483398182532792e-06,
"loss": 0.8459,
"step": 1060
},
{
"epoch": 1.284513805522209,
"grad_norm": 0.045645780861377716,
"learning_rate": 9.470795125695704e-06,
"loss": 0.8316,
"step": 1070
},
{
"epoch": 1.296518607442977,
"grad_norm": 0.0422181636095047,
"learning_rate": 9.458048766313246e-06,
"loss": 0.8144,
"step": 1080
},
{
"epoch": 1.3085234093637454,
"grad_norm": 0.04134783521294594,
"learning_rate": 9.445159512944906e-06,
"loss": 0.8307,
"step": 1090
},
{
"epoch": 1.3205282112845138,
"grad_norm": 0.04174504429101944,
"learning_rate": 9.43212777873035e-06,
"loss": 0.8179,
"step": 1100
},
{
"epoch": 1.3325330132052822,
"grad_norm": 0.0407400019466877,
"learning_rate": 9.418953981376188e-06,
"loss": 0.8141,
"step": 1110
},
{
"epoch": 1.3445378151260505,
"grad_norm": 0.04575135558843613,
"learning_rate": 9.405638543142589e-06,
"loss": 0.7888,
"step": 1120
},
{
"epoch": 1.3565426170468187,
"grad_norm": 0.042352344840765,
"learning_rate": 9.392181890829739e-06,
"loss": 0.8257,
"step": 1130
},
{
"epoch": 1.368547418967587,
"grad_norm": 0.046745285391807556,
"learning_rate": 9.378584455764161e-06,
"loss": 0.816,
"step": 1140
},
{
"epoch": 1.3805522208883554,
"grad_norm": 0.04341897740960121,
"learning_rate": 9.364846673784897e-06,
"loss": 0.7988,
"step": 1150
},
{
"epoch": 1.3925570228091235,
"grad_norm": 0.04738551750779152,
"learning_rate": 9.350968985229532e-06,
"loss": 0.8764,
"step": 1160
},
{
"epoch": 1.4045618247298919,
"grad_norm": 0.03800031542778015,
"learning_rate": 9.336951834920081e-06,
"loss": 0.8228,
"step": 1170
},
{
"epoch": 1.4165666266506602,
"grad_norm": 0.04292312636971474,
"learning_rate": 9.322795672148726e-06,
"loss": 0.7931,
"step": 1180
},
{
"epoch": 1.4285714285714286,
"grad_norm": 0.04134088754653931,
"learning_rate": 9.30850095066343e-06,
"loss": 0.8238,
"step": 1190
},
{
"epoch": 1.440576230492197,
"grad_norm": 0.04374649375677109,
"learning_rate": 9.294068128653373e-06,
"loss": 0.8494,
"step": 1200
},
{
"epoch": 1.452581032412965,
"grad_norm": 0.04475456848740578,
"learning_rate": 9.279497668734286e-06,
"loss": 0.8213,
"step": 1210
},
{
"epoch": 1.4645858343337335,
"grad_norm": 0.04506712034344673,
"learning_rate": 9.264790037933604e-06,
"loss": 0.8117,
"step": 1220
},
{
"epoch": 1.4765906362545018,
"grad_norm": 0.042217791080474854,
"learning_rate": 9.249945707675506e-06,
"loss": 0.8387,
"step": 1230
},
{
"epoch": 1.4885954381752702,
"grad_norm": 0.043338458985090256,
"learning_rate": 9.234965153765809e-06,
"loss": 0.8345,
"step": 1240
},
{
"epoch": 1.5006002400960385,
"grad_norm": 0.04206259921193123,
"learning_rate": 9.219848856376706e-06,
"loss": 0.8062,
"step": 1250
},
{
"epoch": 1.5126050420168067,
"grad_norm": 0.0476699024438858,
"learning_rate": 9.204597300031379e-06,
"loss": 0.8626,
"step": 1260
},
{
"epoch": 1.524609843937575,
"grad_norm": 0.04465340077877045,
"learning_rate": 9.189210973588476e-06,
"loss": 0.8379,
"step": 1270
},
{
"epoch": 1.5366146458583434,
"grad_norm": 0.04375738650560379,
"learning_rate": 9.173690370226432e-06,
"loss": 0.7904,
"step": 1280
},
{
"epoch": 1.5486194477791115,
"grad_norm": 0.04561934620141983,
"learning_rate": 9.158035987427662e-06,
"loss": 0.8152,
"step": 1290
},
{
"epoch": 1.5606242496998801,
"grad_norm": 0.04491221904754639,
"learning_rate": 9.142248326962624e-06,
"loss": 0.7844,
"step": 1300
},
{
"epoch": 1.5726290516206483,
"grad_norm": 0.04250409081578255,
"learning_rate": 9.12632789487372e-06,
"loss": 0.8156,
"step": 1310
},
{
"epoch": 1.5846338535414166,
"grad_norm": 0.04542950913310051,
"learning_rate": 9.110275201459099e-06,
"loss": 0.8496,
"step": 1320
},
{
"epoch": 1.596638655462185,
"grad_norm": 0.04421227052807808,
"learning_rate": 9.09409076125627e-06,
"loss": 0.817,
"step": 1330
},
{
"epoch": 1.6086434573829531,
"grad_norm": 0.04881924018263817,
"learning_rate": 9.07777509302564e-06,
"loss": 0.8047,
"step": 1340
},
{
"epoch": 1.6206482593037215,
"grad_norm": 0.04575929790735245,
"learning_rate": 9.061328719733867e-06,
"loss": 0.818,
"step": 1350
},
{
"epoch": 1.6326530612244898,
"grad_norm": 0.047302693128585815,
"learning_rate": 9.044752168537104e-06,
"loss": 0.8255,
"step": 1360
},
{
"epoch": 1.644657863145258,
"grad_norm": 0.048905596137046814,
"learning_rate": 9.028045970764095e-06,
"loss": 0.8033,
"step": 1370
},
{
"epoch": 1.6566626650660266,
"grad_norm": 0.04502958804368973,
"learning_rate": 9.01121066189916e-06,
"loss": 0.8292,
"step": 1380
},
{
"epoch": 1.6686674669867947,
"grad_norm": 0.04705483838915825,
"learning_rate": 8.994246781565012e-06,
"loss": 0.8343,
"step": 1390
},
{
"epoch": 1.680672268907563,
"grad_norm": 0.04515758529305458,
"learning_rate": 8.977154873505478e-06,
"loss": 0.8458,
"step": 1400
},
{
"epoch": 1.6926770708283314,
"grad_norm": 0.05474863573908806,
"learning_rate": 8.959935485568054e-06,
"loss": 0.8044,
"step": 1410
},
{
"epoch": 1.7046818727490995,
"grad_norm": 0.044155530631542206,
"learning_rate": 8.942589169686356e-06,
"loss": 0.8499,
"step": 1420
},
{
"epoch": 1.716686674669868,
"grad_norm": 0.049613580107688904,
"learning_rate": 8.925116481862426e-06,
"loss": 0.8643,
"step": 1430
},
{
"epoch": 1.7286914765906363,
"grad_norm": 0.05003758892416954,
"learning_rate": 8.907517982148913e-06,
"loss": 0.8944,
"step": 1440
},
{
"epoch": 1.7406962785114044,
"grad_norm": 0.048006169497966766,
"learning_rate": 8.889794234631112e-06,
"loss": 0.8063,
"step": 1450
},
{
"epoch": 1.752701080432173,
"grad_norm": 0.04439476877450943,
"learning_rate": 8.871945807408895e-06,
"loss": 0.8679,
"step": 1460
},
{
"epoch": 1.7647058823529411,
"grad_norm": 0.04884785786271095,
"learning_rate": 8.853973272578495e-06,
"loss": 0.8433,
"step": 1470
},
{
"epoch": 1.7767106842737095,
"grad_norm": 0.050317153334617615,
"learning_rate": 8.835877206214172e-06,
"loss": 0.8394,
"step": 1480
},
{
"epoch": 1.7887154861944778,
"grad_norm": 0.04775743559002876,
"learning_rate": 8.817658188349745e-06,
"loss": 0.8309,
"step": 1490
},
{
"epoch": 1.800720288115246,
"grad_norm": 0.04834270477294922,
"learning_rate": 8.79931680296e-06,
"loss": 0.8414,
"step": 1500
},
{
"epoch": 1.8127250900360146,
"grad_norm": 0.04563102871179581,
"learning_rate": 8.780853637941976e-06,
"loss": 0.8934,
"step": 1510
},
{
"epoch": 1.8247298919567827,
"grad_norm": 0.04749139025807381,
"learning_rate": 8.762269285096117e-06,
"loss": 0.8946,
"step": 1520
},
{
"epoch": 1.836734693877551,
"grad_norm": 0.04547303169965744,
"learning_rate": 8.743564340107305e-06,
"loss": 0.7862,
"step": 1530
},
{
"epoch": 1.8487394957983194,
"grad_norm": 0.043644532561302185,
"learning_rate": 8.724739402525765e-06,
"loss": 0.8096,
"step": 1540
},
{
"epoch": 1.8607442977190876,
"grad_norm": 0.050219349563121796,
"learning_rate": 8.705795075747854e-06,
"loss": 0.798,
"step": 1550
},
{
"epoch": 1.872749099639856,
"grad_norm": 0.04704947769641876,
"learning_rate": 8.686731966996707e-06,
"loss": 0.809,
"step": 1560
},
{
"epoch": 1.8847539015606243,
"grad_norm": 0.05036606639623642,
"learning_rate": 8.667550687302784e-06,
"loss": 0.8156,
"step": 1570
},
{
"epoch": 1.8967587034813924,
"grad_norm": 0.05243535712361336,
"learning_rate": 8.648251851484284e-06,
"loss": 0.772,
"step": 1580
},
{
"epoch": 1.908763505402161,
"grad_norm": 0.0510157085955143,
"learning_rate": 8.628836078127435e-06,
"loss": 0.836,
"step": 1590
},
{
"epoch": 1.9207683073229291,
"grad_norm": 0.04631471633911133,
"learning_rate": 8.609303989566673e-06,
"loss": 0.8281,
"step": 1600
},
{
"epoch": 1.9327731092436975,
"grad_norm": 0.04962284117937088,
"learning_rate": 8.589656211864678e-06,
"loss": 0.831,
"step": 1610
},
{
"epoch": 1.9447779111644659,
"grad_norm": 0.054218314588069916,
"learning_rate": 8.569893374792328e-06,
"loss": 0.7904,
"step": 1620
},
{
"epoch": 1.956782713085234,
"grad_norm": 0.05222290754318237,
"learning_rate": 8.5500161118085e-06,
"loss": 0.8141,
"step": 1630
},
{
"epoch": 1.9687875150060024,
"grad_norm": 0.04537142068147659,
"learning_rate": 8.530025060039763e-06,
"loss": 0.8605,
"step": 1640
},
{
"epoch": 1.9807923169267707,
"grad_norm": 0.05147700011730194,
"learning_rate": 8.509920860259973e-06,
"loss": 0.8549,
"step": 1650
},
{
"epoch": 1.9927971188475389,
"grad_norm": 0.052020113915205,
"learning_rate": 8.489704156869711e-06,
"loss": 0.8093,
"step": 1660
},
{
"epoch": 2.0048019207683074,
"grad_norm": 0.05193515121936798,
"learning_rate": 8.469375597875648e-06,
"loss": 0.8481,
"step": 1670
},
{
"epoch": 2.0168067226890756,
"grad_norm": 0.04633639007806778,
"learning_rate": 8.448935834869763e-06,
"loss": 0.8627,
"step": 1680
},
{
"epoch": 2.028811524609844,
"grad_norm": 0.05221736803650856,
"learning_rate": 8.428385523008463e-06,
"loss": 0.8339,
"step": 1690
},
{
"epoch": 2.0408163265306123,
"grad_norm": 0.04951602965593338,
"learning_rate": 8.40772532099158e-06,
"loss": 0.8483,
"step": 1700
},
{
"epoch": 2.0528211284513804,
"grad_norm": 0.053726278245449066,
"learning_rate": 8.386955891041255e-06,
"loss": 0.8191,
"step": 1710
},
{
"epoch": 2.064825930372149,
"grad_norm": 0.04682547226548195,
"learning_rate": 8.366077898880725e-06,
"loss": 0.8037,
"step": 1720
},
{
"epoch": 2.076830732292917,
"grad_norm": 0.04982489347457886,
"learning_rate": 8.345092013712966e-06,
"loss": 0.7827,
"step": 1730
},
{
"epoch": 2.0888355342136853,
"grad_norm": 0.05112074315547943,
"learning_rate": 8.323998908199256e-06,
"loss": 0.8163,
"step": 1740
},
{
"epoch": 2.100840336134454,
"grad_norm": 0.05606475844979286,
"learning_rate": 8.302799258437611e-06,
"loss": 0.8447,
"step": 1750
},
{
"epoch": 2.112845138055222,
"grad_norm": 0.05271725356578827,
"learning_rate": 8.281493743941112e-06,
"loss": 0.8125,
"step": 1760
},
{
"epoch": 2.1248499399759906,
"grad_norm": 0.05468921363353729,
"learning_rate": 8.260083047616126e-06,
"loss": 0.8326,
"step": 1770
},
{
"epoch": 2.1368547418967587,
"grad_norm": 0.05409565195441246,
"learning_rate": 8.238567855740413e-06,
"loss": 0.8082,
"step": 1780
},
{
"epoch": 2.148859543817527,
"grad_norm": 0.05312246084213257,
"learning_rate": 8.216948857941143e-06,
"loss": 0.8137,
"step": 1790
},
{
"epoch": 2.1608643457382954,
"grad_norm": 0.05429564416408539,
"learning_rate": 8.19522674717277e-06,
"loss": 0.8161,
"step": 1800
},
{
"epoch": 2.1728691476590636,
"grad_norm": 0.049361683428287506,
"learning_rate": 8.173402219694838e-06,
"loss": 0.8547,
"step": 1810
},
{
"epoch": 2.184873949579832,
"grad_norm": 0.050867076963186264,
"learning_rate": 8.151475975049653e-06,
"loss": 0.788,
"step": 1820
},
{
"epoch": 2.1968787515006003,
"grad_norm": 0.051545899361371994,
"learning_rate": 8.129448716039863e-06,
"loss": 0.8079,
"step": 1830
},
{
"epoch": 2.2088835534213684,
"grad_norm": 0.05476554483175278,
"learning_rate": 8.107321148705941e-06,
"loss": 0.8123,
"step": 1840
},
{
"epoch": 2.220888355342137,
"grad_norm": 0.05693934112787247,
"learning_rate": 8.085093982303539e-06,
"loss": 0.815,
"step": 1850
},
{
"epoch": 2.232893157262905,
"grad_norm": 0.057461485266685486,
"learning_rate": 8.062767929280764e-06,
"loss": 0.7947,
"step": 1860
},
{
"epoch": 2.2448979591836733,
"grad_norm": 0.057199206203222275,
"learning_rate": 8.040343705255338e-06,
"loss": 0.815,
"step": 1870
},
{
"epoch": 2.256902761104442,
"grad_norm": 0.05683717131614685,
"learning_rate": 8.017822028991666e-06,
"loss": 0.8979,
"step": 1880
},
{
"epoch": 2.26890756302521,
"grad_norm": 0.05384090170264244,
"learning_rate": 7.995203622377783e-06,
"loss": 0.8424,
"step": 1890
},
{
"epoch": 2.280912364945978,
"grad_norm": 0.05016244202852249,
"learning_rate": 7.972489210402232e-06,
"loss": 0.7551,
"step": 1900
},
{
"epoch": 2.2929171668667467,
"grad_norm": 0.05437694489955902,
"learning_rate": 7.949679521130823e-06,
"loss": 0.8446,
"step": 1910
},
{
"epoch": 2.304921968787515,
"grad_norm": 0.055492572486400604,
"learning_rate": 7.926775285683281e-06,
"loss": 0.8182,
"step": 1920
},
{
"epoch": 2.3169267707082835,
"grad_norm": 0.05502455681562424,
"learning_rate": 7.903777238209835e-06,
"loss": 0.8181,
"step": 1930
},
{
"epoch": 2.3289315726290516,
"grad_norm": 0.05110107734799385,
"learning_rate": 7.880686115867664e-06,
"loss": 0.8521,
"step": 1940
},
{
"epoch": 2.3409363745498197,
"grad_norm": 0.0539153516292572,
"learning_rate": 7.857502658797284e-06,
"loss": 0.8272,
"step": 1950
},
{
"epoch": 2.3529411764705883,
"grad_norm": 0.05461420118808746,
"learning_rate": 7.83422761009882e-06,
"loss": 0.8413,
"step": 1960
},
{
"epoch": 2.3649459783913565,
"grad_norm": 0.05569763854146004,
"learning_rate": 7.810861715808184e-06,
"loss": 0.8687,
"step": 1970
},
{
"epoch": 2.376950780312125,
"grad_norm": 0.055215220898389816,
"learning_rate": 7.787405724873162e-06,
"loss": 0.8196,
"step": 1980
},
{
"epoch": 2.388955582232893,
"grad_norm": 0.06094464287161827,
"learning_rate": 7.76386038912942e-06,
"loss": 0.7921,
"step": 1990
},
{
"epoch": 2.4009603841536613,
"grad_norm": 0.05856522545218468,
"learning_rate": 7.740226463276385e-06,
"loss": 0.8332,
"step": 2000
},
{
"epoch": 2.41296518607443,
"grad_norm": 0.05313005670905113,
"learning_rate": 7.716504704853076e-06,
"loss": 0.8076,
"step": 2010
},
{
"epoch": 2.424969987995198,
"grad_norm": 0.05347991734743118,
"learning_rate": 7.692695874213802e-06,
"loss": 0.8328,
"step": 2020
},
{
"epoch": 2.4369747899159666,
"grad_norm": 0.058754704892635345,
"learning_rate": 7.668800734503812e-06,
"loss": 0.8445,
"step": 2030
},
{
"epoch": 2.4489795918367347,
"grad_norm": 0.05653975158929825,
"learning_rate": 7.644820051634813e-06,
"loss": 0.8378,
"step": 2040
},
{
"epoch": 2.460984393757503,
"grad_norm": 0.05895698443055153,
"learning_rate": 7.620754594260433e-06,
"loss": 0.8214,
"step": 2050
},
{
"epoch": 2.4729891956782715,
"grad_norm": 0.056038543581962585,
"learning_rate": 7.596605133751583e-06,
"loss": 0.792,
"step": 2060
},
{
"epoch": 2.4849939975990396,
"grad_norm": 0.0582413524389267,
"learning_rate": 7.5723724441717225e-06,
"loss": 0.8292,
"step": 2070
},
{
"epoch": 2.4969987995198077,
"grad_norm": 0.05528351292014122,
"learning_rate": 7.548057302252063e-06,
"loss": 0.7871,
"step": 2080
},
{
"epoch": 2.5090036014405763,
"grad_norm": 0.05144486203789711,
"learning_rate": 7.523660487366658e-06,
"loss": 0.7738,
"step": 2090
},
{
"epoch": 2.5210084033613445,
"grad_norm": 0.05947820469737053,
"learning_rate": 7.4991827815074306e-06,
"loss": 0.7955,
"step": 2100
},
{
"epoch": 2.5330132052821126,
"grad_norm": 0.05376136675477028,
"learning_rate": 7.474624969259101e-06,
"loss": 0.8145,
"step": 2110
},
{
"epoch": 2.545018007202881,
"grad_norm": 0.05993662402033806,
"learning_rate": 7.449987837774049e-06,
"loss": 0.8357,
"step": 2120
},
{
"epoch": 2.5570228091236493,
"grad_norm": 0.05299674719572067,
"learning_rate": 7.425272176747068e-06,
"loss": 0.8185,
"step": 2130
},
{
"epoch": 2.569027611044418,
"grad_norm": 0.06037457659840584,
"learning_rate": 7.400478778390068e-06,
"loss": 0.8327,
"step": 2140
},
{
"epoch": 2.581032412965186,
"grad_norm": 0.054685529321432114,
"learning_rate": 7.375608437406674e-06,
"loss": 0.7932,
"step": 2150
},
{
"epoch": 2.593037214885954,
"grad_norm": 0.061256833374500275,
"learning_rate": 7.350661950966755e-06,
"loss": 0.8395,
"step": 2160
},
{
"epoch": 2.6050420168067228,
"grad_norm": 0.056471362709999084,
"learning_rate": 7.325640118680872e-06,
"loss": 0.7806,
"step": 2170
},
{
"epoch": 2.617046818727491,
"grad_norm": 0.0574592649936676,
"learning_rate": 7.300543742574654e-06,
"loss": 0.8243,
"step": 2180
},
{
"epoch": 2.6290516206482595,
"grad_norm": 0.05507262796163559,
"learning_rate": 7.275373627063079e-06,
"loss": 0.8201,
"step": 2190
},
{
"epoch": 2.6410564225690276,
"grad_norm": 0.05754077062010765,
"learning_rate": 7.2501305789247e-06,
"loss": 0.8003,
"step": 2200
},
{
"epoch": 2.6530612244897958,
"grad_norm": 0.055398665368556976,
"learning_rate": 7.224815407275782e-06,
"loss": 0.8291,
"step": 2210
},
{
"epoch": 2.6650660264105643,
"grad_norm": 0.057937685400247574,
"learning_rate": 7.199428923544366e-06,
"loss": 0.7827,
"step": 2220
},
{
"epoch": 2.6770708283313325,
"grad_norm": 0.05905699357390404,
"learning_rate": 7.173971941444262e-06,
"loss": 0.8482,
"step": 2230
},
{
"epoch": 2.689075630252101,
"grad_norm": 0.05891605094075203,
"learning_rate": 7.148445276948968e-06,
"loss": 0.848,
"step": 2240
},
{
"epoch": 2.701080432172869,
"grad_norm": 0.05597429722547531,
"learning_rate": 7.122849748265512e-06,
"loss": 0.8288,
"step": 2250
},
{
"epoch": 2.7130852340936373,
"grad_norm": 0.0589454285800457,
"learning_rate": 7.097186175808229e-06,
"loss": 0.8711,
"step": 2260
},
{
"epoch": 2.725090036014406,
"grad_norm": 0.056011561304330826,
"learning_rate": 7.071455382172465e-06,
"loss": 0.7916,
"step": 2270
},
{
"epoch": 2.737094837935174,
"grad_norm": 0.06104077771306038,
"learning_rate": 7.045658192108204e-06,
"loss": 0.8378,
"step": 2280
},
{
"epoch": 2.7490996398559426,
"grad_norm": 0.060269393026828766,
"learning_rate": 7.019795432493644e-06,
"loss": 0.8296,
"step": 2290
},
{
"epoch": 2.7611044417767108,
"grad_norm": 0.05378441885113716,
"learning_rate": 6.9938679323086785e-06,
"loss": 0.8312,
"step": 2300
},
{
"epoch": 2.773109243697479,
"grad_norm": 0.05314185470342636,
"learning_rate": 6.967876522608339e-06,
"loss": 0.8127,
"step": 2310
},
{
"epoch": 2.785114045618247,
"grad_norm": 0.056222010403871536,
"learning_rate": 6.941822036496146e-06,
"loss": 0.8065,
"step": 2320
},
{
"epoch": 2.7971188475390156,
"grad_norm": 0.05297733470797539,
"learning_rate": 6.915705309097413e-06,
"loss": 0.776,
"step": 2330
},
{
"epoch": 2.8091236494597838,
"grad_norm": 0.05919932574033737,
"learning_rate": 6.889527177532476e-06,
"loss": 0.8115,
"step": 2340
},
{
"epoch": 2.8211284513805523,
"grad_norm": 0.05913585424423218,
"learning_rate": 6.863288480889858e-06,
"loss": 0.8291,
"step": 2350
},
{
"epoch": 2.8331332533013205,
"grad_norm": 0.05690290406346321,
"learning_rate": 6.836990060199379e-06,
"loss": 0.7931,
"step": 2360
},
{
"epoch": 2.8451380552220886,
"grad_norm": 0.058122627437114716,
"learning_rate": 6.810632758405195e-06,
"loss": 0.8436,
"step": 2370
},
{
"epoch": 2.857142857142857,
"grad_norm": 0.0562375970184803,
"learning_rate": 6.7842174203387804e-06,
"loss": 0.8147,
"step": 2380
},
{
"epoch": 2.8691476590636253,
"grad_norm": 0.06252553313970566,
"learning_rate": 6.757744892691847e-06,
"loss": 0.8408,
"step": 2390
},
{
"epoch": 2.881152460984394,
"grad_norm": 0.06220783293247223,
"learning_rate": 6.731216023989207e-06,
"loss": 0.8358,
"step": 2400
},
{
"epoch": 2.893157262905162,
"grad_norm": 0.05928228050470352,
"learning_rate": 6.704631664561574e-06,
"loss": 0.835,
"step": 2410
},
{
"epoch": 2.90516206482593,
"grad_norm": 0.06004854291677475,
"learning_rate": 6.67799266651831e-06,
"loss": 0.8417,
"step": 2420
},
{
"epoch": 2.917166866746699,
"grad_norm": 0.05948818475008011,
"learning_rate": 6.651299883720106e-06,
"loss": 0.8026,
"step": 2430
},
{
"epoch": 2.929171668667467,
"grad_norm": 0.05721951648592949,
"learning_rate": 6.62455417175162e-06,
"loss": 0.8143,
"step": 2440
},
{
"epoch": 2.9411764705882355,
"grad_norm": 0.060878921300172806,
"learning_rate": 6.597756387894051e-06,
"loss": 0.7804,
"step": 2450
},
{
"epoch": 2.9531812725090036,
"grad_norm": 0.06494616717100143,
"learning_rate": 6.57090739109766e-06,
"loss": 0.8528,
"step": 2460
},
{
"epoch": 2.965186074429772,
"grad_norm": 0.061719998717308044,
"learning_rate": 6.544008041954236e-06,
"loss": 0.8341,
"step": 2470
},
{
"epoch": 2.9771908763505404,
"grad_norm": 0.06074754521250725,
"learning_rate": 6.517059202669516e-06,
"loss": 0.8163,
"step": 2480
},
{
"epoch": 2.9891956782713085,
"grad_norm": 0.05759371444582939,
"learning_rate": 6.4900617370355435e-06,
"loss": 0.797,
"step": 2490
},
{
"epoch": 3.0012004801920766,
"grad_norm": 0.06447486579418182,
"learning_rate": 6.4630165104029875e-06,
"loss": 0.8478,
"step": 2500
},
{
"epoch": 3.013205282112845,
"grad_norm": 0.05972028896212578,
"learning_rate": 6.435924389653398e-06,
"loss": 0.7434,
"step": 2510
},
{
"epoch": 3.0252100840336134,
"grad_norm": 0.06719402968883514,
"learning_rate": 6.408786243171426e-06,
"loss": 0.8223,
"step": 2520
},
{
"epoch": 3.037214885954382,
"grad_norm": 0.06629668921232224,
"learning_rate": 6.3816029408169846e-06,
"loss": 0.8201,
"step": 2530
},
{
"epoch": 3.04921968787515,
"grad_norm": 0.06306376308202744,
"learning_rate": 6.354375353897376e-06,
"loss": 0.8387,
"step": 2540
},
{
"epoch": 3.061224489795918,
"grad_norm": 0.05803842842578888,
"learning_rate": 6.3271043551393485e-06,
"loss": 0.8517,
"step": 2550
},
{
"epoch": 3.073229291716687,
"grad_norm": 0.05941968038678169,
"learning_rate": 6.299790818661135e-06,
"loss": 0.8243,
"step": 2560
},
{
"epoch": 3.085234093637455,
"grad_norm": 0.06141958758234978,
"learning_rate": 6.272435619944436e-06,
"loss": 0.8409,
"step": 2570
},
{
"epoch": 3.097238895558223,
"grad_norm": 0.05685436725616455,
"learning_rate": 6.245039635806345e-06,
"loss": 0.8159,
"step": 2580
},
{
"epoch": 3.1092436974789917,
"grad_norm": 0.058166082948446274,
"learning_rate": 6.2176037443712575e-06,
"loss": 0.8486,
"step": 2590
},
{
"epoch": 3.12124849939976,
"grad_norm": 0.05954616516828537,
"learning_rate": 6.190128825042716e-06,
"loss": 0.8417,
"step": 2600
},
{
"epoch": 3.1332533013205284,
"grad_norm": 0.05970808118581772,
"learning_rate": 6.1626157584752285e-06,
"loss": 0.7822,
"step": 2610
},
{
"epoch": 3.1452581032412965,
"grad_norm": 0.06996554881334305,
"learning_rate": 6.135065426546034e-06,
"loss": 0.8074,
"step": 2620
},
{
"epoch": 3.1572629051620646,
"grad_norm": 0.0612507127225399,
"learning_rate": 6.107478712326843e-06,
"loss": 0.819,
"step": 2630
},
{
"epoch": 3.1692677070828332,
"grad_norm": 0.06266125291585922,
"learning_rate": 6.079856500055526e-06,
"loss": 0.8005,
"step": 2640
},
{
"epoch": 3.1812725090036014,
"grad_norm": 0.06757565587759018,
"learning_rate": 6.052199675107773e-06,
"loss": 0.8025,
"step": 2650
},
{
"epoch": 3.19327731092437,
"grad_norm": 0.060953736305236816,
"learning_rate": 6.024509123968722e-06,
"loss": 0.8195,
"step": 2660
},
{
"epoch": 3.205282112845138,
"grad_norm": 0.074392169713974,
"learning_rate": 5.996785734204528e-06,
"loss": 0.8411,
"step": 2670
},
{
"epoch": 3.2172869147659062,
"grad_norm": 0.06586959213018417,
"learning_rate": 5.969030394433928e-06,
"loss": 0.8246,
"step": 2680
},
{
"epoch": 3.229291716686675,
"grad_norm": 0.06062298268079758,
"learning_rate": 5.941243994299758e-06,
"loss": 0.8361,
"step": 2690
},
{
"epoch": 3.241296518607443,
"grad_norm": 0.06378082185983658,
"learning_rate": 5.913427424440428e-06,
"loss": 0.8121,
"step": 2700
},
{
"epoch": 3.2533013205282115,
"grad_norm": 0.061464712023735046,
"learning_rate": 5.885581576461376e-06,
"loss": 0.813,
"step": 2710
},
{
"epoch": 3.2653061224489797,
"grad_norm": 0.06229786574840546,
"learning_rate": 5.857707342906501e-06,
"loss": 0.7757,
"step": 2720
},
{
"epoch": 3.277310924369748,
"grad_norm": 0.0691327154636383,
"learning_rate": 5.82980561722954e-06,
"loss": 0.8063,
"step": 2730
},
{
"epoch": 3.2893157262905164,
"grad_norm": 0.06736263632774353,
"learning_rate": 5.8018772937654366e-06,
"loss": 0.8352,
"step": 2740
},
{
"epoch": 3.3013205282112845,
"grad_norm": 0.06093461439013481,
"learning_rate": 5.773923267701676e-06,
"loss": 0.8216,
"step": 2750
},
{
"epoch": 3.3133253301320527,
"grad_norm": 0.06377444416284561,
"learning_rate": 5.745944435049586e-06,
"loss": 0.8095,
"step": 2760
},
{
"epoch": 3.3253301320528212,
"grad_norm": 0.06320587545633316,
"learning_rate": 5.717941692615626e-06,
"loss": 0.828,
"step": 2770
},
{
"epoch": 3.3373349339735894,
"grad_norm": 0.06745033711194992,
"learning_rate": 5.689915937972634e-06,
"loss": 0.8277,
"step": 2780
},
{
"epoch": 3.3493397358943575,
"grad_norm": 0.06507708877325058,
"learning_rate": 5.661868069431056e-06,
"loss": 0.8177,
"step": 2790
},
{
"epoch": 3.361344537815126,
"grad_norm": 0.0653938576579094,
"learning_rate": 5.633798986010159e-06,
"loss": 0.8302,
"step": 2800
},
{
"epoch": 3.3733493397358942,
"grad_norm": 0.06798886507749557,
"learning_rate": 5.605709587409216e-06,
"loss": 0.7741,
"step": 2810
},
{
"epoch": 3.385354141656663,
"grad_norm": 0.06364308297634125,
"learning_rate": 5.577600773978651e-06,
"loss": 0.7791,
"step": 2820
},
{
"epoch": 3.397358943577431,
"grad_norm": 0.06712163984775543,
"learning_rate": 5.5494734466912005e-06,
"loss": 0.8316,
"step": 2830
},
{
"epoch": 3.409363745498199,
"grad_norm": 0.06373037397861481,
"learning_rate": 5.521328507113027e-06,
"loss": 0.7897,
"step": 2840
},
{
"epoch": 3.4213685474189677,
"grad_norm": 0.06734546273946762,
"learning_rate": 5.493166857374818e-06,
"loss": 0.8325,
"step": 2850
},
{
"epoch": 3.433373349339736,
"grad_norm": 0.0662597045302391,
"learning_rate": 5.464989400142868e-06,
"loss": 0.8304,
"step": 2860
},
{
"epoch": 3.4453781512605044,
"grad_norm": 0.06525112688541412,
"learning_rate": 5.436797038590159e-06,
"loss": 0.869,
"step": 2870
},
{
"epoch": 3.4573829531812725,
"grad_norm": 0.06612803786993027,
"learning_rate": 5.408590676367395e-06,
"loss": 0.8372,
"step": 2880
},
{
"epoch": 3.4693877551020407,
"grad_norm": 0.07390609383583069,
"learning_rate": 5.380371217574046e-06,
"loss": 0.8011,
"step": 2890
},
{
"epoch": 3.4813925570228093,
"grad_norm": 0.06323985010385513,
"learning_rate": 5.352139566729366e-06,
"loss": 0.7972,
"step": 2900
},
{
"epoch": 3.4933973589435774,
"grad_norm": 0.06164005026221275,
"learning_rate": 5.323896628743401e-06,
"loss": 0.8099,
"step": 2910
},
{
"epoch": 3.505402160864346,
"grad_norm": 0.06231848523020744,
"learning_rate": 5.295643308887986e-06,
"loss": 0.8172,
"step": 2920
},
{
"epoch": 3.517406962785114,
"grad_norm": 0.06743638217449188,
"learning_rate": 5.2673805127677255e-06,
"loss": 0.7886,
"step": 2930
},
{
"epoch": 3.5294117647058822,
"grad_norm": 0.0683428943157196,
"learning_rate": 5.239109146290965e-06,
"loss": 0.8155,
"step": 2940
},
{
"epoch": 3.5414165666266504,
"grad_norm": 0.06549340486526489,
"learning_rate": 5.210830115640758e-06,
"loss": 0.7952,
"step": 2950
},
{
"epoch": 3.553421368547419,
"grad_norm": 0.07066480070352554,
"learning_rate": 5.1825443272458194e-06,
"loss": 0.8514,
"step": 2960
},
{
"epoch": 3.5654261704681876,
"grad_norm": 0.0690867155790329,
"learning_rate": 5.154252687751469e-06,
"loss": 0.837,
"step": 2970
},
{
"epoch": 3.5774309723889557,
"grad_norm": 0.07441065460443497,
"learning_rate": 5.125956103990567e-06,
"loss": 0.7944,
"step": 2980
},
{
"epoch": 3.589435774309724,
"grad_norm": 0.06854850053787231,
"learning_rate": 5.0976554829544626e-06,
"loss": 0.8022,
"step": 2990
},
{
"epoch": 3.601440576230492,
"grad_norm": 0.06906180083751678,
"learning_rate": 5.069351731763903e-06,
"loss": 0.7914,
"step": 3000
},
{
"epoch": 3.6134453781512605,
"grad_norm": 0.06705962121486664,
"learning_rate": 5.041045757639972e-06,
"loss": 0.8043,
"step": 3010
},
{
"epoch": 3.6254501800720287,
"grad_norm": 0.07068238407373428,
"learning_rate": 5.0127384678750005e-06,
"loss": 0.8737,
"step": 3020
},
{
"epoch": 3.6374549819927973,
"grad_norm": 0.07324163615703583,
"learning_rate": 4.984430769803493e-06,
"loss": 0.8224,
"step": 3030
},
{
"epoch": 3.6494597839135654,
"grad_norm": 0.06777600198984146,
"learning_rate": 4.95612357077304e-06,
"loss": 0.8424,
"step": 3040
},
{
"epoch": 3.6614645858343335,
"grad_norm": 0.07339445501565933,
"learning_rate": 4.927817778115239e-06,
"loss": 0.8345,
"step": 3050
},
{
"epoch": 3.673469387755102,
"grad_norm": 0.07379196584224701,
"learning_rate": 4.899514299116603e-06,
"loss": 0.7919,
"step": 3060
},
{
"epoch": 3.6854741896758703,
"grad_norm": 0.06788371503353119,
"learning_rate": 4.871214040989492e-06,
"loss": 0.7914,
"step": 3070
},
{
"epoch": 3.697478991596639,
"grad_norm": 0.07233492285013199,
"learning_rate": 4.8429179108430246e-06,
"loss": 0.8388,
"step": 3080
},
{
"epoch": 3.709483793517407,
"grad_norm": 0.07236211746931076,
"learning_rate": 4.814626815654e-06,
"loss": 0.8619,
"step": 3090
},
{
"epoch": 3.721488595438175,
"grad_norm": 0.06984652578830719,
"learning_rate": 4.7863416622378426e-06,
"loss": 0.8038,
"step": 3100
},
{
"epoch": 3.7334933973589437,
"grad_norm": 0.06518954783678055,
"learning_rate": 4.7580633572195156e-06,
"loss": 0.8219,
"step": 3110
},
{
"epoch": 3.745498199279712,
"grad_norm": 0.06867209821939468,
"learning_rate": 4.729792807004472e-06,
"loss": 0.852,
"step": 3120
},
{
"epoch": 3.7575030012004804,
"grad_norm": 0.06609649956226349,
"learning_rate": 4.701530917749604e-06,
"loss": 0.803,
"step": 3130
},
{
"epoch": 3.7695078031212486,
"grad_norm": 0.06952491402626038,
"learning_rate": 4.673278595334191e-06,
"loss": 0.8405,
"step": 3140
},
{
"epoch": 3.7815126050420167,
"grad_norm": 0.0671219453215599,
"learning_rate": 4.645036745330863e-06,
"loss": 0.7897,
"step": 3150
},
{
"epoch": 3.7935174069627853,
"grad_norm": 0.06863446533679962,
"learning_rate": 4.616806272976584e-06,
"loss": 0.8096,
"step": 3160
},
{
"epoch": 3.8055222088835534,
"grad_norm": 0.07853756099939346,
"learning_rate": 4.588588083143625e-06,
"loss": 0.8143,
"step": 3170
},
{
"epoch": 3.817527010804322,
"grad_norm": 0.07341323047876358,
"learning_rate": 4.56038308031057e-06,
"loss": 0.8079,
"step": 3180
},
{
"epoch": 3.82953181272509,
"grad_norm": 0.06746179610490799,
"learning_rate": 4.5321921685333105e-06,
"loss": 0.8208,
"step": 3190
},
{
"epoch": 3.8415366146458583,
"grad_norm": 0.07139655202627182,
"learning_rate": 4.504016251416085e-06,
"loss": 0.8224,
"step": 3200
},
{
"epoch": 3.8535414165666264,
"grad_norm": 0.0728779211640358,
"learning_rate": 4.475856232082506e-06,
"loss": 0.8249,
"step": 3210
},
{
"epoch": 3.865546218487395,
"grad_norm": 0.06635065376758575,
"learning_rate": 4.447713013146612e-06,
"loss": 0.8219,
"step": 3220
},
{
"epoch": 3.877551020408163,
"grad_norm": 0.06454911828041077,
"learning_rate": 4.419587496683939e-06,
"loss": 0.8435,
"step": 3230
},
{
"epoch": 3.8895558223289317,
"grad_norm": 0.07131163030862808,
"learning_rate": 4.391480584202605e-06,
"loss": 0.8159,
"step": 3240
},
{
"epoch": 3.9015606242497,
"grad_norm": 0.07186614722013474,
"learning_rate": 4.363393176614411e-06,
"loss": 0.7746,
"step": 3250
},
{
"epoch": 3.913565426170468,
"grad_norm": 0.07490645349025726,
"learning_rate": 4.335326174205976e-06,
"loss": 0.8523,
"step": 3260
},
{
"epoch": 3.9255702280912366,
"grad_norm": 0.07318370789289474,
"learning_rate": 4.3072804766098605e-06,
"loss": 0.8141,
"step": 3270
},
{
"epoch": 3.9375750300120047,
"grad_norm": 0.07088599354028702,
"learning_rate": 4.279256982775744e-06,
"loss": 0.7967,
"step": 3280
},
{
"epoch": 3.9495798319327733,
"grad_norm": 0.07442790269851685,
"learning_rate": 4.251256590941615e-06,
"loss": 0.8148,
"step": 3290
},
{
"epoch": 3.9615846338535414,
"grad_norm": 0.07056485116481781,
"learning_rate": 4.223280198604962e-06,
"loss": 0.8109,
"step": 3300
},
{
"epoch": 3.9735894357743096,
"grad_norm": 0.07370737940073013,
"learning_rate": 4.195328702494024e-06,
"loss": 0.7949,
"step": 3310
},
{
"epoch": 3.985594237695078,
"grad_norm": 0.07084988802671432,
"learning_rate": 4.167402998539041e-06,
"loss": 0.8082,
"step": 3320
},
{
"epoch": 3.9975990396158463,
"grad_norm": 0.06937604397535324,
"learning_rate": 4.13950398184353e-06,
"loss": 0.8696,
"step": 3330
},
{
"epoch": 4.009603841536615,
"grad_norm": 0.07246571779251099,
"learning_rate": 4.111632546655608e-06,
"loss": 0.8017,
"step": 3340
},
{
"epoch": 4.021608643457383,
"grad_norm": 0.07561132311820984,
"learning_rate": 4.083789586339315e-06,
"loss": 0.801,
"step": 3350
},
{
"epoch": 4.033613445378151,
"grad_norm": 0.07453512400388718,
"learning_rate": 4.055975993345986e-06,
"loss": 0.8244,
"step": 3360
},
{
"epoch": 4.045618247298919,
"grad_norm": 0.0704115629196167,
"learning_rate": 4.028192659185649e-06,
"loss": 0.8294,
"step": 3370
},
{
"epoch": 4.057623049219688,
"grad_norm": 0.06777546554803848,
"learning_rate": 4.000440474398439e-06,
"loss": 0.8299,
"step": 3380
},
{
"epoch": 4.069627851140456,
"grad_norm": 0.08122120797634125,
"learning_rate": 3.972720328526059e-06,
"loss": 0.7945,
"step": 3390
},
{
"epoch": 4.081632653061225,
"grad_norm": 0.0788722038269043,
"learning_rate": 3.945033110083273e-06,
"loss": 0.8077,
"step": 3400
},
{
"epoch": 4.093637454981993,
"grad_norm": 0.0742715448141098,
"learning_rate": 3.917379706529416e-06,
"loss": 0.8234,
"step": 3410
},
{
"epoch": 4.105642256902761,
"grad_norm": 0.07386495918035507,
"learning_rate": 3.889761004239958e-06,
"loss": 0.8523,
"step": 3420
},
{
"epoch": 4.117647058823529,
"grad_norm": 0.07414324581623077,
"learning_rate": 3.862177888478082e-06,
"loss": 0.8185,
"step": 3430
},
{
"epoch": 4.129651860744298,
"grad_norm": 0.07434631884098053,
"learning_rate": 3.834631243366318e-06,
"loss": 0.8613,
"step": 3440
},
{
"epoch": 4.141656662665066,
"grad_norm": 0.07150658965110779,
"learning_rate": 3.8071219518582032e-06,
"loss": 0.7821,
"step": 3450
},
{
"epoch": 4.153661464585834,
"grad_norm": 0.06888636946678162,
"learning_rate": 3.7796508957099744e-06,
"loss": 0.7794,
"step": 3460
},
{
"epoch": 4.165666266506602,
"grad_norm": 0.07071466743946075,
"learning_rate": 3.752218955452311e-06,
"loss": 0.8367,
"step": 3470
},
{
"epoch": 4.177671068427371,
"grad_norm": 0.07348918169736862,
"learning_rate": 3.7248270103621128e-06,
"loss": 0.8085,
"step": 3480
},
{
"epoch": 4.18967587034814,
"grad_norm": 0.07582999765872955,
"learning_rate": 3.697475938434305e-06,
"loss": 0.837,
"step": 3490
},
{
"epoch": 4.201680672268908,
"grad_norm": 0.06943266838788986,
"learning_rate": 3.6701666163537125e-06,
"loss": 0.8058,
"step": 3500
},
{
"epoch": 4.213685474189676,
"grad_norm": 0.08077062666416168,
"learning_rate": 3.6428999194669478e-06,
"loss": 0.8466,
"step": 3510
},
{
"epoch": 4.225690276110444,
"grad_norm": 0.07306414097547531,
"learning_rate": 3.6156767217543507e-06,
"loss": 0.8098,
"step": 3520
},
{
"epoch": 4.237695078031212,
"grad_norm": 0.07767776399850845,
"learning_rate": 3.588497895801994e-06,
"loss": 0.8227,
"step": 3530
},
{
"epoch": 4.249699879951981,
"grad_norm": 0.06969287246465683,
"learning_rate": 3.5613643127736884e-06,
"loss": 0.8048,
"step": 3540
},
{
"epoch": 4.261704681872749,
"grad_norm": 0.08369370549917221,
"learning_rate": 3.534276842383074e-06,
"loss": 0.8361,
"step": 3550
},
{
"epoch": 4.2737094837935174,
"grad_norm": 0.07350631803274155,
"learning_rate": 3.5072363528657488e-06,
"loss": 0.8241,
"step": 3560
},
{
"epoch": 4.285714285714286,
"grad_norm": 0.09979286789894104,
"learning_rate": 3.480243710951422e-06,
"loss": 0.8196,
"step": 3570
},
{
"epoch": 4.297719087635054,
"grad_norm": 0.07020698487758636,
"learning_rate": 3.4532997818361456e-06,
"loss": 0.8213,
"step": 3580
},
{
"epoch": 4.309723889555823,
"grad_norm": 0.0725114569067955,
"learning_rate": 3.426405429154581e-06,
"loss": 0.8232,
"step": 3590
},
{
"epoch": 4.321728691476591,
"grad_norm": 0.07283388823270798,
"learning_rate": 3.3995615149523102e-06,
"loss": 0.855,
"step": 3600
},
{
"epoch": 4.333733493397359,
"grad_norm": 0.0759081020951271,
"learning_rate": 3.372768899658212e-06,
"loss": 0.8116,
"step": 3610
},
{
"epoch": 4.345738295318127,
"grad_norm": 0.07164635509252548,
"learning_rate": 3.3460284420568802e-06,
"loss": 0.8147,
"step": 3620
},
{
"epoch": 4.357743097238895,
"grad_norm": 0.07087879627943039,
"learning_rate": 3.3193409992610897e-06,
"loss": 0.7556,
"step": 3630
},
{
"epoch": 4.369747899159664,
"grad_norm": 0.07007823884487152,
"learning_rate": 3.2927074266843394e-06,
"loss": 0.8307,
"step": 3640
},
{
"epoch": 4.3817527010804325,
"grad_norm": 0.07446113228797913,
"learning_rate": 3.26612857801342e-06,
"loss": 0.8155,
"step": 3650
},
{
"epoch": 4.393757503001201,
"grad_norm": 0.07006789743900299,
"learning_rate": 3.2396053051810515e-06,
"loss": 0.8118,
"step": 3660
},
{
"epoch": 4.405762304921969,
"grad_norm": 0.07662977278232574,
"learning_rate": 3.213138458338587e-06,
"loss": 0.8394,
"step": 3670
},
{
"epoch": 4.417767106842737,
"grad_norm": 0.08047199994325638,
"learning_rate": 3.186728885828748e-06,
"loss": 0.8321,
"step": 3680
},
{
"epoch": 4.429771908763505,
"grad_norm": 0.08414141833782196,
"learning_rate": 3.160377434158448e-06,
"loss": 0.7842,
"step": 3690
},
{
"epoch": 4.441776710684274,
"grad_norm": 0.06629359722137451,
"learning_rate": 3.134084947971644e-06,
"loss": 0.7735,
"step": 3700
},
{
"epoch": 4.453781512605042,
"grad_norm": 0.07507917284965515,
"learning_rate": 3.1078522700222713e-06,
"loss": 0.8031,
"step": 3710
},
{
"epoch": 4.46578631452581,
"grad_norm": 0.07528281211853027,
"learning_rate": 3.0816802411472362e-06,
"loss": 0.817,
"step": 3720
},
{
"epoch": 4.4777911164465785,
"grad_norm": 0.07056977599859238,
"learning_rate": 3.05556970023945e-06,
"loss": 0.8379,
"step": 3730
},
{
"epoch": 4.489795918367347,
"grad_norm": 0.07075227051973343,
"learning_rate": 3.029521484220953e-06,
"loss": 0.7759,
"step": 3740
},
{
"epoch": 4.501800720288116,
"grad_norm": 0.08120349049568176,
"learning_rate": 3.003536428016084e-06,
"loss": 0.812,
"step": 3750
},
{
"epoch": 4.513805522208884,
"grad_norm": 0.0773065984249115,
"learning_rate": 2.9776153645247135e-06,
"loss": 0.7805,
"step": 3760
},
{
"epoch": 4.525810324129652,
"grad_norm": 0.08427418768405914,
"learning_rate": 2.951759124595558e-06,
"loss": 0.8452,
"step": 3770
},
{
"epoch": 4.53781512605042,
"grad_norm": 0.07738559693098068,
"learning_rate": 2.9259685369995372e-06,
"loss": 0.8379,
"step": 3780
},
{
"epoch": 4.549819927971188,
"grad_norm": 0.07159377634525299,
"learning_rate": 2.900244428403214e-06,
"loss": 0.8248,
"step": 3790
},
{
"epoch": 4.561824729891956,
"grad_norm": 0.08146795630455017,
"learning_rate": 2.874587623342305e-06,
"loss": 0.7939,
"step": 3800
},
{
"epoch": 4.573829531812725,
"grad_norm": 0.06939620524644852,
"learning_rate": 2.8489989441952357e-06,
"loss": 0.8221,
"step": 3810
},
{
"epoch": 4.5858343337334935,
"grad_norm": 0.0722632110118866,
"learning_rate": 2.8234792111567915e-06,
"loss": 0.8351,
"step": 3820
},
{
"epoch": 4.597839135654262,
"grad_norm": 0.07618506252765656,
"learning_rate": 2.7980292422118282e-06,
"loss": 0.8298,
"step": 3830
},
{
"epoch": 4.60984393757503,
"grad_norm": 0.07517799735069275,
"learning_rate": 2.7726498531090507e-06,
"loss": 0.8025,
"step": 3840
},
{
"epoch": 4.621848739495798,
"grad_norm": 0.07529459148645401,
"learning_rate": 2.747341857334866e-06,
"loss": 0.8004,
"step": 3850
},
{
"epoch": 4.633853541416567,
"grad_norm": 0.07275144010782242,
"learning_rate": 2.7221060660873057e-06,
"loss": 0.8117,
"step": 3860
},
{
"epoch": 4.645858343337335,
"grad_norm": 0.0774557814002037,
"learning_rate": 2.6969432882500266e-06,
"loss": 0.846,
"step": 3870
},
{
"epoch": 4.657863145258103,
"grad_norm": 0.07570381462574005,
"learning_rate": 2.6718543303663912e-06,
"loss": 0.8258,
"step": 3880
},
{
"epoch": 4.669867947178871,
"grad_norm": 0.07305650413036346,
"learning_rate": 2.646839996613599e-06,
"loss": 0.8034,
"step": 3890
},
{
"epoch": 4.6818727490996395,
"grad_norm": 0.0736539289355278,
"learning_rate": 2.621901088776927e-06,
"loss": 0.7702,
"step": 3900
},
{
"epoch": 4.6938775510204085,
"grad_norm": 0.07207903265953064,
"learning_rate": 2.597038406224022e-06,
"loss": 0.7824,
"step": 3910
},
{
"epoch": 4.705882352941177,
"grad_norm": 0.07334660738706589,
"learning_rate": 2.5722527458792743e-06,
"loss": 0.8283,
"step": 3920
},
{
"epoch": 4.717887154861945,
"grad_norm": 0.06604504585266113,
"learning_rate": 2.5475449021982855e-06,
"loss": 0.8122,
"step": 3930
},
{
"epoch": 4.729891956782713,
"grad_norm": 0.07602079212665558,
"learning_rate": 2.5229156671423915e-06,
"loss": 0.7998,
"step": 3940
},
{
"epoch": 4.741896758703481,
"grad_norm": 0.08058540523052216,
"learning_rate": 2.498365830153285e-06,
"loss": 0.8088,
"step": 3950
},
{
"epoch": 4.75390156062425,
"grad_norm": 0.07190197706222534,
"learning_rate": 2.4738961781277097e-06,
"loss": 0.8067,
"step": 3960
},
{
"epoch": 4.765906362545018,
"grad_norm": 0.07564514130353928,
"learning_rate": 2.4495074953922425e-06,
"loss": 0.8065,
"step": 3970
},
{
"epoch": 4.777911164465786,
"grad_norm": 0.0754808560013771,
"learning_rate": 2.425200563678141e-06,
"loss": 0.8003,
"step": 3980
},
{
"epoch": 4.7899159663865545,
"grad_norm": 0.07958831638097763,
"learning_rate": 2.400976162096302e-06,
"loss": 0.8669,
"step": 3990
},
{
"epoch": 4.801920768307323,
"grad_norm": 0.07733064144849777,
"learning_rate": 2.376835067112276e-06,
"loss": 0.8245,
"step": 4000
},
{
"epoch": 4.813925570228092,
"grad_norm": 0.07412835210561752,
"learning_rate": 2.352778052521384e-06,
"loss": 0.8312,
"step": 4010
},
{
"epoch": 4.82593037214886,
"grad_norm": 0.07761096954345703,
"learning_rate": 2.3288058894239202e-06,
"loss": 0.8025,
"step": 4020
},
{
"epoch": 4.837935174069628,
"grad_norm": 0.06937912851572037,
"learning_rate": 2.3049193462004277e-06,
"loss": 0.8182,
"step": 4030
},
{
"epoch": 4.849939975990396,
"grad_norm": 0.07850659638643265,
"learning_rate": 2.2811191884870755e-06,
"loss": 0.8176,
"step": 4040
},
{
"epoch": 4.861944777911164,
"grad_norm": 0.0807727575302124,
"learning_rate": 2.2574061791511114e-06,
"loss": 0.8194,
"step": 4050
},
{
"epoch": 4.873949579831933,
"grad_norm": 0.07483908534049988,
"learning_rate": 2.2337810782664138e-06,
"loss": 0.7926,
"step": 4060
},
{
"epoch": 4.885954381752701,
"grad_norm": 0.08230816572904587,
"learning_rate": 2.2102446430891323e-06,
"loss": 0.758,
"step": 4070
},
{
"epoch": 4.8979591836734695,
"grad_norm": 0.0813629999756813,
"learning_rate": 2.186797628033406e-06,
"loss": 0.7916,
"step": 4080
},
{
"epoch": 4.909963985594238,
"grad_norm": 0.0788528099656105,
"learning_rate": 2.163440784647193e-06,
"loss": 0.8624,
"step": 4090
},
{
"epoch": 4.921968787515006,
"grad_norm": 0.07894974946975708,
"learning_rate": 2.140174861588175e-06,
"loss": 0.8324,
"step": 4100
},
{
"epoch": 4.933973589435775,
"grad_norm": 0.08242552727460861,
"learning_rate": 2.1170006045997577e-06,
"loss": 0.8335,
"step": 4110
},
{
"epoch": 4.945978391356543,
"grad_norm": 0.08206847310066223,
"learning_rate": 2.0939187564871768e-06,
"loss": 0.7888,
"step": 4120
},
{
"epoch": 4.957983193277311,
"grad_norm": 0.07950755953788757,
"learning_rate": 2.070930057093676e-06,
"loss": 0.815,
"step": 4130
},
{
"epoch": 4.969987995198079,
"grad_norm": 0.07717156410217285,
"learning_rate": 2.0480352432768063e-06,
"loss": 0.8451,
"step": 4140
},
{
"epoch": 4.981992797118847,
"grad_norm": 0.08424907177686691,
"learning_rate": 2.0252350488848015e-06,
"loss": 0.821,
"step": 4150
},
{
"epoch": 4.9939975990396155,
"grad_norm": 0.06911070644855499,
"learning_rate": 2.002530204733051e-06,
"loss": 0.8038,
"step": 4160
},
{
"epoch": 5.0060024009603845,
"grad_norm": 0.0768696665763855,
"learning_rate": 1.9799214385806804e-06,
"loss": 0.8111,
"step": 4170
},
{
"epoch": 5.018007202881153,
"grad_norm": 0.07615645974874496,
"learning_rate": 1.9574094751072287e-06,
"loss": 0.8087,
"step": 4180
},
{
"epoch": 5.030012004801921,
"grad_norm": 0.08393856883049011,
"learning_rate": 1.93499503588941e-06,
"loss": 0.8233,
"step": 4190
},
{
"epoch": 5.042016806722689,
"grad_norm": 0.07502923160791397,
"learning_rate": 1.9126788393779935e-06,
"loss": 0.836,
"step": 4200
},
{
"epoch": 5.054021608643457,
"grad_norm": 0.07897184789180756,
"learning_rate": 1.8904616008747728e-06,
"loss": 0.7748,
"step": 4210
},
{
"epoch": 5.066026410564226,
"grad_norm": 0.0756894201040268,
"learning_rate": 1.8683440325096314e-06,
"loss": 0.7697,
"step": 4220
},
{
"epoch": 5.078031212484994,
"grad_norm": 0.07758835703134537,
"learning_rate": 1.8463268432177301e-06,
"loss": 0.809,
"step": 4230
},
{
"epoch": 5.090036014405762,
"grad_norm": 0.08467192202806473,
"learning_rate": 1.8244107387167697e-06,
"loss": 0.81,
"step": 4240
},
{
"epoch": 5.1020408163265305,
"grad_norm": 0.07697489112615585,
"learning_rate": 1.8025964214843783e-06,
"loss": 0.836,
"step": 4250
},
{
"epoch": 5.114045618247299,
"grad_norm": 0.07720308005809784,
"learning_rate": 1.780884590735595e-06,
"loss": 0.8154,
"step": 4260
},
{
"epoch": 5.126050420168067,
"grad_norm": 0.0763421505689621,
"learning_rate": 1.7592759424004558e-06,
"loss": 0.8531,
"step": 4270
},
{
"epoch": 5.138055222088836,
"grad_norm": 0.08317775279283524,
"learning_rate": 1.7377711691016885e-06,
"loss": 0.7965,
"step": 4280
},
{
"epoch": 5.150060024009604,
"grad_norm": 0.08028178662061691,
"learning_rate": 1.7163709601325078e-06,
"loss": 0.7991,
"step": 4290
},
{
"epoch": 5.162064825930372,
"grad_norm": 0.07481025159358978,
"learning_rate": 1.695076001434524e-06,
"loss": 0.7828,
"step": 4300
},
{
"epoch": 5.17406962785114,
"grad_norm": 0.07857973873615265,
"learning_rate": 1.673886975575763e-06,
"loss": 0.8107,
"step": 4310
},
{
"epoch": 5.186074429771908,
"grad_norm": 0.08358076214790344,
"learning_rate": 1.6528045617287736e-06,
"loss": 0.8227,
"step": 4320
},
{
"epoch": 5.198079231692677,
"grad_norm": 0.08463145047426224,
"learning_rate": 1.6318294356488723e-06,
"loss": 0.8329,
"step": 4330
},
{
"epoch": 5.2100840336134455,
"grad_norm": 0.08062751591205597,
"learning_rate": 1.6109622696524775e-06,
"loss": 0.8005,
"step": 4340
},
{
"epoch": 5.222088835534214,
"grad_norm": 0.07316137105226517,
"learning_rate": 1.5902037325955533e-06,
"loss": 0.7609,
"step": 4350
},
{
"epoch": 5.234093637454982,
"grad_norm": 0.08057930320501328,
"learning_rate": 1.5695544898521841e-06,
"loss": 0.8192,
"step": 4360
},
{
"epoch": 5.24609843937575,
"grad_norm": 0.07895936816930771,
"learning_rate": 1.5490152032932338e-06,
"loss": 0.836,
"step": 4370
},
{
"epoch": 5.258103241296519,
"grad_norm": 0.0799291655421257,
"learning_rate": 1.5285865312651417e-06,
"loss": 0.8181,
"step": 4380
},
{
"epoch": 5.270108043217287,
"grad_norm": 0.07720459997653961,
"learning_rate": 1.508269128568815e-06,
"loss": 0.786,
"step": 4390
},
{
"epoch": 5.282112845138055,
"grad_norm": 0.08408314734697342,
"learning_rate": 1.4880636464386388e-06,
"loss": 0.7868,
"step": 4400
},
{
"epoch": 5.294117647058823,
"grad_norm": 0.08219724148511887,
"learning_rate": 1.4679707325216057e-06,
"loss": 0.8084,
"step": 4410
},
{
"epoch": 5.3061224489795915,
"grad_norm": 0.07679635286331177,
"learning_rate": 1.447991030856558e-06,
"loss": 0.829,
"step": 4420
},
{
"epoch": 5.3181272509003605,
"grad_norm": 0.07849011570215225,
"learning_rate": 1.4281251818535368e-06,
"loss": 0.7877,
"step": 4430
},
{
"epoch": 5.330132052821129,
"grad_norm": 0.08212018013000488,
"learning_rate": 1.4083738222732656e-06,
"loss": 0.8482,
"step": 4440
},
{
"epoch": 5.342136854741897,
"grad_norm": 0.07957368344068527,
"learning_rate": 1.388737585206733e-06,
"loss": 0.7826,
"step": 4450
},
{
"epoch": 5.354141656662665,
"grad_norm": 0.07521677762269974,
"learning_rate": 1.3692171000548987e-06,
"loss": 0.7929,
"step": 4460
},
{
"epoch": 5.366146458583433,
"grad_norm": 0.08291694521903992,
"learning_rate": 1.349812992508527e-06,
"loss": 0.7702,
"step": 4470
},
{
"epoch": 5.378151260504202,
"grad_norm": 0.08579208701848984,
"learning_rate": 1.3305258845281232e-06,
"loss": 0.7703,
"step": 4480
},
{
"epoch": 5.39015606242497,
"grad_norm": 0.07739798724651337,
"learning_rate": 1.311356394324002e-06,
"loss": 0.8072,
"step": 4490
},
{
"epoch": 5.402160864345738,
"grad_norm": 0.07492494583129883,
"learning_rate": 1.2923051363364736e-06,
"loss": 0.824,
"step": 4500
},
{
"epoch": 5.4141656662665065,
"grad_norm": 0.0814451351761818,
"learning_rate": 1.2733727212161478e-06,
"loss": 0.8135,
"step": 4510
},
{
"epoch": 5.426170468187275,
"grad_norm": 0.0821041613817215,
"learning_rate": 1.2545597558043559e-06,
"loss": 0.8094,
"step": 4520
},
{
"epoch": 5.438175270108043,
"grad_norm": 0.08446614444255829,
"learning_rate": 1.2358668431137082e-06,
"loss": 0.799,
"step": 4530
},
{
"epoch": 5.450180072028812,
"grad_norm": 0.08138373494148254,
"learning_rate": 1.2172945823087562e-06,
"loss": 0.8269,
"step": 4540
},
{
"epoch": 5.46218487394958,
"grad_norm": 0.08267953246831894,
"learning_rate": 1.1988435686867988e-06,
"loss": 0.7813,
"step": 4550
},
{
"epoch": 5.474189675870348,
"grad_norm": 0.08228688687086105,
"learning_rate": 1.1805143936587881e-06,
"loss": 0.7929,
"step": 4560
},
{
"epoch": 5.486194477791116,
"grad_norm": 0.07604286819696426,
"learning_rate": 1.1623076447303839e-06,
"loss": 0.7689,
"step": 4570
},
{
"epoch": 5.498199279711884,
"grad_norm": 0.08142969757318497,
"learning_rate": 1.1442239054831194e-06,
"loss": 0.8335,
"step": 4580
},
{
"epoch": 5.510204081632653,
"grad_norm": 0.07931126654148102,
"learning_rate": 1.1262637555556905e-06,
"loss": 0.8089,
"step": 4590
},
{
"epoch": 5.5222088835534215,
"grad_norm": 0.07624116539955139,
"learning_rate": 1.108427770625381e-06,
"loss": 0.8337,
"step": 4600
},
{
"epoch": 5.53421368547419,
"grad_norm": 0.07351070642471313,
"learning_rate": 1.0907165223896138e-06,
"loss": 0.8302,
"step": 4610
},
{
"epoch": 5.546218487394958,
"grad_norm": 0.08688168227672577,
"learning_rate": 1.0731305785476147e-06,
"loss": 0.8255,
"step": 4620
},
{
"epoch": 5.558223289315726,
"grad_norm": 0.08103056997060776,
"learning_rate": 1.0556705027822356e-06,
"loss": 0.7938,
"step": 4630
},
{
"epoch": 5.570228091236495,
"grad_norm": 0.08347798883914948,
"learning_rate": 1.0383368547418638e-06,
"loss": 0.8092,
"step": 4640
},
{
"epoch": 5.582232893157263,
"grad_norm": 0.07991039007902145,
"learning_rate": 1.0211301900224973e-06,
"loss": 0.8539,
"step": 4650
},
{
"epoch": 5.594237695078031,
"grad_norm": 0.07759847491979599,
"learning_rate": 1.0040510601499375e-06,
"loss": 0.8124,
"step": 4660
},
{
"epoch": 5.606242496998799,
"grad_norm": 0.08721420913934708,
"learning_rate": 9.871000125621017e-07,
"loss": 0.7767,
"step": 4670
},
{
"epoch": 5.6182472989195675,
"grad_norm": 0.08467090129852295,
"learning_rate": 9.702775905914847e-07,
"loss": 0.8203,
"step": 4680
},
{
"epoch": 5.630252100840336,
"grad_norm": 0.0824965164065361,
"learning_rate": 9.535843334477413e-07,
"loss": 0.8575,
"step": 4690
},
{
"epoch": 5.642256902761105,
"grad_norm": 0.07862916588783264,
"learning_rate": 9.370207762003963e-07,
"loss": 0.806,
"step": 4700
},
{
"epoch": 5.654261704681873,
"grad_norm": 0.07857292890548706,
"learning_rate": 9.205874497617063e-07,
"loss": 0.7803,
"step": 4710
},
{
"epoch": 5.666266506602641,
"grad_norm": 0.08207172900438309,
"learning_rate": 9.042848808696303e-07,
"loss": 0.815,
"step": 4720
},
{
"epoch": 5.678271308523409,
"grad_norm": 0.07979180663824081,
"learning_rate": 8.881135920709521e-07,
"loss": 0.8307,
"step": 4730
},
{
"epoch": 5.690276110444177,
"grad_norm": 0.07480289041996002,
"learning_rate": 8.720741017045325e-07,
"loss": 0.8415,
"step": 4740
},
{
"epoch": 5.702280912364946,
"grad_norm": 0.08345764875411987,
"learning_rate": 8.561669238846926e-07,
"loss": 0.8413,
"step": 4750
},
{
"epoch": 5.714285714285714,
"grad_norm": 0.08144385367631912,
"learning_rate": 8.403925684847319e-07,
"loss": 0.8402,
"step": 4760
},
{
"epoch": 5.7262905162064826,
"grad_norm": 0.07664457708597183,
"learning_rate": 8.24751541120592e-07,
"loss": 0.8109,
"step": 4770
},
{
"epoch": 5.738295318127251,
"grad_norm": 0.07589486986398697,
"learning_rate": 8.092443431346447e-07,
"loss": 0.7965,
"step": 4780
},
{
"epoch": 5.750300120048019,
"grad_norm": 0.07511530071496964,
"learning_rate": 7.938714715796225e-07,
"loss": 0.8361,
"step": 4790
},
{
"epoch": 5.762304921968788,
"grad_norm": 0.07903044670820236,
"learning_rate": 7.786334192026929e-07,
"loss": 0.8345,
"step": 4800
},
{
"epoch": 5.774309723889556,
"grad_norm": 0.07598318159580231,
"learning_rate": 7.635306744296555e-07,
"loss": 0.7717,
"step": 4810
},
{
"epoch": 5.786314525810324,
"grad_norm": 0.08294142037630081,
"learning_rate": 7.485637213492952e-07,
"loss": 0.8084,
"step": 4820
},
{
"epoch": 5.798319327731092,
"grad_norm": 0.07683376967906952,
"learning_rate": 7.337330396978565e-07,
"loss": 0.8092,
"step": 4830
},
{
"epoch": 5.81032412965186,
"grad_norm": 0.08927746117115021,
"learning_rate": 7.190391048436729e-07,
"loss": 0.8369,
"step": 4840
},
{
"epoch": 5.822328931572629,
"grad_norm": 0.0817866250872612,
"learning_rate": 7.044823877719303e-07,
"loss": 0.8217,
"step": 4850
},
{
"epoch": 5.834333733493398,
"grad_norm": 0.08067484200000763,
"learning_rate": 6.900633550695651e-07,
"loss": 0.8313,
"step": 4860
},
{
"epoch": 5.846338535414166,
"grad_norm": 0.07704111933708191,
"learning_rate": 6.757824689103143e-07,
"loss": 0.8036,
"step": 4870
},
{
"epoch": 5.858343337334934,
"grad_norm": 0.07738319784402847,
"learning_rate": 6.616401870398997e-07,
"loss": 0.839,
"step": 4880
},
{
"epoch": 5.870348139255702,
"grad_norm": 0.07391506433486938,
"learning_rate": 6.476369627613522e-07,
"loss": 0.828,
"step": 4890
},
{
"epoch": 5.882352941176471,
"grad_norm": 0.08492206037044525,
"learning_rate": 6.337732449204886e-07,
"loss": 0.8255,
"step": 4900
},
{
"epoch": 5.894357743097239,
"grad_norm": 0.08384674787521362,
"learning_rate": 6.200494778915184e-07,
"loss": 0.7918,
"step": 4910
},
{
"epoch": 5.906362545018007,
"grad_norm": 0.09019272029399872,
"learning_rate": 6.064661015628043e-07,
"loss": 0.8298,
"step": 4920
},
{
"epoch": 5.918367346938775,
"grad_norm": 0.08015299588441849,
"learning_rate": 5.930235513227622e-07,
"loss": 0.854,
"step": 4930
},
{
"epoch": 5.930372148859544,
"grad_norm": 0.0807531401515007,
"learning_rate": 5.797222580459027e-07,
"loss": 0.8293,
"step": 4940
},
{
"epoch": 5.942376950780313,
"grad_norm": 0.08178652077913284,
"learning_rate": 5.665626480790221e-07,
"loss": 0.8492,
"step": 4950
},
{
"epoch": 5.954381752701081,
"grad_norm": 0.08085218816995621,
"learning_rate": 5.535451432275402e-07,
"loss": 0.82,
"step": 4960
},
{
"epoch": 5.966386554621849,
"grad_norm": 0.07937034964561462,
"learning_rate": 5.406701607419718e-07,
"loss": 0.8534,
"step": 4970
},
{
"epoch": 5.978391356542617,
"grad_norm": 0.07996084541082382,
"learning_rate": 5.27938113304562e-07,
"loss": 0.8082,
"step": 4980
},
{
"epoch": 5.990396158463385,
"grad_norm": 0.07725504040718079,
"learning_rate": 5.153494090160522e-07,
"loss": 0.8334,
"step": 4990
},
{
"epoch": 6.002400960384153,
"grad_norm": 0.08056992292404175,
"learning_rate": 5.029044513825992e-07,
"loss": 0.8501,
"step": 5000
},
{
"epoch": 6.014405762304922,
"grad_norm": 0.08236386626958847,
"learning_rate": 4.906036393028485e-07,
"loss": 0.8664,
"step": 5010
},
{
"epoch": 6.02641056422569,
"grad_norm": 0.08499747514724731,
"learning_rate": 4.784473670551376e-07,
"loss": 0.8105,
"step": 5020
},
{
"epoch": 6.038415366146459,
"grad_norm": 0.07700789719820023,
"learning_rate": 4.66436024284867e-07,
"loss": 0.7831,
"step": 5030
},
{
"epoch": 6.050420168067227,
"grad_norm": 0.07973281294107437,
"learning_rate": 4.54569995992008e-07,
"loss": 0.8325,
"step": 5040
},
{
"epoch": 6.062424969987995,
"grad_norm": 0.08407416939735413,
"learning_rate": 4.428496625187606e-07,
"loss": 0.8581,
"step": 5050
},
{
"epoch": 6.074429771908764,
"grad_norm": 0.08415237069129944,
"learning_rate": 4.312753995373664e-07,
"loss": 0.811,
"step": 5060
},
{
"epoch": 6.086434573829532,
"grad_norm": 0.08722592890262604,
"learning_rate": 4.198475780380623e-07,
"loss": 0.8302,
"step": 5070
},
{
"epoch": 6.0984393757503,
"grad_norm": 0.07541859149932861,
"learning_rate": 4.0856656431719043e-07,
"loss": 0.7965,
"step": 5080
},
{
"epoch": 6.110444177671068,
"grad_norm": 0.08787133544683456,
"learning_rate": 3.974327199654632e-07,
"loss": 0.8184,
"step": 5090
},
{
"epoch": 6.122448979591836,
"grad_norm": 0.08842716366052628,
"learning_rate": 3.8644640185636184e-07,
"loss": 0.8253,
"step": 5100
},
{
"epoch": 6.1344537815126055,
"grad_norm": 0.07971406728029251,
"learning_rate": 3.756079621347097e-07,
"loss": 0.8103,
"step": 5110
},
{
"epoch": 6.146458583433374,
"grad_norm": 0.079940065741539,
"learning_rate": 3.6491774820537616e-07,
"loss": 0.7491,
"step": 5120
},
{
"epoch": 6.158463385354142,
"grad_norm": 0.08296643197536469,
"learning_rate": 3.5437610272214484e-07,
"loss": 0.7997,
"step": 5130
},
{
"epoch": 6.17046818727491,
"grad_norm": 0.08030758798122406,
"learning_rate": 3.4398336357673146e-07,
"loss": 0.8038,
"step": 5140
},
{
"epoch": 6.182472989195678,
"grad_norm": 0.09067036956548691,
"learning_rate": 3.337398638879491e-07,
"loss": 0.8053,
"step": 5150
},
{
"epoch": 6.194477791116446,
"grad_norm": 0.08207783848047256,
"learning_rate": 3.2364593199103456e-07,
"loss": 0.7983,
"step": 5160
},
{
"epoch": 6.206482593037215,
"grad_norm": 0.0827014297246933,
"learning_rate": 3.1370189142712673e-07,
"loss": 0.812,
"step": 5170
},
{
"epoch": 6.218487394957983,
"grad_norm": 0.08526603877544403,
"learning_rate": 3.0390806093288707e-07,
"loss": 0.8212,
"step": 5180
},
{
"epoch": 6.230492196878751,
"grad_norm": 0.0812913328409195,
"learning_rate": 2.942647544302918e-07,
"loss": 0.7901,
"step": 5190
},
{
"epoch": 6.24249699879952,
"grad_norm": 0.07846683263778687,
"learning_rate": 2.8477228101656764e-07,
"loss": 0.7913,
"step": 5200
},
{
"epoch": 6.254501800720288,
"grad_norm": 0.08263833075761795,
"learning_rate": 2.7543094495427913e-07,
"loss": 0.8711,
"step": 5210
},
{
"epoch": 6.266506602641057,
"grad_norm": 0.07612549513578415,
"learning_rate": 2.6624104566158383e-07,
"loss": 0.8158,
"step": 5220
},
{
"epoch": 6.278511404561825,
"grad_norm": 0.07775219529867172,
"learning_rate": 2.57202877702632e-07,
"loss": 0.8169,
"step": 5230
},
{
"epoch": 6.290516206482593,
"grad_norm": 0.07751983404159546,
"learning_rate": 2.4831673077812034e-07,
"loss": 0.838,
"step": 5240
},
{
"epoch": 6.302521008403361,
"grad_norm": 0.0775042176246643,
"learning_rate": 2.395828897160152e-07,
"loss": 0.7884,
"step": 5250
},
{
"epoch": 6.314525810324129,
"grad_norm": 0.08516095578670502,
"learning_rate": 2.310016344624144e-07,
"loss": 0.7672,
"step": 5260
},
{
"epoch": 6.326530612244898,
"grad_norm": 0.07730519026517868,
"learning_rate": 2.225732400725783e-07,
"loss": 0.814,
"step": 5270
},
{
"epoch": 6.3385354141656665,
"grad_norm": 0.08212218433618546,
"learning_rate": 2.1429797670211416e-07,
"loss": 0.8149,
"step": 5280
},
{
"epoch": 6.350540216086435,
"grad_norm": 0.08002620935440063,
"learning_rate": 2.0617610959831613e-07,
"loss": 0.825,
"step": 5290
},
{
"epoch": 6.362545018007203,
"grad_norm": 0.08387167006731033,
"learning_rate": 1.982078990916597e-07,
"loss": 0.801,
"step": 5300
},
{
"epoch": 6.374549819927971,
"grad_norm": 0.07701138406991959,
"learning_rate": 1.9039360058746347e-07,
"loss": 0.844,
"step": 5310
},
{
"epoch": 6.38655462184874,
"grad_norm": 0.08538337796926498,
"learning_rate": 1.827334645576967e-07,
"loss": 0.8474,
"step": 5320
},
{
"epoch": 6.398559423769508,
"grad_norm": 0.08097740262746811,
"learning_rate": 1.7522773653295687e-07,
"loss": 0.7721,
"step": 5330
},
{
"epoch": 6.410564225690276,
"grad_norm": 0.08466079086065292,
"learning_rate": 1.6787665709459387e-07,
"loss": 0.8222,
"step": 5340
},
{
"epoch": 6.422569027611044,
"grad_norm": 0.07942117750644684,
"learning_rate": 1.6068046186700271e-07,
"loss": 0.8039,
"step": 5350
},
{
"epoch": 6.4345738295318124,
"grad_norm": 0.08644157648086548,
"learning_rate": 1.5363938151006897e-07,
"loss": 0.8215,
"step": 5360
},
{
"epoch": 6.4465786314525815,
"grad_norm": 0.07266967743635178,
"learning_rate": 1.4675364171177652e-07,
"loss": 0.801,
"step": 5370
},
{
"epoch": 6.45858343337335,
"grad_norm": 0.07520122081041336,
"learning_rate": 1.400234631809727e-07,
"loss": 0.8319,
"step": 5380
},
{
"epoch": 6.470588235294118,
"grad_norm": 0.08132954686880112,
"learning_rate": 1.3344906164029447e-07,
"loss": 0.8102,
"step": 5390
},
{
"epoch": 6.482593037214886,
"grad_norm": 0.07629051059484482,
"learning_rate": 1.2703064781925345e-07,
"loss": 0.8258,
"step": 5400
},
{
"epoch": 6.494597839135654,
"grad_norm": 0.08365026861429214,
"learning_rate": 1.2076842744748295e-07,
"loss": 0.8448,
"step": 5410
},
{
"epoch": 6.506602641056423,
"grad_norm": 0.07794215530157089,
"learning_rate": 1.146626012481411e-07,
"loss": 0.7922,
"step": 5420
},
{
"epoch": 6.518607442977191,
"grad_norm": 0.08589312434196472,
"learning_rate": 1.0871336493147866e-07,
"loss": 0.7865,
"step": 5430
},
{
"epoch": 6.530612244897959,
"grad_norm": 0.08183544129133224,
"learning_rate": 1.0292090918856634e-07,
"loss": 0.8062,
"step": 5440
},
{
"epoch": 6.5426170468187275,
"grad_norm": 0.08870046585798264,
"learning_rate": 9.728541968518135e-08,
"loss": 0.806,
"step": 5450
},
{
"epoch": 6.554621848739496,
"grad_norm": 0.08433903008699417,
"learning_rate": 9.180707705585545e-08,
"loss": 0.8488,
"step": 5460
},
{
"epoch": 6.566626650660264,
"grad_norm": 0.07865394651889801,
"learning_rate": 8.648605689808965e-08,
"loss": 0.8068,
"step": 5470
},
{
"epoch": 6.578631452581033,
"grad_norm": 0.08350817114114761,
"learning_rate": 8.132252976671972e-08,
"loss": 0.8063,
"step": 5480
},
{
"epoch": 6.590636254501801,
"grad_norm": 0.08394873887300491,
"learning_rate": 7.631666116845349e-08,
"loss": 0.815,
"step": 5490
},
{
"epoch": 6.602641056422569,
"grad_norm": 0.08824454993009567,
"learning_rate": 7.146861155656381e-08,
"loss": 0.8267,
"step": 5500
},
{
"epoch": 6.614645858343337,
"grad_norm": 0.08010157197713852,
"learning_rate": 6.677853632574616e-08,
"loss": 0.8069,
"step": 5510
},
{
"epoch": 6.626650660264105,
"grad_norm": 0.07938225567340851,
"learning_rate": 6.224658580713971e-08,
"loss": 0.773,
"step": 5520
},
{
"epoch": 6.6386554621848735,
"grad_norm": 0.0801059678196907,
"learning_rate": 5.787290526350464e-08,
"loss": 0.7957,
"step": 5530
},
{
"epoch": 6.6506602641056425,
"grad_norm": 0.07904977351427078,
"learning_rate": 5.365763488457021e-08,
"loss": 0.7944,
"step": 5540
},
{
"epoch": 6.662665066026411,
"grad_norm": 0.0811835378408432,
"learning_rate": 4.96009097825384e-08,
"loss": 0.8019,
"step": 5550
},
{
"epoch": 6.674669867947179,
"grad_norm": 0.07907839119434357,
"learning_rate": 4.5702859987753476e-08,
"loss": 0.802,
"step": 5560
},
{
"epoch": 6.686674669867947,
"grad_norm": 0.08255333453416824,
"learning_rate": 4.196361044453645e-08,
"loss": 0.7887,
"step": 5570
},
{
"epoch": 6.698679471788715,
"grad_norm": 0.07908835262060165,
"learning_rate": 3.838328100717714e-08,
"loss": 0.8445,
"step": 5580
},
{
"epoch": 6.710684273709484,
"grad_norm": 0.08409386873245239,
"learning_rate": 3.49619864360945e-08,
"loss": 0.8007,
"step": 5590
},
{
"epoch": 6.722689075630252,
"grad_norm": 0.0717078447341919,
"learning_rate": 3.169983639415841e-08,
"loss": 0.8396,
"step": 5600
},
{
"epoch": 6.73469387755102,
"grad_norm": 0.08110038191080093,
"learning_rate": 2.8596935443172545e-08,
"loss": 0.8113,
"step": 5610
},
{
"epoch": 6.7466986794717885,
"grad_norm": 0.07791571319103241,
"learning_rate": 2.5653383040524228e-08,
"loss": 0.8015,
"step": 5620
},
{
"epoch": 6.758703481392557,
"grad_norm": 0.0744730681180954,
"learning_rate": 2.2869273535997548e-08,
"loss": 0.8109,
"step": 5630
},
{
"epoch": 6.770708283313326,
"grad_norm": 0.08296327292919159,
"learning_rate": 2.0244696168746912e-08,
"loss": 0.8064,
"step": 5640
},
{
"epoch": 6.782713085234094,
"grad_norm": 0.07782948017120361,
"learning_rate": 1.7779735064438196e-08,
"loss": 0.8212,
"step": 5650
},
{
"epoch": 6.794717887154862,
"grad_norm": 0.08207013458013535,
"learning_rate": 1.5474469232552026e-08,
"loss": 0.8189,
"step": 5660
},
{
"epoch": 6.80672268907563,
"grad_norm": 0.08035318553447723,
"learning_rate": 1.332897256385135e-08,
"loss": 0.7571,
"step": 5670
},
{
"epoch": 6.818727490996398,
"grad_norm": 0.0768115296959877,
"learning_rate": 1.1343313828012237e-08,
"loss": 0.8449,
"step": 5680
},
{
"epoch": 6.830732292917167,
"grad_norm": 0.08141221851110458,
"learning_rate": 9.517556671420624e-09,
"loss": 0.8121,
"step": 5690
},
{
"epoch": 6.842737094837935,
"grad_norm": 0.08427503705024719,
"learning_rate": 7.851759615131182e-09,
"loss": 0.8023,
"step": 5700
},
{
"epoch": 6.8547418967587035,
"grad_norm": 0.07593482732772827,
"learning_rate": 6.345976052993252e-09,
"loss": 0.8071,
"step": 5710
},
{
"epoch": 6.866746698679472,
"grad_norm": 0.07990961521863937,
"learning_rate": 5.0002542499366645e-09,
"loss": 0.8203,
"step": 5720
},
{
"epoch": 6.87875150060024,
"grad_norm": 0.07622530311346054,
"learning_rate": 3.814637340427418e-09,
"loss": 0.8464,
"step": 5730
},
{
"epoch": 6.890756302521009,
"grad_norm": 0.08417756855487823,
"learning_rate": 2.7891633270848985e-09,
"loss": 0.8526,
"step": 5740
},
{
"epoch": 6.902761104441777,
"grad_norm": 0.08052320778369904,
"learning_rate": 1.9238650794611848e-09,
"loss": 0.8285,
"step": 5750
},
{
"epoch": 6.914765906362545,
"grad_norm": 0.08085649460554123,
"learning_rate": 1.2187703329891166e-09,
"loss": 0.8666,
"step": 5760
},
{
"epoch": 6.926770708283313,
"grad_norm": 0.08147014677524567,
"learning_rate": 6.739016880957794e-10,
"loss": 0.8447,
"step": 5770
},
{
"epoch": 6.938775510204081,
"grad_norm": 0.08438652008771896,
"learning_rate": 2.89276609472533e-10,
"loss": 0.8355,
"step": 5780
},
{
"epoch": 6.95078031212485,
"grad_norm": 0.08426124602556229,
"learning_rate": 6.490742552101026e-11,
"loss": 0.8033,
"step": 5790
}
],
"logging_steps": 10,
"max_steps": 5799,
"num_input_tokens_seen": 0,
"num_train_epochs": 7,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 9.653539147717018e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}