limernyou's picture
Upload folder using huggingface_hub
7140a44 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.75,
"eval_steps": 100.0,
"global_step": 1500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0025,
"grad_norm": 0.09379793703556061,
"learning_rate": 5.999999999999999e-06,
"loss": 0.6799,
"step": 5
},
{
"epoch": 0.005,
"grad_norm": 0.1399833709001541,
"learning_rate": 1.3499999999999998e-05,
"loss": 0.6954,
"step": 10
},
{
"epoch": 0.0075,
"grad_norm": 0.08632303029298782,
"learning_rate": 2.1e-05,
"loss": 0.6921,
"step": 15
},
{
"epoch": 0.01,
"grad_norm": 0.10006701201200485,
"learning_rate": 2.8499999999999998e-05,
"loss": 0.69,
"step": 20
},
{
"epoch": 0.0125,
"grad_norm": 0.07633858919143677,
"learning_rate": 3.5999999999999994e-05,
"loss": 0.6722,
"step": 25
},
{
"epoch": 0.015,
"grad_norm": 0.09399061650037766,
"learning_rate": 4.3499999999999993e-05,
"loss": 0.6453,
"step": 30
},
{
"epoch": 0.0175,
"grad_norm": 0.0843738541007042,
"learning_rate": 5.1e-05,
"loss": 0.6276,
"step": 35
},
{
"epoch": 0.02,
"grad_norm": 0.08583351224660873,
"learning_rate": 5.85e-05,
"loss": 0.58,
"step": 40
},
{
"epoch": 0.0225,
"grad_norm": 0.09571370482444763,
"learning_rate": 6.599999999999999e-05,
"loss": 0.6355,
"step": 45
},
{
"epoch": 0.025,
"grad_norm": 0.1083935871720314,
"learning_rate": 7.35e-05,
"loss": 0.589,
"step": 50
},
{
"epoch": 0.0275,
"grad_norm": 0.10387319326400757,
"learning_rate": 8.1e-05,
"loss": 0.6061,
"step": 55
},
{
"epoch": 0.03,
"grad_norm": 0.11083361506462097,
"learning_rate": 8.849999999999998e-05,
"loss": 0.572,
"step": 60
},
{
"epoch": 0.0325,
"grad_norm": 0.12665686011314392,
"learning_rate": 9.599999999999999e-05,
"loss": 0.5442,
"step": 65
},
{
"epoch": 0.035,
"grad_norm": 0.1308053582906723,
"learning_rate": 0.00010349999999999998,
"loss": 0.6524,
"step": 70
},
{
"epoch": 0.0375,
"grad_norm": 0.13535510003566742,
"learning_rate": 0.00011099999999999999,
"loss": 0.6404,
"step": 75
},
{
"epoch": 0.04,
"grad_norm": 0.12833671271800995,
"learning_rate": 0.0001185,
"loss": 0.5717,
"step": 80
},
{
"epoch": 0.0425,
"grad_norm": 0.11962099373340607,
"learning_rate": 0.00012599999999999997,
"loss": 0.6098,
"step": 85
},
{
"epoch": 0.045,
"grad_norm": 0.13898271322250366,
"learning_rate": 0.0001335,
"loss": 0.6099,
"step": 90
},
{
"epoch": 0.0475,
"grad_norm": 0.14486610889434814,
"learning_rate": 0.00014099999999999998,
"loss": 0.5744,
"step": 95
},
{
"epoch": 0.05,
"grad_norm": 0.1432138830423355,
"learning_rate": 0.00014849999999999998,
"loss": 0.5659,
"step": 100
},
{
"epoch": 0.0525,
"grad_norm": 0.13487878441810608,
"learning_rate": 0.000156,
"loss": 0.5622,
"step": 105
},
{
"epoch": 0.055,
"grad_norm": 0.12495309859514236,
"learning_rate": 0.0001635,
"loss": 0.5951,
"step": 110
},
{
"epoch": 0.0575,
"grad_norm": 0.13011734187602997,
"learning_rate": 0.00017099999999999998,
"loss": 0.6249,
"step": 115
},
{
"epoch": 0.06,
"grad_norm": 0.13987745344638824,
"learning_rate": 0.00017849999999999997,
"loss": 0.559,
"step": 120
},
{
"epoch": 0.0625,
"grad_norm": 0.13373605906963348,
"learning_rate": 0.000186,
"loss": 0.5475,
"step": 125
},
{
"epoch": 0.065,
"grad_norm": 0.12433867901563644,
"learning_rate": 0.0001935,
"loss": 0.5274,
"step": 130
},
{
"epoch": 0.0675,
"grad_norm": 0.11097615957260132,
"learning_rate": 0.000201,
"loss": 0.678,
"step": 135
},
{
"epoch": 0.07,
"grad_norm": 0.1155027225613594,
"learning_rate": 0.00020849999999999997,
"loss": 0.5611,
"step": 140
},
{
"epoch": 0.0725,
"grad_norm": 0.11431068181991577,
"learning_rate": 0.00021599999999999996,
"loss": 0.6054,
"step": 145
},
{
"epoch": 0.075,
"grad_norm": 0.09796140342950821,
"learning_rate": 0.00022349999999999998,
"loss": 0.5472,
"step": 150
},
{
"epoch": 0.0775,
"grad_norm": 0.09489257633686066,
"learning_rate": 0.00023099999999999998,
"loss": 0.4636,
"step": 155
},
{
"epoch": 0.08,
"grad_norm": 0.10787788033485413,
"learning_rate": 0.0002385,
"loss": 0.6164,
"step": 160
},
{
"epoch": 0.0825,
"grad_norm": 0.10261733084917068,
"learning_rate": 0.00024599999999999996,
"loss": 0.5408,
"step": 165
},
{
"epoch": 0.085,
"grad_norm": 0.11870352178812027,
"learning_rate": 0.0002535,
"loss": 0.5268,
"step": 170
},
{
"epoch": 0.0875,
"grad_norm": 0.11910569667816162,
"learning_rate": 0.000261,
"loss": 0.5461,
"step": 175
},
{
"epoch": 0.09,
"grad_norm": 0.10083702206611633,
"learning_rate": 0.00026849999999999997,
"loss": 0.4794,
"step": 180
},
{
"epoch": 0.0925,
"grad_norm": 0.10453511029481888,
"learning_rate": 0.000276,
"loss": 0.5539,
"step": 185
},
{
"epoch": 0.095,
"grad_norm": 0.101403146982193,
"learning_rate": 0.00028349999999999995,
"loss": 0.5346,
"step": 190
},
{
"epoch": 0.0975,
"grad_norm": 0.10724789649248123,
"learning_rate": 0.00029099999999999997,
"loss": 0.6026,
"step": 195
},
{
"epoch": 0.1,
"grad_norm": 0.1140277311205864,
"learning_rate": 0.0002985,
"loss": 0.5193,
"step": 200
},
{
"epoch": 0.1025,
"grad_norm": 0.09706108272075653,
"learning_rate": 0.0002999963446058092,
"loss": 0.54,
"step": 205
},
{
"epoch": 0.105,
"grad_norm": 0.10003062337636948,
"learning_rate": 0.0002999814948722491,
"loss": 0.5365,
"step": 210
},
{
"epoch": 0.1075,
"grad_norm": 0.1078687533736229,
"learning_rate": 0.00029995522346717746,
"loss": 0.5889,
"step": 215
},
{
"epoch": 0.11,
"grad_norm": 0.10538115352392197,
"learning_rate": 0.0002999175323912636,
"loss": 0.5611,
"step": 220
},
{
"epoch": 0.1125,
"grad_norm": 0.1020808294415474,
"learning_rate": 0.00029986842451482874,
"loss": 0.6103,
"step": 225
},
{
"epoch": 0.115,
"grad_norm": 0.09635835886001587,
"learning_rate": 0.0002998079035776279,
"loss": 0.5229,
"step": 230
},
{
"epoch": 0.1175,
"grad_norm": 0.10287190228700638,
"learning_rate": 0.0002997359741885648,
"loss": 0.5312,
"step": 235
},
{
"epoch": 0.12,
"grad_norm": 0.09160075336694717,
"learning_rate": 0.0002996526418253408,
"loss": 0.5673,
"step": 240
},
{
"epoch": 0.1225,
"grad_norm": 0.08691006153821945,
"learning_rate": 0.000299557912834038,
"loss": 0.5326,
"step": 245
},
{
"epoch": 0.125,
"grad_norm": 0.10096988826990128,
"learning_rate": 0.00029945179442863594,
"loss": 0.6004,
"step": 250
},
{
"epoch": 0.1275,
"grad_norm": 0.09594204276800156,
"learning_rate": 0.000299334294690462,
"loss": 0.5516,
"step": 255
},
{
"epoch": 0.13,
"grad_norm": 0.10281919687986374,
"learning_rate": 0.00029920542256757607,
"loss": 0.5515,
"step": 260
},
{
"epoch": 0.1325,
"grad_norm": 0.08547840267419815,
"learning_rate": 0.00029906518787408944,
"loss": 0.5243,
"step": 265
},
{
"epoch": 0.135,
"grad_norm": 0.10161560773849487,
"learning_rate": 0.0002989136012894168,
"loss": 0.5096,
"step": 270
},
{
"epoch": 0.1375,
"grad_norm": 0.09101904183626175,
"learning_rate": 0.0002987506743574635,
"loss": 0.553,
"step": 275
},
{
"epoch": 0.14,
"grad_norm": 0.09769442677497864,
"learning_rate": 0.0002985764194857463,
"loss": 0.4953,
"step": 280
},
{
"epoch": 0.1425,
"grad_norm": 0.10991579294204712,
"learning_rate": 0.00029839084994444826,
"loss": 0.5152,
"step": 285
},
{
"epoch": 0.145,
"grad_norm": 0.09450916200876236,
"learning_rate": 0.00029819397986540836,
"loss": 0.5397,
"step": 290
},
{
"epoch": 0.1475,
"grad_norm": 0.10876069217920303,
"learning_rate": 0.0002979858242410454,
"loss": 0.4858,
"step": 295
},
{
"epoch": 0.15,
"grad_norm": 0.097995825111866,
"learning_rate": 0.00029776639892321606,
"loss": 0.5566,
"step": 300
},
{
"epoch": 0.1525,
"grad_norm": 0.1145048514008522,
"learning_rate": 0.0002975357206220079,
"loss": 0.4531,
"step": 305
},
{
"epoch": 0.155,
"grad_norm": 0.10271880775690079,
"learning_rate": 0.00029729380690446654,
"loss": 0.5199,
"step": 310
},
{
"epoch": 0.1575,
"grad_norm": 0.11095371842384338,
"learning_rate": 0.0002970406761932583,
"loss": 0.5416,
"step": 315
},
{
"epoch": 0.16,
"grad_norm": 0.09949438273906708,
"learning_rate": 0.00029677634776526673,
"loss": 0.4841,
"step": 320
},
{
"epoch": 0.1625,
"grad_norm": 0.1163724958896637,
"learning_rate": 0.00029650084175012517,
"loss": 0.4913,
"step": 325
},
{
"epoch": 0.165,
"grad_norm": 0.10726840049028397,
"learning_rate": 0.00029621417912868323,
"loss": 0.5203,
"step": 330
},
{
"epoch": 0.1675,
"grad_norm": 0.09609931707382202,
"learning_rate": 0.00029591638173140947,
"loss": 0.5607,
"step": 335
},
{
"epoch": 0.17,
"grad_norm": 0.10824442654848099,
"learning_rate": 0.0002956074722367286,
"loss": 0.6004,
"step": 340
},
{
"epoch": 0.1725,
"grad_norm": 0.10465679317712784,
"learning_rate": 0.00029528747416929463,
"loss": 0.5216,
"step": 345
},
{
"epoch": 0.175,
"grad_norm": 0.10518354922533035,
"learning_rate": 0.0002949564118981994,
"loss": 0.499,
"step": 350
},
{
"epoch": 0.1775,
"grad_norm": 0.0955279991030693,
"learning_rate": 0.0002946143106351165,
"loss": 0.5607,
"step": 355
},
{
"epoch": 0.18,
"grad_norm": 0.11159654706716537,
"learning_rate": 0.0002942611964323817,
"loss": 0.5204,
"step": 360
},
{
"epoch": 0.1825,
"grad_norm": 0.09571187198162079,
"learning_rate": 0.0002938970961810086,
"loss": 0.6113,
"step": 365
},
{
"epoch": 0.185,
"grad_norm": 0.11854679882526398,
"learning_rate": 0.0002935220376086411,
"loss": 0.5639,
"step": 370
},
{
"epoch": 0.1875,
"grad_norm": 0.1050512045621872,
"learning_rate": 0.0002931360492774415,
"loss": 0.548,
"step": 375
},
{
"epoch": 0.19,
"grad_norm": 0.1053968220949173,
"learning_rate": 0.0002927391605819157,
"loss": 0.5507,
"step": 380
},
{
"epoch": 0.1925,
"grad_norm": 0.10567320138216019,
"learning_rate": 0.00029233140174667445,
"loss": 0.5312,
"step": 385
},
{
"epoch": 0.195,
"grad_norm": 0.11914283782243729,
"learning_rate": 0.0002919128038241318,
"loss": 0.5961,
"step": 390
},
{
"epoch": 0.1975,
"grad_norm": 0.09915795922279358,
"learning_rate": 0.0002914833986921401,
"loss": 0.5086,
"step": 395
},
{
"epoch": 0.2,
"grad_norm": 0.10796502232551575,
"learning_rate": 0.0002910432190515628,
"loss": 0.5585,
"step": 400
},
{
"epoch": 0.2025,
"grad_norm": 0.10748997330665588,
"learning_rate": 0.00029059229842378373,
"loss": 0.5466,
"step": 405
},
{
"epoch": 0.205,
"grad_norm": 0.10696308314800262,
"learning_rate": 0.0002901306711481544,
"loss": 0.5513,
"step": 410
},
{
"epoch": 0.2075,
"grad_norm": 0.10418657958507538,
"learning_rate": 0.0002896583723793792,
"loss": 0.5391,
"step": 415
},
{
"epoch": 0.21,
"grad_norm": 0.16421550512313843,
"learning_rate": 0.00028917543808483796,
"loss": 0.4699,
"step": 420
},
{
"epoch": 0.2125,
"grad_norm": 0.12929962575435638,
"learning_rate": 0.00028868190504184696,
"loss": 0.4984,
"step": 425
},
{
"epoch": 0.215,
"grad_norm": 0.10469454526901245,
"learning_rate": 0.00028817781083485816,
"loss": 0.5119,
"step": 430
},
{
"epoch": 0.2175,
"grad_norm": 0.0964970663189888,
"learning_rate": 0.00028766319385259713,
"loss": 0.5167,
"step": 435
},
{
"epoch": 0.22,
"grad_norm": 0.12395574152469635,
"learning_rate": 0.00028713809328513953,
"loss": 0.5692,
"step": 440
},
{
"epoch": 0.2225,
"grad_norm": 0.10189738124608994,
"learning_rate": 0.0002866025491209265,
"loss": 0.4628,
"step": 445
},
{
"epoch": 0.225,
"grad_norm": 0.10433454066514969,
"learning_rate": 0.0002860566021437197,
"loss": 0.4869,
"step": 450
},
{
"epoch": 0.2275,
"grad_norm": 0.13003456592559814,
"learning_rate": 0.0002855002939294951,
"loss": 0.5291,
"step": 455
},
{
"epoch": 0.23,
"grad_norm": 0.11692202836275101,
"learning_rate": 0.000284933666843277,
"loss": 0.5229,
"step": 460
},
{
"epoch": 0.2325,
"grad_norm": 0.10757846385240555,
"learning_rate": 0.0002843567640359119,
"loss": 0.435,
"step": 465
},
{
"epoch": 0.235,
"grad_norm": 0.10775501281023026,
"learning_rate": 0.00028376962944078206,
"loss": 0.4418,
"step": 470
},
{
"epoch": 0.2375,
"grad_norm": 0.11543692648410797,
"learning_rate": 0.00028317230777046015,
"loss": 0.4204,
"step": 475
},
{
"epoch": 0.24,
"grad_norm": 0.10946698486804962,
"learning_rate": 0.00028256484451330403,
"loss": 0.49,
"step": 480
},
{
"epoch": 0.2425,
"grad_norm": 0.11528221517801285,
"learning_rate": 0.00028194728592999247,
"loss": 0.4752,
"step": 485
},
{
"epoch": 0.245,
"grad_norm": 0.10474205762147903,
"learning_rate": 0.0002813196790500027,
"loss": 0.4847,
"step": 490
},
{
"epoch": 0.2475,
"grad_norm": 0.10768820345401764,
"learning_rate": 0.00028068207166802837,
"loss": 0.4664,
"step": 495
},
{
"epoch": 0.25,
"grad_norm": 0.12158560007810593,
"learning_rate": 0.00028003451234034037,
"loss": 0.4741,
"step": 500
},
{
"epoch": 0.2525,
"grad_norm": 0.11635497957468033,
"learning_rate": 0.0002793770503810886,
"loss": 0.4969,
"step": 505
},
{
"epoch": 0.255,
"grad_norm": 0.12205849587917328,
"learning_rate": 0.00027870973585854665,
"loss": 0.4798,
"step": 510
},
{
"epoch": 0.2575,
"grad_norm": 0.10270871222019196,
"learning_rate": 0.00027803261959129905,
"loss": 0.3888,
"step": 515
},
{
"epoch": 0.26,
"grad_norm": 0.11313367635011673,
"learning_rate": 0.0002773457531443712,
"loss": 0.4759,
"step": 520
},
{
"epoch": 0.2625,
"grad_norm": 0.12905193865299225,
"learning_rate": 0.00027664918882530225,
"loss": 0.4442,
"step": 525
},
{
"epoch": 0.265,
"grad_norm": 0.11690939962863922,
"learning_rate": 0.00027594297968016197,
"loss": 0.5535,
"step": 530
},
{
"epoch": 0.2675,
"grad_norm": 0.10021405667066574,
"learning_rate": 0.00027522717948951094,
"loss": 0.4717,
"step": 535
},
{
"epoch": 0.27,
"grad_norm": 0.10104178637266159,
"learning_rate": 0.0002745018427643051,
"loss": 0.4906,
"step": 540
},
{
"epoch": 0.2725,
"grad_norm": 0.12113891541957855,
"learning_rate": 0.00027376702474174425,
"loss": 0.5674,
"step": 545
},
{
"epoch": 0.275,
"grad_norm": 0.11330476403236389,
"learning_rate": 0.0002730227813810658,
"loss": 0.5184,
"step": 550
},
{
"epoch": 0.2775,
"grad_norm": 0.1025850847363472,
"learning_rate": 0.0002722691693592831,
"loss": 0.4395,
"step": 555
},
{
"epoch": 0.28,
"grad_norm": 0.11591499298810959,
"learning_rate": 0.0002715062460668694,
"loss": 0.5003,
"step": 560
},
{
"epoch": 0.2825,
"grad_norm": 0.11281153559684753,
"learning_rate": 0.0002707340696033871,
"loss": 0.4672,
"step": 565
},
{
"epoch": 0.285,
"grad_norm": 0.1123538464307785,
"learning_rate": 0.00026995269877306356,
"loss": 0.513,
"step": 570
},
{
"epoch": 0.2875,
"grad_norm": 0.10776390135288239,
"learning_rate": 0.0002691621930803127,
"loss": 0.4572,
"step": 575
},
{
"epoch": 0.29,
"grad_norm": 0.10008667409420013,
"learning_rate": 0.0002683626127252036,
"loss": 0.4618,
"step": 580
},
{
"epoch": 0.2925,
"grad_norm": 0.13961340487003326,
"learning_rate": 0.00026755401859887595,
"loss": 0.4819,
"step": 585
},
{
"epoch": 0.295,
"grad_norm": 0.1476685106754303,
"learning_rate": 0.00026673647227890316,
"loss": 0.4964,
"step": 590
},
{
"epoch": 0.2975,
"grad_norm": 0.09795507788658142,
"learning_rate": 0.00026591003602460263,
"loss": 0.4796,
"step": 595
},
{
"epoch": 0.3,
"grad_norm": 0.10903532058000565,
"learning_rate": 0.00026507477277229496,
"loss": 0.4775,
"step": 600
},
{
"epoch": 0.3025,
"grad_norm": 0.10258448123931885,
"learning_rate": 0.0002642307461305105,
"loss": 0.4519,
"step": 605
},
{
"epoch": 0.305,
"grad_norm": 0.11204435676336288,
"learning_rate": 0.0002633780203751459,
"loss": 0.4451,
"step": 610
},
{
"epoch": 0.3075,
"grad_norm": 0.10147629678249359,
"learning_rate": 0.0002625166604445689,
"loss": 0.4256,
"step": 615
},
{
"epoch": 0.31,
"grad_norm": 0.10481107234954834,
"learning_rate": 0.00026164673193467306,
"loss": 0.4381,
"step": 620
},
{
"epoch": 0.3125,
"grad_norm": 0.10856641829013824,
"learning_rate": 0.00026076830109388255,
"loss": 0.4958,
"step": 625
},
{
"epoch": 0.315,
"grad_norm": 0.09918677806854248,
"learning_rate": 0.0002598814348181068,
"loss": 0.4335,
"step": 630
},
{
"epoch": 0.3175,
"grad_norm": 0.10417389869689941,
"learning_rate": 0.00025898620064564637,
"loss": 0.4603,
"step": 635
},
{
"epoch": 0.32,
"grad_norm": 0.0903329998254776,
"learning_rate": 0.00025808266675204954,
"loss": 0.3932,
"step": 640
},
{
"epoch": 0.3225,
"grad_norm": 0.11511855572462082,
"learning_rate": 0.0002571709019449205,
"loss": 0.4169,
"step": 645
},
{
"epoch": 0.325,
"grad_norm": 0.11355557292699814,
"learning_rate": 0.0002562509756586793,
"loss": 0.4455,
"step": 650
},
{
"epoch": 0.3275,
"grad_norm": 0.1271187961101532,
"learning_rate": 0.00025532295794927437,
"loss": 0.4902,
"step": 655
},
{
"epoch": 0.33,
"grad_norm": 0.11936645954847336,
"learning_rate": 0.0002543869194888471,
"loss": 0.4843,
"step": 660
},
{
"epoch": 0.3325,
"grad_norm": 0.11935465037822723,
"learning_rate": 0.00025344293156035044,
"loss": 0.4402,
"step": 665
},
{
"epoch": 0.335,
"grad_norm": 0.13073407113552094,
"learning_rate": 0.00025249106605211986,
"loss": 0.467,
"step": 670
},
{
"epoch": 0.3375,
"grad_norm": 0.10340435802936554,
"learning_rate": 0.0002515313954523991,
"loss": 0.4827,
"step": 675
},
{
"epoch": 0.34,
"grad_norm": 0.11634550243616104,
"learning_rate": 0.00025056399284381983,
"loss": 0.466,
"step": 680
},
{
"epoch": 0.3425,
"grad_norm": 0.10582319647073746,
"learning_rate": 0.0002495889318978362,
"loss": 0.4751,
"step": 685
},
{
"epoch": 0.345,
"grad_norm": 0.16781780123710632,
"learning_rate": 0.00024860628686911436,
"loss": 0.4717,
"step": 690
},
{
"epoch": 0.3475,
"grad_norm": 0.11522196233272552,
"learning_rate": 0.0002476161325898776,
"loss": 0.4687,
"step": 695
},
{
"epoch": 0.35,
"grad_norm": 0.11830449104309082,
"learning_rate": 0.000246618544464208,
"loss": 0.436,
"step": 700
},
{
"epoch": 0.3525,
"grad_norm": 0.17485427856445312,
"learning_rate": 0.0002456135984623034,
"loss": 0.4284,
"step": 705
},
{
"epoch": 0.355,
"grad_norm": 0.12288108468055725,
"learning_rate": 0.00024460137111469296,
"loss": 0.4261,
"step": 710
},
{
"epoch": 0.3575,
"grad_norm": 0.11587081104516983,
"learning_rate": 0.0002435819395064079,
"loss": 0.4493,
"step": 715
},
{
"epoch": 0.36,
"grad_norm": 0.10690271109342575,
"learning_rate": 0.0002425553812711123,
"loss": 0.4648,
"step": 720
},
{
"epoch": 0.3625,
"grad_norm": 0.10404397547245026,
"learning_rate": 0.00024152177458519014,
"loss": 0.4634,
"step": 725
},
{
"epoch": 0.365,
"grad_norm": 0.11986954510211945,
"learning_rate": 0.00024048119816179236,
"loss": 0.4525,
"step": 730
},
{
"epoch": 0.3675,
"grad_norm": 0.10243026167154312,
"learning_rate": 0.00023943373124484234,
"loss": 0.4572,
"step": 735
},
{
"epoch": 0.37,
"grad_norm": 0.10386748611927032,
"learning_rate": 0.00023837945360300129,
"loss": 0.3884,
"step": 740
},
{
"epoch": 0.3725,
"grad_norm": 0.11165735125541687,
"learning_rate": 0.0002373184455235934,
"loss": 0.4902,
"step": 745
},
{
"epoch": 0.375,
"grad_norm": 0.09951601922512054,
"learning_rate": 0.00023625078780649178,
"loss": 0.4541,
"step": 750
},
{
"epoch": 0.3775,
"grad_norm": 0.10347504913806915,
"learning_rate": 0.00023517656175796518,
"loss": 0.3871,
"step": 755
},
{
"epoch": 0.38,
"grad_norm": 0.10478132963180542,
"learning_rate": 0.00023409584918448627,
"loss": 0.4329,
"step": 760
},
{
"epoch": 0.3825,
"grad_norm": 0.1198212131857872,
"learning_rate": 0.00023300873238650159,
"loss": 0.425,
"step": 765
},
{
"epoch": 0.385,
"grad_norm": 0.1103711724281311,
"learning_rate": 0.00023191529415216434,
"loss": 0.4274,
"step": 770
},
{
"epoch": 0.3875,
"grad_norm": 0.09940385073423386,
"learning_rate": 0.00023081561775102944,
"loss": 0.4368,
"step": 775
},
{
"epoch": 0.39,
"grad_norm": 0.11599268019199371,
"learning_rate": 0.00022970978692771242,
"loss": 0.4386,
"step": 780
},
{
"epoch": 0.3925,
"grad_norm": 0.10101296752691269,
"learning_rate": 0.00022859788589551188,
"loss": 0.4696,
"step": 785
},
{
"epoch": 0.395,
"grad_norm": 0.10112808644771576,
"learning_rate": 0.00022747999932999624,
"loss": 0.4066,
"step": 790
},
{
"epoch": 0.3975,
"grad_norm": 0.09595459699630737,
"learning_rate": 0.00022635621236255567,
"loss": 0.4837,
"step": 795
},
{
"epoch": 0.4,
"grad_norm": 0.10761380940675735,
"learning_rate": 0.00022522661057391857,
"loss": 0.5446,
"step": 800
},
{
"epoch": 0.4025,
"grad_norm": 0.11919954419136047,
"learning_rate": 0.00022409127998763463,
"loss": 0.5027,
"step": 805
},
{
"epoch": 0.405,
"grad_norm": 0.10851597785949707,
"learning_rate": 0.00022295030706352356,
"loss": 0.4481,
"step": 810
},
{
"epoch": 0.4075,
"grad_norm": 0.10030311346054077,
"learning_rate": 0.00022180377869109104,
"loss": 0.4709,
"step": 815
},
{
"epoch": 0.41,
"grad_norm": 0.111280657351017,
"learning_rate": 0.00022065178218291147,
"loss": 0.4423,
"step": 820
},
{
"epoch": 0.4125,
"grad_norm": 0.11253602802753448,
"learning_rate": 0.00021949440526797926,
"loss": 0.4136,
"step": 825
},
{
"epoch": 0.415,
"grad_norm": 0.10805424302816391,
"learning_rate": 0.00021833173608502732,
"loss": 0.4656,
"step": 830
},
{
"epoch": 0.4175,
"grad_norm": 0.10983198881149292,
"learning_rate": 0.00021716386317581542,
"loss": 0.3687,
"step": 835
},
{
"epoch": 0.42,
"grad_norm": 0.10653118044137955,
"learning_rate": 0.00021599087547838727,
"loss": 0.4654,
"step": 840
},
{
"epoch": 0.4225,
"grad_norm": 0.10856354981660843,
"learning_rate": 0.00021481286232029735,
"loss": 0.4298,
"step": 845
},
{
"epoch": 0.425,
"grad_norm": 0.11233706772327423,
"learning_rate": 0.0002136299134118085,
"loss": 0.4484,
"step": 850
},
{
"epoch": 0.4275,
"grad_norm": 0.1085442528128624,
"learning_rate": 0.00021244211883906017,
"loss": 0.4776,
"step": 855
},
{
"epoch": 0.43,
"grad_norm": 0.12297824025154114,
"learning_rate": 0.0002112495690572077,
"loss": 0.4029,
"step": 860
},
{
"epoch": 0.4325,
"grad_norm": 0.10838114470243454,
"learning_rate": 0.00021005235488353428,
"loss": 0.4848,
"step": 865
},
{
"epoch": 0.435,
"grad_norm": 0.10273341834545135,
"learning_rate": 0.0002088505674905342,
"loss": 0.3989,
"step": 870
},
{
"epoch": 0.4375,
"grad_norm": 0.11189126968383789,
"learning_rate": 0.0002076442983989705,
"loss": 0.438,
"step": 875
},
{
"epoch": 0.44,
"grad_norm": 0.11592905968427658,
"learning_rate": 0.0002064336394709048,
"loss": 0.4786,
"step": 880
},
{
"epoch": 0.4425,
"grad_norm": 0.11230389773845673,
"learning_rate": 0.0002052186829027017,
"loss": 0.3999,
"step": 885
},
{
"epoch": 0.445,
"grad_norm": 0.12455113977193832,
"learning_rate": 0.00020399952121800767,
"loss": 0.4856,
"step": 890
},
{
"epoch": 0.4475,
"grad_norm": 0.1001812294125557,
"learning_rate": 0.00020277624726070526,
"loss": 0.4689,
"step": 895
},
{
"epoch": 0.45,
"grad_norm": 0.11319112777709961,
"learning_rate": 0.00020154895418784242,
"loss": 0.3998,
"step": 900
},
{
"epoch": 0.4525,
"grad_norm": 0.11322236061096191,
"learning_rate": 0.00020031773546253824,
"loss": 0.4321,
"step": 905
},
{
"epoch": 0.455,
"grad_norm": 0.12924689054489136,
"learning_rate": 0.00019908268484686558,
"loss": 0.4208,
"step": 910
},
{
"epoch": 0.4575,
"grad_norm": 0.11435618251562119,
"learning_rate": 0.00019784389639471048,
"loss": 0.4682,
"step": 915
},
{
"epoch": 0.46,
"grad_norm": 0.10801081359386444,
"learning_rate": 0.00019660146444460975,
"loss": 0.428,
"step": 920
},
{
"epoch": 0.4625,
"grad_norm": 0.10906939953565598,
"learning_rate": 0.0001953554836125667,
"loss": 0.4455,
"step": 925
},
{
"epoch": 0.465,
"grad_norm": 0.10790123790502548,
"learning_rate": 0.00019410604878484556,
"loss": 0.4544,
"step": 930
},
{
"epoch": 0.4675,
"grad_norm": 0.10536376386880875,
"learning_rate": 0.000192853255110746,
"loss": 0.376,
"step": 935
},
{
"epoch": 0.47,
"grad_norm": 0.11744682490825653,
"learning_rate": 0.00019159719799535668,
"loss": 0.3887,
"step": 940
},
{
"epoch": 0.4725,
"grad_norm": 0.12954068183898926,
"learning_rate": 0.00019033797309228983,
"loss": 0.4075,
"step": 945
},
{
"epoch": 0.475,
"grad_norm": 0.1401606798171997,
"learning_rate": 0.00018907567629639725,
"loss": 0.4454,
"step": 950
},
{
"epoch": 0.4775,
"grad_norm": 0.12059322744607925,
"learning_rate": 0.00018781040373646706,
"loss": 0.4339,
"step": 955
},
{
"epoch": 0.48,
"grad_norm": 0.11798987537622452,
"learning_rate": 0.00018654225176790336,
"loss": 0.4405,
"step": 960
},
{
"epoch": 0.4825,
"grad_norm": 0.11344211548566818,
"learning_rate": 0.00018527131696538846,
"loss": 0.4124,
"step": 965
},
{
"epoch": 0.485,
"grad_norm": 0.10373330116271973,
"learning_rate": 0.00018399769611552824,
"loss": 0.4329,
"step": 970
},
{
"epoch": 0.4875,
"grad_norm": 0.12053704261779785,
"learning_rate": 0.0001827214862094814,
"loss": 0.4944,
"step": 975
},
{
"epoch": 0.49,
"grad_norm": 0.141033336520195,
"learning_rate": 0.00018144278443557328,
"loss": 0.4569,
"step": 980
},
{
"epoch": 0.4925,
"grad_norm": 0.10922867804765701,
"learning_rate": 0.0001801616881718947,
"loss": 0.3879,
"step": 985
},
{
"epoch": 0.495,
"grad_norm": 0.09843657910823822,
"learning_rate": 0.00017887829497888612,
"loss": 0.4106,
"step": 990
},
{
"epoch": 0.4975,
"grad_norm": 0.12131062150001526,
"learning_rate": 0.000177592702591908,
"loss": 0.4023,
"step": 995
},
{
"epoch": 0.5,
"grad_norm": 0.11343283206224442,
"learning_rate": 0.00017630500891379806,
"loss": 0.4824,
"step": 1000
},
{
"epoch": 0.5025,
"grad_norm": 0.11050508171319962,
"learning_rate": 0.00017501531200741534,
"loss": 0.4098,
"step": 1005
},
{
"epoch": 0.505,
"grad_norm": 0.11737144738435745,
"learning_rate": 0.00017372371008817256,
"loss": 0.3943,
"step": 1010
},
{
"epoch": 0.5075,
"grad_norm": 0.11473528295755386,
"learning_rate": 0.00017243030151655643,
"loss": 0.3796,
"step": 1015
},
{
"epoch": 0.51,
"grad_norm": 0.13086555898189545,
"learning_rate": 0.00017113518479063738,
"loss": 0.4367,
"step": 1020
},
{
"epoch": 0.5125,
"grad_norm": 0.11752833425998688,
"learning_rate": 0.00016983845853856837,
"loss": 0.4097,
"step": 1025
},
{
"epoch": 0.515,
"grad_norm": 0.11596900969743729,
"learning_rate": 0.0001685402215110739,
"loss": 0.3812,
"step": 1030
},
{
"epoch": 0.5175,
"grad_norm": 0.11850260943174362,
"learning_rate": 0.00016724057257392998,
"loss": 0.4354,
"step": 1035
},
{
"epoch": 0.52,
"grad_norm": 0.12466365844011307,
"learning_rate": 0.00016593961070043498,
"loss": 0.4317,
"step": 1040
},
{
"epoch": 0.5225,
"grad_norm": 0.11178991943597794,
"learning_rate": 0.0001646374349638724,
"loss": 0.3936,
"step": 1045
},
{
"epoch": 0.525,
"grad_norm": 0.11252165585756302,
"learning_rate": 0.00016333414452996623,
"loss": 0.386,
"step": 1050
},
{
"epoch": 0.5275,
"grad_norm": 0.12886975705623627,
"learning_rate": 0.0001620298386493288,
"loss": 0.3965,
"step": 1055
},
{
"epoch": 0.53,
"grad_norm": 0.11716549098491669,
"learning_rate": 0.00016072461664990288,
"loss": 0.3924,
"step": 1060
},
{
"epoch": 0.5325,
"grad_norm": 0.11604485660791397,
"learning_rate": 0.000159418577929397,
"loss": 0.3624,
"step": 1065
},
{
"epoch": 0.535,
"grad_norm": 0.11538460850715637,
"learning_rate": 0.00015811182194771633,
"loss": 0.4338,
"step": 1070
},
{
"epoch": 0.5375,
"grad_norm": 0.11618762463331223,
"learning_rate": 0.00015680444821938804,
"loss": 0.4058,
"step": 1075
},
{
"epoch": 0.54,
"grad_norm": 0.11750835925340652,
"learning_rate": 0.00015549655630598343,
"loss": 0.4422,
"step": 1080
},
{
"epoch": 0.5425,
"grad_norm": 0.12725204229354858,
"learning_rate": 0.00015418824580853535,
"loss": 0.4422,
"step": 1085
},
{
"epoch": 0.545,
"grad_norm": 0.11274927109479904,
"learning_rate": 0.00015287961635995347,
"loss": 0.4229,
"step": 1090
},
{
"epoch": 0.5475,
"grad_norm": 0.11833129078149796,
"learning_rate": 0.00015157076761743686,
"loss": 0.4442,
"step": 1095
},
{
"epoch": 0.55,
"grad_norm": 0.11384794861078262,
"learning_rate": 0.00015026179925488475,
"loss": 0.4528,
"step": 1100
},
{
"epoch": 0.5525,
"grad_norm": 0.11864661425352097,
"learning_rate": 0.00014895281095530575,
"loss": 0.3988,
"step": 1105
},
{
"epoch": 0.555,
"grad_norm": 0.11673832684755325,
"learning_rate": 0.00014764390240322691,
"loss": 0.3544,
"step": 1110
},
{
"epoch": 0.5575,
"grad_norm": 0.1174502745270729,
"learning_rate": 0.00014633517327710202,
"loss": 0.4034,
"step": 1115
},
{
"epoch": 0.56,
"grad_norm": 0.12685547769069672,
"learning_rate": 0.00014502672324172107,
"loss": 0.3595,
"step": 1120
},
{
"epoch": 0.5625,
"grad_norm": 0.12368053942918777,
"learning_rate": 0.00014371865194062007,
"loss": 0.3395,
"step": 1125
},
{
"epoch": 0.565,
"grad_norm": 0.1077839657664299,
"learning_rate": 0.000142411058988493,
"loss": 0.4199,
"step": 1130
},
{
"epoch": 0.5675,
"grad_norm": 0.11699855327606201,
"learning_rate": 0.00014110404396360576,
"loss": 0.3443,
"step": 1135
},
{
"epoch": 0.57,
"grad_norm": 0.13238464295864105,
"learning_rate": 0.0001397977064002128,
"loss": 0.3499,
"step": 1140
},
{
"epoch": 0.5725,
"grad_norm": 0.11482933163642883,
"learning_rate": 0.0001384921457809772,
"loss": 0.3619,
"step": 1145
},
{
"epoch": 0.575,
"grad_norm": 0.13390353322029114,
"learning_rate": 0.00013718746152939487,
"loss": 0.3684,
"step": 1150
},
{
"epoch": 0.5775,
"grad_norm": 0.11464900523424149,
"learning_rate": 0.00013588375300222283,
"loss": 0.3313,
"step": 1155
},
{
"epoch": 0.58,
"grad_norm": 0.10367871820926666,
"learning_rate": 0.00013458111948191296,
"loss": 0.3323,
"step": 1160
},
{
"epoch": 0.5825,
"grad_norm": 0.12259294092655182,
"learning_rate": 0.0001332796601690512,
"loss": 0.3986,
"step": 1165
},
{
"epoch": 0.585,
"grad_norm": 0.10923358052968979,
"learning_rate": 0.00013197947417480292,
"loss": 0.3808,
"step": 1170
},
{
"epoch": 0.5875,
"grad_norm": 0.12479504942893982,
"learning_rate": 0.0001306806605133656,
"loss": 0.4429,
"step": 1175
},
{
"epoch": 0.59,
"grad_norm": 0.11521733552217484,
"learning_rate": 0.000129383318094428,
"loss": 0.4778,
"step": 1180
},
{
"epoch": 0.5925,
"grad_norm": 0.14112086594104767,
"learning_rate": 0.00012808754571563827,
"loss": 0.4634,
"step": 1185
},
{
"epoch": 0.595,
"grad_norm": 0.12947902083396912,
"learning_rate": 0.00012679344205507981,
"loss": 0.4439,
"step": 1190
},
{
"epoch": 0.5975,
"grad_norm": 0.13288578391075134,
"learning_rate": 0.0001255011056637567,
"loss": 0.4402,
"step": 1195
},
{
"epoch": 0.6,
"grad_norm": 0.1216069906949997,
"learning_rate": 0.00012421063495808853,
"loss": 0.4203,
"step": 1200
},
{
"epoch": 0.6025,
"grad_norm": 0.11649637669324875,
"learning_rate": 0.000122922128212416,
"loss": 0.4512,
"step": 1205
},
{
"epoch": 0.605,
"grad_norm": 0.1201406940817833,
"learning_rate": 0.00012163568355151628,
"loss": 0.3725,
"step": 1210
},
{
"epoch": 0.6075,
"grad_norm": 0.12117727100849152,
"learning_rate": 0.00012035139894313107,
"loss": 0.4352,
"step": 1215
},
{
"epoch": 0.61,
"grad_norm": 0.11709322035312653,
"learning_rate": 0.00011906937219050556,
"loss": 0.4189,
"step": 1220
},
{
"epoch": 0.6125,
"grad_norm": 0.11865726858377457,
"learning_rate": 0.0001177897009249405,
"loss": 0.3796,
"step": 1225
},
{
"epoch": 0.615,
"grad_norm": 0.10807759314775467,
"learning_rate": 0.0001165124825983573,
"loss": 0.4465,
"step": 1230
},
{
"epoch": 0.6175,
"grad_norm": 0.13788209855556488,
"learning_rate": 0.00011523781447587641,
"loss": 0.4994,
"step": 1235
},
{
"epoch": 0.62,
"grad_norm": 0.12921364605426788,
"learning_rate": 0.00011396579362841044,
"loss": 0.4251,
"step": 1240
},
{
"epoch": 0.6225,
"grad_norm": 0.12162365019321442,
"learning_rate": 0.0001126965169252718,
"loss": 0.3864,
"step": 1245
},
{
"epoch": 0.625,
"grad_norm": 0.12897826731204987,
"learning_rate": 0.00011143008102679559,
"loss": 0.3753,
"step": 1250
},
{
"epoch": 0.6275,
"grad_norm": 0.116109699010849,
"learning_rate": 0.00011016658237697866,
"loss": 0.3296,
"step": 1255
},
{
"epoch": 0.63,
"grad_norm": 0.12935414910316467,
"learning_rate": 0.00010890611719613512,
"loss": 0.3797,
"step": 1260
},
{
"epoch": 0.6325,
"grad_norm": 0.13730891048908234,
"learning_rate": 0.0001076487814735685,
"loss": 0.3711,
"step": 1265
},
{
"epoch": 0.635,
"grad_norm": 0.13870631158351898,
"learning_rate": 0.00010639467096026211,
"loss": 0.4328,
"step": 1270
},
{
"epoch": 0.6375,
"grad_norm": 0.11644043773412704,
"learning_rate": 0.00010514388116158701,
"loss": 0.3283,
"step": 1275
},
{
"epoch": 0.64,
"grad_norm": 0.12221091985702515,
"learning_rate": 0.00010389650733002894,
"loss": 0.3898,
"step": 1280
},
{
"epoch": 0.6425,
"grad_norm": 0.12048634141683578,
"learning_rate": 0.00010265264445793464,
"loss": 0.3256,
"step": 1285
},
{
"epoch": 0.645,
"grad_norm": 0.1250566840171814,
"learning_rate": 0.00010141238727027761,
"loss": 0.408,
"step": 1290
},
{
"epoch": 0.6475,
"grad_norm": 0.13518592715263367,
"learning_rate": 0.00010017583021744454,
"loss": 0.3763,
"step": 1295
},
{
"epoch": 0.65,
"grad_norm": 0.13047736883163452,
"learning_rate": 9.89430674680425e-05,
"loss": 0.3989,
"step": 1300
},
{
"epoch": 0.6525,
"grad_norm": 0.11474955826997757,
"learning_rate": 9.771419290172773e-05,
"loss": 0.3374,
"step": 1305
},
{
"epoch": 0.655,
"grad_norm": 0.11670063436031342,
"learning_rate": 9.648930010205619e-05,
"loss": 0.3343,
"step": 1310
},
{
"epoch": 0.6575,
"grad_norm": 0.15385080873966217,
"learning_rate": 9.526848234935704e-05,
"loss": 0.3432,
"step": 1315
},
{
"epoch": 0.66,
"grad_norm": 0.13441519439220428,
"learning_rate": 9.405183261362863e-05,
"loss": 0.3116,
"step": 1320
},
{
"epoch": 0.6625,
"grad_norm": 0.14772167801856995,
"learning_rate": 9.283944354745888e-05,
"loss": 0.3613,
"step": 1325
},
{
"epoch": 0.665,
"grad_norm": 0.12146154791116714,
"learning_rate": 9.163140747896907e-05,
"loss": 0.3411,
"step": 1330
},
{
"epoch": 0.6675,
"grad_norm": 0.1333102583885193,
"learning_rate": 9.042781640478291e-05,
"loss": 0.396,
"step": 1335
},
{
"epoch": 0.67,
"grad_norm": 0.12051521986722946,
"learning_rate": 8.922876198302062e-05,
"loss": 0.3837,
"step": 1340
},
{
"epoch": 0.6725,
"grad_norm": 0.12071400880813599,
"learning_rate": 8.803433552631874e-05,
"loss": 0.354,
"step": 1345
},
{
"epoch": 0.675,
"grad_norm": 0.11258620023727417,
"learning_rate": 8.684462799487635e-05,
"loss": 0.3197,
"step": 1350
},
{
"epoch": 0.6775,
"grad_norm": 0.11908067762851715,
"learning_rate": 8.565972998952814e-05,
"loss": 0.377,
"step": 1355
},
{
"epoch": 0.68,
"grad_norm": 0.1252991259098053,
"learning_rate": 8.447973174484469e-05,
"loss": 0.3438,
"step": 1360
},
{
"epoch": 0.6825,
"grad_norm": 0.12832245230674744,
"learning_rate": 8.330472312226091e-05,
"loss": 0.346,
"step": 1365
},
{
"epoch": 0.685,
"grad_norm": 0.1396942287683487,
"learning_rate": 8.213479360323258e-05,
"loss": 0.3886,
"step": 1370
},
{
"epoch": 0.6875,
"grad_norm": 0.12938210368156433,
"learning_rate": 8.097003228242225e-05,
"loss": 0.3699,
"step": 1375
},
{
"epoch": 0.69,
"grad_norm": 0.12459377944469452,
"learning_rate": 7.9810527860914e-05,
"loss": 0.3892,
"step": 1380
},
{
"epoch": 0.6925,
"grad_norm": 0.1360333263874054,
"learning_rate": 7.86563686394587e-05,
"loss": 0.3423,
"step": 1385
},
{
"epoch": 0.695,
"grad_norm": 0.1357765644788742,
"learning_rate": 7.750764251174963e-05,
"loss": 0.408,
"step": 1390
},
{
"epoch": 0.6975,
"grad_norm": 0.14453718066215515,
"learning_rate": 7.636443695772887e-05,
"loss": 0.3398,
"step": 1395
},
{
"epoch": 0.7,
"grad_norm": 0.11541519314050674,
"learning_rate": 7.522683903692547e-05,
"loss": 0.4203,
"step": 1400
},
{
"epoch": 0.7025,
"grad_norm": 0.13344840705394745,
"learning_rate": 7.409493538182545e-05,
"loss": 0.3694,
"step": 1405
},
{
"epoch": 0.705,
"grad_norm": 0.13069866597652435,
"learning_rate": 7.296881219127452e-05,
"loss": 0.3889,
"step": 1410
},
{
"epoch": 0.7075,
"grad_norm": 0.12457838654518127,
"learning_rate": 7.184855522391359e-05,
"loss": 0.3342,
"step": 1415
},
{
"epoch": 0.71,
"grad_norm": 0.11990659683942795,
"learning_rate": 7.073424979164794e-05,
"loss": 0.3855,
"step": 1420
},
{
"epoch": 0.7125,
"grad_norm": 0.1389523446559906,
"learning_rate": 6.962598075315046e-05,
"loss": 0.3943,
"step": 1425
},
{
"epoch": 0.715,
"grad_norm": 0.14108599722385406,
"learning_rate": 6.852383250739938e-05,
"loss": 0.388,
"step": 1430
},
{
"epoch": 0.7175,
"grad_norm": 0.1342005580663681,
"learning_rate": 6.742788898725065e-05,
"loss": 0.3602,
"step": 1435
},
{
"epoch": 0.72,
"grad_norm": 0.13516324758529663,
"learning_rate": 6.633823365304648e-05,
"loss": 0.3935,
"step": 1440
},
{
"epoch": 0.7225,
"grad_norm": 0.1302197426557541,
"learning_rate": 6.52549494862593e-05,
"loss": 0.3618,
"step": 1445
},
{
"epoch": 0.725,
"grad_norm": 0.12428996711969376,
"learning_rate": 6.417811898317259e-05,
"loss": 0.3338,
"step": 1450
},
{
"epoch": 0.7275,
"grad_norm": 0.11249776184558868,
"learning_rate": 6.31078241485982e-05,
"loss": 0.3819,
"step": 1455
},
{
"epoch": 0.73,
"grad_norm": 0.1359994113445282,
"learning_rate": 6.204414648963159e-05,
"loss": 0.3356,
"step": 1460
},
{
"epoch": 0.7325,
"grad_norm": 0.1118568629026413,
"learning_rate": 6.098716700944479e-05,
"loss": 0.3223,
"step": 1465
},
{
"epoch": 0.735,
"grad_norm": 0.12038140743970871,
"learning_rate": 5.993696620111741e-05,
"loss": 0.3481,
"step": 1470
},
{
"epoch": 0.7375,
"grad_norm": 0.12787550687789917,
"learning_rate": 5.889362404150703e-05,
"loss": 0.3766,
"step": 1475
},
{
"epoch": 0.74,
"grad_norm": 0.12134893983602524,
"learning_rate": 5.7857219985158506e-05,
"loss": 0.2916,
"step": 1480
},
{
"epoch": 0.7425,
"grad_norm": 0.1274223029613495,
"learning_rate": 5.682783295825345e-05,
"loss": 0.3095,
"step": 1485
},
{
"epoch": 0.745,
"grad_norm": 0.11817299574613571,
"learning_rate": 5.580554135259932e-05,
"loss": 0.3422,
"step": 1490
},
{
"epoch": 0.7475,
"grad_norm": 0.1348387748003006,
"learning_rate": 5.479042301965987e-05,
"loss": 0.4044,
"step": 1495
},
{
"epoch": 0.75,
"grad_norm": 0.14032681286334991,
"learning_rate": 5.378255526462631e-05,
"loss": 0.337,
"step": 1500
}
],
"logging_steps": 5,
"max_steps": 2000,
"num_input_tokens_seen": 0,
"num_train_epochs": 9223372036854775807,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.972183618289664e+18,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}