ReDI_Decomposition / trainer_state.json
moshesbeta's picture
Add files using upload-large-folder tool
04196d2 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 1278,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.023508668821627974,
"grad_norm": 84.37700653076172,
"learning_rate": 7.031250000000001e-06,
"loss": 115.85,
"step": 10
},
{
"epoch": 0.04701733764325595,
"grad_norm": 106.4520034790039,
"learning_rate": 1.484375e-05,
"loss": 116.5781,
"step": 20
},
{
"epoch": 0.07052600646488393,
"grad_norm": 130.89254760742188,
"learning_rate": 2.2656250000000002e-05,
"loss": 103.3281,
"step": 30
},
{
"epoch": 0.0940346752865119,
"grad_norm": 37.089256286621094,
"learning_rate": 3.0468750000000002e-05,
"loss": 60.0406,
"step": 40
},
{
"epoch": 0.11754334410813988,
"grad_norm": 44.051109313964844,
"learning_rate": 3.828125e-05,
"loss": 39.75,
"step": 50
},
{
"epoch": 0.14105201292976785,
"grad_norm": 43.561981201171875,
"learning_rate": 4.609375e-05,
"loss": 24.7984,
"step": 60
},
{
"epoch": 0.16456068175139582,
"grad_norm": 34.622520446777344,
"learning_rate": 5.3906250000000006e-05,
"loss": 11.9723,
"step": 70
},
{
"epoch": 0.1880693505730238,
"grad_norm": 22.26849937438965,
"learning_rate": 6.171875e-05,
"loss": 5.4578,
"step": 80
},
{
"epoch": 0.2115780193946518,
"grad_norm": 9.059814453125,
"learning_rate": 6.953125e-05,
"loss": 3.992,
"step": 90
},
{
"epoch": 0.23508668821627976,
"grad_norm": 4.332883834838867,
"learning_rate": 7.734375e-05,
"loss": 3.1058,
"step": 100
},
{
"epoch": 0.2585953570379077,
"grad_norm": 6.844908237457275,
"learning_rate": 8.515625e-05,
"loss": 2.6654,
"step": 110
},
{
"epoch": 0.2821040258595357,
"grad_norm": 4.539117813110352,
"learning_rate": 9.296875e-05,
"loss": 2.422,
"step": 120
},
{
"epoch": 0.3056126946811637,
"grad_norm": 3.098025321960449,
"learning_rate": 9.999981342914437e-05,
"loss": 2.1623,
"step": 130
},
{
"epoch": 0.32912136350279164,
"grad_norm": 1.3452321290969849,
"learning_rate": 9.997742661115932e-05,
"loss": 2.1429,
"step": 140
},
{
"epoch": 0.35263003232441964,
"grad_norm": 0.37226417660713196,
"learning_rate": 9.991774476447404e-05,
"loss": 2.1347,
"step": 150
},
{
"epoch": 0.3761387011460476,
"grad_norm": 3.523218870162964,
"learning_rate": 9.982081242591919e-05,
"loss": 2.0512,
"step": 160
},
{
"epoch": 0.3996473699676756,
"grad_norm": 0.33414289355278015,
"learning_rate": 9.968670193003843e-05,
"loss": 2.047,
"step": 170
},
{
"epoch": 0.4231560387893036,
"grad_norm": 0.24106919765472412,
"learning_rate": 9.951551335510978e-05,
"loss": 2.1046,
"step": 180
},
{
"epoch": 0.4466647076109315,
"grad_norm": 0.328752726316452,
"learning_rate": 9.930737444846331e-05,
"loss": 2.1397,
"step": 190
},
{
"epoch": 0.4701733764325595,
"grad_norm": 1.1409285068511963,
"learning_rate": 9.906244053115143e-05,
"loss": 2.1377,
"step": 200
},
{
"epoch": 0.49368204525418746,
"grad_norm": 0.0690290704369545,
"learning_rate": 9.87808943820424e-05,
"loss": 2.0992,
"step": 210
},
{
"epoch": 0.5171907140758154,
"grad_norm": 0.20527280867099762,
"learning_rate": 9.846294610142398e-05,
"loss": 2.1339,
"step": 220
},
{
"epoch": 0.5406993828974435,
"grad_norm": 0.3991909623146057,
"learning_rate": 9.810883295421864e-05,
"loss": 2.1176,
"step": 230
},
{
"epoch": 0.5642080517190714,
"grad_norm": 0.08950258791446686,
"learning_rate": 9.771881919292765e-05,
"loss": 2.0895,
"step": 240
},
{
"epoch": 0.5877167205406993,
"grad_norm": 0.17848969995975494,
"learning_rate": 9.729319586043591e-05,
"loss": 2.0877,
"step": 250
},
{
"epoch": 0.6112253893623274,
"grad_norm": 0.30160897970199585,
"learning_rate": 9.683228057282483e-05,
"loss": 2.0648,
"step": 260
},
{
"epoch": 0.6347340581839553,
"grad_norm": 0.11883804202079773,
"learning_rate": 9.63364172823554e-05,
"loss": 2.0591,
"step": 270
},
{
"epoch": 0.6582427270055833,
"grad_norm": 0.1619461327791214,
"learning_rate": 9.580597602079802e-05,
"loss": 2.1386,
"step": 280
},
{
"epoch": 0.6817513958272113,
"grad_norm": 0.11473017930984497,
"learning_rate": 9.524135262330098e-05,
"loss": 2.1028,
"step": 290
},
{
"epoch": 0.7052600646488393,
"grad_norm": 0.059143248945474625,
"learning_rate": 9.464296843300342e-05,
"loss": 2.0881,
"step": 300
},
{
"epoch": 0.7287687334704672,
"grad_norm": 0.1624162495136261,
"learning_rate": 9.401126998661328e-05,
"loss": 2.0971,
"step": 310
},
{
"epoch": 0.7522774022920952,
"grad_norm": 0.15460653603076935,
"learning_rate": 9.334672868118491e-05,
"loss": 2.1154,
"step": 320
},
{
"epoch": 0.7757860711137232,
"grad_norm": 0.4213317334651947,
"learning_rate": 9.26498404223449e-05,
"loss": 2.1113,
"step": 330
},
{
"epoch": 0.7992947399353512,
"grad_norm": 0.14529550075531006,
"learning_rate": 9.192112525422868e-05,
"loss": 2.0867,
"step": 340
},
{
"epoch": 0.8228034087569791,
"grad_norm": 0.06652400642633438,
"learning_rate": 9.116112697140418e-05,
"loss": 2.0807,
"step": 350
},
{
"epoch": 0.8463120775786072,
"grad_norm": 0.12574820220470428,
"learning_rate": 9.037041271307188e-05,
"loss": 2.1711,
"step": 360
},
{
"epoch": 0.8698207464002351,
"grad_norm": 0.10869150608778,
"learning_rate": 8.954957253984426e-05,
"loss": 2.1252,
"step": 370
},
{
"epoch": 0.893329415221863,
"grad_norm": 0.143524169921875,
"learning_rate": 8.869921899342056e-05,
"loss": 2.0468,
"step": 380
},
{
"epoch": 0.916838084043491,
"grad_norm": 0.27682727575302124,
"learning_rate": 8.781998663948513e-05,
"loss": 2.1023,
"step": 390
},
{
"epoch": 0.940346752865119,
"grad_norm": 0.08671136200428009,
"learning_rate": 8.691253159417074e-05,
"loss": 2.0991,
"step": 400
},
{
"epoch": 0.963855421686747,
"grad_norm": 0.14743109047412872,
"learning_rate": 8.597753103444016e-05,
"loss": 2.0858,
"step": 410
},
{
"epoch": 0.9873640905083749,
"grad_norm": 0.14369799196720123,
"learning_rate": 8.501568269275126e-05,
"loss": 2.1057,
"step": 420
},
{
"epoch": 1.0094034675286512,
"grad_norm": 0.09220755100250244,
"learning_rate": 8.40277043363831e-05,
"loss": 1.9405,
"step": 430
},
{
"epoch": 1.0329121363502791,
"grad_norm": 0.25697803497314453,
"learning_rate": 8.301433323181076e-05,
"loss": 2.1221,
"step": 440
},
{
"epoch": 1.056420805171907,
"grad_norm": 0.1321459412574768,
"learning_rate": 8.19763255945298e-05,
"loss": 2.1576,
"step": 450
},
{
"epoch": 1.079929473993535,
"grad_norm": 0.11310122162103653,
"learning_rate": 8.091445602473972e-05,
"loss": 2.0483,
"step": 460
},
{
"epoch": 1.1034381428151632,
"grad_norm": 0.08170254528522491,
"learning_rate": 7.982951692930829e-05,
"loss": 2.1367,
"step": 470
},
{
"epoch": 1.1269468116367911,
"grad_norm": 0.022492246702313423,
"learning_rate": 7.87223179304479e-05,
"loss": 2.1435,
"step": 480
},
{
"epoch": 1.150455480458419,
"grad_norm": 0.015893638134002686,
"learning_rate": 7.759368526154509e-05,
"loss": 2.0666,
"step": 490
},
{
"epoch": 1.173964149280047,
"grad_norm": 0.053727954626083374,
"learning_rate": 7.644446115059425e-05,
"loss": 2.0886,
"step": 500
},
{
"epoch": 1.197472818101675,
"grad_norm": 0.12717482447624207,
"learning_rate": 7.527550319169546e-05,
"loss": 2.1418,
"step": 510
},
{
"epoch": 1.2209814869233029,
"grad_norm": 0.030752327293157578,
"learning_rate": 7.408768370508576e-05,
"loss": 2.0688,
"step": 520
},
{
"epoch": 1.244490155744931,
"grad_norm": 0.07915918529033661,
"learning_rate": 7.288188908618102e-05,
"loss": 2.0857,
"step": 530
},
{
"epoch": 1.267998824566559,
"grad_norm": 0.09853250533342361,
"learning_rate": 7.165901914411435e-05,
"loss": 2.0392,
"step": 540
},
{
"epoch": 1.291507493388187,
"grad_norm": 0.21571961045265198,
"learning_rate": 7.041998643026511e-05,
"loss": 2.0921,
"step": 550
},
{
"epoch": 1.3150161622098149,
"grad_norm": 0.10290564596652985,
"learning_rate": 6.916571555727852e-05,
"loss": 2.0686,
"step": 560
},
{
"epoch": 1.3385248310314428,
"grad_norm": 0.08689925819635391,
"learning_rate": 6.789714250908533e-05,
"loss": 2.0845,
"step": 570
},
{
"epoch": 1.3620334998530708,
"grad_norm": 0.1267048567533493,
"learning_rate": 6.661521394243533e-05,
"loss": 2.1465,
"step": 580
},
{
"epoch": 1.3855421686746987,
"grad_norm": 0.18292003870010376,
"learning_rate": 6.532088648046677e-05,
"loss": 2.067,
"step": 590
},
{
"epoch": 1.4090508374963266,
"grad_norm": 0.051083244383335114,
"learning_rate": 6.401512599883818e-05,
"loss": 2.0697,
"step": 600
},
{
"epoch": 1.4325595063179548,
"grad_norm": 0.0492679663002491,
"learning_rate": 6.269890690495593e-05,
"loss": 2.1064,
"step": 610
},
{
"epoch": 1.4560681751395828,
"grad_norm": 0.051078781485557556,
"learning_rate": 6.137321141083468e-05,
"loss": 2.0533,
"step": 620
},
{
"epoch": 1.4795768439612107,
"grad_norm": 0.11715124547481537,
"learning_rate": 6.003902880013416e-05,
"loss": 2.1156,
"step": 630
},
{
"epoch": 1.5030855127828386,
"grad_norm": 0.041975561529397964,
"learning_rate": 5.869735468991854e-05,
"loss": 2.0543,
"step": 640
},
{
"epoch": 1.5265941816044668,
"grad_norm": 0.10740803927183151,
"learning_rate": 5.73491902876897e-05,
"loss": 2.0539,
"step": 650
},
{
"epoch": 1.5501028504260947,
"grad_norm": 0.07230595499277115,
"learning_rate": 5.599554164424859e-05,
"loss": 2.1078,
"step": 660
},
{
"epoch": 1.5736115192477227,
"grad_norm": 0.010545202530920506,
"learning_rate": 5.463741890294257e-05,
"loss": 2.1434,
"step": 670
},
{
"epoch": 1.5971201880693506,
"grad_norm": 0.0882532000541687,
"learning_rate": 5.327583554585842e-05,
"loss": 2.0696,
"step": 680
},
{
"epoch": 1.6206288568909786,
"grad_norm": 0.15730935335159302,
"learning_rate": 5.191180763752418e-05,
"loss": 2.0743,
"step": 690
},
{
"epoch": 1.6441375257126065,
"grad_norm": 0.16760171949863434,
"learning_rate": 5.054635306668355e-05,
"loss": 2.1303,
"step": 700
},
{
"epoch": 1.6676461945342345,
"grad_norm": 0.12746253609657288,
"learning_rate": 4.9180490786709335e-05,
"loss": 2.1359,
"step": 710
},
{
"epoch": 1.6911548633558624,
"grad_norm": 0.08576823770999908,
"learning_rate": 4.781524005522216e-05,
"loss": 2.0884,
"step": 720
},
{
"epoch": 1.7146635321774903,
"grad_norm": 0.01966356858611107,
"learning_rate": 4.645161967348223e-05,
"loss": 2.1102,
"step": 730
},
{
"epoch": 1.7381722009991183,
"grad_norm": 0.1140172928571701,
"learning_rate": 4.509064722612169e-05,
"loss": 2.134,
"step": 740
},
{
"epoch": 1.7616808698207465,
"grad_norm": 0.11263474822044373,
"learning_rate": 4.373333832178478e-05,
"loss": 2.212,
"step": 750
},
{
"epoch": 1.7851895386423744,
"grad_norm": 0.01955031044781208,
"learning_rate": 4.2380705835242716e-05,
"loss": 2.0876,
"step": 760
},
{
"epoch": 1.8086982074640023,
"grad_norm": 0.02556225284934044,
"learning_rate": 4.10337591515484e-05,
"loss": 2.064,
"step": 770
},
{
"epoch": 1.8322068762856303,
"grad_norm": 0.04094263166189194,
"learning_rate": 3.969350341279566e-05,
"loss": 2.1206,
"step": 780
},
{
"epoch": 1.8557155451072584,
"grad_norm": 0.0680101215839386,
"learning_rate": 3.8360938768044405e-05,
"loss": 2.0975,
"step": 790
},
{
"epoch": 1.8792242139288864,
"grad_norm": 0.08410263061523438,
"learning_rate": 3.7037059626971944e-05,
"loss": 2.0519,
"step": 800
},
{
"epoch": 1.9027328827505143,
"grad_norm": 0.02342092990875244,
"learning_rate": 3.572285391780734e-05,
"loss": 2.019,
"step": 810
},
{
"epoch": 1.9262415515721423,
"grad_norm": 0.02327684499323368,
"learning_rate": 3.441930235010211e-05,
"loss": 2.0966,
"step": 820
},
{
"epoch": 1.9497502203937702,
"grad_norm": 0.0357813723385334,
"learning_rate": 3.312737768288827e-05,
"loss": 2.0924,
"step": 830
},
{
"epoch": 1.9732588892153982,
"grad_norm": 0.07186655700206757,
"learning_rate": 3.184804399876886e-05,
"loss": 2.0315,
"step": 840
},
{
"epoch": 1.996767558037026,
"grad_norm": 0.09240284562110901,
"learning_rate": 3.0582255984483535e-05,
"loss": 2.0409,
"step": 850
},
{
"epoch": 2.0188069350573024,
"grad_norm": 0.04462951049208641,
"learning_rate": 2.9330958218485392e-05,
"loss": 1.9696,
"step": 860
},
{
"epoch": 2.0423156038789303,
"grad_norm": 0.029340475797653198,
"learning_rate": 2.8095084466061367e-05,
"loss": 2.1055,
"step": 870
},
{
"epoch": 2.0658242727005582,
"grad_norm": 0.048022590577602386,
"learning_rate": 2.6875556982521248e-05,
"loss": 2.0865,
"step": 880
},
{
"epoch": 2.089332941522186,
"grad_norm": 0.04670249670743942,
"learning_rate": 2.5673285824976575e-05,
"loss": 2.1081,
"step": 890
},
{
"epoch": 2.112841610343814,
"grad_norm": 0.015516964718699455,
"learning_rate": 2.4489168173221567e-05,
"loss": 2.1023,
"step": 900
},
{
"epoch": 2.136350279165442,
"grad_norm": 0.06762902438640594,
"learning_rate": 2.33240876602242e-05,
"loss": 2.0929,
"step": 910
},
{
"epoch": 2.15985894798707,
"grad_norm": 0.014990041963756084,
"learning_rate": 2.2178913712726045e-05,
"loss": 2.0817,
"step": 920
},
{
"epoch": 2.1833676168086984,
"grad_norm": 0.07376622408628464,
"learning_rate": 2.1054500902443423e-05,
"loss": 2.0868,
"step": 930
},
{
"epoch": 2.2068762856303263,
"grad_norm": 0.03299334645271301,
"learning_rate": 1.9951688308354133e-05,
"loss": 2.1123,
"step": 940
},
{
"epoch": 2.2303849544519543,
"grad_norm": 0.05409041792154312,
"learning_rate": 1.887129889054503e-05,
"loss": 2.051,
"step": 950
},
{
"epoch": 2.2538936232735822,
"grad_norm": 0.041158534586429596,
"learning_rate": 1.7814138876088566e-05,
"loss": 2.0458,
"step": 960
},
{
"epoch": 2.27740229209521,
"grad_norm": 0.03300878778100014,
"learning_rate": 1.6780997157405846e-05,
"loss": 2.0565,
"step": 970
},
{
"epoch": 2.300910960916838,
"grad_norm": 0.009902500547468662,
"learning_rate": 1.5772644703565565e-05,
"loss": 2.0952,
"step": 980
},
{
"epoch": 2.324419629738466,
"grad_norm": 0.44653964042663574,
"learning_rate": 1.4789833984957962e-05,
"loss": 2.0386,
"step": 990
},
{
"epoch": 2.347928298560094,
"grad_norm": 0.4100548326969147,
"learning_rate": 1.3833298411773276e-05,
"loss": 2.0977,
"step": 1000
},
{
"epoch": 2.371436967381722,
"grad_norm": 0.00745609262958169,
"learning_rate": 1.2903751786703482e-05,
"loss": 2.0503,
"step": 1010
},
{
"epoch": 2.39494563620335,
"grad_norm": 0.05212165042757988,
"learning_rate": 1.2001887772276076e-05,
"loss": 2.1437,
"step": 1020
},
{
"epoch": 2.418454305024978,
"grad_norm": 0.0367308035492897,
"learning_rate": 1.1128379373217058e-05,
"loss": 2.0911,
"step": 1030
},
{
"epoch": 2.4419629738466058,
"grad_norm": 0.02135683037340641,
"learning_rate": 1.028387843422952e-05,
"loss": 2.1183,
"step": 1040
},
{
"epoch": 2.465471642668234,
"grad_norm": 0.02447247877717018,
"learning_rate": 9.469015153562855e-06,
"loss": 2.1205,
"step": 1050
},
{
"epoch": 2.488980311489862,
"grad_norm": 0.011414138600230217,
"learning_rate": 8.684397612735024e-06,
"loss": 2.0027,
"step": 1060
},
{
"epoch": 2.51248898031149,
"grad_norm": 0.0028556822799146175,
"learning_rate": 7.930611322759462e-06,
"loss": 2.0682,
"step": 1070
},
{
"epoch": 2.535997649133118,
"grad_norm": 0.008712991140782833,
"learning_rate": 7.2082187872147935e-06,
"loss": 2.0895,
"step": 1080
},
{
"epoch": 2.559506317954746,
"grad_norm": 0.03148069605231285,
"learning_rate": 6.517759082483543e-06,
"loss": 2.1298,
"step": 1090
},
{
"epoch": 2.583014986776374,
"grad_norm": 0.03679489344358444,
"learning_rate": 5.8597474554731745e-06,
"loss": 2.0837,
"step": 1100
},
{
"epoch": 2.606523655598002,
"grad_norm": 0.05236874148249626,
"learning_rate": 5.234674939119538e-06,
"loss": 2.1486,
"step": 1110
},
{
"epoch": 2.6300323244196298,
"grad_norm": 0.03369889780879021,
"learning_rate": 4.643007985959641e-06,
"loss": 2.0959,
"step": 1120
},
{
"epoch": 2.6535409932412577,
"grad_norm": 0.006195698864758015,
"learning_rate": 4.085188120047362e-06,
"loss": 2.0439,
"step": 1130
},
{
"epoch": 2.6770496620628856,
"grad_norm": 0.03900829330086708,
"learning_rate": 3.5616316074715995e-06,
"loss": 2.111,
"step": 1140
},
{
"epoch": 2.7005583308845136,
"grad_norm": 0.00790059007704258,
"learning_rate": 3.0727291457229303e-06,
"loss": 2.1078,
"step": 1150
},
{
"epoch": 2.7240669997061415,
"grad_norm": 0.017095139250159264,
"learning_rate": 2.618845572140527e-06,
"loss": 2.1248,
"step": 1160
},
{
"epoch": 2.7475756685277695,
"grad_norm": 0.006935155484825373,
"learning_rate": 2.2003195916567855e-06,
"loss": 2.0573,
"step": 1170
},
{
"epoch": 2.7710843373493974,
"grad_norm": 0.01475539617240429,
"learning_rate": 1.8174635240431282e-06,
"loss": 2.0218,
"step": 1180
},
{
"epoch": 2.7945930061710254,
"grad_norm": 0.08390510082244873,
"learning_rate": 1.4705630708451712e-06,
"loss": 2.0729,
"step": 1190
},
{
"epoch": 2.8181016749926533,
"grad_norm": 0.04823232442140579,
"learning_rate": 1.1598771021816058e-06,
"loss": 2.0432,
"step": 1200
},
{
"epoch": 2.8416103438142817,
"grad_norm": 0.017384668812155724,
"learning_rate": 8.856374635655695e-07,
"loss": 2.1005,
"step": 1210
},
{
"epoch": 2.8651190126359096,
"grad_norm": 0.016273437067866325,
"learning_rate": 6.480488028928422e-07,
"loss": 2.1043,
"step": 1220
},
{
"epoch": 2.8886276814575376,
"grad_norm": 0.004406425170600414,
"learning_rate": 4.47288417725944e-07,
"loss": 2.1214,
"step": 1230
},
{
"epoch": 2.9121363502791655,
"grad_norm": 0.04330357536673546,
"learning_rate": 2.8350612298801427e-07,
"loss": 2.0679,
"step": 1240
},
{
"epoch": 2.9356450191007935,
"grad_norm": 0.005912340711802244,
"learning_rate": 1.5682413916531802e-07,
"loss": 2.1937,
"step": 1250
},
{
"epoch": 2.9591536879224214,
"grad_norm": 0.021025869995355606,
"learning_rate": 6.733700110174157e-08,
"loss": 2.1239,
"step": 1260
},
{
"epoch": 2.9826623567440493,
"grad_norm": 0.006763559300452471,
"learning_rate": 1.511148745335089e-08,
"loss": 2.1434,
"step": 1270
},
{
"epoch": 3.0,
"step": 1278,
"total_flos": 2.0454733263908045e+17,
"train_loss": 5.729273645344288,
"train_runtime": 5128.0315,
"train_samples_per_second": 1.991,
"train_steps_per_second": 0.249
}
],
"logging_steps": 10,
"max_steps": 1278,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.0454733263908045e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}