Pair_07_08_Multi_audio / trainer_state.json
VOLBEM's picture
Upload folder using huggingface_hub
5532131 verified
{
"best_global_step": 700,
"best_metric": 0.7593940496444702,
"best_model_checkpoint": "./thinker_output/07-08_multi_audio/checkpoint-700",
"epoch": 2.692489053553385,
"eval_steps": 50,
"global_step": 1000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.02694509936005389,
"grad_norm": 7.9127708678466355,
"learning_rate": 5e-05,
"loss": 2.2253,
"step": 10
},
{
"epoch": 0.05389019872010778,
"grad_norm": 2.129170820552183,
"learning_rate": 4.954792043399639e-05,
"loss": 1.4156,
"step": 20
},
{
"epoch": 0.08083529808016167,
"grad_norm": 2.213231596266166,
"learning_rate": 4.909584086799277e-05,
"loss": 1.2292,
"step": 30
},
{
"epoch": 0.10778039744021556,
"grad_norm": 1.9446166933638902,
"learning_rate": 4.864376130198916e-05,
"loss": 1.1526,
"step": 40
},
{
"epoch": 0.13472549680026946,
"grad_norm": 1.8624996582053734,
"learning_rate": 4.8191681735985535e-05,
"loss": 1.1116,
"step": 50
},
{
"epoch": 0.13472549680026946,
"eval_loss": 1.0801820755004883,
"eval_runtime": 116.0,
"eval_samples_per_second": 43.103,
"eval_steps_per_second": 0.681,
"step": 50
},
{
"epoch": 0.16167059616032334,
"grad_norm": 1.644929460056369,
"learning_rate": 4.773960216998192e-05,
"loss": 1.0736,
"step": 60
},
{
"epoch": 0.18861569552037724,
"grad_norm": 1.9154194240377669,
"learning_rate": 4.7287522603978304e-05,
"loss": 1.0532,
"step": 70
},
{
"epoch": 0.21556079488043112,
"grad_norm": 1.5827665298198546,
"learning_rate": 4.683544303797468e-05,
"loss": 1.0312,
"step": 80
},
{
"epoch": 0.24250589424048502,
"grad_norm": 1.429127675663522,
"learning_rate": 4.638336347197107e-05,
"loss": 1.0133,
"step": 90
},
{
"epoch": 0.2694509936005389,
"grad_norm": 1.4187637596535974,
"learning_rate": 4.593128390596745e-05,
"loss": 0.9944,
"step": 100
},
{
"epoch": 0.2694509936005389,
"eval_loss": 0.987108588218689,
"eval_runtime": 113.3326,
"eval_samples_per_second": 44.118,
"eval_steps_per_second": 0.697,
"step": 100
},
{
"epoch": 0.29639609296059277,
"grad_norm": 1.6143534620181745,
"learning_rate": 4.547920433996384e-05,
"loss": 0.9861,
"step": 110
},
{
"epoch": 0.3233411923206467,
"grad_norm": 1.470799422498714,
"learning_rate": 4.5027124773960215e-05,
"loss": 0.9813,
"step": 120
},
{
"epoch": 0.3502862916807006,
"grad_norm": 1.5329678036861214,
"learning_rate": 4.45750452079566e-05,
"loss": 0.9725,
"step": 130
},
{
"epoch": 0.3772313910407545,
"grad_norm": 1.5683570099493096,
"learning_rate": 4.4122965641952984e-05,
"loss": 0.9604,
"step": 140
},
{
"epoch": 0.40417649040080833,
"grad_norm": 1.4886000727881106,
"learning_rate": 4.367088607594937e-05,
"loss": 0.9494,
"step": 150
},
{
"epoch": 0.40417649040080833,
"eval_loss": 0.9371287226676941,
"eval_runtime": 147.0071,
"eval_samples_per_second": 34.012,
"eval_steps_per_second": 0.537,
"step": 150
},
{
"epoch": 0.43112158976086223,
"grad_norm": 1.7059179320901345,
"learning_rate": 4.3218806509945754e-05,
"loss": 0.9409,
"step": 160
},
{
"epoch": 0.45806668912091614,
"grad_norm": 1.3194635069630314,
"learning_rate": 4.276672694394214e-05,
"loss": 0.9244,
"step": 170
},
{
"epoch": 0.48501178848097004,
"grad_norm": 1.3060771848163875,
"learning_rate": 4.2314647377938523e-05,
"loss": 0.9157,
"step": 180
},
{
"epoch": 0.5119568878410239,
"grad_norm": 1.4749915311735549,
"learning_rate": 4.186256781193491e-05,
"loss": 0.9119,
"step": 190
},
{
"epoch": 0.5389019872010778,
"grad_norm": 1.4869922949493382,
"learning_rate": 4.1410488245931286e-05,
"loss": 0.913,
"step": 200
},
{
"epoch": 0.5389019872010778,
"eval_loss": 0.907091498374939,
"eval_runtime": 111.8122,
"eval_samples_per_second": 44.718,
"eval_steps_per_second": 0.707,
"step": 200
},
{
"epoch": 0.5658470865611317,
"grad_norm": 1.4160894102096444,
"learning_rate": 4.095840867992767e-05,
"loss": 0.901,
"step": 210
},
{
"epoch": 0.5927921859211855,
"grad_norm": 1.5792102697843888,
"learning_rate": 4.050632911392405e-05,
"loss": 0.89,
"step": 220
},
{
"epoch": 0.6197372852812395,
"grad_norm": 1.2974993115890197,
"learning_rate": 4.0054249547920434e-05,
"loss": 0.8863,
"step": 230
},
{
"epoch": 0.6466823846412934,
"grad_norm": 1.3514939321309911,
"learning_rate": 3.960216998191682e-05,
"loss": 0.8854,
"step": 240
},
{
"epoch": 0.6736274840013473,
"grad_norm": 1.266918300985221,
"learning_rate": 3.9150090415913203e-05,
"loss": 0.8845,
"step": 250
},
{
"epoch": 0.6736274840013473,
"eval_loss": 0.8746693134307861,
"eval_runtime": 144.3762,
"eval_samples_per_second": 34.632,
"eval_steps_per_second": 0.547,
"step": 250
},
{
"epoch": 0.7005725833614012,
"grad_norm": 1.3519572583606572,
"learning_rate": 3.869801084990959e-05,
"loss": 0.8721,
"step": 260
},
{
"epoch": 0.727517682721455,
"grad_norm": 1.2840451675298208,
"learning_rate": 3.8245931283905966e-05,
"loss": 0.8736,
"step": 270
},
{
"epoch": 0.754462782081509,
"grad_norm": 1.2178137913182443,
"learning_rate": 3.779385171790235e-05,
"loss": 0.8664,
"step": 280
},
{
"epoch": 0.7814078814415628,
"grad_norm": 1.371387951800443,
"learning_rate": 3.7341772151898736e-05,
"loss": 0.8554,
"step": 290
},
{
"epoch": 0.8083529808016167,
"grad_norm": 1.344734634165615,
"learning_rate": 3.688969258589512e-05,
"loss": 0.8638,
"step": 300
},
{
"epoch": 0.8083529808016167,
"eval_loss": 0.8491566181182861,
"eval_runtime": 146.4614,
"eval_samples_per_second": 34.139,
"eval_steps_per_second": 0.539,
"step": 300
},
{
"epoch": 0.8352980801616706,
"grad_norm": 1.272788750854563,
"learning_rate": 3.6437613019891505e-05,
"loss": 0.8541,
"step": 310
},
{
"epoch": 0.8622431795217245,
"grad_norm": 1.3792910383877415,
"learning_rate": 3.598553345388789e-05,
"loss": 0.8481,
"step": 320
},
{
"epoch": 0.8891882788817784,
"grad_norm": 1.203782825434394,
"learning_rate": 3.553345388788427e-05,
"loss": 0.8376,
"step": 330
},
{
"epoch": 0.9161333782418323,
"grad_norm": 1.2646961187289552,
"learning_rate": 3.508137432188065e-05,
"loss": 0.839,
"step": 340
},
{
"epoch": 0.9430784776018861,
"grad_norm": 1.24393119289486,
"learning_rate": 3.462929475587703e-05,
"loss": 0.8381,
"step": 350
},
{
"epoch": 0.9430784776018861,
"eval_loss": 0.8305084109306335,
"eval_runtime": 135.5403,
"eval_samples_per_second": 36.889,
"eval_steps_per_second": 0.583,
"step": 350
},
{
"epoch": 0.9700235769619401,
"grad_norm": 1.2087402616528558,
"learning_rate": 3.4177215189873416e-05,
"loss": 0.8264,
"step": 360
},
{
"epoch": 0.9969686763219939,
"grad_norm": 1.2771110484442496,
"learning_rate": 3.37251356238698e-05,
"loss": 0.8242,
"step": 370
},
{
"epoch": 1.0215560794880432,
"grad_norm": 1.332831610585688,
"learning_rate": 3.3273056057866185e-05,
"loss": 0.6297,
"step": 380
},
{
"epoch": 1.048501178848097,
"grad_norm": 1.2917409109187712,
"learning_rate": 3.282097649186257e-05,
"loss": 0.6539,
"step": 390
},
{
"epoch": 1.0754462782081509,
"grad_norm": 1.3245733128879162,
"learning_rate": 3.2368896925858955e-05,
"loss": 0.6544,
"step": 400
},
{
"epoch": 1.0754462782081509,
"eval_loss": 0.8334468007087708,
"eval_runtime": 148.2965,
"eval_samples_per_second": 33.716,
"eval_steps_per_second": 0.533,
"step": 400
},
{
"epoch": 1.1023913775682048,
"grad_norm": 1.1724140281525666,
"learning_rate": 3.191681735985534e-05,
"loss": 0.6473,
"step": 410
},
{
"epoch": 1.1293364769282586,
"grad_norm": 1.2278740780504742,
"learning_rate": 3.146473779385172e-05,
"loss": 0.6459,
"step": 420
},
{
"epoch": 1.1562815762883125,
"grad_norm": 1.1783670358458123,
"learning_rate": 3.10126582278481e-05,
"loss": 0.6573,
"step": 430
},
{
"epoch": 1.1832266756483665,
"grad_norm": 1.29897205539554,
"learning_rate": 3.056057866184449e-05,
"loss": 0.6548,
"step": 440
},
{
"epoch": 1.2101717750084204,
"grad_norm": 1.1568119957121505,
"learning_rate": 3.010849909584087e-05,
"loss": 0.65,
"step": 450
},
{
"epoch": 1.2101717750084204,
"eval_loss": 0.8229044675827026,
"eval_runtime": 113.6433,
"eval_samples_per_second": 43.997,
"eval_steps_per_second": 0.695,
"step": 450
},
{
"epoch": 1.2371168743684742,
"grad_norm": 1.3004916817149637,
"learning_rate": 2.9656419529837253e-05,
"loss": 0.6474,
"step": 460
},
{
"epoch": 1.2640619737285281,
"grad_norm": 1.212521588358061,
"learning_rate": 2.9204339963833638e-05,
"loss": 0.6518,
"step": 470
},
{
"epoch": 1.291007073088582,
"grad_norm": 1.2553077221366877,
"learning_rate": 2.8752260397830023e-05,
"loss": 0.6456,
"step": 480
},
{
"epoch": 1.3179521724486358,
"grad_norm": 1.184218802614123,
"learning_rate": 2.83001808318264e-05,
"loss": 0.6546,
"step": 490
},
{
"epoch": 1.3448972718086898,
"grad_norm": 1.136986753079325,
"learning_rate": 2.7848101265822786e-05,
"loss": 0.6415,
"step": 500
},
{
"epoch": 1.3448972718086898,
"eval_loss": 0.8110851049423218,
"eval_runtime": 137.5711,
"eval_samples_per_second": 36.345,
"eval_steps_per_second": 0.574,
"step": 500
},
{
"epoch": 1.3718423711687437,
"grad_norm": 1.2143646330707367,
"learning_rate": 2.7396021699819167e-05,
"loss": 0.6558,
"step": 510
},
{
"epoch": 1.3987874705287977,
"grad_norm": 1.265789637913618,
"learning_rate": 2.6943942133815552e-05,
"loss": 0.6505,
"step": 520
},
{
"epoch": 1.4257325698888514,
"grad_norm": 1.2191031554639078,
"learning_rate": 2.6491862567811937e-05,
"loss": 0.6426,
"step": 530
},
{
"epoch": 1.4526776692489054,
"grad_norm": 1.222440942937553,
"learning_rate": 2.603978300180832e-05,
"loss": 0.644,
"step": 540
},
{
"epoch": 1.4796227686089591,
"grad_norm": 1.1495752286958087,
"learning_rate": 2.5587703435804706e-05,
"loss": 0.6415,
"step": 550
},
{
"epoch": 1.4796227686089591,
"eval_loss": 0.7958658337593079,
"eval_runtime": 142.3592,
"eval_samples_per_second": 35.122,
"eval_steps_per_second": 0.555,
"step": 550
},
{
"epoch": 1.506567867969013,
"grad_norm": 1.2049482092443289,
"learning_rate": 2.5135623869801084e-05,
"loss": 0.6395,
"step": 560
},
{
"epoch": 1.533512967329067,
"grad_norm": 1.1839061997290048,
"learning_rate": 2.468354430379747e-05,
"loss": 0.6441,
"step": 570
},
{
"epoch": 1.560458066689121,
"grad_norm": 1.2057965805365276,
"learning_rate": 2.423146473779385e-05,
"loss": 0.6335,
"step": 580
},
{
"epoch": 1.5874031660491748,
"grad_norm": 1.1650282316989717,
"learning_rate": 2.3779385171790235e-05,
"loss": 0.6339,
"step": 590
},
{
"epoch": 1.6143482654092287,
"grad_norm": 1.153394411032144,
"learning_rate": 2.332730560578662e-05,
"loss": 0.6311,
"step": 600
},
{
"epoch": 1.6143482654092287,
"eval_loss": 0.784841775894165,
"eval_runtime": 153.2005,
"eval_samples_per_second": 32.637,
"eval_steps_per_second": 0.516,
"step": 600
},
{
"epoch": 1.6412933647692824,
"grad_norm": 1.1963943501637924,
"learning_rate": 2.2875226039783005e-05,
"loss": 0.6267,
"step": 610
},
{
"epoch": 1.6682384641293364,
"grad_norm": 1.093066884488607,
"learning_rate": 2.2423146473779386e-05,
"loss": 0.6289,
"step": 620
},
{
"epoch": 1.6951835634893904,
"grad_norm": 1.115011570991967,
"learning_rate": 2.197106690777577e-05,
"loss": 0.6299,
"step": 630
},
{
"epoch": 1.7221286628494443,
"grad_norm": 1.1700606931618611,
"learning_rate": 2.1518987341772153e-05,
"loss": 0.6233,
"step": 640
},
{
"epoch": 1.7490737622094983,
"grad_norm": 1.1565551201360744,
"learning_rate": 2.1066907775768534e-05,
"loss": 0.624,
"step": 650
},
{
"epoch": 1.7490737622094983,
"eval_loss": 0.7739421129226685,
"eval_runtime": 148.2001,
"eval_samples_per_second": 33.738,
"eval_steps_per_second": 0.533,
"step": 650
},
{
"epoch": 1.776018861569552,
"grad_norm": 1.2558394571278007,
"learning_rate": 2.061482820976492e-05,
"loss": 0.621,
"step": 660
},
{
"epoch": 1.802963960929606,
"grad_norm": 1.1487412967519839,
"learning_rate": 2.0162748643761304e-05,
"loss": 0.6201,
"step": 670
},
{
"epoch": 1.8299090602896597,
"grad_norm": 1.1633700130604714,
"learning_rate": 1.971066907775769e-05,
"loss": 0.6182,
"step": 680
},
{
"epoch": 1.8568541596497137,
"grad_norm": 1.0892080498580619,
"learning_rate": 1.925858951175407e-05,
"loss": 0.6218,
"step": 690
},
{
"epoch": 1.8837992590097676,
"grad_norm": 1.162193434275119,
"learning_rate": 1.8806509945750454e-05,
"loss": 0.612,
"step": 700
},
{
"epoch": 1.8837992590097676,
"eval_loss": 0.7593940496444702,
"eval_runtime": 113.5973,
"eval_samples_per_second": 44.015,
"eval_steps_per_second": 0.695,
"step": 700
},
{
"epoch": 1.9107443583698216,
"grad_norm": 1.477614030615022,
"learning_rate": 1.8354430379746836e-05,
"loss": 0.4361,
"step": 710
},
{
"epoch": 1.9376894577298753,
"grad_norm": 1.211596696990758,
"learning_rate": 1.7902350813743217e-05,
"loss": 0.4174,
"step": 720
},
{
"epoch": 1.9646345570899293,
"grad_norm": 1.208841417920371,
"learning_rate": 1.7450271247739602e-05,
"loss": 0.4162,
"step": 730
},
{
"epoch": 1.991579656449983,
"grad_norm": 1.2740412194613278,
"learning_rate": 1.6998191681735987e-05,
"loss": 0.4115,
"step": 740
},
{
"epoch": 2.0188615695520378,
"grad_norm": 1.1897059830697447,
"learning_rate": 1.654611211573237e-05,
"loss": 0.4224,
"step": 750
},
{
"epoch": 2.0188615695520378,
"eval_loss": 0.8485522270202637,
"eval_runtime": 115.805,
"eval_samples_per_second": 43.176,
"eval_steps_per_second": 0.682,
"step": 750
},
{
"epoch": 2.0458066689120917,
"grad_norm": 1.2120571201563564,
"learning_rate": 1.6094032549728753e-05,
"loss": 0.4192,
"step": 760
},
{
"epoch": 2.0727517682721457,
"grad_norm": 1.2006852199531266,
"learning_rate": 1.5641952983725134e-05,
"loss": 0.4101,
"step": 770
},
{
"epoch": 2.099696867632199,
"grad_norm": 1.2005979870584054,
"learning_rate": 1.5189873417721521e-05,
"loss": 0.4164,
"step": 780
},
{
"epoch": 2.126641966992253,
"grad_norm": 1.148665043003728,
"learning_rate": 1.4737793851717904e-05,
"loss": 0.4194,
"step": 790
},
{
"epoch": 2.153587066352307,
"grad_norm": 1.134303382163784,
"learning_rate": 1.4285714285714285e-05,
"loss": 0.4213,
"step": 800
},
{
"epoch": 2.153587066352307,
"eval_loss": 0.838707447052002,
"eval_runtime": 139.7297,
"eval_samples_per_second": 35.783,
"eval_steps_per_second": 0.565,
"step": 800
},
{
"epoch": 2.180532165712361,
"grad_norm": 1.2301409422954543,
"learning_rate": 1.383363471971067e-05,
"loss": 0.4179,
"step": 810
},
{
"epoch": 2.207477265072415,
"grad_norm": 1.1312377076286286,
"learning_rate": 1.3381555153707053e-05,
"loss": 0.4178,
"step": 820
},
{
"epoch": 2.234422364432469,
"grad_norm": 1.2707527556350058,
"learning_rate": 1.2929475587703435e-05,
"loss": 0.4125,
"step": 830
},
{
"epoch": 2.2613674637925225,
"grad_norm": 1.185885154215189,
"learning_rate": 1.247739602169982e-05,
"loss": 0.414,
"step": 840
},
{
"epoch": 2.2883125631525765,
"grad_norm": 1.1012295953218187,
"learning_rate": 1.2025316455696203e-05,
"loss": 0.414,
"step": 850
},
{
"epoch": 2.2883125631525765,
"eval_loss": 0.831937313079834,
"eval_runtime": 144.9314,
"eval_samples_per_second": 34.499,
"eval_steps_per_second": 0.545,
"step": 850
},
{
"epoch": 2.3152576625126304,
"grad_norm": 1.1388819607995708,
"learning_rate": 1.1573236889692586e-05,
"loss": 0.4151,
"step": 860
},
{
"epoch": 2.3422027618726844,
"grad_norm": 1.2140213674018405,
"learning_rate": 1.112115732368897e-05,
"loss": 0.4159,
"step": 870
},
{
"epoch": 2.3691478612327384,
"grad_norm": 1.1631774357185438,
"learning_rate": 1.0669077757685354e-05,
"loss": 0.4136,
"step": 880
},
{
"epoch": 2.3960929605927923,
"grad_norm": 1.24334716793248,
"learning_rate": 1.0216998191681737e-05,
"loss": 0.4076,
"step": 890
},
{
"epoch": 2.4230380599528463,
"grad_norm": 1.156296740645565,
"learning_rate": 9.76491862567812e-06,
"loss": 0.4145,
"step": 900
},
{
"epoch": 2.4230380599528463,
"eval_loss": 0.823464572429657,
"eval_runtime": 141.3831,
"eval_samples_per_second": 35.365,
"eval_steps_per_second": 0.559,
"step": 900
},
{
"epoch": 2.4499831593129,
"grad_norm": 1.2360058092342656,
"learning_rate": 9.312839059674505e-06,
"loss": 0.4071,
"step": 910
},
{
"epoch": 2.4769282586729537,
"grad_norm": 1.1672627428975908,
"learning_rate": 8.860759493670886e-06,
"loss": 0.4052,
"step": 920
},
{
"epoch": 2.5038733580330077,
"grad_norm": 1.142198240950353,
"learning_rate": 8.408679927667269e-06,
"loss": 0.3995,
"step": 930
},
{
"epoch": 2.5308184573930617,
"grad_norm": 1.1336163759218327,
"learning_rate": 7.956600361663654e-06,
"loss": 0.4109,
"step": 940
},
{
"epoch": 2.5577635567531156,
"grad_norm": 1.1520513715336371,
"learning_rate": 7.504520795660036e-06,
"loss": 0.4021,
"step": 950
},
{
"epoch": 2.5577635567531156,
"eval_loss": 0.8257409930229187,
"eval_runtime": 145.0235,
"eval_samples_per_second": 34.477,
"eval_steps_per_second": 0.545,
"step": 950
},
{
"epoch": 2.5847086561131696,
"grad_norm": 1.1137257622476067,
"learning_rate": 7.05244122965642e-06,
"loss": 0.3979,
"step": 960
},
{
"epoch": 2.611653755473223,
"grad_norm": 1.107368997807173,
"learning_rate": 6.600361663652803e-06,
"loss": 0.4011,
"step": 970
},
{
"epoch": 2.638598854833277,
"grad_norm": 1.1780589934268402,
"learning_rate": 6.148282097649186e-06,
"loss": 0.4013,
"step": 980
},
{
"epoch": 2.665543954193331,
"grad_norm": 1.174979594904003,
"learning_rate": 5.69620253164557e-06,
"loss": 0.3992,
"step": 990
},
{
"epoch": 2.692489053553385,
"grad_norm": 1.1164061245257322,
"learning_rate": 5.244122965641953e-06,
"loss": 0.393,
"step": 1000
},
{
"epoch": 2.692489053553385,
"eval_loss": 0.8172587156295776,
"eval_runtime": 147.8118,
"eval_samples_per_second": 33.827,
"eval_steps_per_second": 0.534,
"step": 1000
}
],
"logging_steps": 10,
"max_steps": 1116,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 50,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 588328797732864.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}