gr00tfoldingrobot / trainer_state.json
maxzand's picture
Upload folder using huggingface_hub
95760ff verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0395136778115504,
"eval_steps": 500,
"global_step": 1000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.030395136778115502,
"grad_norm": 9.889811515808105,
"learning_rate": 2.0000000000000003e-06,
"loss": 1.6102,
"step": 10
},
{
"epoch": 0.060790273556231005,
"grad_norm": 11.921159744262695,
"learning_rate": 4.000000000000001e-06,
"loss": 1.4517,
"step": 20
},
{
"epoch": 0.0911854103343465,
"grad_norm": 8.763319969177246,
"learning_rate": 6e-06,
"loss": 1.3953,
"step": 30
},
{
"epoch": 0.12158054711246201,
"grad_norm": 4.841191291809082,
"learning_rate": 8.000000000000001e-06,
"loss": 1.0262,
"step": 40
},
{
"epoch": 0.1519756838905775,
"grad_norm": 4.8184733390808105,
"learning_rate": 1e-05,
"loss": 0.6875,
"step": 50
},
{
"epoch": 0.182370820668693,
"grad_norm": 4.519493579864502,
"learning_rate": 1.2e-05,
"loss": 0.5443,
"step": 60
},
{
"epoch": 0.2127659574468085,
"grad_norm": 2.340122699737549,
"learning_rate": 1.4000000000000001e-05,
"loss": 0.3662,
"step": 70
},
{
"epoch": 0.24316109422492402,
"grad_norm": 1.4288235902786255,
"learning_rate": 1.6000000000000003e-05,
"loss": 0.2864,
"step": 80
},
{
"epoch": 0.2735562310030395,
"grad_norm": 1.979983925819397,
"learning_rate": 1.8e-05,
"loss": 0.2239,
"step": 90
},
{
"epoch": 0.303951367781155,
"grad_norm": 1.3639392852783203,
"learning_rate": 2e-05,
"loss": 0.166,
"step": 100
},
{
"epoch": 0.3343465045592705,
"grad_norm": 0.9854844808578491,
"learning_rate": 2.2000000000000003e-05,
"loss": 0.1515,
"step": 110
},
{
"epoch": 0.364741641337386,
"grad_norm": 1.101383924484253,
"learning_rate": 2.4e-05,
"loss": 0.1637,
"step": 120
},
{
"epoch": 0.3951367781155015,
"grad_norm": 1.665946125984192,
"learning_rate": 2.6000000000000002e-05,
"loss": 0.1045,
"step": 130
},
{
"epoch": 0.425531914893617,
"grad_norm": 0.7495555877685547,
"learning_rate": 2.8000000000000003e-05,
"loss": 0.1058,
"step": 140
},
{
"epoch": 0.45592705167173253,
"grad_norm": 1.0226916074752808,
"learning_rate": 3e-05,
"loss": 0.1079,
"step": 150
},
{
"epoch": 0.48632218844984804,
"grad_norm": 0.6161096096038818,
"learning_rate": 3.2000000000000005e-05,
"loss": 0.0993,
"step": 160
},
{
"epoch": 0.5167173252279635,
"grad_norm": 0.7106040120124817,
"learning_rate": 3.4000000000000007e-05,
"loss": 0.1079,
"step": 170
},
{
"epoch": 0.547112462006079,
"grad_norm": 1.0743619203567505,
"learning_rate": 3.6e-05,
"loss": 0.0882,
"step": 180
},
{
"epoch": 0.5775075987841946,
"grad_norm": 0.85890132188797,
"learning_rate": 3.8e-05,
"loss": 0.0943,
"step": 190
},
{
"epoch": 0.60790273556231,
"grad_norm": 1.4179412126541138,
"learning_rate": 4e-05,
"loss": 0.0969,
"step": 200
},
{
"epoch": 0.6382978723404256,
"grad_norm": 2.021270513534546,
"learning_rate": 4.2e-05,
"loss": 0.0934,
"step": 210
},
{
"epoch": 0.668693009118541,
"grad_norm": 1.4204037189483643,
"learning_rate": 4.4000000000000006e-05,
"loss": 0.0992,
"step": 220
},
{
"epoch": 0.6990881458966566,
"grad_norm": 1.0484898090362549,
"learning_rate": 4.600000000000001e-05,
"loss": 0.0807,
"step": 230
},
{
"epoch": 0.729483282674772,
"grad_norm": 1.0355311632156372,
"learning_rate": 4.8e-05,
"loss": 0.0827,
"step": 240
},
{
"epoch": 0.7598784194528876,
"grad_norm": 0.7956388592720032,
"learning_rate": 5e-05,
"loss": 0.0818,
"step": 250
},
{
"epoch": 0.790273556231003,
"grad_norm": 0.8189113736152649,
"learning_rate": 5.2000000000000004e-05,
"loss": 0.076,
"step": 260
},
{
"epoch": 0.8206686930091185,
"grad_norm": 0.6537955403327942,
"learning_rate": 5.4000000000000005e-05,
"loss": 0.0738,
"step": 270
},
{
"epoch": 0.851063829787234,
"grad_norm": 0.8595532774925232,
"learning_rate": 5.6000000000000006e-05,
"loss": 0.0797,
"step": 280
},
{
"epoch": 0.8814589665653495,
"grad_norm": 0.9948979020118713,
"learning_rate": 5.8e-05,
"loss": 0.0674,
"step": 290
},
{
"epoch": 0.9118541033434651,
"grad_norm": 0.7822934985160828,
"learning_rate": 6e-05,
"loss": 0.0839,
"step": 300
},
{
"epoch": 0.9422492401215805,
"grad_norm": 0.9252318143844604,
"learning_rate": 6.2e-05,
"loss": 0.0657,
"step": 310
},
{
"epoch": 0.9726443768996961,
"grad_norm": 1.2264695167541504,
"learning_rate": 6.400000000000001e-05,
"loss": 0.0751,
"step": 320
},
{
"epoch": 1.0030395136778116,
"grad_norm": 0.6637468338012695,
"learning_rate": 6.6e-05,
"loss": 0.0766,
"step": 330
},
{
"epoch": 1.033434650455927,
"grad_norm": 1.3035773038864136,
"learning_rate": 6.800000000000001e-05,
"loss": 0.0637,
"step": 340
},
{
"epoch": 1.0638297872340425,
"grad_norm": 0.8334704637527466,
"learning_rate": 7e-05,
"loss": 0.0753,
"step": 350
},
{
"epoch": 1.094224924012158,
"grad_norm": 1.2079942226409912,
"learning_rate": 7.2e-05,
"loss": 0.0729,
"step": 360
},
{
"epoch": 1.1246200607902737,
"grad_norm": 0.8719915747642517,
"learning_rate": 7.4e-05,
"loss": 0.0577,
"step": 370
},
{
"epoch": 1.155015197568389,
"grad_norm": 0.9174433350563049,
"learning_rate": 7.6e-05,
"loss": 0.0705,
"step": 380
},
{
"epoch": 1.1854103343465046,
"grad_norm": 1.2214233875274658,
"learning_rate": 7.800000000000001e-05,
"loss": 0.0837,
"step": 390
},
{
"epoch": 1.21580547112462,
"grad_norm": 1.2708603143692017,
"learning_rate": 8e-05,
"loss": 0.0658,
"step": 400
},
{
"epoch": 1.2462006079027357,
"grad_norm": 1.0723034143447876,
"learning_rate": 8.2e-05,
"loss": 0.0625,
"step": 410
},
{
"epoch": 1.2765957446808511,
"grad_norm": 1.0916433334350586,
"learning_rate": 8.4e-05,
"loss": 0.087,
"step": 420
},
{
"epoch": 1.3069908814589666,
"grad_norm": 0.6322633624076843,
"learning_rate": 8.6e-05,
"loss": 0.0722,
"step": 430
},
{
"epoch": 1.337386018237082,
"grad_norm": 0.5855675339698792,
"learning_rate": 8.800000000000001e-05,
"loss": 0.0635,
"step": 440
},
{
"epoch": 1.3677811550151975,
"grad_norm": 1.0976413488388062,
"learning_rate": 9e-05,
"loss": 0.0664,
"step": 450
},
{
"epoch": 1.3981762917933132,
"grad_norm": 0.9359951019287109,
"learning_rate": 9.200000000000001e-05,
"loss": 0.0647,
"step": 460
},
{
"epoch": 1.4285714285714286,
"grad_norm": 0.7038888335227966,
"learning_rate": 9.4e-05,
"loss": 0.0716,
"step": 470
},
{
"epoch": 1.458966565349544,
"grad_norm": 0.5396299958229065,
"learning_rate": 9.6e-05,
"loss": 0.0527,
"step": 480
},
{
"epoch": 1.4893617021276595,
"grad_norm": 0.5245558619499207,
"learning_rate": 9.8e-05,
"loss": 0.0627,
"step": 490
},
{
"epoch": 1.5197568389057752,
"grad_norm": 1.350913405418396,
"learning_rate": 0.0001,
"loss": 0.0555,
"step": 500
},
{
"epoch": 1.5501519756838906,
"grad_norm": 0.6774181723594666,
"learning_rate": 9.999972660400536e-05,
"loss": 0.061,
"step": 510
},
{
"epoch": 1.580547112462006,
"grad_norm": 0.5291458964347839,
"learning_rate": 9.999890641901125e-05,
"loss": 0.0497,
"step": 520
},
{
"epoch": 1.6109422492401215,
"grad_norm": 0.8055920600891113,
"learning_rate": 9.999753945398704e-05,
"loss": 0.0612,
"step": 530
},
{
"epoch": 1.641337386018237,
"grad_norm": 0.6707597374916077,
"learning_rate": 9.99956257238817e-05,
"loss": 0.0618,
"step": 540
},
{
"epoch": 1.6717325227963524,
"grad_norm": 0.6728675365447998,
"learning_rate": 9.999316524962345e-05,
"loss": 0.0569,
"step": 550
},
{
"epoch": 1.702127659574468,
"grad_norm": 0.5206692814826965,
"learning_rate": 9.999015805811965e-05,
"loss": 0.0574,
"step": 560
},
{
"epoch": 1.7325227963525835,
"grad_norm": 0.7465855479240417,
"learning_rate": 9.998660418225645e-05,
"loss": 0.0665,
"step": 570
},
{
"epoch": 1.7629179331306992,
"grad_norm": 0.5522341728210449,
"learning_rate": 9.998250366089848e-05,
"loss": 0.072,
"step": 580
},
{
"epoch": 1.7933130699088147,
"grad_norm": 0.6498285531997681,
"learning_rate": 9.997785653888835e-05,
"loss": 0.0471,
"step": 590
},
{
"epoch": 1.8237082066869301,
"grad_norm": 0.8599438071250916,
"learning_rate": 9.997266286704631e-05,
"loss": 0.056,
"step": 600
},
{
"epoch": 1.8541033434650456,
"grad_norm": 2.1041059494018555,
"learning_rate": 9.996692270216947e-05,
"loss": 0.0732,
"step": 610
},
{
"epoch": 1.884498480243161,
"grad_norm": 1.016334891319275,
"learning_rate": 9.996063610703137e-05,
"loss": 0.0657,
"step": 620
},
{
"epoch": 1.9148936170212765,
"grad_norm": 0.6369310021400452,
"learning_rate": 9.995380315038119e-05,
"loss": 0.0499,
"step": 630
},
{
"epoch": 1.9452887537993921,
"grad_norm": 0.6747097969055176,
"learning_rate": 9.994642390694308e-05,
"loss": 0.0552,
"step": 640
},
{
"epoch": 1.9756838905775076,
"grad_norm": 0.8461899161338806,
"learning_rate": 9.993849845741524e-05,
"loss": 0.051,
"step": 650
},
{
"epoch": 2.0060790273556233,
"grad_norm": 1.1829544305801392,
"learning_rate": 9.993002688846913e-05,
"loss": 0.0548,
"step": 660
},
{
"epoch": 2.0364741641337387,
"grad_norm": 0.5973241329193115,
"learning_rate": 9.992100929274846e-05,
"loss": 0.0505,
"step": 670
},
{
"epoch": 2.066869300911854,
"grad_norm": 0.4595549404621124,
"learning_rate": 9.991144576886823e-05,
"loss": 0.064,
"step": 680
},
{
"epoch": 2.0972644376899696,
"grad_norm": 0.5466217994689941,
"learning_rate": 9.990133642141359e-05,
"loss": 0.0475,
"step": 690
},
{
"epoch": 2.127659574468085,
"grad_norm": 0.4098721146583557,
"learning_rate": 9.989068136093873e-05,
"loss": 0.0459,
"step": 700
},
{
"epoch": 2.1580547112462005,
"grad_norm": 0.5519854426383972,
"learning_rate": 9.987948070396571e-05,
"loss": 0.0463,
"step": 710
},
{
"epoch": 2.188449848024316,
"grad_norm": 0.5170401930809021,
"learning_rate": 9.986773457298311e-05,
"loss": 0.0484,
"step": 720
},
{
"epoch": 2.2188449848024314,
"grad_norm": 0.3473241925239563,
"learning_rate": 9.985544309644475e-05,
"loss": 0.0546,
"step": 730
},
{
"epoch": 2.2492401215805473,
"grad_norm": 0.38744667172431946,
"learning_rate": 9.984260640876821e-05,
"loss": 0.0428,
"step": 740
},
{
"epoch": 2.2796352583586628,
"grad_norm": 0.7359247803688049,
"learning_rate": 9.98292246503335e-05,
"loss": 0.0592,
"step": 750
},
{
"epoch": 2.310030395136778,
"grad_norm": 0.5814300179481506,
"learning_rate": 9.981529796748134e-05,
"loss": 0.0399,
"step": 760
},
{
"epoch": 2.3404255319148937,
"grad_norm": 0.5898028016090393,
"learning_rate": 9.980082651251175e-05,
"loss": 0.046,
"step": 770
},
{
"epoch": 2.370820668693009,
"grad_norm": 0.4852890968322754,
"learning_rate": 9.97858104436822e-05,
"loss": 0.046,
"step": 780
},
{
"epoch": 2.4012158054711246,
"grad_norm": 0.4703480899333954,
"learning_rate": 9.977024992520602e-05,
"loss": 0.0374,
"step": 790
},
{
"epoch": 2.43161094224924,
"grad_norm": 0.43933120369911194,
"learning_rate": 9.975414512725057e-05,
"loss": 0.0454,
"step": 800
},
{
"epoch": 2.4620060790273555,
"grad_norm": 0.639541745185852,
"learning_rate": 9.973749622593534e-05,
"loss": 0.0379,
"step": 810
},
{
"epoch": 2.4924012158054714,
"grad_norm": 0.6174078583717346,
"learning_rate": 9.972030340333001e-05,
"loss": 0.0539,
"step": 820
},
{
"epoch": 2.522796352583587,
"grad_norm": 0.49048295617103577,
"learning_rate": 9.970256684745258e-05,
"loss": 0.045,
"step": 830
},
{
"epoch": 2.5531914893617023,
"grad_norm": 0.3859421908855438,
"learning_rate": 9.968428675226714e-05,
"loss": 0.038,
"step": 840
},
{
"epoch": 2.5835866261398177,
"grad_norm": 0.5661002993583679,
"learning_rate": 9.966546331768191e-05,
"loss": 0.0377,
"step": 850
},
{
"epoch": 2.613981762917933,
"grad_norm": 0.6421768665313721,
"learning_rate": 9.964609674954696e-05,
"loss": 0.0516,
"step": 860
},
{
"epoch": 2.6443768996960486,
"grad_norm": 0.36835750937461853,
"learning_rate": 9.962618725965196e-05,
"loss": 0.0354,
"step": 870
},
{
"epoch": 2.674772036474164,
"grad_norm": 0.34994593262672424,
"learning_rate": 9.96057350657239e-05,
"loss": 0.0391,
"step": 880
},
{
"epoch": 2.7051671732522795,
"grad_norm": 0.501876175403595,
"learning_rate": 9.95847403914247e-05,
"loss": 0.0459,
"step": 890
},
{
"epoch": 2.735562310030395,
"grad_norm": 0.4514513611793518,
"learning_rate": 9.956320346634876e-05,
"loss": 0.0368,
"step": 900
},
{
"epoch": 2.7659574468085104,
"grad_norm": 0.4428149461746216,
"learning_rate": 9.954112452602045e-05,
"loss": 0.033,
"step": 910
},
{
"epoch": 2.7963525835866263,
"grad_norm": 0.7399154305458069,
"learning_rate": 9.95185038118915e-05,
"loss": 0.0443,
"step": 920
},
{
"epoch": 2.8267477203647418,
"grad_norm": 0.4058426320552826,
"learning_rate": 9.949534157133844e-05,
"loss": 0.051,
"step": 930
},
{
"epoch": 2.857142857142857,
"grad_norm": 0.3738716244697571,
"learning_rate": 9.94716380576598e-05,
"loss": 0.0318,
"step": 940
},
{
"epoch": 2.8875379939209727,
"grad_norm": 0.5219188332557678,
"learning_rate": 9.944739353007344e-05,
"loss": 0.0383,
"step": 950
},
{
"epoch": 2.917933130699088,
"grad_norm": 0.28857406973838806,
"learning_rate": 9.942260825371358e-05,
"loss": 0.0413,
"step": 960
},
{
"epoch": 2.9483282674772036,
"grad_norm": 0.4435504674911499,
"learning_rate": 9.939728249962807e-05,
"loss": 0.0331,
"step": 970
},
{
"epoch": 2.978723404255319,
"grad_norm": 0.6941442489624023,
"learning_rate": 9.937141654477528e-05,
"loss": 0.039,
"step": 980
},
{
"epoch": 3.0091185410334345,
"grad_norm": 0.5373895764350891,
"learning_rate": 9.934501067202117e-05,
"loss": 0.0434,
"step": 990
},
{
"epoch": 3.0395136778115504,
"grad_norm": 0.3684053122997284,
"learning_rate": 9.931806517013612e-05,
"loss": 0.0367,
"step": 1000
}
],
"logging_steps": 10,
"max_steps": 10000,
"num_input_tokens_seen": 0,
"num_train_epochs": 31,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.301578581152896e+16,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}