AgThinker-32B-mid / trainer_state.json
chegde's picture
Add files using upload-large-folder tool
3109990 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.149184149184149,
"eval_steps": 500,
"global_step": 1500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.014344629729245113,
"grad_norm": 2.203942058202339,
"learning_rate": 4.285714285714286e-06,
"loss": 1.1704,
"step": 10
},
{
"epoch": 0.028689259458490227,
"grad_norm": 1.1497467131489452,
"learning_rate": 9.047619047619047e-06,
"loss": 0.9166,
"step": 20
},
{
"epoch": 0.04303388918773534,
"grad_norm": 0.8889709375752792,
"learning_rate": 1.3809523809523811e-05,
"loss": 0.7948,
"step": 30
},
{
"epoch": 0.05737851891698045,
"grad_norm": 0.9291104053668569,
"learning_rate": 1.8571428571428572e-05,
"loss": 0.7447,
"step": 40
},
{
"epoch": 0.07172314864622557,
"grad_norm": 1.1881747732049386,
"learning_rate": 2.3333333333333336e-05,
"loss": 0.729,
"step": 50
},
{
"epoch": 0.08606777837547068,
"grad_norm": 0.9032703792102055,
"learning_rate": 2.8095238095238096e-05,
"loss": 0.7177,
"step": 60
},
{
"epoch": 0.1004124081047158,
"grad_norm": 0.9447848465031972,
"learning_rate": 3.285714285714286e-05,
"loss": 0.7048,
"step": 70
},
{
"epoch": 0.1147570378339609,
"grad_norm": 0.9807912049624671,
"learning_rate": 3.761904761904762e-05,
"loss": 0.7048,
"step": 80
},
{
"epoch": 0.129101667563206,
"grad_norm": 0.8678523409043055,
"learning_rate": 4.2380952380952385e-05,
"loss": 0.7065,
"step": 90
},
{
"epoch": 0.14344629729245115,
"grad_norm": 1.095952408037045,
"learning_rate": 4.714285714285714e-05,
"loss": 0.7098,
"step": 100
},
{
"epoch": 0.15779092702169625,
"grad_norm": 0.932556439833824,
"learning_rate": 5.1904761904761913e-05,
"loss": 0.7047,
"step": 110
},
{
"epoch": 0.17213555675094136,
"grad_norm": 0.716991228951169,
"learning_rate": 5.666666666666667e-05,
"loss": 0.704,
"step": 120
},
{
"epoch": 0.1864801864801865,
"grad_norm": 0.7487957467977157,
"learning_rate": 6.142857142857143e-05,
"loss": 0.7013,
"step": 130
},
{
"epoch": 0.2008248162094316,
"grad_norm": 0.7206473376735171,
"learning_rate": 6.619047619047619e-05,
"loss": 0.7038,
"step": 140
},
{
"epoch": 0.2151694459386767,
"grad_norm": 0.6694694288559293,
"learning_rate": 7.095238095238096e-05,
"loss": 0.7119,
"step": 150
},
{
"epoch": 0.2295140756679218,
"grad_norm": 0.7219727300855844,
"learning_rate": 7.571428571428571e-05,
"loss": 0.7068,
"step": 160
},
{
"epoch": 0.24385870539716695,
"grad_norm": 0.6818226409038356,
"learning_rate": 8.047619047619048e-05,
"loss": 0.7088,
"step": 170
},
{
"epoch": 0.258203335126412,
"grad_norm": 0.6977379507391648,
"learning_rate": 8.523809523809524e-05,
"loss": 0.7032,
"step": 180
},
{
"epoch": 0.2725479648556572,
"grad_norm": 0.6553925195793743,
"learning_rate": 9e-05,
"loss": 0.7081,
"step": 190
},
{
"epoch": 0.2868925945849023,
"grad_norm": 0.5491527719782942,
"learning_rate": 9.476190476190476e-05,
"loss": 0.7079,
"step": 200
},
{
"epoch": 0.3012372243141474,
"grad_norm": 0.6640413812930087,
"learning_rate": 9.952380952380953e-05,
"loss": 0.7153,
"step": 210
},
{
"epoch": 0.3155818540433925,
"grad_norm": 0.5699857338148432,
"learning_rate": 9.999436939807164e-05,
"loss": 0.7235,
"step": 220
},
{
"epoch": 0.3299264837726376,
"grad_norm": 0.6222679010251994,
"learning_rate": 9.99749072170404e-05,
"loss": 0.7215,
"step": 230
},
{
"epoch": 0.3442711135018827,
"grad_norm": 0.529203333569492,
"learning_rate": 9.994154935353517e-05,
"loss": 0.7074,
"step": 240
},
{
"epoch": 0.3586157432311278,
"grad_norm": 0.524281983121501,
"learning_rate": 9.98943050828164e-05,
"loss": 0.7086,
"step": 250
},
{
"epoch": 0.372960372960373,
"grad_norm": 0.5037155864766709,
"learning_rate": 9.983318754130435e-05,
"loss": 0.7004,
"step": 260
},
{
"epoch": 0.3873050026896181,
"grad_norm": 0.5619940203814628,
"learning_rate": 9.975821372292653e-05,
"loss": 0.6961,
"step": 270
},
{
"epoch": 0.4016496324188632,
"grad_norm": 0.46139196683094363,
"learning_rate": 9.966940447439245e-05,
"loss": 0.6914,
"step": 280
},
{
"epoch": 0.4159942621481083,
"grad_norm": 0.5701814263114097,
"learning_rate": 9.956678448939718e-05,
"loss": 0.6805,
"step": 290
},
{
"epoch": 0.4303388918773534,
"grad_norm": 0.5822888434573161,
"learning_rate": 9.945038230175509e-05,
"loss": 0.6918,
"step": 300
},
{
"epoch": 0.4446835216065985,
"grad_norm": 0.44360325798902245,
"learning_rate": 9.932023027746602e-05,
"loss": 0.6798,
"step": 310
},
{
"epoch": 0.4590281513358436,
"grad_norm": 0.4584389136933647,
"learning_rate": 9.917636460571578e-05,
"loss": 0.6848,
"step": 320
},
{
"epoch": 0.47337278106508873,
"grad_norm": 0.42186486444358706,
"learning_rate": 9.901882528881363e-05,
"loss": 0.6856,
"step": 330
},
{
"epoch": 0.4877174107943339,
"grad_norm": 0.4547332297314441,
"learning_rate": 9.884765613106948e-05,
"loss": 0.6761,
"step": 340
},
{
"epoch": 0.502062040523579,
"grad_norm": 0.5669430722485119,
"learning_rate": 9.866290472661406e-05,
"loss": 0.6795,
"step": 350
},
{
"epoch": 0.516406670252824,
"grad_norm": 0.4480667658595798,
"learning_rate": 9.846462244616508e-05,
"loss": 0.6701,
"step": 360
},
{
"epoch": 0.5307512999820692,
"grad_norm": 0.4330593851481768,
"learning_rate": 9.825286442274357e-05,
"loss": 0.6785,
"step": 370
},
{
"epoch": 0.5450959297113144,
"grad_norm": 0.43973236782853137,
"learning_rate": 9.802768953634388e-05,
"loss": 0.6692,
"step": 380
},
{
"epoch": 0.5594405594405595,
"grad_norm": 0.4800141760306811,
"learning_rate": 9.778916039756193e-05,
"loss": 0.6644,
"step": 390
},
{
"epoch": 0.5737851891698046,
"grad_norm": 0.4370272358079981,
"learning_rate": 9.753734333018616e-05,
"loss": 0.6582,
"step": 400
},
{
"epoch": 0.5881298188990497,
"grad_norm": 0.46636751943765076,
"learning_rate": 9.727230835275598e-05,
"loss": 0.6606,
"step": 410
},
{
"epoch": 0.6024744486282948,
"grad_norm": 0.44398226842129596,
"learning_rate": 9.699412915909284e-05,
"loss": 0.6591,
"step": 420
},
{
"epoch": 0.6168190783575399,
"grad_norm": 0.417529788133883,
"learning_rate": 9.670288309780953e-05,
"loss": 0.6502,
"step": 430
},
{
"epoch": 0.631163708086785,
"grad_norm": 0.425244481854258,
"learning_rate": 9.639865115080304e-05,
"loss": 0.6601,
"step": 440
},
{
"epoch": 0.6455083378160301,
"grad_norm": 0.4266753216646151,
"learning_rate": 9.608151791073737e-05,
"loss": 0.6552,
"step": 450
},
{
"epoch": 0.6598529675452752,
"grad_norm": 0.4060939108779214,
"learning_rate": 9.575157155752222e-05,
"loss": 0.6483,
"step": 460
},
{
"epoch": 0.6741975972745203,
"grad_norm": 0.44733731016902734,
"learning_rate": 9.54089038337943e-05,
"loss": 0.6414,
"step": 470
},
{
"epoch": 0.6885422270037654,
"grad_norm": 0.4343263599516913,
"learning_rate": 9.5053610019408e-05,
"loss": 0.6452,
"step": 480
},
{
"epoch": 0.7028868567330105,
"grad_norm": 0.40484892671799194,
"learning_rate": 9.468578890494256e-05,
"loss": 0.6399,
"step": 490
},
{
"epoch": 0.7172314864622557,
"grad_norm": 0.3977086895873736,
"learning_rate": 9.430554276423292e-05,
"loss": 0.6427,
"step": 500
},
{
"epoch": 0.7315761161915008,
"grad_norm": 0.4059393728648637,
"learning_rate": 9.391297732593229e-05,
"loss": 0.6405,
"step": 510
},
{
"epoch": 0.745920745920746,
"grad_norm": 0.3918015124178049,
"learning_rate": 9.350820174411386e-05,
"loss": 0.6428,
"step": 520
},
{
"epoch": 0.7602653756499911,
"grad_norm": 0.397284227604159,
"learning_rate": 9.309132856792023e-05,
"loss": 0.6379,
"step": 530
},
{
"epoch": 0.7746100053792362,
"grad_norm": 0.423979379524061,
"learning_rate": 9.266247371026873e-05,
"loss": 0.6346,
"step": 540
},
{
"epoch": 0.7889546351084813,
"grad_norm": 0.3974744664267281,
"learning_rate": 9.222175641562143e-05,
"loss": 0.6389,
"step": 550
},
{
"epoch": 0.8032992648377264,
"grad_norm": 0.40660189053377105,
"learning_rate": 9.176929922682891e-05,
"loss": 0.6328,
"step": 560
},
{
"epoch": 0.8176438945669715,
"grad_norm": 0.39183846193855565,
"learning_rate": 9.130522795105676e-05,
"loss": 0.6332,
"step": 570
},
{
"epoch": 0.8319885242962166,
"grad_norm": 0.3794359883589157,
"learning_rate": 9.082967162480459e-05,
"loss": 0.6248,
"step": 580
},
{
"epoch": 0.8463331540254617,
"grad_norm": 0.3757192118653646,
"learning_rate": 9.034276247802688e-05,
"loss": 0.6204,
"step": 590
},
{
"epoch": 0.8606777837547068,
"grad_norm": 0.4344414187841982,
"learning_rate": 8.984463589736614e-05,
"loss": 0.6253,
"step": 600
},
{
"epoch": 0.8750224134839519,
"grad_norm": 0.4124738293723295,
"learning_rate": 8.933543038850816e-05,
"loss": 0.6254,
"step": 610
},
{
"epoch": 0.889367043213197,
"grad_norm": 0.3729179255559684,
"learning_rate": 8.881528753767007e-05,
"loss": 0.6225,
"step": 620
},
{
"epoch": 0.9037116729424421,
"grad_norm": 0.3999425136973519,
"learning_rate": 8.82843519722319e-05,
"loss": 0.6219,
"step": 630
},
{
"epoch": 0.9180563026716873,
"grad_norm": 0.3613041053584094,
"learning_rate": 8.774277132052237e-05,
"loss": 0.6151,
"step": 640
},
{
"epoch": 0.9324009324009324,
"grad_norm": 0.3606937460298364,
"learning_rate": 8.719069617077046e-05,
"loss": 0.6216,
"step": 650
},
{
"epoch": 0.9467455621301775,
"grad_norm": 0.42662552468134024,
"learning_rate": 8.662828002923378e-05,
"loss": 0.6179,
"step": 660
},
{
"epoch": 0.9610901918594227,
"grad_norm": 0.3630873900729852,
"learning_rate": 8.605567927751576e-05,
"loss": 0.6171,
"step": 670
},
{
"epoch": 0.9754348215886678,
"grad_norm": 0.3905293019503008,
"learning_rate": 8.547305312908318e-05,
"loss": 0.6085,
"step": 680
},
{
"epoch": 0.9897794513179129,
"grad_norm": 0.36834813483218043,
"learning_rate": 8.48805635849964e-05,
"loss": 0.6074,
"step": 690
},
{
"epoch": 1.002868925945849,
"grad_norm": 0.39960980145566394,
"learning_rate": 8.427837538886437e-05,
"loss": 0.5784,
"step": 700
},
{
"epoch": 1.017213555675094,
"grad_norm": 0.38333026599469155,
"learning_rate": 8.366665598103727e-05,
"loss": 0.4367,
"step": 710
},
{
"epoch": 1.0315581854043392,
"grad_norm": 0.3516312791289111,
"learning_rate": 8.304557545204908e-05,
"loss": 0.4364,
"step": 720
},
{
"epoch": 1.0459028151335843,
"grad_norm": 0.33320216950958276,
"learning_rate": 8.241530649532339e-05,
"loss": 0.4252,
"step": 730
},
{
"epoch": 1.0602474448628294,
"grad_norm": 0.3367270532117864,
"learning_rate": 8.177602435915546e-05,
"loss": 0.4245,
"step": 740
},
{
"epoch": 1.0745920745920745,
"grad_norm": 0.36314384787697035,
"learning_rate": 8.11279067979839e-05,
"loss": 0.4172,
"step": 750
},
{
"epoch": 1.0889367043213196,
"grad_norm": 0.3331116126500675,
"learning_rate": 8.04711340229654e-05,
"loss": 0.4117,
"step": 760
},
{
"epoch": 1.1032813340505647,
"grad_norm": 0.3316800823164157,
"learning_rate": 7.980588865186649e-05,
"loss": 0.4217,
"step": 770
},
{
"epoch": 1.11762596377981,
"grad_norm": 0.37471291784176136,
"learning_rate": 7.913235565828613e-05,
"loss": 0.4178,
"step": 780
},
{
"epoch": 1.1319705935090552,
"grad_norm": 0.3734129628024853,
"learning_rate": 7.845072232022311e-05,
"loss": 0.4173,
"step": 790
},
{
"epoch": 1.1463152232383003,
"grad_norm": 0.33461223045242555,
"learning_rate": 7.776117816800288e-05,
"loss": 0.4258,
"step": 800
},
{
"epoch": 1.1606598529675454,
"grad_norm": 0.3149294832232542,
"learning_rate": 7.706391493157805e-05,
"loss": 0.4194,
"step": 810
},
{
"epoch": 1.1750044826967905,
"grad_norm": 0.33310067113179365,
"learning_rate": 7.635912648721718e-05,
"loss": 0.4159,
"step": 820
},
{
"epoch": 1.1893491124260356,
"grad_norm": 0.34107614209430825,
"learning_rate": 7.564700880359696e-05,
"loss": 0.4193,
"step": 830
},
{
"epoch": 1.2036937421552807,
"grad_norm": 0.3033843674878663,
"learning_rate": 7.492775988731243e-05,
"loss": 0.4096,
"step": 840
},
{
"epoch": 1.2180383718845258,
"grad_norm": 0.36776234094697624,
"learning_rate": 7.420157972782063e-05,
"loss": 0.4179,
"step": 850
},
{
"epoch": 1.232383001613771,
"grad_norm": 0.32016874059260647,
"learning_rate": 7.346867024183291e-05,
"loss": 0.4178,
"step": 860
},
{
"epoch": 1.246727631343016,
"grad_norm": 0.3461545416340315,
"learning_rate": 7.272923521717133e-05,
"loss": 0.4178,
"step": 870
},
{
"epoch": 1.2610722610722611,
"grad_norm": 0.3239854507991741,
"learning_rate": 7.198348025610481e-05,
"loss": 0.4173,
"step": 880
},
{
"epoch": 1.2754168908015062,
"grad_norm": 0.35967022774745094,
"learning_rate": 7.12316127181808e-05,
"loss": 0.4128,
"step": 890
},
{
"epoch": 1.2897615205307513,
"grad_norm": 0.33647912683040776,
"learning_rate": 7.047384166256815e-05,
"loss": 0.4206,
"step": 900
},
{
"epoch": 1.3041061502599964,
"grad_norm": 0.34026212090737123,
"learning_rate": 6.971037778992775e-05,
"loss": 0.4196,
"step": 910
},
{
"epoch": 1.3184507799892415,
"grad_norm": 0.31786823344694176,
"learning_rate": 6.894143338382639e-05,
"loss": 0.4217,
"step": 920
},
{
"epoch": 1.3327954097184866,
"grad_norm": 0.3189700837737566,
"learning_rate": 6.81672222517107e-05,
"loss": 0.4225,
"step": 930
},
{
"epoch": 1.3471400394477318,
"grad_norm": 0.3275473534649835,
"learning_rate": 6.73879596654573e-05,
"loss": 0.4119,
"step": 940
},
{
"epoch": 1.3614846691769769,
"grad_norm": 0.3420058843642656,
"learning_rate": 6.660386230151571e-05,
"loss": 0.4153,
"step": 950
},
{
"epoch": 1.375829298906222,
"grad_norm": 0.36451277332483983,
"learning_rate": 6.581514818066088e-05,
"loss": 0.4176,
"step": 960
},
{
"epoch": 1.390173928635467,
"grad_norm": 0.3392630329016263,
"learning_rate": 6.502203660737169e-05,
"loss": 0.4146,
"step": 970
},
{
"epoch": 1.4045185583647122,
"grad_norm": 0.32756084558577897,
"learning_rate": 6.422474810885278e-05,
"loss": 0.4093,
"step": 980
},
{
"epoch": 1.4188631880939573,
"grad_norm": 0.31495606064539233,
"learning_rate": 6.342350437371614e-05,
"loss": 0.4127,
"step": 990
},
{
"epoch": 1.4332078178232024,
"grad_norm": 0.3106906576340927,
"learning_rate": 6.26185281903399e-05,
"loss": 0.4109,
"step": 1000
},
{
"epoch": 1.4475524475524475,
"grad_norm": 0.34231901054792907,
"learning_rate": 6.181004338492141e-05,
"loss": 0.4115,
"step": 1010
},
{
"epoch": 1.4618970772816926,
"grad_norm": 0.2998753745826404,
"learning_rate": 6.09982747592415e-05,
"loss": 0.4063,
"step": 1020
},
{
"epoch": 1.4762417070109377,
"grad_norm": 0.3030265142945988,
"learning_rate": 6.018344802815778e-05,
"loss": 0.4096,
"step": 1030
},
{
"epoch": 1.4905863367401828,
"grad_norm": 0.3209761231799724,
"learning_rate": 5.936578975684378e-05,
"loss": 0.4115,
"step": 1040
},
{
"epoch": 1.504930966469428,
"grad_norm": 0.3117296901575659,
"learning_rate": 5.854552729779184e-05,
"loss": 0.4083,
"step": 1050
},
{
"epoch": 1.519275596198673,
"grad_norm": 0.32123211938760937,
"learning_rate": 5.772288872759702e-05,
"loss": 0.4068,
"step": 1060
},
{
"epoch": 1.5336202259279181,
"grad_norm": 0.32646003873154356,
"learning_rate": 5.6898102783539665e-05,
"loss": 0.4037,
"step": 1070
},
{
"epoch": 1.5479648556571632,
"grad_norm": 0.31930171711882954,
"learning_rate": 5.607139879998427e-05,
"loss": 0.4052,
"step": 1080
},
{
"epoch": 1.5623094853864083,
"grad_norm": 0.2987369966742978,
"learning_rate": 5.524300664461235e-05,
"loss": 0.4041,
"step": 1090
},
{
"epoch": 1.5766541151156535,
"grad_norm": 0.29715944144054796,
"learning_rate": 5.441315665450697e-05,
"loss": 0.408,
"step": 1100
},
{
"epoch": 1.5909987448448986,
"grad_norm": 0.300825472928675,
"learning_rate": 5.3582079572106794e-05,
"loss": 0.3996,
"step": 1110
},
{
"epoch": 1.6053433745741437,
"grad_norm": 0.321003899975276,
"learning_rate": 5.275000648104743e-05,
"loss": 0.4012,
"step": 1120
},
{
"epoch": 1.6196880043033888,
"grad_norm": 0.3068035924351862,
"learning_rate": 5.191716874190785e-05,
"loss": 0.4056,
"step": 1130
},
{
"epoch": 1.6340326340326339,
"grad_norm": 0.3172811217753757,
"learning_rate": 5.1083797927879896e-05,
"loss": 0.3995,
"step": 1140
},
{
"epoch": 1.648377263761879,
"grad_norm": 0.3135234713854417,
"learning_rate": 5.025012576037855e-05,
"loss": 0.4058,
"step": 1150
},
{
"epoch": 1.6627218934911243,
"grad_norm": 0.2927786895264955,
"learning_rate": 4.9416384044611124e-05,
"loss": 0.4015,
"step": 1160
},
{
"epoch": 1.6770665232203694,
"grad_norm": 0.29358825034715325,
"learning_rate": 4.858280460512302e-05,
"loss": 0.4013,
"step": 1170
},
{
"epoch": 1.6914111529496145,
"grad_norm": 0.3100930655471993,
"learning_rate": 4.7749619221338227e-05,
"loss": 0.4009,
"step": 1180
},
{
"epoch": 1.7057557826788596,
"grad_norm": 0.307169278710133,
"learning_rate": 4.691705956311225e-05,
"loss": 0.3992,
"step": 1190
},
{
"epoch": 1.7201004124081047,
"grad_norm": 0.29726825931152523,
"learning_rate": 4.608535712631566e-05,
"loss": 0.3941,
"step": 1200
},
{
"epoch": 1.7344450421373498,
"grad_norm": 0.2930544538896363,
"learning_rate": 4.525474316846581e-05,
"loss": 0.3916,
"step": 1210
},
{
"epoch": 1.748789671866595,
"grad_norm": 0.3040077165902238,
"learning_rate": 4.4425448644425066e-05,
"loss": 0.3923,
"step": 1220
},
{
"epoch": 1.76313430159584,
"grad_norm": 0.29459907021508275,
"learning_rate": 4.359770414218296e-05,
"loss": 0.3875,
"step": 1230
},
{
"epoch": 1.7774789313250852,
"grad_norm": 0.30065987027605934,
"learning_rate": 4.2771739818740565e-05,
"loss": 0.3961,
"step": 1240
},
{
"epoch": 1.7918235610543303,
"grad_norm": 0.2920496264668077,
"learning_rate": 4.194778533611451e-05,
"loss": 0.3852,
"step": 1250
},
{
"epoch": 1.8061681907835754,
"grad_norm": 0.2888363234117279,
"learning_rate": 4.112606979747881e-05,
"loss": 0.3937,
"step": 1260
},
{
"epoch": 1.8205128205128205,
"grad_norm": 0.30497579196679886,
"learning_rate": 4.030682168346192e-05,
"loss": 0.3817,
"step": 1270
},
{
"epoch": 1.8348574502420656,
"grad_norm": 0.28969651963041293,
"learning_rate": 3.949026878861704e-05,
"loss": 0.383,
"step": 1280
},
{
"epoch": 1.8492020799713107,
"grad_norm": 0.28259658296458734,
"learning_rate": 3.867663815808303e-05,
"loss": 0.3832,
"step": 1290
},
{
"epoch": 1.8635467097005558,
"grad_norm": 0.30667496660431287,
"learning_rate": 3.78661560244539e-05,
"loss": 0.3833,
"step": 1300
},
{
"epoch": 1.8778913394298011,
"grad_norm": 0.3058644823146065,
"learning_rate": 3.705904774487396e-05,
"loss": 0.3794,
"step": 1310
},
{
"epoch": 1.8922359691590462,
"grad_norm": 0.2990291886718597,
"learning_rate": 3.6255537738376706e-05,
"loss": 0.3874,
"step": 1320
},
{
"epoch": 1.9065805988882913,
"grad_norm": 0.2863080421019278,
"learning_rate": 3.545584942348426e-05,
"loss": 0.3778,
"step": 1330
},
{
"epoch": 1.9209252286175365,
"grad_norm": 0.2900728201465024,
"learning_rate": 3.466020515608525e-05,
"loss": 0.3755,
"step": 1340
},
{
"epoch": 1.9352698583467816,
"grad_norm": 0.28225638086220933,
"learning_rate": 3.386882616760794e-05,
"loss": 0.3794,
"step": 1350
},
{
"epoch": 1.9496144880760267,
"grad_norm": 0.29192196316916985,
"learning_rate": 3.30819325035062e-05,
"loss": 0.378,
"step": 1360
},
{
"epoch": 1.9639591178052718,
"grad_norm": 0.28745199551725825,
"learning_rate": 3.229974296207513e-05,
"loss": 0.3812,
"step": 1370
},
{
"epoch": 1.9783037475345169,
"grad_norm": 0.29973956860014883,
"learning_rate": 3.152247503361353e-05,
"loss": 0.3724,
"step": 1380
},
{
"epoch": 1.992648377263762,
"grad_norm": 0.31079745266426567,
"learning_rate": 3.075034483994997e-05,
"loss": 0.3718,
"step": 1390
},
{
"epoch": 2.005737851891698,
"grad_norm": 0.33118419222329254,
"learning_rate": 2.998356707434947e-05,
"loss": 0.2803,
"step": 1400
},
{
"epoch": 2.020082481620943,
"grad_norm": 0.2667436179131777,
"learning_rate": 2.9222354941817375e-05,
"loss": 0.1535,
"step": 1410
},
{
"epoch": 2.034427111350188,
"grad_norm": 0.2667296842605953,
"learning_rate": 2.846692009981693e-05,
"loss": 0.1438,
"step": 1420
},
{
"epoch": 2.0487717410794333,
"grad_norm": 0.2687925364002219,
"learning_rate": 2.771747259941734e-05,
"loss": 0.1383,
"step": 1430
},
{
"epoch": 2.0631163708086784,
"grad_norm": 0.258692098758459,
"learning_rate": 2.6974220826888374e-05,
"loss": 0.1348,
"step": 1440
},
{
"epoch": 2.0774610005379235,
"grad_norm": 0.27396537684578004,
"learning_rate": 2.623737144575787e-05,
"loss": 0.1371,
"step": 1450
},
{
"epoch": 2.0918056302671686,
"grad_norm": 0.26487337871277594,
"learning_rate": 2.5507129339348335e-05,
"loss": 0.132,
"step": 1460
},
{
"epoch": 2.1061502599964137,
"grad_norm": 0.26513450922087345,
"learning_rate": 2.478369755380839e-05,
"loss": 0.1335,
"step": 1470
},
{
"epoch": 2.120494889725659,
"grad_norm": 0.25972172147117933,
"learning_rate": 2.406727724165524e-05,
"loss": 0.134,
"step": 1480
},
{
"epoch": 2.134839519454904,
"grad_norm": 0.2524599507939378,
"learning_rate": 2.3358067605843537e-05,
"loss": 0.1307,
"step": 1490
},
{
"epoch": 2.149184149184149,
"grad_norm": 0.8197351837947908,
"learning_rate": 2.2656265844376367e-05,
"loss": 0.1332,
"step": 1500
}
],
"logging_steps": 10,
"max_steps": 2094,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1017570026586112.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}