AgThinker-3B / trainer_state.json
chegde's picture
Add files using upload-large-folder tool
94fc864 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 2094,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.014344629729245113,
"grad_norm": 3.3866311674749072,
"learning_rate": 4.285714285714286e-06,
"loss": 1.5266,
"step": 10
},
{
"epoch": 0.028689259458490227,
"grad_norm": 1.8728384521077752,
"learning_rate": 9.047619047619047e-06,
"loss": 1.3066,
"step": 20
},
{
"epoch": 0.04303388918773534,
"grad_norm": 1.1126841032649692,
"learning_rate": 1.3809523809523811e-05,
"loss": 1.1436,
"step": 30
},
{
"epoch": 0.05737851891698045,
"grad_norm": 1.1207104465369435,
"learning_rate": 1.8571428571428572e-05,
"loss": 1.0535,
"step": 40
},
{
"epoch": 0.07172314864622557,
"grad_norm": 1.1391226479086436,
"learning_rate": 2.3333333333333336e-05,
"loss": 1.0172,
"step": 50
},
{
"epoch": 0.08606777837547068,
"grad_norm": 1.146545389971907,
"learning_rate": 2.8095238095238096e-05,
"loss": 0.9859,
"step": 60
},
{
"epoch": 0.1004124081047158,
"grad_norm": 1.1888015259911366,
"learning_rate": 3.285714285714286e-05,
"loss": 0.9528,
"step": 70
},
{
"epoch": 0.1147570378339609,
"grad_norm": 1.2903823343286538,
"learning_rate": 3.761904761904762e-05,
"loss": 0.9397,
"step": 80
},
{
"epoch": 0.129101667563206,
"grad_norm": 1.2943713329149742,
"learning_rate": 4.2380952380952385e-05,
"loss": 0.9282,
"step": 90
},
{
"epoch": 0.14344629729245115,
"grad_norm": 1.2718909502396083,
"learning_rate": 4.714285714285714e-05,
"loss": 0.9186,
"step": 100
},
{
"epoch": 0.15779092702169625,
"grad_norm": 1.4791012823315328,
"learning_rate": 5.1904761904761913e-05,
"loss": 0.9074,
"step": 110
},
{
"epoch": 0.17213555675094136,
"grad_norm": 1.336532460075517,
"learning_rate": 5.666666666666667e-05,
"loss": 0.9002,
"step": 120
},
{
"epoch": 0.1864801864801865,
"grad_norm": 1.6094057366302217,
"learning_rate": 6.142857142857143e-05,
"loss": 0.8887,
"step": 130
},
{
"epoch": 0.2008248162094316,
"grad_norm": 1.070539230983606,
"learning_rate": 6.619047619047619e-05,
"loss": 0.884,
"step": 140
},
{
"epoch": 0.2151694459386767,
"grad_norm": 1.1200792510458928,
"learning_rate": 7.095238095238096e-05,
"loss": 0.8809,
"step": 150
},
{
"epoch": 0.2295140756679218,
"grad_norm": 1.087537431311211,
"learning_rate": 7.571428571428571e-05,
"loss": 0.8799,
"step": 160
},
{
"epoch": 0.24385870539716695,
"grad_norm": 0.9999716597796505,
"learning_rate": 8.047619047619048e-05,
"loss": 0.8802,
"step": 170
},
{
"epoch": 0.258203335126412,
"grad_norm": 1.2050995451812405,
"learning_rate": 8.523809523809524e-05,
"loss": 0.8692,
"step": 180
},
{
"epoch": 0.2725479648556572,
"grad_norm": 0.9483149209977663,
"learning_rate": 9e-05,
"loss": 0.8675,
"step": 190
},
{
"epoch": 0.2868925945849023,
"grad_norm": 0.9047494721372505,
"learning_rate": 9.476190476190476e-05,
"loss": 0.8651,
"step": 200
},
{
"epoch": 0.3012372243141474,
"grad_norm": 0.9614336061705705,
"learning_rate": 9.952380952380953e-05,
"loss": 0.8743,
"step": 210
},
{
"epoch": 0.3155818540433925,
"grad_norm": 0.840798357905234,
"learning_rate": 9.999436939807164e-05,
"loss": 0.8831,
"step": 220
},
{
"epoch": 0.3299264837726376,
"grad_norm": 0.894805005888125,
"learning_rate": 9.99749072170404e-05,
"loss": 0.8753,
"step": 230
},
{
"epoch": 0.3442711135018827,
"grad_norm": 0.9621522988496558,
"learning_rate": 9.994154935353517e-05,
"loss": 0.8553,
"step": 240
},
{
"epoch": 0.3586157432311278,
"grad_norm": 0.8338221919665476,
"learning_rate": 9.98943050828164e-05,
"loss": 0.8577,
"step": 250
},
{
"epoch": 0.372960372960373,
"grad_norm": 0.8401540225438021,
"learning_rate": 9.983318754130435e-05,
"loss": 0.842,
"step": 260
},
{
"epoch": 0.3873050026896181,
"grad_norm": 0.7967631358112257,
"learning_rate": 9.975821372292653e-05,
"loss": 0.8346,
"step": 270
},
{
"epoch": 0.4016496324188632,
"grad_norm": 0.7572726108984663,
"learning_rate": 9.966940447439245e-05,
"loss": 0.8272,
"step": 280
},
{
"epoch": 0.4159942621481083,
"grad_norm": 0.7862046452006615,
"learning_rate": 9.956678448939718e-05,
"loss": 0.8161,
"step": 290
},
{
"epoch": 0.4303388918773534,
"grad_norm": 0.8007930832524255,
"learning_rate": 9.945038230175509e-05,
"loss": 0.8263,
"step": 300
},
{
"epoch": 0.4446835216065985,
"grad_norm": 0.6882798627741967,
"learning_rate": 9.932023027746602e-05,
"loss": 0.8119,
"step": 310
},
{
"epoch": 0.4590281513358436,
"grad_norm": 0.7191586339334445,
"learning_rate": 9.917636460571578e-05,
"loss": 0.8195,
"step": 320
},
{
"epoch": 0.47337278106508873,
"grad_norm": 0.6479226154771014,
"learning_rate": 9.901882528881363e-05,
"loss": 0.8152,
"step": 330
},
{
"epoch": 0.4877174107943339,
"grad_norm": 0.7520756414640943,
"learning_rate": 9.884765613106948e-05,
"loss": 0.8056,
"step": 340
},
{
"epoch": 0.502062040523579,
"grad_norm": 0.7496882676872738,
"learning_rate": 9.866290472661406e-05,
"loss": 0.8143,
"step": 350
},
{
"epoch": 0.516406670252824,
"grad_norm": 0.8849980877887388,
"learning_rate": 9.846462244616508e-05,
"loss": 0.8139,
"step": 360
},
{
"epoch": 0.5307512999820692,
"grad_norm": 0.689970676364689,
"learning_rate": 9.825286442274357e-05,
"loss": 0.8052,
"step": 370
},
{
"epoch": 0.5450959297113144,
"grad_norm": 0.7192552702117557,
"learning_rate": 9.802768953634388e-05,
"loss": 0.7918,
"step": 380
},
{
"epoch": 0.5594405594405595,
"grad_norm": 0.6975024662772501,
"learning_rate": 9.778916039756193e-05,
"loss": 0.7859,
"step": 390
},
{
"epoch": 0.5737851891698046,
"grad_norm": 0.6553934247634706,
"learning_rate": 9.753734333018616e-05,
"loss": 0.776,
"step": 400
},
{
"epoch": 0.5881298188990497,
"grad_norm": 0.757809076489788,
"learning_rate": 9.727230835275598e-05,
"loss": 0.7811,
"step": 410
},
{
"epoch": 0.6024744486282948,
"grad_norm": 0.7082602077432336,
"learning_rate": 9.699412915909284e-05,
"loss": 0.7819,
"step": 420
},
{
"epoch": 0.6168190783575399,
"grad_norm": 0.6348735880274488,
"learning_rate": 9.670288309780953e-05,
"loss": 0.7679,
"step": 430
},
{
"epoch": 0.631163708086785,
"grad_norm": 0.6405240862226912,
"learning_rate": 9.639865115080304e-05,
"loss": 0.779,
"step": 440
},
{
"epoch": 0.6455083378160301,
"grad_norm": 0.6634767779062684,
"learning_rate": 9.608151791073737e-05,
"loss": 0.7739,
"step": 450
},
{
"epoch": 0.6598529675452752,
"grad_norm": 0.5759274474197833,
"learning_rate": 9.575157155752222e-05,
"loss": 0.7643,
"step": 460
},
{
"epoch": 0.6741975972745203,
"grad_norm": 0.6640199963749864,
"learning_rate": 9.54089038337943e-05,
"loss": 0.7556,
"step": 470
},
{
"epoch": 0.6885422270037654,
"grad_norm": 0.6203004765144989,
"learning_rate": 9.5053610019408e-05,
"loss": 0.7596,
"step": 480
},
{
"epoch": 0.7028868567330105,
"grad_norm": 0.5775087991410168,
"learning_rate": 9.468578890494256e-05,
"loss": 0.7534,
"step": 490
},
{
"epoch": 0.7172314864622557,
"grad_norm": 0.6714966253079345,
"learning_rate": 9.430554276423292e-05,
"loss": 0.7552,
"step": 500
},
{
"epoch": 0.7315761161915008,
"grad_norm": 0.5888288333456809,
"learning_rate": 9.391297732593229e-05,
"loss": 0.7526,
"step": 510
},
{
"epoch": 0.745920745920746,
"grad_norm": 0.6259226612543665,
"learning_rate": 9.350820174411386e-05,
"loss": 0.7537,
"step": 520
},
{
"epoch": 0.7602653756499911,
"grad_norm": 0.5499968625723676,
"learning_rate": 9.309132856792023e-05,
"loss": 0.7499,
"step": 530
},
{
"epoch": 0.7746100053792362,
"grad_norm": 0.6500285934875327,
"learning_rate": 9.266247371026873e-05,
"loss": 0.7447,
"step": 540
},
{
"epoch": 0.7889546351084813,
"grad_norm": 0.5672171674827838,
"learning_rate": 9.222175641562143e-05,
"loss": 0.7496,
"step": 550
},
{
"epoch": 0.8032992648377264,
"grad_norm": 0.6465733807975256,
"learning_rate": 9.176929922682891e-05,
"loss": 0.74,
"step": 560
},
{
"epoch": 0.8176438945669715,
"grad_norm": 0.6101304041389433,
"learning_rate": 9.130522795105676e-05,
"loss": 0.7411,
"step": 570
},
{
"epoch": 0.8319885242962166,
"grad_norm": 0.5489063036723607,
"learning_rate": 9.082967162480459e-05,
"loss": 0.7323,
"step": 580
},
{
"epoch": 0.8463331540254617,
"grad_norm": 0.5516558559909698,
"learning_rate": 9.034276247802688e-05,
"loss": 0.7231,
"step": 590
},
{
"epoch": 0.8606777837547068,
"grad_norm": 0.6477118047968917,
"learning_rate": 8.984463589736614e-05,
"loss": 0.7288,
"step": 600
},
{
"epoch": 0.8750224134839519,
"grad_norm": 0.5962486006428408,
"learning_rate": 8.933543038850816e-05,
"loss": 0.732,
"step": 610
},
{
"epoch": 0.889367043213197,
"grad_norm": 0.5713715183887016,
"learning_rate": 8.881528753767007e-05,
"loss": 0.728,
"step": 620
},
{
"epoch": 0.9037116729424421,
"grad_norm": 0.6301745705273146,
"learning_rate": 8.82843519722319e-05,
"loss": 0.7263,
"step": 630
},
{
"epoch": 0.9180563026716873,
"grad_norm": 0.5819941044165688,
"learning_rate": 8.774277132052237e-05,
"loss": 0.7189,
"step": 640
},
{
"epoch": 0.9324009324009324,
"grad_norm": 0.5336149445366645,
"learning_rate": 8.719069617077046e-05,
"loss": 0.7254,
"step": 650
},
{
"epoch": 0.9467455621301775,
"grad_norm": 0.6083401813961216,
"learning_rate": 8.662828002923378e-05,
"loss": 0.7218,
"step": 660
},
{
"epoch": 0.9610901918594227,
"grad_norm": 0.5371262960983777,
"learning_rate": 8.605567927751576e-05,
"loss": 0.719,
"step": 670
},
{
"epoch": 0.9754348215886678,
"grad_norm": 0.5551868136565808,
"learning_rate": 8.547305312908318e-05,
"loss": 0.7089,
"step": 680
},
{
"epoch": 0.9897794513179129,
"grad_norm": 0.5899796664828114,
"learning_rate": 8.48805635849964e-05,
"loss": 0.7075,
"step": 690
},
{
"epoch": 1.002868925945849,
"grad_norm": 0.608174630463498,
"learning_rate": 8.427837538886437e-05,
"loss": 0.6894,
"step": 700
},
{
"epoch": 1.017213555675094,
"grad_norm": 0.6116990279567732,
"learning_rate": 8.366665598103727e-05,
"loss": 0.5865,
"step": 710
},
{
"epoch": 1.0315581854043392,
"grad_norm": 0.5768126568975168,
"learning_rate": 8.304557545204908e-05,
"loss": 0.5931,
"step": 720
},
{
"epoch": 1.0459028151335843,
"grad_norm": 0.5727241793491287,
"learning_rate": 8.241530649532339e-05,
"loss": 0.5859,
"step": 730
},
{
"epoch": 1.0602474448628294,
"grad_norm": 0.5834066080830065,
"learning_rate": 8.177602435915546e-05,
"loss": 0.5833,
"step": 740
},
{
"epoch": 1.0745920745920745,
"grad_norm": 0.6168394884857555,
"learning_rate": 8.11279067979839e-05,
"loss": 0.5805,
"step": 750
},
{
"epoch": 1.0889367043213196,
"grad_norm": 0.4885078172344663,
"learning_rate": 8.04711340229654e-05,
"loss": 0.5733,
"step": 760
},
{
"epoch": 1.1032813340505647,
"grad_norm": 0.5289947493535304,
"learning_rate": 7.980588865186649e-05,
"loss": 0.5812,
"step": 770
},
{
"epoch": 1.11762596377981,
"grad_norm": 0.5672904537751536,
"learning_rate": 7.913235565828613e-05,
"loss": 0.5811,
"step": 780
},
{
"epoch": 1.1319705935090552,
"grad_norm": 0.6107186114378813,
"learning_rate": 7.845072232022311e-05,
"loss": 0.5755,
"step": 790
},
{
"epoch": 1.1463152232383003,
"grad_norm": 0.6696900109358123,
"learning_rate": 7.776117816800288e-05,
"loss": 0.5916,
"step": 800
},
{
"epoch": 1.1606598529675454,
"grad_norm": 0.5412095989289454,
"learning_rate": 7.706391493157805e-05,
"loss": 0.5822,
"step": 810
},
{
"epoch": 1.1750044826967905,
"grad_norm": 0.5408330429359739,
"learning_rate": 7.635912648721718e-05,
"loss": 0.5811,
"step": 820
},
{
"epoch": 1.1893491124260356,
"grad_norm": 0.5952577265784902,
"learning_rate": 7.564700880359696e-05,
"loss": 0.5799,
"step": 830
},
{
"epoch": 1.2036937421552807,
"grad_norm": 0.5026109808475954,
"learning_rate": 7.492775988731243e-05,
"loss": 0.5673,
"step": 840
},
{
"epoch": 1.2180383718845258,
"grad_norm": 0.5302543029902266,
"learning_rate": 7.420157972782063e-05,
"loss": 0.5754,
"step": 850
},
{
"epoch": 1.232383001613771,
"grad_norm": 0.5532266667153597,
"learning_rate": 7.346867024183291e-05,
"loss": 0.579,
"step": 860
},
{
"epoch": 1.246727631343016,
"grad_norm": 0.5669382050811871,
"learning_rate": 7.272923521717133e-05,
"loss": 0.5756,
"step": 870
},
{
"epoch": 1.2610722610722611,
"grad_norm": 0.5338735116713776,
"learning_rate": 7.198348025610481e-05,
"loss": 0.5769,
"step": 880
},
{
"epoch": 1.2754168908015062,
"grad_norm": 0.5863122040109188,
"learning_rate": 7.12316127181808e-05,
"loss": 0.5685,
"step": 890
},
{
"epoch": 1.2897615205307513,
"grad_norm": 0.5541827820189409,
"learning_rate": 7.047384166256815e-05,
"loss": 0.5801,
"step": 900
},
{
"epoch": 1.3041061502599964,
"grad_norm": 0.5753955062923616,
"learning_rate": 6.971037778992775e-05,
"loss": 0.5795,
"step": 910
},
{
"epoch": 1.3184507799892415,
"grad_norm": 0.47433414173616173,
"learning_rate": 6.894143338382639e-05,
"loss": 0.5768,
"step": 920
},
{
"epoch": 1.3327954097184866,
"grad_norm": 0.5274791181928579,
"learning_rate": 6.81672222517107e-05,
"loss": 0.578,
"step": 930
},
{
"epoch": 1.3471400394477318,
"grad_norm": 0.5774970902851654,
"learning_rate": 6.73879596654573e-05,
"loss": 0.5682,
"step": 940
},
{
"epoch": 1.3614846691769769,
"grad_norm": 0.5662099451295628,
"learning_rate": 6.660386230151571e-05,
"loss": 0.5735,
"step": 950
},
{
"epoch": 1.375829298906222,
"grad_norm": 0.5574084243877258,
"learning_rate": 6.581514818066088e-05,
"loss": 0.5739,
"step": 960
},
{
"epoch": 1.390173928635467,
"grad_norm": 0.539589642853108,
"learning_rate": 6.502203660737169e-05,
"loss": 0.5688,
"step": 970
},
{
"epoch": 1.4045185583647122,
"grad_norm": 0.49293926542133193,
"learning_rate": 6.422474810885278e-05,
"loss": 0.5669,
"step": 980
},
{
"epoch": 1.4188631880939573,
"grad_norm": 0.49620916916034896,
"learning_rate": 6.342350437371614e-05,
"loss": 0.5702,
"step": 990
},
{
"epoch": 1.4332078178232024,
"grad_norm": 0.5152950617397968,
"learning_rate": 6.26185281903399e-05,
"loss": 0.5685,
"step": 1000
},
{
"epoch": 1.4475524475524475,
"grad_norm": 0.545911912027696,
"learning_rate": 6.181004338492141e-05,
"loss": 0.5651,
"step": 1010
},
{
"epoch": 1.4618970772816926,
"grad_norm": 0.5074263813146201,
"learning_rate": 6.09982747592415e-05,
"loss": 0.5615,
"step": 1020
},
{
"epoch": 1.4762417070109377,
"grad_norm": 0.5094909684862107,
"learning_rate": 6.018344802815778e-05,
"loss": 0.5679,
"step": 1030
},
{
"epoch": 1.4905863367401828,
"grad_norm": 0.5103259058296251,
"learning_rate": 5.936578975684378e-05,
"loss": 0.5701,
"step": 1040
},
{
"epoch": 1.504930966469428,
"grad_norm": 0.5278709818430807,
"learning_rate": 5.854552729779184e-05,
"loss": 0.5627,
"step": 1050
},
{
"epoch": 1.519275596198673,
"grad_norm": 0.510686482301255,
"learning_rate": 5.772288872759702e-05,
"loss": 0.5625,
"step": 1060
},
{
"epoch": 1.5336202259279181,
"grad_norm": 0.5278526048784437,
"learning_rate": 5.6898102783539665e-05,
"loss": 0.5606,
"step": 1070
},
{
"epoch": 1.5479648556571632,
"grad_norm": 0.5559268236737518,
"learning_rate": 5.607139879998427e-05,
"loss": 0.5595,
"step": 1080
},
{
"epoch": 1.5623094853864083,
"grad_norm": 0.5583305948616533,
"learning_rate": 5.524300664461235e-05,
"loss": 0.5575,
"step": 1090
},
{
"epoch": 1.5766541151156535,
"grad_norm": 0.5186185461448056,
"learning_rate": 5.441315665450697e-05,
"loss": 0.5589,
"step": 1100
},
{
"epoch": 1.5909987448448986,
"grad_norm": 0.5168193702732372,
"learning_rate": 5.3582079572106794e-05,
"loss": 0.5534,
"step": 1110
},
{
"epoch": 1.6053433745741437,
"grad_norm": 0.5037158486433893,
"learning_rate": 5.275000648104743e-05,
"loss": 0.5569,
"step": 1120
},
{
"epoch": 1.6196880043033888,
"grad_norm": 0.49884254214328216,
"learning_rate": 5.191716874190785e-05,
"loss": 0.5564,
"step": 1130
},
{
"epoch": 1.6340326340326339,
"grad_norm": 0.5071669705468655,
"learning_rate": 5.1083797927879896e-05,
"loss": 0.5552,
"step": 1140
},
{
"epoch": 1.648377263761879,
"grad_norm": 0.5410568223255828,
"learning_rate": 5.025012576037855e-05,
"loss": 0.564,
"step": 1150
},
{
"epoch": 1.6627218934911243,
"grad_norm": 0.4850296619715352,
"learning_rate": 4.9416384044611124e-05,
"loss": 0.5579,
"step": 1160
},
{
"epoch": 1.6770665232203694,
"grad_norm": 0.5039355867374112,
"learning_rate": 4.858280460512302e-05,
"loss": 0.5551,
"step": 1170
},
{
"epoch": 1.6914111529496145,
"grad_norm": 0.5354526153870778,
"learning_rate": 4.7749619221338227e-05,
"loss": 0.5553,
"step": 1180
},
{
"epoch": 1.7057557826788596,
"grad_norm": 0.5149848285267117,
"learning_rate": 4.691705956311225e-05,
"loss": 0.554,
"step": 1190
},
{
"epoch": 1.7201004124081047,
"grad_norm": 0.5076867052953596,
"learning_rate": 4.608535712631566e-05,
"loss": 0.553,
"step": 1200
},
{
"epoch": 1.7344450421373498,
"grad_norm": 0.5245034600250137,
"learning_rate": 4.525474316846581e-05,
"loss": 0.5455,
"step": 1210
},
{
"epoch": 1.748789671866595,
"grad_norm": 0.5044035057693634,
"learning_rate": 4.4425448644425066e-05,
"loss": 0.5487,
"step": 1220
},
{
"epoch": 1.76313430159584,
"grad_norm": 0.5182301310051464,
"learning_rate": 4.359770414218296e-05,
"loss": 0.5401,
"step": 1230
},
{
"epoch": 1.7774789313250852,
"grad_norm": 0.5066048824304953,
"learning_rate": 4.2771739818740565e-05,
"loss": 0.5496,
"step": 1240
},
{
"epoch": 1.7918235610543303,
"grad_norm": 0.5337687629976401,
"learning_rate": 4.194778533611451e-05,
"loss": 0.5413,
"step": 1250
},
{
"epoch": 1.8061681907835754,
"grad_norm": 0.5207289329535878,
"learning_rate": 4.112606979747881e-05,
"loss": 0.544,
"step": 1260
},
{
"epoch": 1.8205128205128205,
"grad_norm": 0.5450234045888507,
"learning_rate": 4.030682168346192e-05,
"loss": 0.5358,
"step": 1270
},
{
"epoch": 1.8348574502420656,
"grad_norm": 0.48504143892984203,
"learning_rate": 3.949026878861704e-05,
"loss": 0.539,
"step": 1280
},
{
"epoch": 1.8492020799713107,
"grad_norm": 0.49111142317889844,
"learning_rate": 3.867663815808303e-05,
"loss": 0.5389,
"step": 1290
},
{
"epoch": 1.8635467097005558,
"grad_norm": 0.5008380643048362,
"learning_rate": 3.78661560244539e-05,
"loss": 0.5363,
"step": 1300
},
{
"epoch": 1.8778913394298011,
"grad_norm": 0.4952709952315973,
"learning_rate": 3.705904774487396e-05,
"loss": 0.5319,
"step": 1310
},
{
"epoch": 1.8922359691590462,
"grad_norm": 0.5164716516662802,
"learning_rate": 3.6255537738376706e-05,
"loss": 0.5402,
"step": 1320
},
{
"epoch": 1.9065805988882913,
"grad_norm": 0.5240016137645628,
"learning_rate": 3.545584942348426e-05,
"loss": 0.5309,
"step": 1330
},
{
"epoch": 1.9209252286175365,
"grad_norm": 0.5058856022275582,
"learning_rate": 3.466020515608525e-05,
"loss": 0.5298,
"step": 1340
},
{
"epoch": 1.9352698583467816,
"grad_norm": 0.5019184399441088,
"learning_rate": 3.386882616760794e-05,
"loss": 0.5319,
"step": 1350
},
{
"epoch": 1.9496144880760267,
"grad_norm": 0.471920743317965,
"learning_rate": 3.30819325035062e-05,
"loss": 0.5287,
"step": 1360
},
{
"epoch": 1.9639591178052718,
"grad_norm": 0.5094525188470136,
"learning_rate": 3.229974296207513e-05,
"loss": 0.5385,
"step": 1370
},
{
"epoch": 1.9783037475345169,
"grad_norm": 0.5606826371060244,
"learning_rate": 3.152247503361353e-05,
"loss": 0.527,
"step": 1380
},
{
"epoch": 1.992648377263762,
"grad_norm": 0.5085670056810321,
"learning_rate": 3.075034483994997e-05,
"loss": 0.5257,
"step": 1390
},
{
"epoch": 2.005737851891698,
"grad_norm": 0.5719992437170487,
"learning_rate": 2.998356707434947e-05,
"loss": 0.4623,
"step": 1400
},
{
"epoch": 2.020082481620943,
"grad_norm": 0.5387964274909228,
"learning_rate": 2.9222354941817375e-05,
"loss": 0.3606,
"step": 1410
},
{
"epoch": 2.034427111350188,
"grad_norm": 0.5698349863645911,
"learning_rate": 2.846692009981693e-05,
"loss": 0.3527,
"step": 1420
},
{
"epoch": 2.0487717410794333,
"grad_norm": 0.5324199413342471,
"learning_rate": 2.771747259941734e-05,
"loss": 0.3462,
"step": 1430
},
{
"epoch": 2.0631163708086784,
"grad_norm": 0.53215786873291,
"learning_rate": 2.6974220826888374e-05,
"loss": 0.342,
"step": 1440
},
{
"epoch": 2.0774610005379235,
"grad_norm": 0.5110033203467198,
"learning_rate": 2.623737144575787e-05,
"loss": 0.3462,
"step": 1450
},
{
"epoch": 2.0918056302671686,
"grad_norm": 0.5246624681955846,
"learning_rate": 2.5507129339348335e-05,
"loss": 0.3487,
"step": 1460
},
{
"epoch": 2.1061502599964137,
"grad_norm": 0.5070575905086573,
"learning_rate": 2.478369755380839e-05,
"loss": 0.3465,
"step": 1470
},
{
"epoch": 2.120494889725659,
"grad_norm": 0.5132555087524301,
"learning_rate": 2.406727724165524e-05,
"loss": 0.3451,
"step": 1480
},
{
"epoch": 2.134839519454904,
"grad_norm": 0.5143913782142077,
"learning_rate": 2.3358067605843537e-05,
"loss": 0.3442,
"step": 1490
},
{
"epoch": 2.149184149184149,
"grad_norm": 0.4894675520933593,
"learning_rate": 2.2656265844376367e-05,
"loss": 0.3449,
"step": 1500
},
{
"epoch": 2.163528778913394,
"grad_norm": 0.5118938379622366,
"learning_rate": 2.1962067095473648e-05,
"loss": 0.3443,
"step": 1510
},
{
"epoch": 2.1778734086426392,
"grad_norm": 0.5264199925745078,
"learning_rate": 2.127566438331345e-05,
"loss": 0.3443,
"step": 1520
},
{
"epoch": 2.1922180383718843,
"grad_norm": 0.5517034745554297,
"learning_rate": 2.059724856436092e-05,
"loss": 0.34,
"step": 1530
},
{
"epoch": 2.2065626681011294,
"grad_norm": 0.5153912300862288,
"learning_rate": 1.992700827430007e-05,
"loss": 0.3447,
"step": 1540
},
{
"epoch": 2.2209072978303745,
"grad_norm": 0.49999210085621454,
"learning_rate": 1.9265129875582954e-05,
"loss": 0.3402,
"step": 1550
},
{
"epoch": 2.23525192755962,
"grad_norm": 0.5117332103404209,
"learning_rate": 1.8611797405611097e-05,
"loss": 0.3434,
"step": 1560
},
{
"epoch": 2.249596557288865,
"grad_norm": 0.5334345620530477,
"learning_rate": 1.7967192525563254e-05,
"loss": 0.3429,
"step": 1570
},
{
"epoch": 2.2639411870181103,
"grad_norm": 0.5052797787039551,
"learning_rate": 1.733149446988394e-05,
"loss": 0.3402,
"step": 1580
},
{
"epoch": 2.2782858167473554,
"grad_norm": 0.5115033156660913,
"learning_rate": 1.670487999644669e-05,
"loss": 0.3413,
"step": 1590
},
{
"epoch": 2.2926304464766005,
"grad_norm": 0.4961174768714891,
"learning_rate": 1.6087523337406024e-05,
"loss": 0.3407,
"step": 1600
},
{
"epoch": 2.3069750762058456,
"grad_norm": 0.5181202897712733,
"learning_rate": 1.547959615075164e-05,
"loss": 0.3343,
"step": 1610
},
{
"epoch": 2.3213197059350907,
"grad_norm": 0.5076475142323975,
"learning_rate": 1.4881267472578325e-05,
"loss": 0.3415,
"step": 1620
},
{
"epoch": 2.335664335664336,
"grad_norm": 0.4921735423850851,
"learning_rate": 1.4292703670084916e-05,
"loss": 0.3392,
"step": 1630
},
{
"epoch": 2.350008965393581,
"grad_norm": 0.5151996030919073,
"learning_rate": 1.3714068395315427e-05,
"loss": 0.3349,
"step": 1640
},
{
"epoch": 2.364353595122826,
"grad_norm": 0.5104074354031792,
"learning_rate": 1.3145522539655041e-05,
"loss": 0.3354,
"step": 1650
},
{
"epoch": 2.378698224852071,
"grad_norm": 0.5380960601841327,
"learning_rate": 1.2587224189093755e-05,
"loss": 0.3302,
"step": 1660
},
{
"epoch": 2.3930428545813163,
"grad_norm": 0.5049847193289703,
"learning_rate": 1.2039328580270065e-05,
"loss": 0.3333,
"step": 1670
},
{
"epoch": 2.4073874843105614,
"grad_norm": 0.5040716684214055,
"learning_rate": 1.150198805730689e-05,
"loss": 0.3329,
"step": 1680
},
{
"epoch": 2.4217321140398065,
"grad_norm": 0.5040857331959766,
"learning_rate": 1.0975352029451863e-05,
"loss": 0.3337,
"step": 1690
},
{
"epoch": 2.4360767437690516,
"grad_norm": 0.5091871791617659,
"learning_rate": 1.0459566929533588e-05,
"loss": 0.3345,
"step": 1700
},
{
"epoch": 2.4504213734982967,
"grad_norm": 0.4986671760440932,
"learning_rate": 9.954776173245511e-06,
"loss": 0.3369,
"step": 1710
},
{
"epoch": 2.464766003227542,
"grad_norm": 0.5146311314437344,
"learning_rate": 9.461120119268713e-06,
"loss": 0.3338,
"step": 1720
},
{
"epoch": 2.479110632956787,
"grad_norm": 0.5052152754810867,
"learning_rate": 8.978736030244783e-06,
"loss": 0.331,
"step": 1730
},
{
"epoch": 2.493455262686032,
"grad_norm": 0.5060074098914322,
"learning_rate": 8.50775803460948e-06,
"loss": 0.3336,
"step": 1740
},
{
"epoch": 2.507799892415277,
"grad_norm": 0.5047637166351485,
"learning_rate": 8.048317089297875e-06,
"loss": 0.331,
"step": 1750
},
{
"epoch": 2.5221445221445222,
"grad_norm": 0.5051085483525463,
"learning_rate": 7.600540943331347e-06,
"loss": 0.3316,
"step": 1760
},
{
"epoch": 2.5364891518737673,
"grad_norm": 0.5291232815833027,
"learning_rate": 7.164554102296617e-06,
"loss": 0.3322,
"step": 1770
},
{
"epoch": 2.5508337816030124,
"grad_norm": 0.4950112167812311,
"learning_rate": 6.740477793726529e-06,
"loss": 0.3302,
"step": 1780
},
{
"epoch": 2.5651784113322575,
"grad_norm": 0.5171790095780296,
"learning_rate": 6.32842993339236e-06,
"loss": 0.3278,
"step": 1790
},
{
"epoch": 2.5795230410615027,
"grad_norm": 0.5041803659831317,
"learning_rate": 5.928525092516934e-06,
"loss": 0.3334,
"step": 1800
},
{
"epoch": 2.5938676707907478,
"grad_norm": 0.5073754782012652,
"learning_rate": 5.540874465917778e-06,
"loss": 0.3325,
"step": 1810
},
{
"epoch": 2.608212300519993,
"grad_norm": 0.48715355421051987,
"learning_rate": 5.165585841089021e-06,
"loss": 0.3246,
"step": 1820
},
{
"epoch": 2.622556930249238,
"grad_norm": 0.5162975920861064,
"learning_rate": 4.8027635682307445e-06,
"loss": 0.325,
"step": 1830
},
{
"epoch": 2.636901559978483,
"grad_norm": 0.5030085273093482,
"learning_rate": 4.45250853123404e-06,
"loss": 0.3267,
"step": 1840
},
{
"epoch": 2.651246189707728,
"grad_norm": 0.5261872357667398,
"learning_rate": 4.1149181196299905e-06,
"loss": 0.3277,
"step": 1850
},
{
"epoch": 2.6655908194369733,
"grad_norm": 0.49678770417028423,
"learning_rate": 3.7900862015101457e-06,
"loss": 0.3285,
"step": 1860
},
{
"epoch": 2.6799354491662184,
"grad_norm": 0.4889164865698146,
"learning_rate": 3.4781030974262108e-06,
"loss": 0.3268,
"step": 1870
},
{
"epoch": 2.6942800788954635,
"grad_norm": 0.4884408260663131,
"learning_rate": 3.1790555552761615e-06,
"loss": 0.3246,
"step": 1880
},
{
"epoch": 2.7086247086247086,
"grad_norm": 0.49292610830996564,
"learning_rate": 2.8930267261836395e-06,
"loss": 0.3231,
"step": 1890
},
{
"epoch": 2.7229693383539537,
"grad_norm": 0.4890311152459734,
"learning_rate": 2.6200961413776094e-06,
"loss": 0.324,
"step": 1900
},
{
"epoch": 2.737313968083199,
"grad_norm": 0.5165429385071639,
"learning_rate": 2.3603396900783724e-06,
"loss": 0.3251,
"step": 1910
},
{
"epoch": 2.751658597812444,
"grad_norm": 0.5024981727502739,
"learning_rate": 2.113829598396383e-06,
"loss": 0.322,
"step": 1920
},
{
"epoch": 2.766003227541689,
"grad_norm": 0.49038298209467507,
"learning_rate": 1.8806344092494932e-06,
"loss": 0.3248,
"step": 1930
},
{
"epoch": 2.780347857270934,
"grad_norm": 0.4917027026533285,
"learning_rate": 1.6608189633044113e-06,
"loss": 0.325,
"step": 1940
},
{
"epoch": 2.7946924870001792,
"grad_norm": 0.4869252686371276,
"learning_rate": 1.4544443809475561e-06,
"loss": 0.3247,
"step": 1950
},
{
"epoch": 2.8090371167294244,
"grad_norm": 0.48253239623178734,
"learning_rate": 1.2615680452903067e-06,
"loss": 0.3234,
"step": 1960
},
{
"epoch": 2.8233817464586695,
"grad_norm": 0.5096600620224961,
"learning_rate": 1.082243586213455e-06,
"loss": 0.3219,
"step": 1970
},
{
"epoch": 2.8377263761879146,
"grad_norm": 0.48655701646749633,
"learning_rate": 9.165208654552671e-07,
"loss": 0.321,
"step": 1980
},
{
"epoch": 2.8520710059171597,
"grad_norm": 0.47440137878879957,
"learning_rate": 7.6444596274724e-07,
"loss": 0.3229,
"step": 1990
},
{
"epoch": 2.866415635646405,
"grad_norm": 0.5034996226878241,
"learning_rate": 6.260611630015067e-07,
"loss": 0.3221,
"step": 2000
},
{
"epoch": 2.88076026537565,
"grad_norm": 0.47683170425958754,
"learning_rate": 5.01404944553363e-07,
"loss": 0.3219,
"step": 2010
},
{
"epoch": 2.895104895104895,
"grad_norm": 0.4909218835510976,
"learning_rate": 3.9051196846225935e-07,
"loss": 0.3251,
"step": 2020
},
{
"epoch": 2.90944952483414,
"grad_norm": 0.48456217220790176,
"learning_rate": 2.9341306887417653e-07,
"loss": 0.3213,
"step": 2030
},
{
"epoch": 2.923794154563385,
"grad_norm": 0.4779578566731188,
"learning_rate": 2.101352444480842e-07,
"loss": 0.3241,
"step": 2040
},
{
"epoch": 2.9381387842926303,
"grad_norm": 0.49966345259743034,
"learning_rate": 1.4070165084889008e-07,
"loss": 0.3199,
"step": 2050
},
{
"epoch": 2.9524834140218754,
"grad_norm": 0.4754785997240394,
"learning_rate": 8.513159430892925e-08,
"loss": 0.3225,
"step": 2060
},
{
"epoch": 2.9668280437511205,
"grad_norm": 0.4903628137644153,
"learning_rate": 4.344052625981365e-08,
"loss": 0.3219,
"step": 2070
},
{
"epoch": 2.9811726734803656,
"grad_norm": 0.4992880216120249,
"learning_rate": 1.5640039036085575e-08,
"loss": 0.3187,
"step": 2080
},
{
"epoch": 2.9955173032096107,
"grad_norm": 0.4963785468111048,
"learning_rate": 1.7378626519626296e-09,
"loss": 0.3238,
"step": 2090
},
{
"epoch": 3.0,
"step": 2094,
"total_flos": 306663055425536.0,
"train_loss": 0.5762376924754328,
"train_runtime": 90627.2697,
"train_samples_per_second": 1.477,
"train_steps_per_second": 0.023
}
],
"logging_steps": 10,
"max_steps": 2094,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 306663055425536.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}