both_capaug_nopack / trainer_state.json
LHL3341's picture
upload checkpoint
bab2864 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 3702,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.008103727714748784,
"grad_norm": 4.41643598484432,
"learning_rate": 4.851752021563343e-07,
"loss": 0.7447,
"step": 10
},
{
"epoch": 0.01620745542949757,
"grad_norm": 2.295952183500694,
"learning_rate": 1.0242587601078167e-06,
"loss": 0.7127,
"step": 20
},
{
"epoch": 0.024311183144246355,
"grad_norm": 1.3769997720542277,
"learning_rate": 1.5633423180592994e-06,
"loss": 0.6431,
"step": 30
},
{
"epoch": 0.03241491085899514,
"grad_norm": 0.9736635011034338,
"learning_rate": 2.1024258760107817e-06,
"loss": 0.5434,
"step": 40
},
{
"epoch": 0.04051863857374392,
"grad_norm": 0.6368736552660994,
"learning_rate": 2.6415094339622644e-06,
"loss": 0.5008,
"step": 50
},
{
"epoch": 0.04862236628849271,
"grad_norm": 0.5434828963597956,
"learning_rate": 3.1805929919137467e-06,
"loss": 0.468,
"step": 60
},
{
"epoch": 0.05672609400324149,
"grad_norm": 0.5103042443609198,
"learning_rate": 3.7196765498652294e-06,
"loss": 0.4627,
"step": 70
},
{
"epoch": 0.06482982171799027,
"grad_norm": 0.4778961353506782,
"learning_rate": 4.258760107816712e-06,
"loss": 0.45,
"step": 80
},
{
"epoch": 0.07293354943273905,
"grad_norm": 0.5031019722668856,
"learning_rate": 4.797843665768194e-06,
"loss": 0.439,
"step": 90
},
{
"epoch": 0.08103727714748785,
"grad_norm": 0.4735493470149646,
"learning_rate": 5.336927223719677e-06,
"loss": 0.4375,
"step": 100
},
{
"epoch": 0.08914100486223663,
"grad_norm": 0.4485376999795482,
"learning_rate": 5.8760107816711595e-06,
"loss": 0.4276,
"step": 110
},
{
"epoch": 0.09724473257698542,
"grad_norm": 0.5105959900024185,
"learning_rate": 6.415094339622642e-06,
"loss": 0.4195,
"step": 120
},
{
"epoch": 0.1053484602917342,
"grad_norm": 0.5002794471107062,
"learning_rate": 6.954177897574125e-06,
"loss": 0.4191,
"step": 130
},
{
"epoch": 0.11345218800648298,
"grad_norm": 0.4775871973775412,
"learning_rate": 7.493261455525606e-06,
"loss": 0.4146,
"step": 140
},
{
"epoch": 0.12155591572123177,
"grad_norm": 0.4914920705583914,
"learning_rate": 8.03234501347709e-06,
"loss": 0.4128,
"step": 150
},
{
"epoch": 0.12965964343598055,
"grad_norm": 0.48908081697569705,
"learning_rate": 8.571428571428571e-06,
"loss": 0.4206,
"step": 160
},
{
"epoch": 0.13776337115072934,
"grad_norm": 0.5007305881001459,
"learning_rate": 9.110512129380054e-06,
"loss": 0.4133,
"step": 170
},
{
"epoch": 0.1458670988654781,
"grad_norm": 0.4914360309893019,
"learning_rate": 9.649595687331537e-06,
"loss": 0.407,
"step": 180
},
{
"epoch": 0.1539708265802269,
"grad_norm": 0.4828448494347662,
"learning_rate": 1.018867924528302e-05,
"loss": 0.4071,
"step": 190
},
{
"epoch": 0.1620745542949757,
"grad_norm": 0.49582749991681885,
"learning_rate": 1.0727762803234503e-05,
"loss": 0.4072,
"step": 200
},
{
"epoch": 0.17017828200972449,
"grad_norm": 0.4699707357588344,
"learning_rate": 1.1266846361185985e-05,
"loss": 0.3974,
"step": 210
},
{
"epoch": 0.17828200972447325,
"grad_norm": 0.4991420223820099,
"learning_rate": 1.1805929919137466e-05,
"loss": 0.405,
"step": 220
},
{
"epoch": 0.18638573743922204,
"grad_norm": 0.5140149870680131,
"learning_rate": 1.234501347708895e-05,
"loss": 0.3983,
"step": 230
},
{
"epoch": 0.19448946515397084,
"grad_norm": 0.4890055950875211,
"learning_rate": 1.2884097035040433e-05,
"loss": 0.3986,
"step": 240
},
{
"epoch": 0.2025931928687196,
"grad_norm": 0.6215519273650058,
"learning_rate": 1.3423180592991916e-05,
"loss": 0.4055,
"step": 250
},
{
"epoch": 0.2106969205834684,
"grad_norm": 0.43686298686328745,
"learning_rate": 1.3962264150943397e-05,
"loss": 0.3902,
"step": 260
},
{
"epoch": 0.2188006482982172,
"grad_norm": 0.45169864689601186,
"learning_rate": 1.4501347708894879e-05,
"loss": 0.3912,
"step": 270
},
{
"epoch": 0.22690437601296595,
"grad_norm": 0.4703662675657425,
"learning_rate": 1.5040431266846362e-05,
"loss": 0.3948,
"step": 280
},
{
"epoch": 0.23500810372771475,
"grad_norm": 0.5077460907087984,
"learning_rate": 1.5579514824797845e-05,
"loss": 0.4029,
"step": 290
},
{
"epoch": 0.24311183144246354,
"grad_norm": 0.5673991220584315,
"learning_rate": 1.6118598382749326e-05,
"loss": 0.3971,
"step": 300
},
{
"epoch": 0.25121555915721233,
"grad_norm": 0.6359275384596881,
"learning_rate": 1.6657681940700808e-05,
"loss": 0.4013,
"step": 310
},
{
"epoch": 0.2593192868719611,
"grad_norm": 0.4456778178081555,
"learning_rate": 1.7196765498652293e-05,
"loss": 0.3919,
"step": 320
},
{
"epoch": 0.26742301458670986,
"grad_norm": 0.4341335587338741,
"learning_rate": 1.7735849056603774e-05,
"loss": 0.3998,
"step": 330
},
{
"epoch": 0.2755267423014587,
"grad_norm": 0.4571901760639164,
"learning_rate": 1.827493261455526e-05,
"loss": 0.3974,
"step": 340
},
{
"epoch": 0.28363047001620745,
"grad_norm": 0.47940673982145304,
"learning_rate": 1.881401617250674e-05,
"loss": 0.3919,
"step": 350
},
{
"epoch": 0.2917341977309562,
"grad_norm": 0.4642039239008883,
"learning_rate": 1.9353099730458222e-05,
"loss": 0.3964,
"step": 360
},
{
"epoch": 0.29983792544570503,
"grad_norm": 0.49129179078057755,
"learning_rate": 1.9892183288409707e-05,
"loss": 0.3963,
"step": 370
},
{
"epoch": 0.3079416531604538,
"grad_norm": 0.46565355070187053,
"learning_rate": 1.999971535838293e-05,
"loss": 0.4005,
"step": 380
},
{
"epoch": 0.3160453808752026,
"grad_norm": 0.4991505049021048,
"learning_rate": 1.999855902958531e-05,
"loss": 0.3986,
"step": 390
},
{
"epoch": 0.3241491085899514,
"grad_norm": 0.450506246985201,
"learning_rate": 1.999651332628271e-05,
"loss": 0.3969,
"step": 400
},
{
"epoch": 0.33225283630470015,
"grad_norm": 0.5440771482828131,
"learning_rate": 1.9993578430440986e-05,
"loss": 0.4017,
"step": 410
},
{
"epoch": 0.34035656401944897,
"grad_norm": 0.46282436229664603,
"learning_rate": 1.9989754603119914e-05,
"loss": 0.3982,
"step": 420
},
{
"epoch": 0.34846029173419774,
"grad_norm": 0.5429960948303321,
"learning_rate": 1.9985042184449953e-05,
"loss": 0.395,
"step": 430
},
{
"epoch": 0.3565640194489465,
"grad_norm": 0.40365597666932407,
"learning_rate": 1.997944159360201e-05,
"loss": 0.3848,
"step": 440
},
{
"epoch": 0.3646677471636953,
"grad_norm": 0.443907942327311,
"learning_rate": 1.997295332875014e-05,
"loss": 0.3963,
"step": 450
},
{
"epoch": 0.3727714748784441,
"grad_norm": 0.5255368584935723,
"learning_rate": 1.9965577967027244e-05,
"loss": 0.3933,
"step": 460
},
{
"epoch": 0.38087520259319285,
"grad_norm": 0.42527977955696455,
"learning_rate": 1.9957316164473717e-05,
"loss": 0.3907,
"step": 470
},
{
"epoch": 0.3889789303079417,
"grad_norm": 0.4749169933273577,
"learning_rate": 1.9948168655979115e-05,
"loss": 0.3916,
"step": 480
},
{
"epoch": 0.39708265802269044,
"grad_norm": 0.44444390734791117,
"learning_rate": 1.993813625521677e-05,
"loss": 0.3842,
"step": 490
},
{
"epoch": 0.4051863857374392,
"grad_norm": 0.4642831318826894,
"learning_rate": 1.9927219854571415e-05,
"loss": 0.388,
"step": 500
},
{
"epoch": 0.413290113452188,
"grad_norm": 0.3960497536000858,
"learning_rate": 1.9915420425059816e-05,
"loss": 0.3827,
"step": 510
},
{
"epoch": 0.4213938411669368,
"grad_norm": 0.5662768786356301,
"learning_rate": 1.9902739016244386e-05,
"loss": 0.3845,
"step": 520
},
{
"epoch": 0.42949756888168555,
"grad_norm": 0.4432787440857732,
"learning_rate": 1.9889176756139838e-05,
"loss": 0.3855,
"step": 530
},
{
"epoch": 0.4376012965964344,
"grad_norm": 0.451796060211973,
"learning_rate": 1.987473485111285e-05,
"loss": 0.3908,
"step": 540
},
{
"epoch": 0.44570502431118314,
"grad_norm": 0.43457237574930996,
"learning_rate": 1.9859414585774735e-05,
"loss": 0.3868,
"step": 550
},
{
"epoch": 0.4538087520259319,
"grad_norm": 0.47083870975317493,
"learning_rate": 1.9843217322867204e-05,
"loss": 0.3886,
"step": 560
},
{
"epoch": 0.4619124797406807,
"grad_norm": 0.5757185193113336,
"learning_rate": 1.9826144503141132e-05,
"loss": 0.3893,
"step": 570
},
{
"epoch": 0.4700162074554295,
"grad_norm": 0.4468495226361423,
"learning_rate": 1.9808197645228408e-05,
"loss": 0.3781,
"step": 580
},
{
"epoch": 0.47811993517017826,
"grad_norm": 0.40873606713135685,
"learning_rate": 1.9789378345506854e-05,
"loss": 0.3817,
"step": 590
},
{
"epoch": 0.4862236628849271,
"grad_norm": 0.41969294252511485,
"learning_rate": 1.976968827795822e-05,
"loss": 0.392,
"step": 600
},
{
"epoch": 0.49432739059967584,
"grad_norm": 0.3955061020676128,
"learning_rate": 1.97491291940193e-05,
"loss": 0.3833,
"step": 610
},
{
"epoch": 0.5024311183144247,
"grad_norm": 0.47471127904705646,
"learning_rate": 1.972770292242611e-05,
"loss": 0.385,
"step": 620
},
{
"epoch": 0.5105348460291734,
"grad_norm": 0.4269035087250598,
"learning_rate": 1.9705411369051252e-05,
"loss": 0.3805,
"step": 630
},
{
"epoch": 0.5186385737439222,
"grad_norm": 0.4380160054225159,
"learning_rate": 1.9682256516734377e-05,
"loss": 0.3919,
"step": 640
},
{
"epoch": 0.526742301458671,
"grad_norm": 0.4802590775403936,
"learning_rate": 1.9658240425105792e-05,
"loss": 0.3807,
"step": 650
},
{
"epoch": 0.5348460291734197,
"grad_norm": 0.41972342740779855,
"learning_rate": 1.9633365230403294e-05,
"loss": 0.3822,
"step": 660
},
{
"epoch": 0.5429497568881686,
"grad_norm": 1.1224896880383322,
"learning_rate": 1.9607633145282117e-05,
"loss": 0.3832,
"step": 670
},
{
"epoch": 0.5510534846029174,
"grad_norm": 0.4434033568303931,
"learning_rate": 1.958104645861812e-05,
"loss": 0.3838,
"step": 680
},
{
"epoch": 0.5591572123176661,
"grad_norm": 0.41274172915536217,
"learning_rate": 1.9553607535304214e-05,
"loss": 0.3853,
"step": 690
},
{
"epoch": 0.5672609400324149,
"grad_norm": 0.5018043496489497,
"learning_rate": 1.9525318816039975e-05,
"loss": 0.3819,
"step": 700
},
{
"epoch": 0.5753646677471637,
"grad_norm": 0.4964263931154507,
"learning_rate": 1.9496182817114567e-05,
"loss": 0.3842,
"step": 710
},
{
"epoch": 0.5834683954619124,
"grad_norm": 0.44925239230925906,
"learning_rate": 1.9466202130182898e-05,
"loss": 0.3818,
"step": 720
},
{
"epoch": 0.5915721231766613,
"grad_norm": 0.41290537140097916,
"learning_rate": 1.9435379422035104e-05,
"loss": 0.3754,
"step": 730
},
{
"epoch": 0.5996758508914101,
"grad_norm": 0.4429728423580205,
"learning_rate": 1.940371743435934e-05,
"loss": 0.3784,
"step": 740
},
{
"epoch": 0.6077795786061588,
"grad_norm": 0.3953775179529158,
"learning_rate": 1.9371218983497888e-05,
"loss": 0.3791,
"step": 750
},
{
"epoch": 0.6158833063209076,
"grad_norm": 0.4034222884371506,
"learning_rate": 1.9337886960196666e-05,
"loss": 0.3823,
"step": 760
},
{
"epoch": 0.6239870340356564,
"grad_norm": 0.4214201185114954,
"learning_rate": 1.9303724329348072e-05,
"loss": 0.379,
"step": 770
},
{
"epoch": 0.6320907617504052,
"grad_norm": 0.43459015373350257,
"learning_rate": 1.926873412972728e-05,
"loss": 0.3679,
"step": 780
},
{
"epoch": 0.640194489465154,
"grad_norm": 0.4011832765597355,
"learning_rate": 1.9232919473721918e-05,
"loss": 0.3677,
"step": 790
},
{
"epoch": 0.6482982171799028,
"grad_norm": 0.4014412249585877,
"learning_rate": 1.919628354705524e-05,
"loss": 0.3786,
"step": 800
},
{
"epoch": 0.6564019448946515,
"grad_norm": 0.37213700520936743,
"learning_rate": 1.915882960850274e-05,
"loss": 0.3747,
"step": 810
},
{
"epoch": 0.6645056726094003,
"grad_norm": 0.33071711798616843,
"learning_rate": 1.9120560989602282e-05,
"loss": 0.3781,
"step": 820
},
{
"epoch": 0.6726094003241491,
"grad_norm": 0.45109358722716725,
"learning_rate": 1.9081481094357778e-05,
"loss": 0.3761,
"step": 830
},
{
"epoch": 0.6807131280388979,
"grad_norm": 0.3788195302814343,
"learning_rate": 1.904159339893638e-05,
"loss": 0.3714,
"step": 840
},
{
"epoch": 0.6888168557536467,
"grad_norm": 0.6175330343756978,
"learning_rate": 1.9000901451359286e-05,
"loss": 0.3724,
"step": 850
},
{
"epoch": 0.6969205834683955,
"grad_norm": 0.43932223458945724,
"learning_rate": 1.8959408871186134e-05,
"loss": 0.3744,
"step": 860
},
{
"epoch": 0.7050243111831442,
"grad_norm": 0.40099546045362755,
"learning_rate": 1.8917119349193054e-05,
"loss": 0.3754,
"step": 870
},
{
"epoch": 0.713128038897893,
"grad_norm": 0.4405742977630959,
"learning_rate": 1.8874036647044356e-05,
"loss": 0.3766,
"step": 880
},
{
"epoch": 0.7212317666126418,
"grad_norm": 0.40738504715105506,
"learning_rate": 1.883016459695794e-05,
"loss": 0.3722,
"step": 890
},
{
"epoch": 0.7293354943273906,
"grad_norm": 0.3956122497267419,
"learning_rate": 1.8785507101364423e-05,
"loss": 0.3729,
"step": 900
},
{
"epoch": 0.7374392220421394,
"grad_norm": 0.40591878329996733,
"learning_rate": 1.8740068132559995e-05,
"loss": 0.3818,
"step": 910
},
{
"epoch": 0.7455429497568882,
"grad_norm": 0.39199283131452295,
"learning_rate": 1.869385173235312e-05,
"loss": 0.377,
"step": 920
},
{
"epoch": 0.7536466774716369,
"grad_norm": 0.42458410626482784,
"learning_rate": 1.864686201170497e-05,
"loss": 0.3687,
"step": 930
},
{
"epoch": 0.7617504051863857,
"grad_norm": 0.401604545594171,
"learning_rate": 1.8599103150363796e-05,
"loss": 0.3708,
"step": 940
},
{
"epoch": 0.7698541329011345,
"grad_norm": 87.81767261332284,
"learning_rate": 1.8550579396493108e-05,
"loss": 0.3653,
"step": 950
},
{
"epoch": 0.7779578606158833,
"grad_norm": 0.4139369558770015,
"learning_rate": 1.8501295066293818e-05,
"loss": 0.377,
"step": 960
},
{
"epoch": 0.7860615883306321,
"grad_norm": 0.4625631547372043,
"learning_rate": 1.84512545436203e-05,
"loss": 0.3656,
"step": 970
},
{
"epoch": 0.7941653160453809,
"grad_norm": 0.499708486150745,
"learning_rate": 1.8400462279590455e-05,
"loss": 0.3759,
"step": 980
},
{
"epoch": 0.8022690437601296,
"grad_norm": 0.41350589433724916,
"learning_rate": 1.8348922792189786e-05,
"loss": 0.3685,
"step": 990
},
{
"epoch": 0.8103727714748784,
"grad_norm": 0.6762985379621705,
"learning_rate": 1.82966406658695e-05,
"loss": 0.3661,
"step": 1000
},
{
"epoch": 0.8184764991896273,
"grad_norm": 0.416220354789335,
"learning_rate": 1.8243620551138752e-05,
"loss": 0.3748,
"step": 1010
},
{
"epoch": 0.826580226904376,
"grad_norm": 0.3805930072320955,
"learning_rate": 1.8189867164150946e-05,
"loss": 0.3676,
"step": 1020
},
{
"epoch": 0.8346839546191248,
"grad_norm": 0.38903904199656336,
"learning_rate": 1.8135385286284263e-05,
"loss": 0.3632,
"step": 1030
},
{
"epoch": 0.8427876823338736,
"grad_norm": 0.4782529960181865,
"learning_rate": 1.8080179763716343e-05,
"loss": 0.3698,
"step": 1040
},
{
"epoch": 0.8508914100486223,
"grad_norm": 0.3712030837456596,
"learning_rate": 1.8024255506993203e-05,
"loss": 0.368,
"step": 1050
},
{
"epoch": 0.8589951377633711,
"grad_norm": 0.36733985835621663,
"learning_rate": 1.796761749059247e-05,
"loss": 0.3726,
"step": 1060
},
{
"epoch": 0.86709886547812,
"grad_norm": 0.444647790232743,
"learning_rate": 1.791027075248088e-05,
"loss": 0.3621,
"step": 1070
},
{
"epoch": 0.8752025931928687,
"grad_norm": 0.3970513835898113,
"learning_rate": 1.7852220393666155e-05,
"loss": 0.3748,
"step": 1080
},
{
"epoch": 0.8833063209076175,
"grad_norm": 0.3988717428880392,
"learning_rate": 1.7793471577743263e-05,
"loss": 0.3682,
"step": 1090
},
{
"epoch": 0.8914100486223663,
"grad_norm": 0.4459200769730532,
"learning_rate": 1.7734029530435128e-05,
"loss": 0.3761,
"step": 1100
},
{
"epoch": 0.899513776337115,
"grad_norm": 0.38898443324355875,
"learning_rate": 1.7673899539127785e-05,
"loss": 0.3696,
"step": 1110
},
{
"epoch": 0.9076175040518638,
"grad_norm": 0.37805917634894226,
"learning_rate": 1.7613086952400072e-05,
"loss": 0.3766,
"step": 1120
},
{
"epoch": 0.9157212317666127,
"grad_norm": 0.41019056181222463,
"learning_rate": 1.7551597179547876e-05,
"loss": 0.3757,
"step": 1130
},
{
"epoch": 0.9238249594813615,
"grad_norm": 0.4209855541387669,
"learning_rate": 1.7489435690102975e-05,
"loss": 0.381,
"step": 1140
},
{
"epoch": 0.9319286871961102,
"grad_norm": 0.4025369455294472,
"learning_rate": 1.7426608013346504e-05,
"loss": 0.3715,
"step": 1150
},
{
"epoch": 0.940032414910859,
"grad_norm": 0.3510393249568996,
"learning_rate": 1.7363119737817148e-05,
"loss": 0.3675,
"step": 1160
},
{
"epoch": 0.9481361426256077,
"grad_norm": 0.3835263547355758,
"learning_rate": 1.7298976510814026e-05,
"loss": 0.378,
"step": 1170
},
{
"epoch": 0.9562398703403565,
"grad_norm": 0.4533397123438105,
"learning_rate": 1.723418403789438e-05,
"loss": 0.3746,
"step": 1180
},
{
"epoch": 0.9643435980551054,
"grad_norm": 0.39846733706234416,
"learning_rate": 1.716874808236602e-05,
"loss": 0.3561,
"step": 1190
},
{
"epoch": 0.9724473257698542,
"grad_norm": 0.40053796669906305,
"learning_rate": 1.710267446477474e-05,
"loss": 0.3766,
"step": 1200
},
{
"epoch": 0.9805510534846029,
"grad_norm": 0.3829527128726701,
"learning_rate": 1.7035969062386527e-05,
"loss": 0.3652,
"step": 1210
},
{
"epoch": 0.9886547811993517,
"grad_norm": 0.4162641923499344,
"learning_rate": 1.69686378086648e-05,
"loss": 0.3794,
"step": 1220
},
{
"epoch": 0.9967585089141004,
"grad_norm": 0.3942994138514852,
"learning_rate": 1.6900686692742617e-05,
"loss": 0.3657,
"step": 1230
},
{
"epoch": 1.0048622366288493,
"grad_norm": 0.35757653557443925,
"learning_rate": 1.683212175888994e-05,
"loss": 0.3374,
"step": 1240
},
{
"epoch": 1.012965964343598,
"grad_norm": 0.40348503930104396,
"learning_rate": 1.6762949105976028e-05,
"loss": 0.3091,
"step": 1250
},
{
"epoch": 1.0210696920583469,
"grad_norm": 0.4042269198828,
"learning_rate": 1.6693174886926878e-05,
"loss": 0.3111,
"step": 1260
},
{
"epoch": 1.0291734197730957,
"grad_norm": 0.4439522801292698,
"learning_rate": 1.6622805308177972e-05,
"loss": 0.307,
"step": 1270
},
{
"epoch": 1.0372771474878444,
"grad_norm": 0.4386336752083495,
"learning_rate": 1.6551846629122204e-05,
"loss": 0.3087,
"step": 1280
},
{
"epoch": 1.0453808752025933,
"grad_norm": 0.388567488696101,
"learning_rate": 1.648030516155309e-05,
"loss": 0.3107,
"step": 1290
},
{
"epoch": 1.053484602917342,
"grad_norm": 0.44064851617021744,
"learning_rate": 1.6408187269103334e-05,
"loss": 0.3044,
"step": 1300
},
{
"epoch": 1.0615883306320908,
"grad_norm": 0.41408251510783445,
"learning_rate": 1.63354993666788e-05,
"loss": 0.3072,
"step": 1310
},
{
"epoch": 1.0696920583468394,
"grad_norm": 0.38002677181649674,
"learning_rate": 1.626224791988789e-05,
"loss": 0.3094,
"step": 1320
},
{
"epoch": 1.0777957860615883,
"grad_norm": 0.4172348362426581,
"learning_rate": 1.618843944446642e-05,
"loss": 0.3053,
"step": 1330
},
{
"epoch": 1.0858995137763372,
"grad_norm": 0.3766654417771684,
"learning_rate": 1.6114080505698057e-05,
"loss": 0.3114,
"step": 1340
},
{
"epoch": 1.0940032414910859,
"grad_norm": 0.3798522719943436,
"learning_rate": 1.6039177717830334e-05,
"loss": 0.3177,
"step": 1350
},
{
"epoch": 1.1021069692058347,
"grad_norm": 0.40719290368835925,
"learning_rate": 1.5963737743486296e-05,
"loss": 0.3113,
"step": 1360
},
{
"epoch": 1.1102106969205834,
"grad_norm": 0.3859375066355612,
"learning_rate": 1.5887767293071886e-05,
"loss": 0.3168,
"step": 1370
},
{
"epoch": 1.1183144246353323,
"grad_norm": 0.44786137434859913,
"learning_rate": 1.581127312417902e-05,
"loss": 0.3095,
"step": 1380
},
{
"epoch": 1.1264181523500811,
"grad_norm": 0.4122166831699624,
"learning_rate": 1.5734262040984518e-05,
"loss": 0.3093,
"step": 1390
},
{
"epoch": 1.1345218800648298,
"grad_norm": 0.36002819869898145,
"learning_rate": 1.565674089364487e-05,
"loss": 0.3074,
"step": 1400
},
{
"epoch": 1.1426256077795787,
"grad_norm": 0.39577499018930706,
"learning_rate": 1.5578716577686906e-05,
"loss": 0.3134,
"step": 1410
},
{
"epoch": 1.1507293354943273,
"grad_norm": 0.38753643436786867,
"learning_rate": 1.550019603339444e-05,
"loss": 0.3056,
"step": 1420
},
{
"epoch": 1.1588330632090762,
"grad_norm": 0.35803720972169256,
"learning_rate": 1.542118624519092e-05,
"loss": 0.3143,
"step": 1430
},
{
"epoch": 1.1669367909238249,
"grad_norm": 0.4019227028306303,
"learning_rate": 1.5341694241018186e-05,
"loss": 0.3078,
"step": 1440
},
{
"epoch": 1.1750405186385737,
"grad_norm": 0.40838651309105123,
"learning_rate": 1.5261727091711304e-05,
"loss": 0.3085,
"step": 1450
},
{
"epoch": 1.1831442463533226,
"grad_norm": 0.40165213026886154,
"learning_rate": 1.5181291910369643e-05,
"loss": 0.3099,
"step": 1460
},
{
"epoch": 1.1912479740680713,
"grad_norm": 0.3759220524819882,
"learning_rate": 1.5100395851724127e-05,
"loss": 0.3062,
"step": 1470
},
{
"epoch": 1.1993517017828201,
"grad_norm": 0.41920709253732197,
"learning_rate": 1.501904611150086e-05,
"loss": 0.3074,
"step": 1480
},
{
"epoch": 1.2074554294975688,
"grad_norm": 0.40333128066808394,
"learning_rate": 1.4937249925781025e-05,
"loss": 0.3085,
"step": 1490
},
{
"epoch": 1.2155591572123177,
"grad_norm": 0.3931274788288254,
"learning_rate": 1.4855014570357266e-05,
"loss": 0.3102,
"step": 1500
},
{
"epoch": 1.2236628849270665,
"grad_norm": 0.4117002507781213,
"learning_rate": 1.477234736008648e-05,
"loss": 0.307,
"step": 1510
},
{
"epoch": 1.2317666126418152,
"grad_norm": 0.38514751693279475,
"learning_rate": 1.4689255648239182e-05,
"loss": 0.3063,
"step": 1520
},
{
"epoch": 1.239870340356564,
"grad_norm": 0.3833840502504147,
"learning_rate": 1.4605746825845394e-05,
"loss": 0.3104,
"step": 1530
},
{
"epoch": 1.2479740680713127,
"grad_norm": 0.40656817719400695,
"learning_rate": 1.452182832103726e-05,
"loss": 0.3042,
"step": 1540
},
{
"epoch": 1.2560777957860616,
"grad_norm": 0.3978789087166176,
"learning_rate": 1.4437507598388256e-05,
"loss": 0.3093,
"step": 1550
},
{
"epoch": 1.2641815235008105,
"grad_norm": 0.40751944678962887,
"learning_rate": 1.4352792158249267e-05,
"loss": 0.3111,
"step": 1560
},
{
"epoch": 1.2722852512155591,
"grad_norm": 0.3686549500370485,
"learning_rate": 1.4267689536081391e-05,
"loss": 0.302,
"step": 1570
},
{
"epoch": 1.280388978930308,
"grad_norm": 0.3584275890847868,
"learning_rate": 1.4182207301785673e-05,
"loss": 0.2987,
"step": 1580
},
{
"epoch": 1.2884927066450567,
"grad_norm": 0.3977017056962877,
"learning_rate": 1.4096353059029759e-05,
"loss": 0.3075,
"step": 1590
},
{
"epoch": 1.2965964343598055,
"grad_norm": 0.3712636385272552,
"learning_rate": 1.4010134444571554e-05,
"loss": 0.305,
"step": 1600
},
{
"epoch": 1.3047001620745542,
"grad_norm": 0.37269512136572663,
"learning_rate": 1.3923559127579917e-05,
"loss": 0.3111,
"step": 1610
},
{
"epoch": 1.312803889789303,
"grad_norm": 0.3936454352334608,
"learning_rate": 1.3836634808952488e-05,
"loss": 0.3123,
"step": 1620
},
{
"epoch": 1.320907617504052,
"grad_norm": 0.4230281324644334,
"learning_rate": 1.3749369220630706e-05,
"loss": 0.3074,
"step": 1630
},
{
"epoch": 1.3290113452188006,
"grad_norm": 0.37222327941648686,
"learning_rate": 1.3661770124912037e-05,
"loss": 0.3126,
"step": 1640
},
{
"epoch": 1.3371150729335495,
"grad_norm": 0.37547495560760374,
"learning_rate": 1.357384531375952e-05,
"loss": 0.3062,
"step": 1650
},
{
"epoch": 1.3452188006482984,
"grad_norm": 0.3740057853136358,
"learning_rate": 1.3485602608108665e-05,
"loss": 0.31,
"step": 1660
},
{
"epoch": 1.353322528363047,
"grad_norm": 0.3763454169937548,
"learning_rate": 1.3397049857171789e-05,
"loss": 0.3136,
"step": 1670
},
{
"epoch": 1.3614262560777957,
"grad_norm": 0.36750506900659763,
"learning_rate": 1.3308194937739811e-05,
"loss": 0.3064,
"step": 1680
},
{
"epoch": 1.3695299837925445,
"grad_norm": 0.34978267737985513,
"learning_rate": 1.3219045753481633e-05,
"loss": 0.3109,
"step": 1690
},
{
"epoch": 1.3776337115072934,
"grad_norm": 0.38012184308401736,
"learning_rate": 1.3129610234241078e-05,
"loss": 0.3074,
"step": 1700
},
{
"epoch": 1.385737439222042,
"grad_norm": 0.3675680062731044,
"learning_rate": 1.3039896335331553e-05,
"loss": 0.3072,
"step": 1710
},
{
"epoch": 1.393841166936791,
"grad_norm": 0.37558022818976783,
"learning_rate": 1.2949912036828402e-05,
"loss": 0.3088,
"step": 1720
},
{
"epoch": 1.4019448946515398,
"grad_norm": 0.44474757366346707,
"learning_rate": 1.2859665342859094e-05,
"loss": 0.3103,
"step": 1730
},
{
"epoch": 1.4100486223662885,
"grad_norm": 0.384724775566465,
"learning_rate": 1.2769164280891232e-05,
"loss": 0.3099,
"step": 1740
},
{
"epoch": 1.4181523500810373,
"grad_norm": 0.3537045686625036,
"learning_rate": 1.2678416901018547e-05,
"loss": 0.316,
"step": 1750
},
{
"epoch": 1.426256077795786,
"grad_norm": 0.36861281422703607,
"learning_rate": 1.2587431275244776e-05,
"loss": 0.309,
"step": 1760
},
{
"epoch": 1.4343598055105349,
"grad_norm": 0.3908436057984349,
"learning_rate": 1.2496215496765725e-05,
"loss": 0.3061,
"step": 1770
},
{
"epoch": 1.4424635332252835,
"grad_norm": 0.3724688967470804,
"learning_rate": 1.2404777679249331e-05,
"loss": 0.3092,
"step": 1780
},
{
"epoch": 1.4505672609400324,
"grad_norm": 0.3926545802632573,
"learning_rate": 1.2313125956113963e-05,
"loss": 0.3093,
"step": 1790
},
{
"epoch": 1.4586709886547813,
"grad_norm": 0.34999013214992414,
"learning_rate": 1.2221268479804952e-05,
"loss": 0.3063,
"step": 1800
},
{
"epoch": 1.46677471636953,
"grad_norm": 0.3790313923510292,
"learning_rate": 1.2129213421069434e-05,
"loss": 0.3054,
"step": 1810
},
{
"epoch": 1.4748784440842788,
"grad_norm": 0.37358025823808116,
"learning_rate": 1.2036968968229553e-05,
"loss": 0.2995,
"step": 1820
},
{
"epoch": 1.4829821717990275,
"grad_norm": 0.36851119500263724,
"learning_rate": 1.1944543326454099e-05,
"loss": 0.3073,
"step": 1830
},
{
"epoch": 1.4910858995137763,
"grad_norm": 0.3829376421748194,
"learning_rate": 1.1851944717028685e-05,
"loss": 0.3026,
"step": 1840
},
{
"epoch": 1.499189627228525,
"grad_norm": 0.36920620624919487,
"learning_rate": 1.1759181376624425e-05,
"loss": 0.307,
"step": 1850
},
{
"epoch": 1.5072933549432739,
"grad_norm": 0.35131939755626207,
"learning_rate": 1.1666261556565311e-05,
"loss": 0.3102,
"step": 1860
},
{
"epoch": 1.5153970826580228,
"grad_norm": 0.34515241407243896,
"learning_rate": 1.1573193522094235e-05,
"loss": 0.3056,
"step": 1870
},
{
"epoch": 1.5235008103727714,
"grad_norm": 0.37318125155154075,
"learning_rate": 1.1479985551637802e-05,
"loss": 0.3102,
"step": 1880
},
{
"epoch": 1.5316045380875203,
"grad_norm": 0.3990883322888776,
"learning_rate": 1.1386645936069959e-05,
"loss": 0.2997,
"step": 1890
},
{
"epoch": 1.5397082658022692,
"grad_norm": 0.444127337726431,
"learning_rate": 1.129318297797453e-05,
"loss": 0.3126,
"step": 1900
},
{
"epoch": 1.5478119935170178,
"grad_norm": 0.3861401484638357,
"learning_rate": 1.1199604990906673e-05,
"loss": 0.3033,
"step": 1910
},
{
"epoch": 1.5559157212317665,
"grad_norm": 0.384631402351326,
"learning_rate": 1.1105920298653424e-05,
"loss": 0.3022,
"step": 1920
},
{
"epoch": 1.5640194489465153,
"grad_norm": 0.3703957202531771,
"learning_rate": 1.1012137234493257e-05,
"loss": 0.3084,
"step": 1930
},
{
"epoch": 1.5721231766612642,
"grad_norm": 0.39826934578530954,
"learning_rate": 1.0918264140454858e-05,
"loss": 0.3083,
"step": 1940
},
{
"epoch": 1.5802269043760129,
"grad_norm": 0.7352916958662763,
"learning_rate": 1.0824309366575102e-05,
"loss": 0.3035,
"step": 1950
},
{
"epoch": 1.5883306320907618,
"grad_norm": 0.397036753855381,
"learning_rate": 1.07302812701563e-05,
"loss": 0.3106,
"step": 1960
},
{
"epoch": 1.5964343598055106,
"grad_norm": 0.3578171904603572,
"learning_rate": 1.0636188215022822e-05,
"loss": 0.3072,
"step": 1970
},
{
"epoch": 1.6045380875202593,
"grad_norm": 0.42176295168410677,
"learning_rate": 1.0542038570777145e-05,
"loss": 0.3005,
"step": 1980
},
{
"epoch": 1.6126418152350082,
"grad_norm": 0.36392753016960694,
"learning_rate": 1.0447840712055346e-05,
"loss": 0.2979,
"step": 1990
},
{
"epoch": 1.620745542949757,
"grad_norm": 0.35565052352071846,
"learning_rate": 1.0353603017782198e-05,
"loss": 0.3067,
"step": 2000
},
{
"epoch": 1.6288492706645057,
"grad_norm": 0.3534123782056326,
"learning_rate": 1.025933387042587e-05,
"loss": 0.3012,
"step": 2010
},
{
"epoch": 1.6369529983792543,
"grad_norm": 0.3712130571783743,
"learning_rate": 1.0165041655252273e-05,
"loss": 0.3077,
"step": 2020
},
{
"epoch": 1.6450567260940032,
"grad_norm": 0.3873748455398449,
"learning_rate": 1.0070734759579213e-05,
"loss": 0.3056,
"step": 2030
},
{
"epoch": 1.653160453808752,
"grad_norm": 0.402229853077094,
"learning_rate": 9.976421572030335e-06,
"loss": 0.3079,
"step": 2040
},
{
"epoch": 1.6612641815235007,
"grad_norm": 0.36166023457576596,
"learning_rate": 9.882110481788943e-06,
"loss": 0.3022,
"step": 2050
},
{
"epoch": 1.6693679092382496,
"grad_norm": 0.38042934153448144,
"learning_rate": 9.787809877851788e-06,
"loss": 0.3052,
"step": 2060
},
{
"epoch": 1.6774716369529985,
"grad_norm": 0.35732867252673295,
"learning_rate": 9.693528148282858e-06,
"loss": 0.3106,
"step": 2070
},
{
"epoch": 1.6855753646677472,
"grad_norm": 0.3557537206036274,
"learning_rate": 9.599273679467261e-06,
"loss": 0.3045,
"step": 2080
},
{
"epoch": 1.6936790923824958,
"grad_norm": 0.38418581588863143,
"learning_rate": 9.505054855365255e-06,
"loss": 0.3048,
"step": 2090
},
{
"epoch": 1.7017828200972447,
"grad_norm": 0.36659241830029776,
"learning_rate": 9.4108800567665e-06,
"loss": 0.3051,
"step": 2100
},
{
"epoch": 1.7098865478119936,
"grad_norm": 0.33583794828920516,
"learning_rate": 9.31675766054456e-06,
"loss": 0.3022,
"step": 2110
},
{
"epoch": 1.7179902755267422,
"grad_norm": 0.3792749151698344,
"learning_rate": 9.222696038911799e-06,
"loss": 0.3067,
"step": 2120
},
{
"epoch": 1.726094003241491,
"grad_norm": 0.3644544219380795,
"learning_rate": 9.128703558674671e-06,
"loss": 0.3046,
"step": 2130
},
{
"epoch": 1.73419773095624,
"grad_norm": 0.34781480850287866,
"learning_rate": 9.034788580489483e-06,
"loss": 0.3035,
"step": 2140
},
{
"epoch": 1.7423014586709886,
"grad_norm": 0.38098830321811866,
"learning_rate": 8.9409594581187e-06,
"loss": 0.2987,
"step": 2150
},
{
"epoch": 1.7504051863857373,
"grad_norm": 0.3928927475654598,
"learning_rate": 8.84722453768791e-06,
"loss": 0.3016,
"step": 2160
},
{
"epoch": 1.7585089141004864,
"grad_norm": 0.3810130176649336,
"learning_rate": 8.753592156943398e-06,
"loss": 0.3008,
"step": 2170
},
{
"epoch": 1.766612641815235,
"grad_norm": 0.3673708859368621,
"learning_rate": 8.660070644510527e-06,
"loss": 0.2996,
"step": 2180
},
{
"epoch": 1.7747163695299837,
"grad_norm": 0.3432831698380344,
"learning_rate": 8.566668319152885e-06,
"loss": 0.3026,
"step": 2190
},
{
"epoch": 1.7828200972447326,
"grad_norm": 0.3539730750475116,
"learning_rate": 8.47339348903234e-06,
"loss": 0.2954,
"step": 2200
},
{
"epoch": 1.7909238249594814,
"grad_norm": 0.38021077491756267,
"learning_rate": 8.38025445097003e-06,
"loss": 0.304,
"step": 2210
},
{
"epoch": 1.79902755267423,
"grad_norm": 0.3772266937147294,
"learning_rate": 8.287259489708345e-06,
"loss": 0.3019,
"step": 2220
},
{
"epoch": 1.807131280388979,
"grad_norm": 0.3600033524881556,
"learning_rate": 8.194416877174011e-06,
"loss": 0.2997,
"step": 2230
},
{
"epoch": 1.8152350081037278,
"grad_norm": 0.40881095918061927,
"learning_rate": 8.101734871742293e-06,
"loss": 0.3036,
"step": 2240
},
{
"epoch": 1.8233387358184765,
"grad_norm": 0.3385939341389464,
"learning_rate": 8.00922171750241e-06,
"loss": 0.3017,
"step": 2250
},
{
"epoch": 1.8314424635332252,
"grad_norm": 0.39501609760805706,
"learning_rate": 7.916885643524218e-06,
"loss": 0.3086,
"step": 2260
},
{
"epoch": 1.839546191247974,
"grad_norm": 0.39922657943009876,
"learning_rate": 7.824734863126242e-06,
"loss": 0.3031,
"step": 2270
},
{
"epoch": 1.847649918962723,
"grad_norm": 0.387442623340492,
"learning_rate": 7.732777573145095e-06,
"loss": 0.3019,
"step": 2280
},
{
"epoch": 1.8557536466774716,
"grad_norm": 0.36454095702438943,
"learning_rate": 7.641021953206365e-06,
"loss": 0.3047,
"step": 2290
},
{
"epoch": 1.8638573743922204,
"grad_norm": 0.3582449767433757,
"learning_rate": 7.549476164997023e-06,
"loss": 0.305,
"step": 2300
},
{
"epoch": 1.8719611021069693,
"grad_norm": 0.3900631140559724,
"learning_rate": 7.458148351539469e-06,
"loss": 0.2993,
"step": 2310
},
{
"epoch": 1.880064829821718,
"grad_norm": 0.3852621286926431,
"learning_rate": 7.367046636467181e-06,
"loss": 0.2997,
"step": 2320
},
{
"epoch": 1.8881685575364666,
"grad_norm": 0.3649720452278969,
"learning_rate": 7.276179123302132e-06,
"loss": 0.3017,
"step": 2330
},
{
"epoch": 1.8962722852512157,
"grad_norm": 0.35300522121938893,
"learning_rate": 7.185553894733961e-06,
"loss": 0.3046,
"step": 2340
},
{
"epoch": 1.9043760129659644,
"grad_norm": 0.3507704303666437,
"learning_rate": 7.0951790119010335e-06,
"loss": 0.3015,
"step": 2350
},
{
"epoch": 1.912479740680713,
"grad_norm": 0.3839261297077148,
"learning_rate": 7.0050625136734e-06,
"loss": 0.2998,
"step": 2360
},
{
"epoch": 1.920583468395462,
"grad_norm": 0.3981276874011154,
"learning_rate": 6.915212415937727e-06,
"loss": 0.2998,
"step": 2370
},
{
"epoch": 1.9286871961102108,
"grad_norm": 0.3528142239405429,
"learning_rate": 6.825636710884276e-06,
"loss": 0.2977,
"step": 2380
},
{
"epoch": 1.9367909238249594,
"grad_norm": 0.3675295952763021,
"learning_rate": 6.736343366296023e-06,
"loss": 0.3091,
"step": 2390
},
{
"epoch": 1.9448946515397083,
"grad_norm": 0.3577900618376749,
"learning_rate": 6.647340324839899e-06,
"loss": 0.3023,
"step": 2400
},
{
"epoch": 1.9529983792544572,
"grad_norm": 0.3674518637690264,
"learning_rate": 6.558635503360282e-06,
"loss": 0.2988,
"step": 2410
},
{
"epoch": 1.9611021069692058,
"grad_norm": 0.3605169758525672,
"learning_rate": 6.470236792174821e-06,
"loss": 0.2987,
"step": 2420
},
{
"epoch": 1.9692058346839545,
"grad_norm": 0.3824360661190583,
"learning_rate": 6.382152054372559e-06,
"loss": 0.2972,
"step": 2430
},
{
"epoch": 1.9773095623987034,
"grad_norm": 0.38147252308460916,
"learning_rate": 6.294389125114533e-06,
"loss": 0.3017,
"step": 2440
},
{
"epoch": 1.9854132901134522,
"grad_norm": 0.3773849249838894,
"learning_rate": 6.206955810936812e-06,
"loss": 0.3023,
"step": 2450
},
{
"epoch": 1.993517017828201,
"grad_norm": 0.3622352292875066,
"learning_rate": 6.1198598890561245e-06,
"loss": 0.3002,
"step": 2460
},
{
"epoch": 2.0016207455429496,
"grad_norm": 0.48414462847535367,
"learning_rate": 6.033109106678061e-06,
"loss": 0.2895,
"step": 2470
},
{
"epoch": 2.0097244732576987,
"grad_norm": 0.4778040530703245,
"learning_rate": 5.946711180307962e-06,
"loss": 0.2487,
"step": 2480
},
{
"epoch": 2.0178282009724473,
"grad_norm": 0.40481507834165836,
"learning_rate": 5.860673795064526e-06,
"loss": 0.2419,
"step": 2490
},
{
"epoch": 2.025931928687196,
"grad_norm": 0.3769283264752454,
"learning_rate": 5.7750046039962256e-06,
"loss": 0.2449,
"step": 2500
},
{
"epoch": 2.034035656401945,
"grad_norm": 0.40225289732048686,
"learning_rate": 5.689711227400565e-06,
"loss": 0.2369,
"step": 2510
},
{
"epoch": 2.0421393841166937,
"grad_norm": 0.3597440140671611,
"learning_rate": 5.604801252146254e-06,
"loss": 0.2449,
"step": 2520
},
{
"epoch": 2.0502431118314424,
"grad_norm": 0.43056264385312676,
"learning_rate": 5.520282230998338e-06,
"loss": 0.2377,
"step": 2530
},
{
"epoch": 2.0583468395461915,
"grad_norm": 0.40672175374448094,
"learning_rate": 5.436161681946395e-06,
"loss": 0.2404,
"step": 2540
},
{
"epoch": 2.06645056726094,
"grad_norm": 0.42926885318716373,
"learning_rate": 5.352447087535806e-06,
"loss": 0.2414,
"step": 2550
},
{
"epoch": 2.0745542949756888,
"grad_norm": 0.5162855996427805,
"learning_rate": 5.2691458942021824e-06,
"loss": 0.245,
"step": 2560
},
{
"epoch": 2.0826580226904374,
"grad_norm": 0.4030059706999796,
"learning_rate": 5.186265511608986e-06,
"loss": 0.2449,
"step": 2570
},
{
"epoch": 2.0907617504051865,
"grad_norm": 0.3640201809957514,
"learning_rate": 5.103813311988464e-06,
"loss": 0.2469,
"step": 2580
},
{
"epoch": 2.098865478119935,
"grad_norm": 0.3871861841266068,
"learning_rate": 5.021796629485864e-06,
"loss": 0.2514,
"step": 2590
},
{
"epoch": 2.106969205834684,
"grad_norm": 0.5492303866157153,
"learning_rate": 4.9402227595070685e-06,
"loss": 0.242,
"step": 2600
},
{
"epoch": 2.115072933549433,
"grad_norm": 0.3835548046205682,
"learning_rate": 4.859098958069676e-06,
"loss": 0.2456,
"step": 2610
},
{
"epoch": 2.1231766612641816,
"grad_norm": 0.39957383330666024,
"learning_rate": 4.778432441157563e-06,
"loss": 0.2427,
"step": 2620
},
{
"epoch": 2.1312803889789302,
"grad_norm": 0.39333044441333004,
"learning_rate": 4.698230384079027e-06,
"loss": 0.2393,
"step": 2630
},
{
"epoch": 2.139384116693679,
"grad_norm": 0.36729336365936005,
"learning_rate": 4.618499920828534e-06,
"loss": 0.2372,
"step": 2640
},
{
"epoch": 2.147487844408428,
"grad_norm": 0.40530532180259077,
"learning_rate": 4.539248143452152e-06,
"loss": 0.2441,
"step": 2650
},
{
"epoch": 2.1555915721231766,
"grad_norm": 0.3633603405343877,
"learning_rate": 4.46048210141672e-06,
"loss": 0.2415,
"step": 2660
},
{
"epoch": 2.1636952998379253,
"grad_norm": 0.38435198476630983,
"learning_rate": 4.382208800982786e-06,
"loss": 0.2485,
"step": 2670
},
{
"epoch": 2.1717990275526744,
"grad_norm": 0.36454635251775136,
"learning_rate": 4.304435204581391e-06,
"loss": 0.2428,
"step": 2680
},
{
"epoch": 2.179902755267423,
"grad_norm": 0.44742997759209024,
"learning_rate": 4.227168230194784e-06,
"loss": 0.2456,
"step": 2690
},
{
"epoch": 2.1880064829821717,
"grad_norm": 0.40837321795280773,
"learning_rate": 4.150414750741034e-06,
"loss": 0.2404,
"step": 2700
},
{
"epoch": 2.1961102106969204,
"grad_norm": 0.3821888357009334,
"learning_rate": 4.074181593462714e-06,
"loss": 0.242,
"step": 2710
},
{
"epoch": 2.2042139384116695,
"grad_norm": 0.40271178342646896,
"learning_rate": 3.998475539319584e-06,
"loss": 0.2425,
"step": 2720
},
{
"epoch": 2.212317666126418,
"grad_norm": 0.38801509503129883,
"learning_rate": 3.92330332238545e-06,
"loss": 0.2442,
"step": 2730
},
{
"epoch": 2.2204213938411668,
"grad_norm": 0.3534415230352728,
"learning_rate": 3.8486716292491535e-06,
"loss": 0.241,
"step": 2740
},
{
"epoch": 2.228525121555916,
"grad_norm": 0.37647910687345254,
"learning_rate": 3.7745870984198008e-06,
"loss": 0.2422,
"step": 2750
},
{
"epoch": 2.2366288492706645,
"grad_norm": 0.3766459600082522,
"learning_rate": 3.701056319736257e-06,
"loss": 0.2448,
"step": 2760
},
{
"epoch": 2.244732576985413,
"grad_norm": 0.3464221374514046,
"learning_rate": 3.6280858337809875e-06,
"loss": 0.2411,
"step": 2770
},
{
"epoch": 2.2528363047001623,
"grad_norm": 0.35349946526860715,
"learning_rate": 3.555682131298279e-06,
"loss": 0.2424,
"step": 2780
},
{
"epoch": 2.260940032414911,
"grad_norm": 0.3746915447251235,
"learning_rate": 3.483851652616862e-06,
"loss": 0.2424,
"step": 2790
},
{
"epoch": 2.2690437601296596,
"grad_norm": 0.6234025063246161,
"learning_rate": 3.412600787077066e-06,
"loss": 0.2406,
"step": 2800
},
{
"epoch": 2.2771474878444082,
"grad_norm": 0.37707590766424204,
"learning_rate": 3.3419358724624785e-06,
"loss": 0.2413,
"step": 2810
},
{
"epoch": 2.2852512155591573,
"grad_norm": 0.3621840927527013,
"learning_rate": 3.2718631944361856e-06,
"loss": 0.2416,
"step": 2820
},
{
"epoch": 2.293354943273906,
"grad_norm": 0.41372028650623555,
"learning_rate": 3.2023889859816737e-06,
"loss": 0.2421,
"step": 2830
},
{
"epoch": 2.3014586709886546,
"grad_norm": 0.3935266800497859,
"learning_rate": 3.1335194268484027e-06,
"loss": 0.2454,
"step": 2840
},
{
"epoch": 2.3095623987034037,
"grad_norm": 0.4248974196750723,
"learning_rate": 3.065260643002115e-06,
"loss": 0.2393,
"step": 2850
},
{
"epoch": 2.3176661264181524,
"grad_norm": 0.3779952532638862,
"learning_rate": 2.9976187060799254e-06,
"loss": 0.2438,
"step": 2860
},
{
"epoch": 2.325769854132901,
"grad_norm": 0.3710708542572521,
"learning_rate": 2.9305996328502473e-06,
"loss": 0.2457,
"step": 2870
},
{
"epoch": 2.3338735818476497,
"grad_norm": 0.48653618881309907,
"learning_rate": 2.8642093846776007e-06,
"loss": 0.237,
"step": 2880
},
{
"epoch": 2.341977309562399,
"grad_norm": 0.37666681552522957,
"learning_rate": 2.7984538669923534e-06,
"loss": 0.2406,
"step": 2890
},
{
"epoch": 2.3500810372771475,
"grad_norm": 0.3696871763042728,
"learning_rate": 2.7333389287654276e-06,
"loss": 0.2406,
"step": 2900
},
{
"epoch": 2.358184764991896,
"grad_norm": 0.40375286405615424,
"learning_rate": 2.6688703619880207e-06,
"loss": 0.2443,
"step": 2910
},
{
"epoch": 2.366288492706645,
"grad_norm": 0.3951499786529782,
"learning_rate": 2.6050539011564256e-06,
"loss": 0.2378,
"step": 2920
},
{
"epoch": 2.374392220421394,
"grad_norm": 0.3455475906909619,
"learning_rate": 2.5418952227619387e-06,
"loss": 0.2431,
"step": 2930
},
{
"epoch": 2.3824959481361425,
"grad_norm": 0.35442046683340545,
"learning_rate": 2.479399944785923e-06,
"loss": 0.2417,
"step": 2940
},
{
"epoch": 2.3905996758508916,
"grad_norm": 0.3790247644949051,
"learning_rate": 2.4175736262001003e-06,
"loss": 0.2424,
"step": 2950
},
{
"epoch": 2.3987034035656403,
"grad_norm": 0.43282934256355926,
"learning_rate": 2.356421766472077e-06,
"loss": 0.2389,
"step": 2960
},
{
"epoch": 2.406807131280389,
"grad_norm": 0.33507324391542254,
"learning_rate": 2.2959498050761665e-06,
"loss": 0.2436,
"step": 2970
},
{
"epoch": 2.4149108589951376,
"grad_norm": 0.37525875829578087,
"learning_rate": 2.236163121009535e-06,
"loss": 0.2359,
"step": 2980
},
{
"epoch": 2.4230145867098867,
"grad_norm": 0.36335267791657855,
"learning_rate": 2.1770670323137543e-06,
"loss": 0.2388,
"step": 2990
},
{
"epoch": 2.4311183144246353,
"grad_norm": 0.41452055541584026,
"learning_rate": 2.118666795601746e-06,
"loss": 0.2456,
"step": 3000
},
{
"epoch": 2.439222042139384,
"grad_norm": 0.3905804922635912,
"learning_rate": 2.0609676055902206e-06,
"loss": 0.2367,
"step": 3010
},
{
"epoch": 2.447325769854133,
"grad_norm": 0.36571192283420584,
"learning_rate": 2.0039745946375843e-06,
"loss": 0.242,
"step": 3020
},
{
"epoch": 2.4554294975688817,
"grad_norm": 0.3675617748803302,
"learning_rate": 1.9476928322874388e-06,
"loss": 0.2432,
"step": 3030
},
{
"epoch": 2.4635332252836304,
"grad_norm": 0.35676767068188675,
"learning_rate": 1.8921273248176274e-06,
"loss": 0.237,
"step": 3040
},
{
"epoch": 2.471636952998379,
"grad_norm": 0.36644299533028146,
"learning_rate": 1.8372830147949273e-06,
"loss": 0.2442,
"step": 3050
},
{
"epoch": 2.479740680713128,
"grad_norm": 0.4160554805432118,
"learning_rate": 1.783164780635409e-06,
"loss": 0.231,
"step": 3060
},
{
"epoch": 2.487844408427877,
"grad_norm": 0.3980921006536382,
"learning_rate": 1.7297774361705067e-06,
"loss": 0.2401,
"step": 3070
},
{
"epoch": 2.4959481361426255,
"grad_norm": 0.3346524587640772,
"learning_rate": 1.6771257302188138e-06,
"loss": 0.2392,
"step": 3080
},
{
"epoch": 2.5040518638573745,
"grad_norm": 0.32559428720426414,
"learning_rate": 1.6252143461636894e-06,
"loss": 0.2365,
"step": 3090
},
{
"epoch": 2.512155591572123,
"grad_norm": 0.3542121354772109,
"learning_rate": 1.5740479015366516e-06,
"loss": 0.2466,
"step": 3100
},
{
"epoch": 2.520259319286872,
"grad_norm": 0.3674306028706605,
"learning_rate": 1.523630947606659e-06,
"loss": 0.2406,
"step": 3110
},
{
"epoch": 2.528363047001621,
"grad_norm": 0.6504296348081645,
"learning_rate": 1.4739679689752761e-06,
"loss": 0.2404,
"step": 3120
},
{
"epoch": 2.5364667747163696,
"grad_norm": 0.37795777647227535,
"learning_rate": 1.4250633831777617e-06,
"loss": 0.2397,
"step": 3130
},
{
"epoch": 2.5445705024311183,
"grad_norm": 0.34829635023141153,
"learning_rate": 1.3769215402901248e-06,
"loss": 0.2481,
"step": 3140
},
{
"epoch": 2.5526742301458674,
"grad_norm": 0.3367601703351784,
"learning_rate": 1.3295467225421922e-06,
"loss": 0.2354,
"step": 3150
},
{
"epoch": 2.560777957860616,
"grad_norm": 0.41258176404249586,
"learning_rate": 1.2829431439366979e-06,
"loss": 0.2379,
"step": 3160
},
{
"epoch": 2.5688816855753647,
"grad_norm": 0.3586025355065321,
"learning_rate": 1.2371149498744483e-06,
"loss": 0.2386,
"step": 3170
},
{
"epoch": 2.5769854132901133,
"grad_norm": 0.3977383494490472,
"learning_rate": 1.1920662167855801e-06,
"loss": 0.2422,
"step": 3180
},
{
"epoch": 2.585089141004862,
"grad_norm": 0.3641837478877437,
"learning_rate": 1.1478009517669786e-06,
"loss": 0.2395,
"step": 3190
},
{
"epoch": 2.593192868719611,
"grad_norm": 0.3871271662932699,
"learning_rate": 1.1043230922258286e-06,
"loss": 0.2425,
"step": 3200
},
{
"epoch": 2.6012965964343597,
"grad_norm": 0.36018092198137386,
"learning_rate": 1.0616365055293832e-06,
"loss": 0.2441,
"step": 3210
},
{
"epoch": 2.6094003241491084,
"grad_norm": 0.3753745984491537,
"learning_rate": 1.0197449886609734e-06,
"loss": 0.2417,
"step": 3220
},
{
"epoch": 2.6175040518638575,
"grad_norm": 0.42776777626884477,
"learning_rate": 9.786522678822441e-07,
"loss": 0.2354,
"step": 3230
},
{
"epoch": 2.625607779578606,
"grad_norm": 0.34671945274097893,
"learning_rate": 9.383619984017256e-07,
"loss": 0.244,
"step": 3240
},
{
"epoch": 2.633711507293355,
"grad_norm": 0.3618790436982639,
"learning_rate": 8.988777640496804e-07,
"loss": 0.2421,
"step": 3250
},
{
"epoch": 2.641815235008104,
"grad_norm": 0.4049583851733269,
"learning_rate": 8.602030769593361e-07,
"loss": 0.2371,
"step": 3260
},
{
"epoch": 2.6499189627228525,
"grad_norm": 0.34242168591352745,
"learning_rate": 8.223413772544753e-07,
"loss": 0.2473,
"step": 3270
},
{
"epoch": 2.658022690437601,
"grad_norm": 0.3499036792341549,
"learning_rate": 7.8529603274343e-07,
"loss": 0.2432,
"step": 3280
},
{
"epoch": 2.6661264181523503,
"grad_norm": 0.36172746060316596,
"learning_rate": 7.490703386195209e-07,
"loss": 0.2384,
"step": 3290
},
{
"epoch": 2.674230145867099,
"grad_norm": 0.35677730811220215,
"learning_rate": 7.136675171679486e-07,
"loss": 0.2375,
"step": 3300
},
{
"epoch": 2.6823338735818476,
"grad_norm": 0.36366157530580207,
"learning_rate": 6.790907174791673e-07,
"loss": 0.2405,
"step": 3310
},
{
"epoch": 2.6904376012965967,
"grad_norm": 0.3539145763518158,
"learning_rate": 6.453430151687778e-07,
"loss": 0.2407,
"step": 3320
},
{
"epoch": 2.6985413290113454,
"grad_norm": 0.3778238325447947,
"learning_rate": 6.124274121039409e-07,
"loss": 0.2378,
"step": 3330
},
{
"epoch": 2.706645056726094,
"grad_norm": 0.38437638190308854,
"learning_rate": 5.803468361363718e-07,
"loss": 0.237,
"step": 3340
},
{
"epoch": 2.7147487844408427,
"grad_norm": 0.34533885030316197,
"learning_rate": 5.491041408418973e-07,
"loss": 0.2375,
"step": 3350
},
{
"epoch": 2.7228525121555913,
"grad_norm": 0.35887290078699746,
"learning_rate": 5.187021052666364e-07,
"loss": 0.2394,
"step": 3360
},
{
"epoch": 2.7309562398703404,
"grad_norm": 0.3469607617700606,
"learning_rate": 4.891434336797995e-07,
"loss": 0.2395,
"step": 3370
},
{
"epoch": 2.739059967585089,
"grad_norm": 0.3679709706559888,
"learning_rate": 4.604307553331466e-07,
"loss": 0.2353,
"step": 3380
},
{
"epoch": 2.7471636952998377,
"grad_norm": 0.37439295175686094,
"learning_rate": 4.3256662422711203e-07,
"loss": 0.2415,
"step": 3390
},
{
"epoch": 2.755267423014587,
"grad_norm": 0.3464818073633876,
"learning_rate": 4.0555351888362126e-07,
"loss": 0.2375,
"step": 3400
},
{
"epoch": 2.7633711507293355,
"grad_norm": 0.38460802340884914,
"learning_rate": 3.793938421256349e-07,
"loss": 0.243,
"step": 3410
},
{
"epoch": 2.771474878444084,
"grad_norm": 0.3784763205025592,
"learning_rate": 3.5408992086340963e-07,
"loss": 0.2436,
"step": 3420
},
{
"epoch": 2.7795786061588332,
"grad_norm": 0.36507578916154754,
"learning_rate": 3.296440058875239e-07,
"loss": 0.2438,
"step": 3430
},
{
"epoch": 2.787682333873582,
"grad_norm": 0.41031945985509827,
"learning_rate": 3.060582716686633e-07,
"loss": 0.2429,
"step": 3440
},
{
"epoch": 2.7957860615883305,
"grad_norm": 0.3549499553755001,
"learning_rate": 2.833348161642091e-07,
"loss": 0.2418,
"step": 3450
},
{
"epoch": 2.8038897893030796,
"grad_norm": 0.36952966030768436,
"learning_rate": 2.61475660631616e-07,
"loss": 0.2406,
"step": 3460
},
{
"epoch": 2.8119935170178283,
"grad_norm": 0.35964152966221336,
"learning_rate": 2.4048274944862837e-07,
"loss": 0.2393,
"step": 3470
},
{
"epoch": 2.820097244732577,
"grad_norm": 0.43135260965676503,
"learning_rate": 2.2035794994031856e-07,
"loss": 0.2399,
"step": 3480
},
{
"epoch": 2.828200972447326,
"grad_norm": 0.5065451500877334,
"learning_rate": 2.0110305221299641e-07,
"loss": 0.2415,
"step": 3490
},
{
"epoch": 2.8363047001620747,
"grad_norm": 0.3454128864980503,
"learning_rate": 1.8271976899497555e-07,
"loss": 0.2448,
"step": 3500
},
{
"epoch": 2.8444084278768234,
"grad_norm": 0.3406341708191905,
"learning_rate": 1.6520973548422526e-07,
"loss": 0.2397,
"step": 3510
},
{
"epoch": 2.852512155591572,
"grad_norm": 0.363239272589924,
"learning_rate": 1.485745092029145e-07,
"loss": 0.2385,
"step": 3520
},
{
"epoch": 2.8606158833063207,
"grad_norm": 0.6575701078321629,
"learning_rate": 1.328155698588829e-07,
"loss": 0.2375,
"step": 3530
},
{
"epoch": 2.8687196110210698,
"grad_norm": 0.34359980127668627,
"learning_rate": 1.1793431921400589e-07,
"loss": 0.2389,
"step": 3540
},
{
"epoch": 2.8768233387358184,
"grad_norm": 0.38043150212923815,
"learning_rate": 1.0393208095951569e-07,
"loss": 0.2416,
"step": 3550
},
{
"epoch": 2.884927066450567,
"grad_norm": 0.3512211277422864,
"learning_rate": 9.081010059825868e-08,
"loss": 0.237,
"step": 3560
},
{
"epoch": 2.893030794165316,
"grad_norm": 0.37245699031048846,
"learning_rate": 7.856954533390305e-08,
"loss": 0.2378,
"step": 3570
},
{
"epoch": 2.901134521880065,
"grad_norm": 0.35845897853920533,
"learning_rate": 6.721150396712062e-08,
"loss": 0.237,
"step": 3580
},
{
"epoch": 2.9092382495948135,
"grad_norm": 0.35955619474371414,
"learning_rate": 5.67369867987344e-08,
"loss": 0.2382,
"step": 3590
},
{
"epoch": 2.9173419773095626,
"grad_norm": 0.3514639483532844,
"learning_rate": 4.7146925539852675e-08,
"loss": 0.2376,
"step": 3600
},
{
"epoch": 2.9254457050243112,
"grad_norm": 1.9726171105605486,
"learning_rate": 3.8442173228994126e-08,
"loss": 0.2438,
"step": 3610
},
{
"epoch": 2.93354943273906,
"grad_norm": 0.32818993394512846,
"learning_rate": 3.0623504156210846e-08,
"loss": 0.2377,
"step": 3620
},
{
"epoch": 2.941653160453809,
"grad_norm": 0.3586707061758846,
"learning_rate": 2.3691613794210034e-08,
"loss": 0.2353,
"step": 3630
},
{
"epoch": 2.9497568881685576,
"grad_norm": 0.36206597653859174,
"learning_rate": 1.7647118736499046e-08,
"loss": 0.2401,
"step": 3640
},
{
"epoch": 2.9578606158833063,
"grad_norm": 0.35445601697521645,
"learning_rate": 1.2490556642530404e-08,
"loss": 0.2344,
"step": 3650
},
{
"epoch": 2.965964343598055,
"grad_norm": 0.3374877564626054,
"learning_rate": 8.222386189886688e-09,
"loss": 0.2351,
"step": 3660
},
{
"epoch": 2.974068071312804,
"grad_norm": 0.3530874111127371,
"learning_rate": 4.84298703347319e-09,
"loss": 0.2407,
"step": 3670
},
{
"epoch": 2.9821717990275527,
"grad_norm": 0.35497167072248725,
"learning_rate": 2.3526597717482558e-09,
"loss": 0.2438,
"step": 3680
},
{
"epoch": 2.9902755267423013,
"grad_norm": 0.34905140416452324,
"learning_rate": 7.516259199913389e-10,
"loss": 0.2402,
"step": 3690
},
{
"epoch": 2.99837925445705,
"grad_norm": 0.3498948265589354,
"learning_rate": 4.002789059098788e-11,
"loss": 0.2425,
"step": 3700
}
],
"logging_steps": 10,
"max_steps": 3702,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 10000000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5702183913259008.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}