GUI-AIMA-3B-kl / trainer_state.json
smz8599's picture
Initial upload
52bb4f1 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.5023863350916855,
"eval_steps": 500,
"global_step": 2000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0025119316754584277,
"grad_norm": 244.78928615182636,
"learning_rate": 3.75e-07,
"loss": 14.0363,
"step": 10
},
{
"epoch": 0.005023863350916855,
"grad_norm": 318.1891497601652,
"learning_rate": 7.916666666666667e-07,
"loss": 13.4987,
"step": 20
},
{
"epoch": 0.007535795026375283,
"grad_norm": 211.3992791294665,
"learning_rate": 1.2083333333333333e-06,
"loss": 11.4405,
"step": 30
},
{
"epoch": 0.01004772670183371,
"grad_norm": 425.154938297453,
"learning_rate": 1.6250000000000001e-06,
"loss": 8.7926,
"step": 40
},
{
"epoch": 0.012559658377292138,
"grad_norm": 136.69306564524402,
"learning_rate": 2.041666666666667e-06,
"loss": 7.1601,
"step": 50
},
{
"epoch": 0.015071590052750565,
"grad_norm": 173.2049647337299,
"learning_rate": 2.4583333333333332e-06,
"loss": 6.2254,
"step": 60
},
{
"epoch": 0.017583521728208994,
"grad_norm": 163.0490968387617,
"learning_rate": 2.875e-06,
"loss": 5.7153,
"step": 70
},
{
"epoch": 0.02009545340366742,
"grad_norm": 70.98311735934806,
"learning_rate": 3.2916666666666668e-06,
"loss": 5.3196,
"step": 80
},
{
"epoch": 0.02260738507912585,
"grad_norm": 101.87087924315038,
"learning_rate": 3.708333333333334e-06,
"loss": 5.1087,
"step": 90
},
{
"epoch": 0.025119316754584276,
"grad_norm": 215.4312626681897,
"learning_rate": 4.125e-06,
"loss": 4.9029,
"step": 100
},
{
"epoch": 0.027631248430042703,
"grad_norm": 78.6441119779914,
"learning_rate": 4.541666666666667e-06,
"loss": 4.8007,
"step": 110
},
{
"epoch": 0.03014318010550113,
"grad_norm": 112.82426912718005,
"learning_rate": 4.958333333333334e-06,
"loss": 4.8032,
"step": 120
},
{
"epoch": 0.032655111780959555,
"grad_norm": 44.43109383036881,
"learning_rate": 4.999932966293553e-06,
"loss": 4.5469,
"step": 130
},
{
"epoch": 0.03516704345641799,
"grad_norm": 42.94277182614008,
"learning_rate": 4.999701249454364e-06,
"loss": 4.6328,
"step": 140
},
{
"epoch": 0.037678975131876416,
"grad_norm": 102.19912456135354,
"learning_rate": 4.999304037243215e-06,
"loss": 4.5481,
"step": 150
},
{
"epoch": 0.04019090680733484,
"grad_norm": 103.48520496328905,
"learning_rate": 4.998741355957963e-06,
"loss": 4.0883,
"step": 160
},
{
"epoch": 0.04270283848279327,
"grad_norm": 59.75671346819062,
"learning_rate": 4.998013242851519e-06,
"loss": 4.09,
"step": 170
},
{
"epoch": 0.0452147701582517,
"grad_norm": 54.56786450011987,
"learning_rate": 4.997119746129383e-06,
"loss": 3.8574,
"step": 180
},
{
"epoch": 0.047726701833710125,
"grad_norm": 41.17219526900339,
"learning_rate": 4.996060924946452e-06,
"loss": 3.3506,
"step": 190
},
{
"epoch": 0.05023863350916855,
"grad_norm": 64.19755013449358,
"learning_rate": 4.994836849403107e-06,
"loss": 3.1208,
"step": 200
},
{
"epoch": 0.05275056518462698,
"grad_norm": 45.5637995880343,
"learning_rate": 4.9934476005405665e-06,
"loss": 2.9927,
"step": 210
},
{
"epoch": 0.05526249686008541,
"grad_norm": 45.04513390690093,
"learning_rate": 4.991893270335526e-06,
"loss": 2.6818,
"step": 220
},
{
"epoch": 0.057774428535543834,
"grad_norm": 46.258300578654584,
"learning_rate": 4.990173961694062e-06,
"loss": 2.4673,
"step": 230
},
{
"epoch": 0.06028636021100226,
"grad_norm": 53.325400128100654,
"learning_rate": 4.988289788444829e-06,
"loss": 2.1375,
"step": 240
},
{
"epoch": 0.06279829188646069,
"grad_norm": 28.819163526585655,
"learning_rate": 4.986240875331513e-06,
"loss": 1.9451,
"step": 250
},
{
"epoch": 0.06531022356191911,
"grad_norm": 63.6642536784459,
"learning_rate": 4.984027358004582e-06,
"loss": 1.7398,
"step": 260
},
{
"epoch": 0.06782215523737754,
"grad_norm": 64.052007917374,
"learning_rate": 4.981649383012295e-06,
"loss": 1.5556,
"step": 270
},
{
"epoch": 0.07033408691283598,
"grad_norm": 41.550046723661794,
"learning_rate": 4.979107107791009e-06,
"loss": 1.5029,
"step": 280
},
{
"epoch": 0.0728460185882944,
"grad_norm": 28.030603583599877,
"learning_rate": 4.976400700654752e-06,
"loss": 1.4048,
"step": 290
},
{
"epoch": 0.07535795026375283,
"grad_norm": 21.7743052603298,
"learning_rate": 4.9735303407840775e-06,
"loss": 1.3447,
"step": 300
},
{
"epoch": 0.07786988193921125,
"grad_norm": 27.724313133031774,
"learning_rate": 4.970496218214205e-06,
"loss": 1.369,
"step": 310
},
{
"epoch": 0.08038181361466969,
"grad_norm": 39.89486925273176,
"learning_rate": 4.9672985338224355e-06,
"loss": 1.2511,
"step": 320
},
{
"epoch": 0.0828937452901281,
"grad_norm": 32.71736411454423,
"learning_rate": 4.963937499314857e-06,
"loss": 1.1617,
"step": 330
},
{
"epoch": 0.08540567696558654,
"grad_norm": 17.47130667747679,
"learning_rate": 4.960413337212321e-06,
"loss": 1.1987,
"step": 340
},
{
"epoch": 0.08791760864104496,
"grad_norm": 17.44490963530016,
"learning_rate": 4.956726280835718e-06,
"loss": 1.2543,
"step": 350
},
{
"epoch": 0.0904295403165034,
"grad_norm": 52.67566693581501,
"learning_rate": 4.952876574290525e-06,
"loss": 1.1756,
"step": 360
},
{
"epoch": 0.09294147199196182,
"grad_norm": 36.08376015056633,
"learning_rate": 4.948864472450646e-06,
"loss": 1.1497,
"step": 370
},
{
"epoch": 0.09545340366742025,
"grad_norm": 33.06467813616241,
"learning_rate": 4.9446902409415386e-06,
"loss": 1.1862,
"step": 380
},
{
"epoch": 0.09796533534287867,
"grad_norm": 79.8613840478269,
"learning_rate": 4.940354156122624e-06,
"loss": 1.1443,
"step": 390
},
{
"epoch": 0.1004772670183371,
"grad_norm": 28.636783892356842,
"learning_rate": 4.935856505068999e-06,
"loss": 1.0705,
"step": 400
},
{
"epoch": 0.10298919869379553,
"grad_norm": 28.248493771801982,
"learning_rate": 4.931197585552418e-06,
"loss": 1.0342,
"step": 410
},
{
"epoch": 0.10550113036925396,
"grad_norm": 35.58893396345627,
"learning_rate": 4.92637770602159e-06,
"loss": 1.0984,
"step": 420
},
{
"epoch": 0.10801306204471238,
"grad_norm": 29.5108168082451,
"learning_rate": 4.92139718558175e-06,
"loss": 1.121,
"step": 430
},
{
"epoch": 0.11052499372017081,
"grad_norm": 17.994080670599825,
"learning_rate": 4.916256353973535e-06,
"loss": 1.1537,
"step": 440
},
{
"epoch": 0.11303692539562923,
"grad_norm": 32.844817441284924,
"learning_rate": 4.910955551551153e-06,
"loss": 1.1486,
"step": 450
},
{
"epoch": 0.11554885707108767,
"grad_norm": 30.914619782560816,
"learning_rate": 4.90549512925985e-06,
"loss": 1.0481,
"step": 460
},
{
"epoch": 0.11806078874654609,
"grad_norm": 25.55475057225604,
"learning_rate": 4.899875448612672e-06,
"loss": 1.0072,
"step": 470
},
{
"epoch": 0.12057272042200452,
"grad_norm": 201.00243368632337,
"learning_rate": 4.894096881666538e-06,
"loss": 1.1146,
"step": 480
},
{
"epoch": 0.12308465209746294,
"grad_norm": 51.017788501532685,
"learning_rate": 4.8881598109976e-06,
"loss": 0.9959,
"step": 490
},
{
"epoch": 0.12559658377292138,
"grad_norm": 30.076609186117565,
"learning_rate": 4.882064629675917e-06,
"loss": 1.0998,
"step": 500
},
{
"epoch": 0.1281085154483798,
"grad_norm": 25.914605989320286,
"learning_rate": 4.875811741239431e-06,
"loss": 0.9488,
"step": 510
},
{
"epoch": 0.13062044712383822,
"grad_norm": 174.16033120144348,
"learning_rate": 4.869401559667253e-06,
"loss": 1.0104,
"step": 520
},
{
"epoch": 0.13313237879929665,
"grad_norm": 24.53663814061994,
"learning_rate": 4.862834509352251e-06,
"loss": 0.9866,
"step": 530
},
{
"epoch": 0.1356443104747551,
"grad_norm": 106.69096534898281,
"learning_rate": 4.8561110250729546e-06,
"loss": 1.0145,
"step": 540
},
{
"epoch": 0.13815624215021352,
"grad_norm": 13.875443655979048,
"learning_rate": 4.849231551964771e-06,
"loss": 1.0714,
"step": 550
},
{
"epoch": 0.14066817382567195,
"grad_norm": 21.267058863118724,
"learning_rate": 4.842196545490512e-06,
"loss": 1.0451,
"step": 560
},
{
"epoch": 0.14318010550113036,
"grad_norm": 28.25986035068646,
"learning_rate": 4.8350064714102415e-06,
"loss": 0.9979,
"step": 570
},
{
"epoch": 0.1456920371765888,
"grad_norm": 18.771189948926747,
"learning_rate": 4.827661805750438e-06,
"loss": 1.0446,
"step": 580
},
{
"epoch": 0.14820396885204723,
"grad_norm": 33.442572448255376,
"learning_rate": 4.82016303477248e-06,
"loss": 1.0067,
"step": 590
},
{
"epoch": 0.15071590052750566,
"grad_norm": 18.758813933167882,
"learning_rate": 4.812510654940453e-06,
"loss": 0.9925,
"step": 600
},
{
"epoch": 0.15322783220296407,
"grad_norm": 18.44228949074116,
"learning_rate": 4.804705172888277e-06,
"loss": 0.9199,
"step": 610
},
{
"epoch": 0.1557397638784225,
"grad_norm": 18.907319724853284,
"learning_rate": 4.796747105386169e-06,
"loss": 0.9933,
"step": 620
},
{
"epoch": 0.15825169555388094,
"grad_norm": 24.999775435118767,
"learning_rate": 4.788636979306428e-06,
"loss": 0.946,
"step": 630
},
{
"epoch": 0.16076362722933937,
"grad_norm": 20.69959166678649,
"learning_rate": 4.78037533158855e-06,
"loss": 1.0123,
"step": 640
},
{
"epoch": 0.16327555890479778,
"grad_norm": 88.2316410192213,
"learning_rate": 4.771962709203684e-06,
"loss": 0.9311,
"step": 650
},
{
"epoch": 0.1657874905802562,
"grad_norm": 18.849075967996672,
"learning_rate": 4.763399669118414e-06,
"loss": 0.9532,
"step": 660
},
{
"epoch": 0.16829942225571465,
"grad_norm": 9.168139689519153,
"learning_rate": 4.754686778257891e-06,
"loss": 0.8347,
"step": 670
},
{
"epoch": 0.17081135393117308,
"grad_norm": 57.688726700615845,
"learning_rate": 4.745824613468293e-06,
"loss": 0.892,
"step": 680
},
{
"epoch": 0.1733232856066315,
"grad_norm": 34.38325408055489,
"learning_rate": 4.736813761478638e-06,
"loss": 0.941,
"step": 690
},
{
"epoch": 0.17583521728208992,
"grad_norm": 18.535830277017187,
"learning_rate": 4.727654818861937e-06,
"loss": 0.9297,
"step": 700
},
{
"epoch": 0.17834714895754836,
"grad_norm": 48.80919542270613,
"learning_rate": 4.718348391995698e-06,
"loss": 0.9512,
"step": 710
},
{
"epoch": 0.1808590806330068,
"grad_norm": 10.424093008575271,
"learning_rate": 4.7088950970217825e-06,
"loss": 0.9061,
"step": 720
},
{
"epoch": 0.1833710123084652,
"grad_norm": 13.175028312423365,
"learning_rate": 4.699295559805606e-06,
"loss": 0.9454,
"step": 730
},
{
"epoch": 0.18588294398392363,
"grad_norm": 65.05485953479175,
"learning_rate": 4.68955041589471e-06,
"loss": 0.9451,
"step": 740
},
{
"epoch": 0.18839487565938207,
"grad_norm": 25.29768837033324,
"learning_rate": 4.679660310476682e-06,
"loss": 0.831,
"step": 750
},
{
"epoch": 0.1909068073348405,
"grad_norm": 19.717151893244445,
"learning_rate": 4.669625898336439e-06,
"loss": 0.8221,
"step": 760
},
{
"epoch": 0.1934187390102989,
"grad_norm": 41.31445498318851,
"learning_rate": 4.659447843812876e-06,
"loss": 0.8756,
"step": 770
},
{
"epoch": 0.19593067068575734,
"grad_norm": 14.81608234505718,
"learning_rate": 4.649126820754887e-06,
"loss": 0.9191,
"step": 780
},
{
"epoch": 0.19844260236121578,
"grad_norm": 19.95005110698172,
"learning_rate": 4.638663512476748e-06,
"loss": 0.9372,
"step": 790
},
{
"epoch": 0.2009545340366742,
"grad_norm": 20.065073396172792,
"learning_rate": 4.628058611712879e-06,
"loss": 1.0179,
"step": 800
},
{
"epoch": 0.20346646571213264,
"grad_norm": 16.42772007372194,
"learning_rate": 4.617312820571981e-06,
"loss": 0.9572,
"step": 810
},
{
"epoch": 0.20597839738759105,
"grad_norm": 15.309541598862848,
"learning_rate": 4.606426850490551e-06,
"loss": 0.9536,
"step": 820
},
{
"epoch": 0.20849032906304948,
"grad_norm": 15.317369217882426,
"learning_rate": 4.595401422185783e-06,
"loss": 0.8942,
"step": 830
},
{
"epoch": 0.21100226073850792,
"grad_norm": 12.500948604145567,
"learning_rate": 4.584237265607847e-06,
"loss": 0.8162,
"step": 840
},
{
"epoch": 0.21351419241396635,
"grad_norm": 37.708490424859214,
"learning_rate": 4.5729351198915715e-06,
"loss": 0.9038,
"step": 850
},
{
"epoch": 0.21602612408942476,
"grad_norm": 34.796355808216596,
"learning_rate": 4.561495733307496e-06,
"loss": 0.9207,
"step": 860
},
{
"epoch": 0.2185380557648832,
"grad_norm": 26.298494542320878,
"learning_rate": 4.549919863212341e-06,
"loss": 0.9457,
"step": 870
},
{
"epoch": 0.22104998744034163,
"grad_norm": 16.186659690778107,
"learning_rate": 4.538208275998861e-06,
"loss": 1.0715,
"step": 880
},
{
"epoch": 0.22356191911580006,
"grad_norm": 24.662608067526172,
"learning_rate": 4.526361747045108e-06,
"loss": 0.9606,
"step": 890
},
{
"epoch": 0.22607385079125847,
"grad_norm": 40.835876651445666,
"learning_rate": 4.514381060663092e-06,
"loss": 0.933,
"step": 900
},
{
"epoch": 0.2285857824667169,
"grad_norm": 21.30586385592757,
"learning_rate": 4.50226701004686e-06,
"loss": 0.8754,
"step": 910
},
{
"epoch": 0.23109771414217534,
"grad_norm": 16.612023821577484,
"learning_rate": 4.49002039721998e-06,
"loss": 0.7735,
"step": 920
},
{
"epoch": 0.23360964581763377,
"grad_norm": 21.74905032903839,
"learning_rate": 4.477642032982439e-06,
"loss": 0.856,
"step": 930
},
{
"epoch": 0.23612157749309218,
"grad_norm": 59.36191090209393,
"learning_rate": 4.4651327368569695e-06,
"loss": 0.9044,
"step": 940
},
{
"epoch": 0.2386335091685506,
"grad_norm": 15.835736575328223,
"learning_rate": 4.452493337034785e-06,
"loss": 0.9162,
"step": 950
},
{
"epoch": 0.24114544084400905,
"grad_norm": 15.54957354753592,
"learning_rate": 4.439724670320755e-06,
"loss": 0.9259,
"step": 960
},
{
"epoch": 0.24365737251946748,
"grad_norm": 28.022932138978874,
"learning_rate": 4.426827582077999e-06,
"loss": 0.8854,
"step": 970
},
{
"epoch": 0.2461693041949259,
"grad_norm": 13.702537050157368,
"learning_rate": 4.413802926171921e-06,
"loss": 0.8087,
"step": 980
},
{
"epoch": 0.24868123587038432,
"grad_norm": 30.82271055352541,
"learning_rate": 4.400651564913676e-06,
"loss": 0.9174,
"step": 990
},
{
"epoch": 0.25119316754584275,
"grad_norm": 19.961883729903487,
"learning_rate": 4.387374369003083e-06,
"loss": 0.8261,
"step": 1000
},
{
"epoch": 0.2537050992213012,
"grad_norm": 13.910420016384712,
"learning_rate": 4.373972217470976e-06,
"loss": 0.8063,
"step": 1010
},
{
"epoch": 0.2562170308967596,
"grad_norm": 15.74547796378487,
"learning_rate": 4.36044599762101e-06,
"loss": 0.893,
"step": 1020
},
{
"epoch": 0.25872896257221806,
"grad_norm": 22.817146904726147,
"learning_rate": 4.346796604970913e-06,
"loss": 0.825,
"step": 1030
},
{
"epoch": 0.26124089424767644,
"grad_norm": 34.97383754680491,
"learning_rate": 4.333024943193198e-06,
"loss": 0.8637,
"step": 1040
},
{
"epoch": 0.26375282592313487,
"grad_norm": 14.442348394650033,
"learning_rate": 4.31913192405534e-06,
"loss": 0.8512,
"step": 1050
},
{
"epoch": 0.2662647575985933,
"grad_norm": 109.10193874564091,
"learning_rate": 4.305118467359402e-06,
"loss": 0.8808,
"step": 1060
},
{
"epoch": 0.26877668927405174,
"grad_norm": 27.206026550683518,
"learning_rate": 4.290985500881143e-06,
"loss": 0.8114,
"step": 1070
},
{
"epoch": 0.2712886209495102,
"grad_norm": 16.12708282763735,
"learning_rate": 4.276733960308599e-06,
"loss": 0.8569,
"step": 1080
},
{
"epoch": 0.2738005526249686,
"grad_norm": 84.431643774864,
"learning_rate": 4.262364789180123e-06,
"loss": 0.9139,
"step": 1090
},
{
"epoch": 0.27631248430042704,
"grad_norm": 14.872893259437928,
"learning_rate": 4.247878938821929e-06,
"loss": 0.7897,
"step": 1100
},
{
"epoch": 0.2788244159758855,
"grad_norm": 44.05664559088591,
"learning_rate": 4.2332773682851e-06,
"loss": 0.8999,
"step": 1110
},
{
"epoch": 0.2813363476513439,
"grad_norm": 14.79051628084432,
"learning_rate": 4.218561044282099e-06,
"loss": 0.8443,
"step": 1120
},
{
"epoch": 0.2838482793268023,
"grad_norm": 33.45294178219554,
"learning_rate": 4.20373094112276e-06,
"loss": 0.8567,
"step": 1130
},
{
"epoch": 0.2863602110022607,
"grad_norm": 46.4659449561201,
"learning_rate": 4.18878804064979e-06,
"loss": 0.9058,
"step": 1140
},
{
"epoch": 0.28887214267771916,
"grad_norm": 30.620320847801434,
"learning_rate": 4.173733332173759e-06,
"loss": 0.8482,
"step": 1150
},
{
"epoch": 0.2913840743531776,
"grad_norm": 32.099385967886,
"learning_rate": 4.158567812407609e-06,
"loss": 0.9224,
"step": 1160
},
{
"epoch": 0.293896006028636,
"grad_norm": 26.04095063399648,
"learning_rate": 4.1432924854006554e-06,
"loss": 0.7986,
"step": 1170
},
{
"epoch": 0.29640793770409446,
"grad_norm": 21.53191952152889,
"learning_rate": 4.127908362472121e-06,
"loss": 0.8285,
"step": 1180
},
{
"epoch": 0.2989198693795529,
"grad_norm": 14.804545412809087,
"learning_rate": 4.112416462144181e-06,
"loss": 0.8189,
"step": 1190
},
{
"epoch": 0.30143180105501133,
"grad_norm": 28.04536852758184,
"learning_rate": 4.096817810074521e-06,
"loss": 0.9102,
"step": 1200
},
{
"epoch": 0.3039437327304697,
"grad_norm": 16.38088643600367,
"learning_rate": 4.081113438988443e-06,
"loss": 0.8876,
"step": 1210
},
{
"epoch": 0.30645566440592814,
"grad_norm": 26.286121352348058,
"learning_rate": 4.065304388610491e-06,
"loss": 0.8078,
"step": 1220
},
{
"epoch": 0.3089675960813866,
"grad_norm": 49.451144669082176,
"learning_rate": 4.049391705595605e-06,
"loss": 0.874,
"step": 1230
},
{
"epoch": 0.311479527756845,
"grad_norm": 43.93733084504646,
"learning_rate": 4.033376443459842e-06,
"loss": 0.8575,
"step": 1240
},
{
"epoch": 0.31399145943230344,
"grad_norm": 33.5919976090378,
"learning_rate": 4.017259662510613e-06,
"loss": 0.8988,
"step": 1250
},
{
"epoch": 0.3165033911077619,
"grad_norm": 41.047784413490206,
"learning_rate": 4.00104242977649e-06,
"loss": 0.8776,
"step": 1260
},
{
"epoch": 0.3190153227832203,
"grad_norm": 21.485503202698204,
"learning_rate": 3.9847258189365664e-06,
"loss": 0.9317,
"step": 1270
},
{
"epoch": 0.32152725445867875,
"grad_norm": 18.745800000012018,
"learning_rate": 3.968310910249364e-06,
"loss": 0.8406,
"step": 1280
},
{
"epoch": 0.3240391861341371,
"grad_norm": 36.22488794899786,
"learning_rate": 3.951798790481318e-06,
"loss": 0.8217,
"step": 1290
},
{
"epoch": 0.32655111780959556,
"grad_norm": 16.051366022833278,
"learning_rate": 3.9351905528348285e-06,
"loss": 0.8654,
"step": 1300
},
{
"epoch": 0.329063049485054,
"grad_norm": 28.70539547850433,
"learning_rate": 3.918487296875881e-06,
"loss": 0.8681,
"step": 1310
},
{
"epoch": 0.3315749811605124,
"grad_norm": 14.816194644055994,
"learning_rate": 3.901690128461248e-06,
"loss": 0.8818,
"step": 1320
},
{
"epoch": 0.33408691283597086,
"grad_norm": 13.007885823046339,
"learning_rate": 3.8848001596652765e-06,
"loss": 0.8706,
"step": 1330
},
{
"epoch": 0.3365988445114293,
"grad_norm": 12.482101372342157,
"learning_rate": 3.867818508706261e-06,
"loss": 0.8012,
"step": 1340
},
{
"epoch": 0.33911077618688773,
"grad_norm": 26.169043780262097,
"learning_rate": 3.850746299872412e-06,
"loss": 0.9217,
"step": 1350
},
{
"epoch": 0.34162270786234616,
"grad_norm": 46.59460344018054,
"learning_rate": 3.833584663447418e-06,
"loss": 0.876,
"step": 1360
},
{
"epoch": 0.3441346395378046,
"grad_norm": 16.84714833096335,
"learning_rate": 3.816334735635618e-06,
"loss": 0.8291,
"step": 1370
},
{
"epoch": 0.346646571213263,
"grad_norm": 37.545682101952806,
"learning_rate": 3.7989976584867743e-06,
"loss": 0.7684,
"step": 1380
},
{
"epoch": 0.3491585028887214,
"grad_norm": 27.139012954460355,
"learning_rate": 3.7815745798204646e-06,
"loss": 0.761,
"step": 1390
},
{
"epoch": 0.35167043456417985,
"grad_norm": 32.23972648827027,
"learning_rate": 3.7640666531500868e-06,
"loss": 0.796,
"step": 1400
},
{
"epoch": 0.3541823662396383,
"grad_norm": 31.529769235569418,
"learning_rate": 3.7464750376064934e-06,
"loss": 0.8033,
"step": 1410
},
{
"epoch": 0.3566942979150967,
"grad_norm": 20.56253529260212,
"learning_rate": 3.7288008978612457e-06,
"loss": 0.8213,
"step": 1420
},
{
"epoch": 0.35920622959055515,
"grad_norm": 10.619817778124629,
"learning_rate": 3.711045404049507e-06,
"loss": 0.829,
"step": 1430
},
{
"epoch": 0.3617181612660136,
"grad_norm": 19.304755989056705,
"learning_rate": 3.693209731692573e-06,
"loss": 0.8201,
"step": 1440
},
{
"epoch": 0.364230092941472,
"grad_norm": 18.94828073608545,
"learning_rate": 3.675295061620047e-06,
"loss": 0.851,
"step": 1450
},
{
"epoch": 0.3667420246169304,
"grad_norm": 29.832008452352895,
"learning_rate": 3.6573025798916566e-06,
"loss": 0.7624,
"step": 1460
},
{
"epoch": 0.36925395629238883,
"grad_norm": 34.613004229596356,
"learning_rate": 3.6392334777187354e-06,
"loss": 0.7536,
"step": 1470
},
{
"epoch": 0.37176588796784726,
"grad_norm": 15.983386653537558,
"learning_rate": 3.621088951385353e-06,
"loss": 0.769,
"step": 1480
},
{
"epoch": 0.3742778196433057,
"grad_norm": 18.541258614645898,
"learning_rate": 3.6028702021691164e-06,
"loss": 0.8561,
"step": 1490
},
{
"epoch": 0.37678975131876413,
"grad_norm": 13.569243914961227,
"learning_rate": 3.5845784362616375e-06,
"loss": 0.8026,
"step": 1500
},
{
"epoch": 0.37930168299422257,
"grad_norm": 13.589477313526036,
"learning_rate": 3.566214864688674e-06,
"loss": 0.7856,
"step": 1510
},
{
"epoch": 0.381813614669681,
"grad_norm": 9.08955551951589,
"learning_rate": 3.5477807032299565e-06,
"loss": 0.8099,
"step": 1520
},
{
"epoch": 0.38432554634513943,
"grad_norm": 30.698836150904377,
"learning_rate": 3.5292771723386916e-06,
"loss": 0.7516,
"step": 1530
},
{
"epoch": 0.3868374780205978,
"grad_norm": 13.203900449832538,
"learning_rate": 3.5107054970607624e-06,
"loss": 0.8014,
"step": 1540
},
{
"epoch": 0.38934940969605625,
"grad_norm": 22.948546424383103,
"learning_rate": 3.4920669069536265e-06,
"loss": 0.8049,
"step": 1550
},
{
"epoch": 0.3918613413715147,
"grad_norm": 31.428429983393126,
"learning_rate": 3.4733626360049065e-06,
"loss": 0.8858,
"step": 1560
},
{
"epoch": 0.3943732730469731,
"grad_norm": 17.236743198133453,
"learning_rate": 3.4545939225506935e-06,
"loss": 0.8511,
"step": 1570
},
{
"epoch": 0.39688520472243155,
"grad_norm": 22.980911438080426,
"learning_rate": 3.4357620091935656e-06,
"loss": 0.7983,
"step": 1580
},
{
"epoch": 0.39939713639789,
"grad_norm": 11.441966137764789,
"learning_rate": 3.416868142720316e-06,
"loss": 0.7239,
"step": 1590
},
{
"epoch": 0.4019090680733484,
"grad_norm": 14.91445014623267,
"learning_rate": 3.39791357401941e-06,
"loss": 0.7582,
"step": 1600
},
{
"epoch": 0.40442099974880685,
"grad_norm": 41.79040186628522,
"learning_rate": 3.3788995579981694e-06,
"loss": 0.8557,
"step": 1610
},
{
"epoch": 0.4069329314242653,
"grad_norm": 14.432086044310571,
"learning_rate": 3.359827353499685e-06,
"loss": 0.7389,
"step": 1620
},
{
"epoch": 0.40944486309972367,
"grad_norm": 15.440731757695739,
"learning_rate": 3.340698223219484e-06,
"loss": 0.7378,
"step": 1630
},
{
"epoch": 0.4119567947751821,
"grad_norm": 9.55068956773867,
"learning_rate": 3.321513433621919e-06,
"loss": 0.7992,
"step": 1640
},
{
"epoch": 0.41446872645064053,
"grad_norm": 21.998448239749745,
"learning_rate": 3.3022742548563293e-06,
"loss": 0.7342,
"step": 1650
},
{
"epoch": 0.41698065812609897,
"grad_norm": 10.474200018473434,
"learning_rate": 3.282981960672948e-06,
"loss": 0.7554,
"step": 1660
},
{
"epoch": 0.4194925898015574,
"grad_norm": 20.24084901441791,
"learning_rate": 3.2636378283385684e-06,
"loss": 0.7177,
"step": 1670
},
{
"epoch": 0.42200452147701584,
"grad_norm": 15.252907769914906,
"learning_rate": 3.2442431385519853e-06,
"loss": 0.8415,
"step": 1680
},
{
"epoch": 0.42451645315247427,
"grad_norm": 31.941655660403704,
"learning_rate": 3.2247991753592018e-06,
"loss": 0.7944,
"step": 1690
},
{
"epoch": 0.4270283848279327,
"grad_norm": 12.168344877357876,
"learning_rate": 3.2053072260684206e-06,
"loss": 0.7722,
"step": 1700
},
{
"epoch": 0.4295403165033911,
"grad_norm": 24.94079803525497,
"learning_rate": 3.1857685811648143e-06,
"loss": 0.7928,
"step": 1710
},
{
"epoch": 0.4320522481788495,
"grad_norm": 15.210129562675117,
"learning_rate": 3.1661845342250874e-06,
"loss": 0.7647,
"step": 1720
},
{
"epoch": 0.43456417985430795,
"grad_norm": 29.737670350280883,
"learning_rate": 3.1465563818318356e-06,
"loss": 0.757,
"step": 1730
},
{
"epoch": 0.4370761115297664,
"grad_norm": 34.11288282754126,
"learning_rate": 3.1268854234877023e-06,
"loss": 0.7198,
"step": 1740
},
{
"epoch": 0.4395880432052248,
"grad_norm": 25.982392999301357,
"learning_rate": 3.107172961529343e-06,
"loss": 0.8304,
"step": 1750
},
{
"epoch": 0.44209997488068326,
"grad_norm": 19.060965746755112,
"learning_rate": 3.0874203010412057e-06,
"loss": 0.7495,
"step": 1760
},
{
"epoch": 0.4446119065561417,
"grad_norm": 9.450290682003281,
"learning_rate": 3.067628749769125e-06,
"loss": 0.6833,
"step": 1770
},
{
"epoch": 0.4471238382316001,
"grad_norm": 81.52825084551178,
"learning_rate": 3.047799618033739e-06,
"loss": 0.7265,
"step": 1780
},
{
"epoch": 0.4496357699070585,
"grad_norm": 112.89167490869183,
"learning_rate": 3.027934218643742e-06,
"loss": 0.7124,
"step": 1790
},
{
"epoch": 0.45214770158251694,
"grad_norm": 12.80679045659251,
"learning_rate": 3.008033866808967e-06,
"loss": 0.795,
"step": 1800
},
{
"epoch": 0.45465963325797537,
"grad_norm": 7.979181961331163,
"learning_rate": 2.9880998800533095e-06,
"loss": 0.7602,
"step": 1810
},
{
"epoch": 0.4571715649334338,
"grad_norm": 104.22429484767227,
"learning_rate": 2.968133578127501e-06,
"loss": 0.7924,
"step": 1820
},
{
"epoch": 0.45968349660889224,
"grad_norm": 25.072483756192565,
"learning_rate": 2.9481362829217335e-06,
"loss": 0.7684,
"step": 1830
},
{
"epoch": 0.4621954282843507,
"grad_norm": 28.08445712014526,
"learning_rate": 2.9281093183781406e-06,
"loss": 0.7569,
"step": 1840
},
{
"epoch": 0.4647073599598091,
"grad_norm": 17.77498223316829,
"learning_rate": 2.9080540104031487e-06,
"loss": 0.8381,
"step": 1850
},
{
"epoch": 0.46721929163526754,
"grad_norm": 32.223471519237506,
"learning_rate": 2.887971686779688e-06,
"loss": 0.7015,
"step": 1860
},
{
"epoch": 0.4697312233107259,
"grad_norm": 48.40756577090116,
"learning_rate": 2.8678636770792907e-06,
"loss": 0.7002,
"step": 1870
},
{
"epoch": 0.47224315498618435,
"grad_norm": 24.520924633617422,
"learning_rate": 2.847731312574061e-06,
"loss": 0.7943,
"step": 1880
},
{
"epoch": 0.4747550866616428,
"grad_norm": 22.88033862082146,
"learning_rate": 2.8275759261485407e-06,
"loss": 0.6797,
"step": 1890
},
{
"epoch": 0.4772670183371012,
"grad_norm": 21.004816938320573,
"learning_rate": 2.80739885221146e-06,
"loss": 0.8262,
"step": 1900
},
{
"epoch": 0.47977895001255966,
"grad_norm": 40.505775555323616,
"learning_rate": 2.7872014266073972e-06,
"loss": 0.856,
"step": 1910
},
{
"epoch": 0.4822908816880181,
"grad_norm": 15.196929474940113,
"learning_rate": 2.7669849865283303e-06,
"loss": 0.6539,
"step": 1920
},
{
"epoch": 0.4848028133634765,
"grad_norm": 21.136530968784285,
"learning_rate": 2.746750870425114e-06,
"loss": 0.6839,
"step": 1930
},
{
"epoch": 0.48731474503893496,
"grad_norm": 66.94533407380263,
"learning_rate": 2.7265004179188607e-06,
"loss": 0.7212,
"step": 1940
},
{
"epoch": 0.4898266767143934,
"grad_norm": 17.537525895857943,
"learning_rate": 2.7062349697122536e-06,
"loss": 0.7541,
"step": 1950
},
{
"epoch": 0.4923386083898518,
"grad_norm": 54.32967278575285,
"learning_rate": 2.68595586750078e-06,
"loss": 0.7619,
"step": 1960
},
{
"epoch": 0.4948505400653102,
"grad_norm": 21.831994544086985,
"learning_rate": 2.665664453883907e-06,
"loss": 0.7713,
"step": 1970
},
{
"epoch": 0.49736247174076864,
"grad_norm": 27.261105082644335,
"learning_rate": 2.6453620722761897e-06,
"loss": 0.7319,
"step": 1980
},
{
"epoch": 0.4998744034162271,
"grad_norm": 303.80383252156145,
"learning_rate": 2.6250500668183325e-06,
"loss": 0.7719,
"step": 1990
},
{
"epoch": 0.5023863350916855,
"grad_norm": 18.975055416205706,
"learning_rate": 2.6047297822881962e-06,
"loss": 0.8072,
"step": 2000
}
],
"logging_steps": 10,
"max_steps": 3981,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 250,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 2517310013177856.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}