FIRM-Edit-8B / trainer_state.json
zpy777's picture
Upload folder using huggingface_hub
0e7ffe6 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 2292,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.004363953742090334,
"grad_norm": 26.499397821507202,
"learning_rate": 3.91304347826087e-07,
"loss": 1.5146,
"step": 10
},
{
"epoch": 0.008727907484180668,
"grad_norm": 18.696221795032976,
"learning_rate": 8.260869565217392e-07,
"loss": 1.3975,
"step": 20
},
{
"epoch": 0.013091861226271002,
"grad_norm": 5.195746812383845,
"learning_rate": 1.2608695652173913e-06,
"loss": 1.0301,
"step": 30
},
{
"epoch": 0.017455814968361336,
"grad_norm": 2.9388986813240305,
"learning_rate": 1.6956521739130435e-06,
"loss": 0.9338,
"step": 40
},
{
"epoch": 0.02181976871045167,
"grad_norm": 2.157896980678217,
"learning_rate": 2.130434782608696e-06,
"loss": 0.834,
"step": 50
},
{
"epoch": 0.026183722452542003,
"grad_norm": 2.030356833343019,
"learning_rate": 2.5652173913043484e-06,
"loss": 0.7817,
"step": 60
},
{
"epoch": 0.030547676194632337,
"grad_norm": 2.010110703082707,
"learning_rate": 3e-06,
"loss": 0.7423,
"step": 70
},
{
"epoch": 0.03491162993672267,
"grad_norm": 2.036582614030934,
"learning_rate": 3.4347826086956526e-06,
"loss": 0.7193,
"step": 80
},
{
"epoch": 0.039275583678813,
"grad_norm": 1.9487163653762136,
"learning_rate": 3.869565217391304e-06,
"loss": 0.6965,
"step": 90
},
{
"epoch": 0.04363953742090334,
"grad_norm": 1.8884072181165537,
"learning_rate": 4.304347826086957e-06,
"loss": 0.6863,
"step": 100
},
{
"epoch": 0.04800349116299367,
"grad_norm": 1.9672176130957406,
"learning_rate": 4.739130434782609e-06,
"loss": 0.6761,
"step": 110
},
{
"epoch": 0.05236744490508401,
"grad_norm": 1.9175860223528627,
"learning_rate": 5.173913043478262e-06,
"loss": 0.669,
"step": 120
},
{
"epoch": 0.05673139864717434,
"grad_norm": 4.64170340673796,
"learning_rate": 5.608695652173914e-06,
"loss": 0.6553,
"step": 130
},
{
"epoch": 0.061095352389264675,
"grad_norm": 2.4419914657221766,
"learning_rate": 6.043478260869565e-06,
"loss": 0.6635,
"step": 140
},
{
"epoch": 0.065459306131355,
"grad_norm": 2.219811974469962,
"learning_rate": 6.478260869565218e-06,
"loss": 0.6417,
"step": 150
},
{
"epoch": 0.06982325987344534,
"grad_norm": 1.8731532270945315,
"learning_rate": 6.91304347826087e-06,
"loss": 0.6427,
"step": 160
},
{
"epoch": 0.07418721361553568,
"grad_norm": 1.9110914927729745,
"learning_rate": 7.347826086956522e-06,
"loss": 0.6442,
"step": 170
},
{
"epoch": 0.078551167357626,
"grad_norm": 1.9295608959806378,
"learning_rate": 7.782608695652174e-06,
"loss": 0.6369,
"step": 180
},
{
"epoch": 0.08291512109971634,
"grad_norm": 1.7914358748612738,
"learning_rate": 8.217391304347827e-06,
"loss": 0.639,
"step": 190
},
{
"epoch": 0.08727907484180668,
"grad_norm": 1.9446891106680222,
"learning_rate": 8.65217391304348e-06,
"loss": 0.6414,
"step": 200
},
{
"epoch": 0.09164302858389702,
"grad_norm": 2.1347110290826032,
"learning_rate": 9.086956521739132e-06,
"loss": 0.6392,
"step": 210
},
{
"epoch": 0.09600698232598734,
"grad_norm": 1.9030077411906519,
"learning_rate": 9.521739130434784e-06,
"loss": 0.6407,
"step": 220
},
{
"epoch": 0.10037093606807768,
"grad_norm": 1.9072699967620688,
"learning_rate": 9.956521739130436e-06,
"loss": 0.6296,
"step": 230
},
{
"epoch": 0.10473488981016801,
"grad_norm": 6.11411147753339,
"learning_rate": 9.999529953691501e-06,
"loss": 0.6364,
"step": 240
},
{
"epoch": 0.10909884355225835,
"grad_norm": 1.9274840356722631,
"learning_rate": 9.997905215717636e-06,
"loss": 0.6271,
"step": 250
},
{
"epoch": 0.11346279729434867,
"grad_norm": 1.6084651209610894,
"learning_rate": 9.995120360083208e-06,
"loss": 0.625,
"step": 260
},
{
"epoch": 0.11782675103643901,
"grad_norm": 1.9201552574955696,
"learning_rate": 9.991176033211077e-06,
"loss": 0.6271,
"step": 270
},
{
"epoch": 0.12219070477852935,
"grad_norm": 1.8527351849291473,
"learning_rate": 9.98607315066148e-06,
"loss": 0.62,
"step": 280
},
{
"epoch": 0.12655465852061967,
"grad_norm": 2.1728262084174137,
"learning_rate": 9.979812896919497e-06,
"loss": 0.6234,
"step": 290
},
{
"epoch": 0.13091861226271,
"grad_norm": 2.5508242501465848,
"learning_rate": 9.97239672512011e-06,
"loss": 0.6228,
"step": 300
},
{
"epoch": 0.13528256600480035,
"grad_norm": 1.8024605834099054,
"learning_rate": 9.963826356710913e-06,
"loss": 0.6189,
"step": 310
},
{
"epoch": 0.13964651974689068,
"grad_norm": 1.7899544274401067,
"learning_rate": 9.954103781052508e-06,
"loss": 0.622,
"step": 320
},
{
"epoch": 0.14401047348898102,
"grad_norm": 2.6634929838451162,
"learning_rate": 9.943231254956749e-06,
"loss": 0.6215,
"step": 330
},
{
"epoch": 0.14837442723107136,
"grad_norm": 1.6887392545009612,
"learning_rate": 9.93121130216288e-06,
"loss": 0.6095,
"step": 340
},
{
"epoch": 0.1527383809731617,
"grad_norm": 1.564527068427684,
"learning_rate": 9.918046712751731e-06,
"loss": 0.6073,
"step": 350
},
{
"epoch": 0.157102334715252,
"grad_norm": 1.7759578419290645,
"learning_rate": 9.903740542498071e-06,
"loss": 0.6095,
"step": 360
},
{
"epoch": 0.16146628845734234,
"grad_norm": 2.604040660939313,
"learning_rate": 9.888296112161308e-06,
"loss": 0.6039,
"step": 370
},
{
"epoch": 0.16583024219943268,
"grad_norm": 1.7552723705971771,
"learning_rate": 9.87171700671467e-06,
"loss": 0.6049,
"step": 380
},
{
"epoch": 0.17019419594152302,
"grad_norm": 2.786695194524141,
"learning_rate": 9.854007074513056e-06,
"loss": 0.6069,
"step": 390
},
{
"epoch": 0.17455814968361336,
"grad_norm": 1.6966428135982474,
"learning_rate": 9.835170426399757e-06,
"loss": 0.6067,
"step": 400
},
{
"epoch": 0.1789221034257037,
"grad_norm": 3.1900936362907073,
"learning_rate": 9.815211434752236e-06,
"loss": 0.6021,
"step": 410
},
{
"epoch": 0.18328605716779403,
"grad_norm": 2.0950790085022795,
"learning_rate": 9.79413473246722e-06,
"loss": 0.6064,
"step": 420
},
{
"epoch": 0.18765001090988437,
"grad_norm": 1.8573324559272892,
"learning_rate": 9.771945211885294e-06,
"loss": 0.5962,
"step": 430
},
{
"epoch": 0.19201396465197468,
"grad_norm": 1.7147053708189852,
"learning_rate": 9.748648023655302e-06,
"loss": 0.6005,
"step": 440
},
{
"epoch": 0.19637791839406502,
"grad_norm": 1.620349780247674,
"learning_rate": 9.724248575538764e-06,
"loss": 0.5922,
"step": 450
},
{
"epoch": 0.20074187213615535,
"grad_norm": 1.654417334838078,
"learning_rate": 9.698752531154623e-06,
"loss": 0.5977,
"step": 460
},
{
"epoch": 0.2051058258782457,
"grad_norm": 3.7007534847341335,
"learning_rate": 9.672165808664609e-06,
"loss": 0.5989,
"step": 470
},
{
"epoch": 0.20946977962033603,
"grad_norm": 1.8306160811104213,
"learning_rate": 9.6444945793995e-06,
"loss": 0.5914,
"step": 480
},
{
"epoch": 0.21383373336242636,
"grad_norm": 1.9518306293763308,
"learning_rate": 9.61574526642664e-06,
"loss": 0.5883,
"step": 490
},
{
"epoch": 0.2181976871045167,
"grad_norm": 1.4825723400886388,
"learning_rate": 9.585924543058997e-06,
"loss": 0.591,
"step": 500
},
{
"epoch": 0.2181976871045167,
"eval_loss": 0.5827435255050659,
"eval_runtime": 69.7858,
"eval_samples_per_second": 53.062,
"eval_steps_per_second": 3.324,
"step": 500
},
{
"epoch": 0.22256164084660704,
"grad_norm": 1.480013652739435,
"learning_rate": 9.555039331306164e-06,
"loss": 0.5893,
"step": 510
},
{
"epoch": 0.22692559458869735,
"grad_norm": 1.697064020550412,
"learning_rate": 9.523096800267602e-06,
"loss": 0.5967,
"step": 520
},
{
"epoch": 0.2312895483307877,
"grad_norm": 1.4577169028061336,
"learning_rate": 9.490104364468557e-06,
"loss": 0.5876,
"step": 530
},
{
"epoch": 0.23565350207287802,
"grad_norm": 1.407863250608549,
"learning_rate": 9.456069682138981e-06,
"loss": 0.5879,
"step": 540
},
{
"epoch": 0.24001745581496836,
"grad_norm": 1.5047099371255703,
"learning_rate": 9.421000653435908e-06,
"loss": 0.5898,
"step": 550
},
{
"epoch": 0.2443814095570587,
"grad_norm": 1.5343210297318903,
"learning_rate": 9.38490541860966e-06,
"loss": 0.5938,
"step": 560
},
{
"epoch": 0.24874536329914904,
"grad_norm": 1.4956667878525625,
"learning_rate": 9.347792356114324e-06,
"loss": 0.5893,
"step": 570
},
{
"epoch": 0.25310931704123935,
"grad_norm": 1.5101062644216703,
"learning_rate": 9.30967008066294e-06,
"loss": 0.5976,
"step": 580
},
{
"epoch": 0.2574732707833297,
"grad_norm": 1.5236342080851415,
"learning_rate": 9.27054744122785e-06,
"loss": 0.592,
"step": 590
},
{
"epoch": 0.26183722452542,
"grad_norm": 1.6140869174258692,
"learning_rate": 9.230433518986664e-06,
"loss": 0.5888,
"step": 600
},
{
"epoch": 0.2662011782675104,
"grad_norm": 1.419486174653226,
"learning_rate": 9.189337625214324e-06,
"loss": 0.5842,
"step": 610
},
{
"epoch": 0.2705651320096007,
"grad_norm": 1.5576946265406408,
"learning_rate": 9.147269299121782e-06,
"loss": 0.582,
"step": 620
},
{
"epoch": 0.274929085751691,
"grad_norm": 1.4457336642256946,
"learning_rate": 9.104238305641731e-06,
"loss": 0.579,
"step": 630
},
{
"epoch": 0.27929303949378137,
"grad_norm": 1.5422960552682006,
"learning_rate": 9.060254633161973e-06,
"loss": 0.5889,
"step": 640
},
{
"epoch": 0.2836569932358717,
"grad_norm": 1.5024597086356501,
"learning_rate": 9.015328491206901e-06,
"loss": 0.5793,
"step": 650
},
{
"epoch": 0.28802094697796204,
"grad_norm": 1.3686070974882203,
"learning_rate": 8.969470308067662e-06,
"loss": 0.5807,
"step": 660
},
{
"epoch": 0.29238490072005235,
"grad_norm": 2.6998427178862894,
"learning_rate": 8.922690728381527e-06,
"loss": 0.5748,
"step": 670
},
{
"epoch": 0.2967488544621427,
"grad_norm": 2.065649261696864,
"learning_rate": 8.875000610661052e-06,
"loss": 0.5688,
"step": 680
},
{
"epoch": 0.30111280820423303,
"grad_norm": 1.5958849520041398,
"learning_rate": 8.826411024773595e-06,
"loss": 0.5727,
"step": 690
},
{
"epoch": 0.3054767619463234,
"grad_norm": 1.4305834347213016,
"learning_rate": 8.776933249371769e-06,
"loss": 0.5621,
"step": 700
},
{
"epoch": 0.3098407156884137,
"grad_norm": 1.4826756109906936,
"learning_rate": 8.726578769275426e-06,
"loss": 0.57,
"step": 710
},
{
"epoch": 0.314204669430504,
"grad_norm": 1.4919661778521895,
"learning_rate": 8.675359272805796e-06,
"loss": 0.5733,
"step": 720
},
{
"epoch": 0.3185686231725944,
"grad_norm": 1.539486119661036,
"learning_rate": 8.62328664907239e-06,
"loss": 0.5617,
"step": 730
},
{
"epoch": 0.3229325769146847,
"grad_norm": 2.412337158368947,
"learning_rate": 8.570372985213283e-06,
"loss": 0.5748,
"step": 740
},
{
"epoch": 0.32729653065677505,
"grad_norm": 1.4949607784946515,
"learning_rate": 8.516630563589436e-06,
"loss": 0.5765,
"step": 750
},
{
"epoch": 0.33166048439886536,
"grad_norm": 2.1389988555015527,
"learning_rate": 8.462071858933717e-06,
"loss": 0.5717,
"step": 760
},
{
"epoch": 0.33602443814095573,
"grad_norm": 1.7992243083658257,
"learning_rate": 8.406709535455242e-06,
"loss": 0.5728,
"step": 770
},
{
"epoch": 0.34038839188304604,
"grad_norm": 1.7459966293740241,
"learning_rate": 8.35055644389976e-06,
"loss": 0.5621,
"step": 780
},
{
"epoch": 0.34475234562513635,
"grad_norm": 1.6584257041747206,
"learning_rate": 8.293625618566723e-06,
"loss": 0.5676,
"step": 790
},
{
"epoch": 0.3491162993672267,
"grad_norm": 2.234742145469745,
"learning_rate": 8.23593027428375e-06,
"loss": 0.5657,
"step": 800
},
{
"epoch": 0.353480253109317,
"grad_norm": 1.4160485895062545,
"learning_rate": 8.177483803339203e-06,
"loss": 0.5739,
"step": 810
},
{
"epoch": 0.3578442068514074,
"grad_norm": 1.6236310717790317,
"learning_rate": 8.118299772373546e-06,
"loss": 0.5708,
"step": 820
},
{
"epoch": 0.3622081605934977,
"grad_norm": 1.4409935142518557,
"learning_rate": 8.05839191923025e-06,
"loss": 0.5652,
"step": 830
},
{
"epoch": 0.36657211433558806,
"grad_norm": 1.7422827738179059,
"learning_rate": 7.997774149766962e-06,
"loss": 0.5665,
"step": 840
},
{
"epoch": 0.37093606807767837,
"grad_norm": 1.4654585573940748,
"learning_rate": 7.936460534627648e-06,
"loss": 0.5592,
"step": 850
},
{
"epoch": 0.37530002181976874,
"grad_norm": 1.5684193622827431,
"learning_rate": 7.874465305976527e-06,
"loss": 0.558,
"step": 860
},
{
"epoch": 0.37966397556185905,
"grad_norm": 1.8405640707393334,
"learning_rate": 7.81180285419447e-06,
"loss": 0.5653,
"step": 870
},
{
"epoch": 0.38402792930394936,
"grad_norm": 1.475228530845922,
"learning_rate": 7.748487724538707e-06,
"loss": 0.5587,
"step": 880
},
{
"epoch": 0.3883918830460397,
"grad_norm": 1.331005791046321,
"learning_rate": 7.684534613766566e-06,
"loss": 0.574,
"step": 890
},
{
"epoch": 0.39275583678813003,
"grad_norm": 1.3582718830868108,
"learning_rate": 7.619958366724043e-06,
"loss": 0.5657,
"step": 900
},
{
"epoch": 0.3971197905302204,
"grad_norm": 1.3639931742020133,
"learning_rate": 7.554773972900024e-06,
"loss": 0.5642,
"step": 910
},
{
"epoch": 0.4014837442723107,
"grad_norm": 1.4421529537691287,
"learning_rate": 7.488996562946886e-06,
"loss": 0.5585,
"step": 920
},
{
"epoch": 0.40584769801440107,
"grad_norm": 1.4453139064733957,
"learning_rate": 7.422641405168386e-06,
"loss": 0.5611,
"step": 930
},
{
"epoch": 0.4102116517564914,
"grad_norm": 1.41874403862543,
"learning_rate": 7.355723901975546e-06,
"loss": 0.5556,
"step": 940
},
{
"epoch": 0.4145756054985817,
"grad_norm": 1.569355186279227,
"learning_rate": 7.288259586311455e-06,
"loss": 0.5638,
"step": 950
},
{
"epoch": 0.41893955924067205,
"grad_norm": 1.340010425244894,
"learning_rate": 7.22026411804573e-06,
"loss": 0.554,
"step": 960
},
{
"epoch": 0.42330351298276236,
"grad_norm": 1.4102271417835663,
"learning_rate": 7.151753280339554e-06,
"loss": 0.5574,
"step": 970
},
{
"epoch": 0.42766746672485273,
"grad_norm": 1.4168840313026727,
"learning_rate": 7.082742975982057e-06,
"loss": 0.5613,
"step": 980
},
{
"epoch": 0.43203142046694304,
"grad_norm": 1.4932003760222727,
"learning_rate": 7.0132492236989694e-06,
"loss": 0.5549,
"step": 990
},
{
"epoch": 0.4363953742090334,
"grad_norm": 1.3727246781695526,
"learning_rate": 6.94328815443432e-06,
"loss": 0.5605,
"step": 1000
},
{
"epoch": 0.4363953742090334,
"eval_loss": 0.545971155166626,
"eval_runtime": 68.1299,
"eval_samples_per_second": 54.352,
"eval_steps_per_second": 3.405,
"step": 1000
},
{
"epoch": 0.4407593279511237,
"grad_norm": 1.5610600215808323,
"learning_rate": 6.872876007606127e-06,
"loss": 0.5505,
"step": 1010
},
{
"epoch": 0.4451232816932141,
"grad_norm": 1.2665441451258144,
"learning_rate": 6.802029127336884e-06,
"loss": 0.5511,
"step": 1020
},
{
"epoch": 0.4494872354353044,
"grad_norm": 1.4767565256072326,
"learning_rate": 6.73076395865975e-06,
"loss": 0.5479,
"step": 1030
},
{
"epoch": 0.4538511891773947,
"grad_norm": 1.3450898726192824,
"learning_rate": 6.6590970437013135e-06,
"loss": 0.5549,
"step": 1040
},
{
"epoch": 0.45821514291948506,
"grad_norm": 1.3691285786320522,
"learning_rate": 6.587045017841828e-06,
"loss": 0.5566,
"step": 1050
},
{
"epoch": 0.4625790966615754,
"grad_norm": 1.407701866024824,
"learning_rate": 6.514624605853785e-06,
"loss": 0.5519,
"step": 1060
},
{
"epoch": 0.46694305040366574,
"grad_norm": 1.7016011593861633,
"learning_rate": 6.441852618019757e-06,
"loss": 0.5453,
"step": 1070
},
{
"epoch": 0.47130700414575605,
"grad_norm": 2.0081488788972495,
"learning_rate": 6.368745946230371e-06,
"loss": 0.5509,
"step": 1080
},
{
"epoch": 0.4756709578878464,
"grad_norm": 1.4313272433739415,
"learning_rate": 6.295321560063358e-06,
"loss": 0.5558,
"step": 1090
},
{
"epoch": 0.4800349116299367,
"grad_norm": 1.3908306510503996,
"learning_rate": 6.221596502844558e-06,
"loss": 0.5496,
"step": 1100
},
{
"epoch": 0.48439886537202703,
"grad_norm": 1.4083635646126667,
"learning_rate": 6.147587887691812e-06,
"loss": 0.5447,
"step": 1110
},
{
"epoch": 0.4887628191141174,
"grad_norm": 17.48008864471473,
"learning_rate": 6.073312893542644e-06,
"loss": 0.5507,
"step": 1120
},
{
"epoch": 0.4931267728562077,
"grad_norm": 1.7242442516057832,
"learning_rate": 5.998788761166689e-06,
"loss": 0.5497,
"step": 1130
},
{
"epoch": 0.49749072659829807,
"grad_norm": 1.6226845226734892,
"learning_rate": 5.9240327891637296e-06,
"loss": 0.5521,
"step": 1140
},
{
"epoch": 0.5018546803403884,
"grad_norm": 1.6604002116905412,
"learning_rate": 5.849062329948353e-06,
"loss": 0.5451,
"step": 1150
},
{
"epoch": 0.5062186340824787,
"grad_norm": 1.5873724924653319,
"learning_rate": 5.773894785722082e-06,
"loss": 0.547,
"step": 1160
},
{
"epoch": 0.5105825878245691,
"grad_norm": 1.4609962799328036,
"learning_rate": 5.698547604433963e-06,
"loss": 0.551,
"step": 1170
},
{
"epoch": 0.5149465415666594,
"grad_norm": 2.35781566677412,
"learning_rate": 5.623038275730543e-06,
"loss": 0.5482,
"step": 1180
},
{
"epoch": 0.5193104953087497,
"grad_norm": 1.7956175108962174,
"learning_rate": 5.547384326896152e-06,
"loss": 0.5378,
"step": 1190
},
{
"epoch": 0.52367444905084,
"grad_norm": 1.3754392835784166,
"learning_rate": 5.4716033187844565e-06,
"loss": 0.5289,
"step": 1200
},
{
"epoch": 0.5280384027929304,
"grad_norm": 1.351125298582442,
"learning_rate": 5.3957128417422196e-06,
"loss": 0.5375,
"step": 1210
},
{
"epoch": 0.5324023565350208,
"grad_norm": 1.2885359407090353,
"learning_rate": 5.319730511526225e-06,
"loss": 0.5484,
"step": 1220
},
{
"epoch": 0.5367663102771111,
"grad_norm": 1.4931791483959027,
"learning_rate": 5.243673965214276e-06,
"loss": 0.5352,
"step": 1230
},
{
"epoch": 0.5411302640192014,
"grad_norm": 1.3301082529926556,
"learning_rate": 5.167560857111286e-06,
"loss": 0.5407,
"step": 1240
},
{
"epoch": 0.5454942177612917,
"grad_norm": 1.3023885557832229,
"learning_rate": 5.091408854651327e-06,
"loss": 0.5354,
"step": 1250
},
{
"epoch": 0.549858171503382,
"grad_norm": 1.8840653667484866,
"learning_rate": 5.0152356342966624e-06,
"loss": 0.5425,
"step": 1260
},
{
"epoch": 0.5542221252454724,
"grad_norm": 1.3585936189959857,
"learning_rate": 4.939058877434672e-06,
"loss": 0.5386,
"step": 1270
},
{
"epoch": 0.5585860789875627,
"grad_norm": 1.4612137856640806,
"learning_rate": 4.862896266273627e-06,
"loss": 0.5512,
"step": 1280
},
{
"epoch": 0.562950032729653,
"grad_norm": 1.3030406905808916,
"learning_rate": 4.786765479738293e-06,
"loss": 0.5422,
"step": 1290
},
{
"epoch": 0.5673139864717434,
"grad_norm": 1.3169625647849659,
"learning_rate": 4.7106841893662755e-06,
"loss": 0.5377,
"step": 1300
},
{
"epoch": 0.5716779402138338,
"grad_norm": 1.4365133088202295,
"learning_rate": 4.634670055206092e-06,
"loss": 0.5429,
"step": 1310
},
{
"epoch": 0.5760418939559241,
"grad_norm": 1.400666928678306,
"learning_rate": 4.5587407217179094e-06,
"loss": 0.5441,
"step": 1320
},
{
"epoch": 0.5804058476980144,
"grad_norm": 1.4020862611201026,
"learning_rate": 4.482913813677896e-06,
"loss": 0.5372,
"step": 1330
},
{
"epoch": 0.5847698014401047,
"grad_norm": 1.3899074056942011,
"learning_rate": 4.407206932087143e-06,
"loss": 0.5431,
"step": 1340
},
{
"epoch": 0.589133755182195,
"grad_norm": 1.2780180293099543,
"learning_rate": 4.331637650086111e-06,
"loss": 0.5354,
"step": 1350
},
{
"epoch": 0.5934977089242854,
"grad_norm": 1.3047906646065353,
"learning_rate": 4.256223508875536e-06,
"loss": 0.536,
"step": 1360
},
{
"epoch": 0.5978616626663757,
"grad_norm": 2.2450550137169025,
"learning_rate": 4.180982013644749e-06,
"loss": 0.5345,
"step": 1370
},
{
"epoch": 0.6022256164084661,
"grad_norm": 1.2604878888770283,
"learning_rate": 4.105930629508369e-06,
"loss": 0.5349,
"step": 1380
},
{
"epoch": 0.6065895701505564,
"grad_norm": 1.4408134764916174,
"learning_rate": 4.0310867774522724e-06,
"loss": 0.5338,
"step": 1390
},
{
"epoch": 0.6109535238926468,
"grad_norm": 1.194095025493141,
"learning_rate": 3.95646783028983e-06,
"loss": 0.5332,
"step": 1400
},
{
"epoch": 0.6153174776347371,
"grad_norm": 1.3682237212576729,
"learning_rate": 3.882091108629311e-06,
"loss": 0.5305,
"step": 1410
},
{
"epoch": 0.6196814313768274,
"grad_norm": 1.3808616625009607,
"learning_rate": 3.807973876853414e-06,
"loss": 0.5333,
"step": 1420
},
{
"epoch": 0.6240453851189177,
"grad_norm": 1.3166679869381683,
"learning_rate": 3.734133339111844e-06,
"loss": 0.5305,
"step": 1430
},
{
"epoch": 0.628409338861008,
"grad_norm": 2.075662539665934,
"learning_rate": 3.660586635327869e-06,
"loss": 0.5386,
"step": 1440
},
{
"epoch": 0.6327732926030984,
"grad_norm": 1.3904648380650928,
"learning_rate": 3.587350837219788e-06,
"loss": 0.5396,
"step": 1450
},
{
"epoch": 0.6371372463451888,
"grad_norm": 1.492731925374217,
"learning_rate": 3.5144429443382356e-06,
"loss": 0.5366,
"step": 1460
},
{
"epoch": 0.6415012000872791,
"grad_norm": 1.2451404313444014,
"learning_rate": 3.4418798801202256e-06,
"loss": 0.5255,
"step": 1470
},
{
"epoch": 0.6458651538293694,
"grad_norm": 1.5668762418599276,
"learning_rate": 3.3696784879608747e-06,
"loss": 0.5307,
"step": 1480
},
{
"epoch": 0.6502291075714598,
"grad_norm": 1.4372909935764755,
"learning_rate": 3.2978555273037006e-06,
"loss": 0.5269,
"step": 1490
},
{
"epoch": 0.6545930613135501,
"grad_norm": 1.3776181734494666,
"learning_rate": 3.2264276697504026e-06,
"loss": 0.5252,
"step": 1500
},
{
"epoch": 0.6545930613135501,
"eval_loss": 0.5199297666549683,
"eval_runtime": 69.8992,
"eval_samples_per_second": 52.976,
"eval_steps_per_second": 3.319,
"step": 1500
},
{
"epoch": 0.6589570150556404,
"grad_norm": 1.4365144824102365,
"learning_rate": 3.1554114951910387e-06,
"loss": 0.5312,
"step": 1510
},
{
"epoch": 0.6633209687977307,
"grad_norm": 1.2422304913068782,
"learning_rate": 3.0848234879554916e-06,
"loss": 0.5283,
"step": 1520
},
{
"epoch": 0.667684922539821,
"grad_norm": 1.2495828228531431,
"learning_rate": 3.0146800329871107e-06,
"loss": 0.5359,
"step": 1530
},
{
"epoch": 0.6720488762819115,
"grad_norm": 1.3009383931934748,
"learning_rate": 2.944997412039422e-06,
"loss": 0.5286,
"step": 1540
},
{
"epoch": 0.6764128300240018,
"grad_norm": 1.3215444464354889,
"learning_rate": 2.8757917998968042e-06,
"loss": 0.5345,
"step": 1550
},
{
"epoch": 0.6807767837660921,
"grad_norm": 1.199193600939233,
"learning_rate": 2.807079260619976e-06,
"loss": 0.5226,
"step": 1560
},
{
"epoch": 0.6851407375081824,
"grad_norm": 1.3916367911407932,
"learning_rate": 2.7388757438171953e-06,
"loss": 0.5294,
"step": 1570
},
{
"epoch": 0.6895046912502727,
"grad_norm": 1.2821114411807195,
"learning_rate": 2.6711970809420327e-06,
"loss": 0.5243,
"step": 1580
},
{
"epoch": 0.6938686449923631,
"grad_norm": 1.2276903910497599,
"learning_rate": 2.6040589816185534e-06,
"loss": 0.5227,
"step": 1590
},
{
"epoch": 0.6982325987344534,
"grad_norm": 1.2534508386414063,
"learning_rate": 2.5374770299947837e-06,
"loss": 0.5336,
"step": 1600
},
{
"epoch": 0.7025965524765437,
"grad_norm": 1.2952567981586396,
"learning_rate": 2.471466681125316e-06,
"loss": 0.5242,
"step": 1610
},
{
"epoch": 0.706960506218634,
"grad_norm": 1.2491075248862247,
"learning_rate": 2.4060432573838686e-06,
"loss": 0.5229,
"step": 1620
},
{
"epoch": 0.7113244599607245,
"grad_norm": 1.3070997549733596,
"learning_rate": 2.3412219449066316e-06,
"loss": 0.5238,
"step": 1630
},
{
"epoch": 0.7156884137028148,
"grad_norm": 1.2500083040545036,
"learning_rate": 2.2770177900672658e-06,
"loss": 0.527,
"step": 1640
},
{
"epoch": 0.7200523674449051,
"grad_norm": 1.2464501526769323,
"learning_rate": 2.213445695984318e-06,
"loss": 0.5208,
"step": 1650
},
{
"epoch": 0.7244163211869954,
"grad_norm": 1.2752274285884029,
"learning_rate": 2.150520419061896e-06,
"loss": 0.5268,
"step": 1660
},
{
"epoch": 0.7287802749290857,
"grad_norm": 1.243680862493251,
"learning_rate": 2.0882565655644054e-06,
"loss": 0.5233,
"step": 1670
},
{
"epoch": 0.7331442286711761,
"grad_norm": 1.6747464378055472,
"learning_rate": 2.026668588226133e-06,
"loss": 0.5254,
"step": 1680
},
{
"epoch": 0.7375081824132664,
"grad_norm": 1.2620103056544976,
"learning_rate": 1.965770782896455e-06,
"loss": 0.5224,
"step": 1690
},
{
"epoch": 0.7418721361553567,
"grad_norm": 1.3278419822937977,
"learning_rate": 1.9055772852214916e-06,
"loss": 0.515,
"step": 1700
},
{
"epoch": 0.746236089897447,
"grad_norm": 1.6987330727891572,
"learning_rate": 1.8461020673629172e-06,
"loss": 0.5144,
"step": 1710
},
{
"epoch": 0.7506000436395375,
"grad_norm": 1.3497988296955354,
"learning_rate": 1.7873589347547321e-06,
"loss": 0.5173,
"step": 1720
},
{
"epoch": 0.7549639973816278,
"grad_norm": 1.6208371219517643,
"learning_rate": 1.729361522898737e-06,
"loss": 0.5228,
"step": 1730
},
{
"epoch": 0.7593279511237181,
"grad_norm": 1.308791373132814,
"learning_rate": 1.6721232941994526e-06,
"loss": 0.5282,
"step": 1740
},
{
"epoch": 0.7636919048658084,
"grad_norm": 1.2618118893101176,
"learning_rate": 1.6156575348392122e-06,
"loss": 0.5208,
"step": 1750
},
{
"epoch": 0.7680558586078987,
"grad_norm": 1.3104736125525216,
"learning_rate": 1.559977351694158e-06,
"loss": 0.5208,
"step": 1760
},
{
"epoch": 0.7724198123499891,
"grad_norm": 1.2487367116120625,
"learning_rate": 1.5050956692918739e-06,
"loss": 0.5186,
"step": 1770
},
{
"epoch": 0.7767837660920794,
"grad_norm": 1.156634335046638,
"learning_rate": 1.4510252268113263e-06,
"loss": 0.5192,
"step": 1780
},
{
"epoch": 0.7811477198341698,
"grad_norm": 1.3021570425828364,
"learning_rate": 1.3977785751258345e-06,
"loss": 0.5211,
"step": 1790
},
{
"epoch": 0.7855116735762601,
"grad_norm": 1.2504588787454711,
"learning_rate": 1.34536807388976e-06,
"loss": 0.5159,
"step": 1800
},
{
"epoch": 0.7898756273183504,
"grad_norm": 1.991690391624192,
"learning_rate": 1.2938058886695643e-06,
"loss": 0.5169,
"step": 1810
},
{
"epoch": 0.7942395810604408,
"grad_norm": 1.4484577791210265,
"learning_rate": 1.2431039881199374e-06,
"loss": 0.5246,
"step": 1820
},
{
"epoch": 0.7986035348025311,
"grad_norm": 1.2031874504717428,
"learning_rate": 1.1932741412056187e-06,
"loss": 0.5238,
"step": 1830
},
{
"epoch": 0.8029674885446214,
"grad_norm": 1.4251143516882596,
"learning_rate": 1.1443279144695746e-06,
"loss": 0.5202,
"step": 1840
},
{
"epoch": 0.8073314422867117,
"grad_norm": 1.428433930967254,
"learning_rate": 1.0962766693481686e-06,
"loss": 0.5178,
"step": 1850
},
{
"epoch": 0.8116953960288021,
"grad_norm": 1.3957444052098604,
"learning_rate": 1.049131559533933e-06,
"loss": 0.5142,
"step": 1860
},
{
"epoch": 0.8160593497708925,
"grad_norm": 1.7513113973659105,
"learning_rate": 1.002903528386564e-06,
"loss": 0.5188,
"step": 1870
},
{
"epoch": 0.8204233035129828,
"grad_norm": 1.3254734615581885,
"learning_rate": 9.576033063927398e-07,
"loss": 0.5207,
"step": 1880
},
{
"epoch": 0.8247872572550731,
"grad_norm": 1.3416379717629452,
"learning_rate": 9.132414086753578e-07,
"loss": 0.5134,
"step": 1890
},
{
"epoch": 0.8291512109971634,
"grad_norm": 1.2636084433299684,
"learning_rate": 8.698281325527502e-07,
"loss": 0.5144,
"step": 1900
},
{
"epoch": 0.8335151647392538,
"grad_norm": 1.718038484652991,
"learning_rate": 8.273735551484613e-07,
"loss": 0.5158,
"step": 1910
},
{
"epoch": 0.8378791184813441,
"grad_norm": 1.365124227466459,
"learning_rate": 7.858875310521447e-07,
"loss": 0.5116,
"step": 1920
},
{
"epoch": 0.8422430722234344,
"grad_norm": 1.2857842722903012,
"learning_rate": 7.45379690032102e-07,
"loss": 0.5213,
"step": 1930
},
{
"epoch": 0.8466070259655247,
"grad_norm": 1.8949074875975938,
"learning_rate": 7.058594348000142e-07,
"loss": 0.5126,
"step": 1940
},
{
"epoch": 0.8509709797076151,
"grad_norm": 2.0508297274530425,
"learning_rate": 6.673359388283796e-07,
"loss": 0.5118,
"step": 1950
},
{
"epoch": 0.8553349334497055,
"grad_norm": 1.2982410941631513,
"learning_rate": 6.29818144221161e-07,
"loss": 0.5256,
"step": 1960
},
{
"epoch": 0.8596988871917958,
"grad_norm": 1.3881080528727965,
"learning_rate": 5.933147596381295e-07,
"loss": 0.5178,
"step": 1970
},
{
"epoch": 0.8640628409338861,
"grad_norm": 1.2689567204311927,
"learning_rate": 5.578342582734153e-07,
"loss": 0.5106,
"step": 1980
},
{
"epoch": 0.8684267946759764,
"grad_norm": 1.3514399003622013,
"learning_rate": 5.233848758886945e-07,
"loss": 0.515,
"step": 1990
},
{
"epoch": 0.8727907484180668,
"grad_norm": 1.8878476820927994,
"learning_rate": 4.899746089015006e-07,
"loss": 0.5075,
"step": 2000
},
{
"epoch": 0.8727907484180668,
"eval_loss": 0.5055009722709656,
"eval_runtime": 68.9834,
"eval_samples_per_second": 53.68,
"eval_steps_per_second": 3.363,
"step": 2000
},
{
"epoch": 0.8771547021601571,
"grad_norm": 1.1864343347355735,
"learning_rate": 4.57611212529091e-07,
"loss": 0.5118,
"step": 2010
},
{
"epoch": 0.8815186559022474,
"grad_norm": 1.3391728742114728,
"learning_rate": 4.263021989882965e-07,
"loss": 0.5209,
"step": 2020
},
{
"epoch": 0.8858826096443377,
"grad_norm": 1.265995062918622,
"learning_rate": 3.9605483575178517e-07,
"loss": 0.5172,
"step": 2030
},
{
"epoch": 0.8902465633864282,
"grad_norm": 1.3612996127522512,
"learning_rate": 3.6687614386112403e-07,
"loss": 0.51,
"step": 2040
},
{
"epoch": 0.8946105171285185,
"grad_norm": 1.1822823920287409,
"learning_rate": 3.38772896297056e-07,
"loss": 0.512,
"step": 2050
},
{
"epoch": 0.8989744708706088,
"grad_norm": 1.4239619500028629,
"learning_rate": 3.117516164073459e-07,
"loss": 0.519,
"step": 2060
},
{
"epoch": 0.9033384246126991,
"grad_norm": 1.3503881714270254,
"learning_rate": 2.858185763925764e-07,
"loss": 0.5104,
"step": 2070
},
{
"epoch": 0.9077023783547894,
"grad_norm": 1.3564880256851914,
"learning_rate": 2.6097979585024193e-07,
"loss": 0.5136,
"step": 2080
},
{
"epoch": 0.9120663320968798,
"grad_norm": 1.2482713788944904,
"learning_rate": 2.3724104037747553e-07,
"loss": 0.5085,
"step": 2090
},
{
"epoch": 0.9164302858389701,
"grad_norm": 1.4746577210230283,
"learning_rate": 2.146078202327284e-07,
"loss": 0.5159,
"step": 2100
},
{
"epoch": 0.9207942395810604,
"grad_norm": 2.19277205698307,
"learning_rate": 1.9308538905673135e-07,
"loss": 0.5118,
"step": 2110
},
{
"epoch": 0.9251581933231507,
"grad_norm": 1.3172149163534492,
"learning_rate": 1.7267874265301187e-07,
"loss": 0.5139,
"step": 2120
},
{
"epoch": 0.9295221470652411,
"grad_norm": 1.217984783068752,
"learning_rate": 1.533926178282641e-07,
"loss": 0.5084,
"step": 2130
},
{
"epoch": 0.9338861008073315,
"grad_norm": 1.3602984814635521,
"learning_rate": 1.3523149129283987e-07,
"loss": 0.5222,
"step": 2140
},
{
"epoch": 0.9382500545494218,
"grad_norm": 1.3460153750821495,
"learning_rate": 1.1819957862160835e-07,
"loss": 0.5069,
"step": 2150
},
{
"epoch": 0.9426140082915121,
"grad_norm": 1.4510294571460467,
"learning_rate": 1.0230083327543661e-07,
"loss": 0.5142,
"step": 2160
},
{
"epoch": 0.9469779620336024,
"grad_norm": 1.396143496017602,
"learning_rate": 8.753894568350573e-08,
"loss": 0.5155,
"step": 2170
},
{
"epoch": 0.9513419157756928,
"grad_norm": 1.437628267014867,
"learning_rate": 7.391734238668846e-08,
"loss": 0.5074,
"step": 2180
},
{
"epoch": 0.9557058695177831,
"grad_norm": 1.3148582495813304,
"learning_rate": 6.143918524217696e-08,
"loss": 0.5176,
"step": 2190
},
{
"epoch": 0.9600698232598734,
"grad_norm": 1.267741565404716,
"learning_rate": 5.01073706895483e-08,
"loss": 0.5137,
"step": 2200
},
{
"epoch": 0.9644337770019638,
"grad_norm": 1.2859440475396895,
"learning_rate": 3.992452907844502e-08,
"loss": 0.5142,
"step": 2210
},
{
"epoch": 0.9687977307440541,
"grad_norm": 1.2296422531635047,
"learning_rate": 3.08930240580102e-08,
"loss": 0.5025,
"step": 2220
},
{
"epoch": 0.9731616844861445,
"grad_norm": 1.3046489064561737,
"learning_rate": 2.3014952028240223e-08,
"loss": 0.5138,
"step": 2230
},
{
"epoch": 0.9775256382282348,
"grad_norm": 1.238897807436414,
"learning_rate": 1.6292141653366232e-08,
"loss": 0.5067,
"step": 2240
},
{
"epoch": 0.9818895919703251,
"grad_norm": 1.2689478765861664,
"learning_rate": 1.0726153437379816e-08,
"loss": 0.5188,
"step": 2250
},
{
"epoch": 0.9862535457124154,
"grad_norm": 2.492246786468264,
"learning_rate": 6.3182793618110634e-09,
"loss": 0.5151,
"step": 2260
},
{
"epoch": 0.9906174994545058,
"grad_norm": 1.2722668029711148,
"learning_rate": 3.0695425858329187e-09,
"loss": 0.5197,
"step": 2270
},
{
"epoch": 0.9949814531965961,
"grad_norm": 1.3363964992431479,
"learning_rate": 9.806972087605949e-10,
"loss": 0.5223,
"step": 2280
},
{
"epoch": 0.9993454069386865,
"grad_norm": 1.5575122083398747,
"learning_rate": 5.222809501492787e-11,
"loss": 0.5178,
"step": 2290
},
{
"epoch": 1.0,
"step": 2292,
"total_flos": 2450977789313024.0,
"train_loss": 0.5714389552828737,
"train_runtime": 14627.8115,
"train_samples_per_second": 25.06,
"train_steps_per_second": 0.157
}
],
"logging_steps": 10,
"max_steps": 2292,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 200,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2450977789313024.0,
"train_batch_size": 10,
"trial_name": null,
"trial_params": null
}