SykoLLM-V4.8-Beta / trainer_state.json
SykoSLM's picture
Upload folder using huggingface_hub
57d6efa verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.225,
"eval_steps": 500,
"global_step": 1800,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00125,
"grad_norm": 0.07407524436712265,
"learning_rate": 2.25e-06,
"loss": 2.2311376571655273,
"step": 10
},
{
"epoch": 0.0025,
"grad_norm": 0.07508661597967148,
"learning_rate": 4.75e-06,
"loss": 2.254044532775879,
"step": 20
},
{
"epoch": 0.00375,
"grad_norm": 0.08043646067380905,
"learning_rate": 7.25e-06,
"loss": 2.2791187286376955,
"step": 30
},
{
"epoch": 0.005,
"grad_norm": 0.07152171432971954,
"learning_rate": 9.750000000000002e-06,
"loss": 2.29660701751709,
"step": 40
},
{
"epoch": 0.00625,
"grad_norm": 0.07239466905593872,
"learning_rate": 1.225e-05,
"loss": 2.2157821655273438,
"step": 50
},
{
"epoch": 0.0075,
"grad_norm": 0.07063177227973938,
"learning_rate": 1.475e-05,
"loss": 2.2415660858154296,
"step": 60
},
{
"epoch": 0.00875,
"grad_norm": 0.07137205451726913,
"learning_rate": 1.725e-05,
"loss": 2.188554573059082,
"step": 70
},
{
"epoch": 0.01,
"grad_norm": 0.07028814405202866,
"learning_rate": 1.9750000000000002e-05,
"loss": 2.177708053588867,
"step": 80
},
{
"epoch": 0.01125,
"grad_norm": 0.07327590137720108,
"learning_rate": 2.2250000000000002e-05,
"loss": 2.22082405090332,
"step": 90
},
{
"epoch": 0.0125,
"grad_norm": 0.06442452222108841,
"learning_rate": 2.4750000000000002e-05,
"loss": 2.2364925384521483,
"step": 100
},
{
"epoch": 0.01375,
"grad_norm": 0.06526491045951843,
"learning_rate": 2.725e-05,
"loss": 2.2664451599121094,
"step": 110
},
{
"epoch": 0.015,
"grad_norm": 0.06679920852184296,
"learning_rate": 2.975e-05,
"loss": 2.188519287109375,
"step": 120
},
{
"epoch": 0.01625,
"grad_norm": 0.07065364718437195,
"learning_rate": 3.2250000000000005e-05,
"loss": 2.270803451538086,
"step": 130
},
{
"epoch": 0.0175,
"grad_norm": 0.07348085194826126,
"learning_rate": 3.475e-05,
"loss": 2.213222122192383,
"step": 140
},
{
"epoch": 0.01875,
"grad_norm": 0.0695674866437912,
"learning_rate": 3.7250000000000004e-05,
"loss": 2.208852767944336,
"step": 150
},
{
"epoch": 0.02,
"grad_norm": 0.07005158066749573,
"learning_rate": 3.9750000000000004e-05,
"loss": 2.161180305480957,
"step": 160
},
{
"epoch": 0.02125,
"grad_norm": 0.06879838556051254,
"learning_rate": 4.2250000000000004e-05,
"loss": 2.1557788848876953,
"step": 170
},
{
"epoch": 0.0225,
"grad_norm": 0.07037137448787689,
"learning_rate": 4.4750000000000004e-05,
"loss": 2.2376060485839844,
"step": 180
},
{
"epoch": 0.02375,
"grad_norm": 0.06982705742120743,
"learning_rate": 4.7249999999999997e-05,
"loss": 2.217011260986328,
"step": 190
},
{
"epoch": 0.025,
"grad_norm": 0.06674907356500626,
"learning_rate": 4.975e-05,
"loss": 2.256929397583008,
"step": 200
},
{
"epoch": 0.02625,
"grad_norm": 0.07002667337656021,
"learning_rate": 4.999983575010662e-05,
"loss": 2.179422378540039,
"step": 210
},
{
"epoch": 0.0275,
"grad_norm": 0.0717761367559433,
"learning_rate": 4.9999267975468225e-05,
"loss": 2.2249755859375,
"step": 220
},
{
"epoch": 0.02875,
"grad_norm": 0.07136455923318863,
"learning_rate": 4.9998294657516734e-05,
"loss": 2.222713088989258,
"step": 230
},
{
"epoch": 0.03,
"grad_norm": 0.07328379899263382,
"learning_rate": 4.999691581204152e-05,
"loss": 2.1884521484375,
"step": 240
},
{
"epoch": 0.03125,
"grad_norm": 0.07197773456573486,
"learning_rate": 4.999513146141048e-05,
"loss": 2.2087806701660155,
"step": 250
},
{
"epoch": 0.0325,
"grad_norm": 0.07391443103551865,
"learning_rate": 4.999294163456975e-05,
"loss": 2.2307754516601563,
"step": 260
},
{
"epoch": 0.03375,
"grad_norm": 0.07325068861246109,
"learning_rate": 4.9990346367043114e-05,
"loss": 2.198223114013672,
"step": 270
},
{
"epoch": 0.035,
"grad_norm": 0.07103461772203445,
"learning_rate": 4.9987345700931586e-05,
"loss": 2.228559112548828,
"step": 280
},
{
"epoch": 0.03625,
"grad_norm": 0.07458475232124329,
"learning_rate": 4.998393968491256e-05,
"loss": 2.1933670043945312,
"step": 290
},
{
"epoch": 0.0375,
"grad_norm": 0.06882892549037933,
"learning_rate": 4.9980128374239156e-05,
"loss": 2.2323123931884767,
"step": 300
},
{
"epoch": 0.03875,
"grad_norm": 0.07062669098377228,
"learning_rate": 4.997591183073923e-05,
"loss": 2.2588329315185547,
"step": 310
},
{
"epoch": 0.04,
"grad_norm": 0.07071756571531296,
"learning_rate": 4.997129012281443e-05,
"loss": 2.11534423828125,
"step": 320
},
{
"epoch": 0.04125,
"grad_norm": 0.06557808071374893,
"learning_rate": 4.996626332543907e-05,
"loss": 2.1684688568115233,
"step": 330
},
{
"epoch": 0.0425,
"grad_norm": 0.0680045560002327,
"learning_rate": 4.9960831520158904e-05,
"loss": 2.1954887390136717,
"step": 340
},
{
"epoch": 0.04375,
"grad_norm": 0.07579149305820465,
"learning_rate": 4.9954994795089786e-05,
"loss": 2.120090103149414,
"step": 350
},
{
"epoch": 0.045,
"grad_norm": 0.07108975201845169,
"learning_rate": 4.994875324491629e-05,
"loss": 2.1955142974853517,
"step": 360
},
{
"epoch": 0.04625,
"grad_norm": 0.07410898804664612,
"learning_rate": 4.994210697089014e-05,
"loss": 2.2114063262939454,
"step": 370
},
{
"epoch": 0.0475,
"grad_norm": 0.06744780391454697,
"learning_rate": 4.993505608082857e-05,
"loss": 2.1475910186767577,
"step": 380
},
{
"epoch": 0.04875,
"grad_norm": 0.06901328265666962,
"learning_rate": 4.992760068911258e-05,
"loss": 2.2015354156494142,
"step": 390
},
{
"epoch": 0.05,
"grad_norm": 0.0716853216290474,
"learning_rate": 4.991974091668509e-05,
"loss": 2.2013004302978514,
"step": 400
},
{
"epoch": 0.05125,
"grad_norm": 0.07426904886960983,
"learning_rate": 4.991147689104895e-05,
"loss": 2.170287322998047,
"step": 410
},
{
"epoch": 0.0525,
"grad_norm": 0.06732422858476639,
"learning_rate": 4.990280874626491e-05,
"loss": 2.1896934509277344,
"step": 420
},
{
"epoch": 0.05375,
"grad_norm": 0.06560606509447098,
"learning_rate": 4.98937366229494e-05,
"loss": 2.2104034423828125,
"step": 430
},
{
"epoch": 0.055,
"grad_norm": 0.06944538652896881,
"learning_rate": 4.98842606682723e-05,
"loss": 2.126712989807129,
"step": 440
},
{
"epoch": 0.05625,
"grad_norm": 0.06867734342813492,
"learning_rate": 4.98743810359545e-05,
"loss": 2.234315299987793,
"step": 450
},
{
"epoch": 0.0575,
"grad_norm": 0.07116754353046417,
"learning_rate": 4.986409788626546e-05,
"loss": 2.184938430786133,
"step": 460
},
{
"epoch": 0.05875,
"grad_norm": 0.06651467084884644,
"learning_rate": 4.985341138602056e-05,
"loss": 2.152230453491211,
"step": 470
},
{
"epoch": 0.06,
"grad_norm": 0.06930788606405258,
"learning_rate": 4.984232170857842e-05,
"loss": 2.181766891479492,
"step": 480
},
{
"epoch": 0.06125,
"grad_norm": 0.07047919183969498,
"learning_rate": 4.9830829033838096e-05,
"loss": 2.187181091308594,
"step": 490
},
{
"epoch": 0.0625,
"grad_norm": 0.07572918385267258,
"learning_rate": 4.981893354823614e-05,
"loss": 2.1816110610961914,
"step": 500
},
{
"epoch": 0.06375,
"grad_norm": 0.07347270846366882,
"learning_rate": 4.9806635444743595e-05,
"loss": 2.2242454528808593,
"step": 510
},
{
"epoch": 0.065,
"grad_norm": 0.06983183324337006,
"learning_rate": 4.979393492286284e-05,
"loss": 2.1648427963256838,
"step": 520
},
{
"epoch": 0.06625,
"grad_norm": 0.06904461979866028,
"learning_rate": 4.9780832188624375e-05,
"loss": 2.1993919372558595,
"step": 530
},
{
"epoch": 0.0675,
"grad_norm": 0.0692647248506546,
"learning_rate": 4.976732745458348e-05,
"loss": 2.1890403747558596,
"step": 540
},
{
"epoch": 0.06875,
"grad_norm": 0.06891040503978729,
"learning_rate": 4.975342093981675e-05,
"loss": 2.1854305267333984,
"step": 550
},
{
"epoch": 0.07,
"grad_norm": 0.0714147537946701,
"learning_rate": 4.973911286991856e-05,
"loss": 2.219981002807617,
"step": 560
},
{
"epoch": 0.07125,
"grad_norm": 0.07621099054813385,
"learning_rate": 4.9724403476997384e-05,
"loss": 2.182698440551758,
"step": 570
},
{
"epoch": 0.0725,
"grad_norm": 0.07603556662797928,
"learning_rate": 4.970929299967204e-05,
"loss": 2.1696598052978517,
"step": 580
},
{
"epoch": 0.07375,
"grad_norm": 0.07282973825931549,
"learning_rate": 4.969378168306784e-05,
"loss": 2.1716108322143555,
"step": 590
},
{
"epoch": 0.075,
"grad_norm": 0.06459035724401474,
"learning_rate": 4.967786977881257e-05,
"loss": 2.1681615829467775,
"step": 600
},
{
"epoch": 0.07625,
"grad_norm": 0.07177098840475082,
"learning_rate": 4.966155754503245e-05,
"loss": 2.175799560546875,
"step": 610
},
{
"epoch": 0.0775,
"grad_norm": 0.06652142852544785,
"learning_rate": 4.964484524634792e-05,
"loss": 2.1907752990722655,
"step": 620
},
{
"epoch": 0.07875,
"grad_norm": 0.0727977305650711,
"learning_rate": 4.962773315386935e-05,
"loss": 2.1232975006103514,
"step": 630
},
{
"epoch": 0.08,
"grad_norm": 0.07146522402763367,
"learning_rate": 4.961022154519267e-05,
"loss": 2.1747344970703124,
"step": 640
},
{
"epoch": 0.08125,
"grad_norm": 0.06721191853284836,
"learning_rate": 4.959231070439482e-05,
"loss": 2.2015512466430662,
"step": 650
},
{
"epoch": 0.0825,
"grad_norm": 0.06989260762929916,
"learning_rate": 4.957400092202917e-05,
"loss": 2.187305450439453,
"step": 660
},
{
"epoch": 0.08375,
"grad_norm": 0.06762924790382385,
"learning_rate": 4.955529249512082e-05,
"loss": 2.1369449615478517,
"step": 670
},
{
"epoch": 0.085,
"grad_norm": 0.06524205207824707,
"learning_rate": 4.953618572716175e-05,
"loss": 2.162923049926758,
"step": 680
},
{
"epoch": 0.08625,
"grad_norm": 0.0713450163602829,
"learning_rate": 4.951668092810593e-05,
"loss": 2.2041025161743164,
"step": 690
},
{
"epoch": 0.0875,
"grad_norm": 0.0673929825425148,
"learning_rate": 4.949677841436423e-05,
"loss": 2.124150848388672,
"step": 700
},
{
"epoch": 0.08875,
"grad_norm": 0.06595680117607117,
"learning_rate": 4.947647850879938e-05,
"loss": 2.1866949081420897,
"step": 710
},
{
"epoch": 0.09,
"grad_norm": 0.06935486942529678,
"learning_rate": 4.945578154072065e-05,
"loss": 2.168051528930664,
"step": 720
},
{
"epoch": 0.09125,
"grad_norm": 0.07811375707387924,
"learning_rate": 4.943468784587856e-05,
"loss": 2.2397434234619142,
"step": 730
},
{
"epoch": 0.0925,
"grad_norm": 0.06744036078453064,
"learning_rate": 4.9413197766459394e-05,
"loss": 2.206024169921875,
"step": 740
},
{
"epoch": 0.09375,
"grad_norm": 0.06797294318675995,
"learning_rate": 4.939131165107969e-05,
"loss": 2.16088981628418,
"step": 750
},
{
"epoch": 0.095,
"grad_norm": 0.06975241005420685,
"learning_rate": 4.936902985478055e-05,
"loss": 2.2246139526367186,
"step": 760
},
{
"epoch": 0.09625,
"grad_norm": 0.0649324357509613,
"learning_rate": 4.9346352739021895e-05,
"loss": 2.1035717010498045,
"step": 770
},
{
"epoch": 0.0975,
"grad_norm": 0.0682821199297905,
"learning_rate": 4.93232806716766e-05,
"loss": 2.1523895263671875,
"step": 780
},
{
"epoch": 0.09875,
"grad_norm": 0.06782221794128418,
"learning_rate": 4.9299814027024536e-05,
"loss": 2.199502182006836,
"step": 790
},
{
"epoch": 0.1,
"grad_norm": 0.06508056074380875,
"learning_rate": 4.927595318574649e-05,
"loss": 2.1637832641601564,
"step": 800
},
{
"epoch": 0.10125,
"grad_norm": 0.06774035096168518,
"learning_rate": 4.9251698534917965e-05,
"loss": 2.227591323852539,
"step": 810
},
{
"epoch": 0.1025,
"grad_norm": 0.06584357470273972,
"learning_rate": 4.9227050468002954e-05,
"loss": 2.204555130004883,
"step": 820
},
{
"epoch": 0.10375,
"grad_norm": 0.06837441772222519,
"learning_rate": 4.92020093848475e-05,
"loss": 2.1817317962646485,
"step": 830
},
{
"epoch": 0.105,
"grad_norm": 0.06870482861995697,
"learning_rate": 4.9176575691673265e-05,
"loss": 2.1889968872070313,
"step": 840
},
{
"epoch": 0.10625,
"grad_norm": 0.06784000247716904,
"learning_rate": 4.9150749801070884e-05,
"loss": 2.1590961456298827,
"step": 850
},
{
"epoch": 0.1075,
"grad_norm": 0.06860610842704773,
"learning_rate": 4.912453213199332e-05,
"loss": 2.1836381912231446,
"step": 860
},
{
"epoch": 0.10875,
"grad_norm": 0.06371141225099564,
"learning_rate": 4.909792310974904e-05,
"loss": 2.136852264404297,
"step": 870
},
{
"epoch": 0.11,
"grad_norm": 0.06636381894350052,
"learning_rate": 4.9070923165995135e-05,
"loss": 2.123456573486328,
"step": 880
},
{
"epoch": 0.11125,
"grad_norm": 0.06704560667276382,
"learning_rate": 4.9043532738730284e-05,
"loss": 2.210599899291992,
"step": 890
},
{
"epoch": 0.1125,
"grad_norm": 0.06560536473989487,
"learning_rate": 4.901575227228769e-05,
"loss": 2.1686206817626954,
"step": 900
},
{
"epoch": 0.11375,
"grad_norm": 0.06711073964834213,
"learning_rate": 4.898758221732783e-05,
"loss": 2.1291219711303713,
"step": 910
},
{
"epoch": 0.115,
"grad_norm": 0.07313625514507294,
"learning_rate": 4.89590230308312e-05,
"loss": 2.1620079040527345,
"step": 920
},
{
"epoch": 0.11625,
"grad_norm": 0.07415210455656052,
"learning_rate": 4.8930075176090844e-05,
"loss": 2.2015281677246095,
"step": 930
},
{
"epoch": 0.1175,
"grad_norm": 0.06644522398710251,
"learning_rate": 4.890073912270486e-05,
"loss": 2.154953384399414,
"step": 940
},
{
"epoch": 0.11875,
"grad_norm": 0.0653102844953537,
"learning_rate": 4.887101534656882e-05,
"loss": 2.186296081542969,
"step": 950
},
{
"epoch": 0.12,
"grad_norm": 0.06608347594738007,
"learning_rate": 4.8840904329867955e-05,
"loss": 2.1376876831054688,
"step": 960
},
{
"epoch": 0.12125,
"grad_norm": 0.06143832579255104,
"learning_rate": 4.881040656106944e-05,
"loss": 2.1134971618652343,
"step": 970
},
{
"epoch": 0.1225,
"grad_norm": 0.06736340373754501,
"learning_rate": 4.8779522534914414e-05,
"loss": 2.1506475448608398,
"step": 980
},
{
"epoch": 0.12375,
"grad_norm": 0.06936616450548172,
"learning_rate": 4.874825275240996e-05,
"loss": 2.125407028198242,
"step": 990
},
{
"epoch": 0.125,
"grad_norm": 0.07099438458681107,
"learning_rate": 4.871659772082097e-05,
"loss": 2.2163812637329103,
"step": 1000
},
{
"epoch": 0.12625,
"grad_norm": 0.06724845618009567,
"learning_rate": 4.868455795366195e-05,
"loss": 2.1106163024902345,
"step": 1010
},
{
"epoch": 0.1275,
"grad_norm": 0.0672367662191391,
"learning_rate": 4.8652133970688636e-05,
"loss": 2.071784019470215,
"step": 1020
},
{
"epoch": 0.12875,
"grad_norm": 0.06996046751737595,
"learning_rate": 4.861932629788962e-05,
"loss": 2.198385238647461,
"step": 1030
},
{
"epoch": 0.13,
"grad_norm": 0.06979460269212723,
"learning_rate": 4.858613546747777e-05,
"loss": 2.111052894592285,
"step": 1040
},
{
"epoch": 0.13125,
"grad_norm": 0.06372442096471786,
"learning_rate": 4.8552562017881634e-05,
"loss": 2.133512496948242,
"step": 1050
},
{
"epoch": 0.1325,
"grad_norm": 0.06758279353380203,
"learning_rate": 4.851860649373666e-05,
"loss": 2.1648277282714843,
"step": 1060
},
{
"epoch": 0.13375,
"grad_norm": 0.07092203199863434,
"learning_rate": 4.848426944587644e-05,
"loss": 2.1862049102783203,
"step": 1070
},
{
"epoch": 0.135,
"grad_norm": 0.06371688097715378,
"learning_rate": 4.844955143132366e-05,
"loss": 2.197888946533203,
"step": 1080
},
{
"epoch": 0.13625,
"grad_norm": 0.06636802852153778,
"learning_rate": 4.841445301328117e-05,
"loss": 2.172176742553711,
"step": 1090
},
{
"epoch": 0.1375,
"grad_norm": 0.06878487765789032,
"learning_rate": 4.8378974761122794e-05,
"loss": 2.2152605056762695,
"step": 1100
},
{
"epoch": 0.13875,
"grad_norm": 0.06840436160564423,
"learning_rate": 4.834311725038409e-05,
"loss": 2.2093944549560547,
"step": 1110
},
{
"epoch": 0.14,
"grad_norm": 0.06905412673950195,
"learning_rate": 4.830688106275305e-05,
"loss": 2.1413171768188475,
"step": 1120
},
{
"epoch": 0.14125,
"grad_norm": 0.06591488420963287,
"learning_rate": 4.827026678606063e-05,
"loss": 2.2281288146972655,
"step": 1130
},
{
"epoch": 0.1425,
"grad_norm": 0.06814192980527878,
"learning_rate": 4.823327501427121e-05,
"loss": 2.241679000854492,
"step": 1140
},
{
"epoch": 0.14375,
"grad_norm": 0.06960905343294144,
"learning_rate": 4.8195906347473e-05,
"loss": 2.1834861755371096,
"step": 1150
},
{
"epoch": 0.145,
"grad_norm": 0.07025773823261261,
"learning_rate": 4.8158161391868276e-05,
"loss": 2.2492515563964846,
"step": 1160
},
{
"epoch": 0.14625,
"grad_norm": 0.06904685497283936,
"learning_rate": 4.812004075976352e-05,
"loss": 2.1875804901123046,
"step": 1170
},
{
"epoch": 0.1475,
"grad_norm": 0.06693335622549057,
"learning_rate": 4.808154506955955e-05,
"loss": 2.23748779296875,
"step": 1180
},
{
"epoch": 0.14875,
"grad_norm": 0.06956436485052109,
"learning_rate": 4.804267494574145e-05,
"loss": 2.2373470306396483,
"step": 1190
},
{
"epoch": 0.15,
"grad_norm": 0.07944408059120178,
"learning_rate": 4.800343101886842e-05,
"loss": 2.2595752716064452,
"step": 1200
},
{
"epoch": 0.15125,
"grad_norm": 0.06570149958133698,
"learning_rate": 4.7963813925563586e-05,
"loss": 2.2755199432373048,
"step": 1210
},
{
"epoch": 0.1525,
"grad_norm": 0.06871885061264038,
"learning_rate": 4.792382430850366e-05,
"loss": 2.2775409698486326,
"step": 1220
},
{
"epoch": 0.15375,
"grad_norm": 0.06904297322034836,
"learning_rate": 4.788346281640852e-05,
"loss": 2.2760135650634767,
"step": 1230
},
{
"epoch": 0.155,
"grad_norm": 0.06859485059976578,
"learning_rate": 4.784273010403065e-05,
"loss": 2.259684753417969,
"step": 1240
},
{
"epoch": 0.15625,
"grad_norm": 0.06464747339487076,
"learning_rate": 4.780162683214457e-05,
"loss": 2.308213806152344,
"step": 1250
},
{
"epoch": 0.1575,
"grad_norm": 0.06627296656370163,
"learning_rate": 4.77601536675361e-05,
"loss": 2.3249114990234374,
"step": 1260
},
{
"epoch": 0.15875,
"grad_norm": 0.06741499155759811,
"learning_rate": 4.77183112829915e-05,
"loss": 2.350922393798828,
"step": 1270
},
{
"epoch": 0.16,
"grad_norm": 0.06964308768510818,
"learning_rate": 4.7676100357286624e-05,
"loss": 2.244986915588379,
"step": 1280
},
{
"epoch": 0.16125,
"grad_norm": 0.0634695515036583,
"learning_rate": 4.7633521575175874e-05,
"loss": 2.2805355072021483,
"step": 1290
},
{
"epoch": 0.1625,
"grad_norm": 0.06915149837732315,
"learning_rate": 4.759057562738109e-05,
"loss": 2.3919076919555664,
"step": 1300
},
{
"epoch": 0.16375,
"grad_norm": 0.06478839367628098,
"learning_rate": 4.754726321058034e-05,
"loss": 2.259754180908203,
"step": 1310
},
{
"epoch": 0.165,
"grad_norm": 0.06912291049957275,
"learning_rate": 4.7503585027396646e-05,
"loss": 2.2800014495849608,
"step": 1320
},
{
"epoch": 0.16625,
"grad_norm": 0.06265991181135178,
"learning_rate": 4.745954178638656e-05,
"loss": 2.291925811767578,
"step": 1330
},
{
"epoch": 0.1675,
"grad_norm": 0.06898439675569534,
"learning_rate": 4.741513420202867e-05,
"loss": 2.360172080993652,
"step": 1340
},
{
"epoch": 0.16875,
"grad_norm": 0.06620923429727554,
"learning_rate": 4.7370362994712036e-05,
"loss": 2.368211364746094,
"step": 1350
},
{
"epoch": 0.17,
"grad_norm": 0.07054638862609863,
"learning_rate": 4.732522889072447e-05,
"loss": 2.3170215606689455,
"step": 1360
},
{
"epoch": 0.17125,
"grad_norm": 0.07074211537837982,
"learning_rate": 4.7279732622240766e-05,
"loss": 2.276025581359863,
"step": 1370
},
{
"epoch": 0.1725,
"grad_norm": 0.06905434280633926,
"learning_rate": 4.723387492731084e-05,
"loss": 2.2982337951660154,
"step": 1380
},
{
"epoch": 0.17375,
"grad_norm": 0.06408198177814484,
"learning_rate": 4.718765654984773e-05,
"loss": 2.3407756805419924,
"step": 1390
},
{
"epoch": 0.175,
"grad_norm": 0.06612099707126617,
"learning_rate": 4.7141078239615566e-05,
"loss": 2.2696388244628904,
"step": 1400
},
{
"epoch": 0.17625,
"grad_norm": 0.06839219480752945,
"learning_rate": 4.709414075221734e-05,
"loss": 2.227347755432129,
"step": 1410
},
{
"epoch": 0.1775,
"grad_norm": 0.06748591363430023,
"learning_rate": 4.7046844849082725e-05,
"loss": 2.310345458984375,
"step": 1420
},
{
"epoch": 0.17875,
"grad_norm": 0.06979987025260925,
"learning_rate": 4.699919129745567e-05,
"loss": 2.3062816619873048,
"step": 1430
},
{
"epoch": 0.18,
"grad_norm": 0.06655360758304596,
"learning_rate": 4.695118087038198e-05,
"loss": 2.30033016204834,
"step": 1440
},
{
"epoch": 0.18125,
"grad_norm": 0.0673886388540268,
"learning_rate": 4.690281434669677e-05,
"loss": 2.3160564422607424,
"step": 1450
},
{
"epoch": 0.1825,
"grad_norm": 0.0680936723947525,
"learning_rate": 4.685409251101183e-05,
"loss": 2.378013801574707,
"step": 1460
},
{
"epoch": 0.18375,
"grad_norm": 0.07122587412595749,
"learning_rate": 4.680501615370288e-05,
"loss": 2.2907699584960937,
"step": 1470
},
{
"epoch": 0.185,
"grad_norm": 0.07071765512228012,
"learning_rate": 4.675558607089677e-05,
"loss": 2.3167457580566406,
"step": 1480
},
{
"epoch": 0.18625,
"grad_norm": 0.07135665416717529,
"learning_rate": 4.6705803064458575e-05,
"loss": 2.36193962097168,
"step": 1490
},
{
"epoch": 0.1875,
"grad_norm": 0.07396227866411209,
"learning_rate": 4.665566794197854e-05,
"loss": 2.3736957550048827,
"step": 1500
},
{
"epoch": 0.18875,
"grad_norm": 0.06626343727111816,
"learning_rate": 4.6605181516759047e-05,
"loss": 2.3442031860351564,
"step": 1510
},
{
"epoch": 0.19,
"grad_norm": 0.06523022055625916,
"learning_rate": 4.6554344607801335e-05,
"loss": 2.3258100509643556,
"step": 1520
},
{
"epoch": 0.19125,
"grad_norm": 0.06743918359279633,
"learning_rate": 4.6503158039792324e-05,
"loss": 2.286594009399414,
"step": 1530
},
{
"epoch": 0.1925,
"grad_norm": 0.06837300956249237,
"learning_rate": 4.645162264309112e-05,
"loss": 2.320156288146973,
"step": 1540
},
{
"epoch": 0.19375,
"grad_norm": 0.06472747027873993,
"learning_rate": 4.639973925371562e-05,
"loss": 2.39737548828125,
"step": 1550
},
{
"epoch": 0.195,
"grad_norm": 0.06730551272630692,
"learning_rate": 4.634750871332896e-05,
"loss": 2.3250579833984375,
"step": 1560
},
{
"epoch": 0.19625,
"grad_norm": 0.06307782232761383,
"learning_rate": 4.6294931869225774e-05,
"loss": 2.300210189819336,
"step": 1570
},
{
"epoch": 0.1975,
"grad_norm": 0.06609486788511276,
"learning_rate": 4.6242009574318554e-05,
"loss": 2.2991945266723635,
"step": 1580
},
{
"epoch": 0.19875,
"grad_norm": 0.0668780580163002,
"learning_rate": 4.6188742687123754e-05,
"loss": 2.2811546325683594,
"step": 1590
},
{
"epoch": 0.2,
"grad_norm": 0.07088489830493927,
"learning_rate": 4.6135132071747864e-05,
"loss": 2.32708740234375,
"step": 1600
},
{
"epoch": 0.20125,
"grad_norm": 0.06599114835262299,
"learning_rate": 4.608117859787342e-05,
"loss": 2.26273250579834,
"step": 1610
},
{
"epoch": 0.2025,
"grad_norm": 0.06767764687538147,
"learning_rate": 4.602688314074487e-05,
"loss": 2.3097578048706056,
"step": 1620
},
{
"epoch": 0.20375,
"grad_norm": 0.06805308163166046,
"learning_rate": 4.597224658115438e-05,
"loss": 2.322467231750488,
"step": 1630
},
{
"epoch": 0.205,
"grad_norm": 0.06872619688510895,
"learning_rate": 4.5917269805427567e-05,
"loss": 2.303860092163086,
"step": 1640
},
{
"epoch": 0.20625,
"grad_norm": 0.06846166402101517,
"learning_rate": 4.5861953705409086e-05,
"loss": 2.2648983001708984,
"step": 1650
},
{
"epoch": 0.2075,
"grad_norm": 0.06966902315616608,
"learning_rate": 4.58062991784482e-05,
"loss": 2.3156864166259767,
"step": 1660
},
{
"epoch": 0.20875,
"grad_norm": 0.06632312387228012,
"learning_rate": 4.575030712738419e-05,
"loss": 2.302859878540039,
"step": 1670
},
{
"epoch": 0.21,
"grad_norm": 0.07102814316749573,
"learning_rate": 4.5693978460531725e-05,
"loss": 2.3624195098876952,
"step": 1680
},
{
"epoch": 0.21125,
"grad_norm": 0.06857810169458389,
"learning_rate": 4.563731409166615e-05,
"loss": 2.360994338989258,
"step": 1690
},
{
"epoch": 0.2125,
"grad_norm": 0.06877714395523071,
"learning_rate": 4.558031494000858e-05,
"loss": 2.322928237915039,
"step": 1700
},
{
"epoch": 0.21375,
"grad_norm": 0.07134252786636353,
"learning_rate": 4.5522981930211114e-05,
"loss": 2.2678380966186524,
"step": 1710
},
{
"epoch": 0.215,
"grad_norm": 0.07036204636096954,
"learning_rate": 4.546531599234173e-05,
"loss": 2.314472770690918,
"step": 1720
},
{
"epoch": 0.21625,
"grad_norm": 0.06794637441635132,
"learning_rate": 4.540731806186922e-05,
"loss": 2.385559844970703,
"step": 1730
},
{
"epoch": 0.2175,
"grad_norm": 0.06780719757080078,
"learning_rate": 4.5348989079648065e-05,
"loss": 2.261928939819336,
"step": 1740
},
{
"epoch": 0.21875,
"grad_norm": 0.06747139990329742,
"learning_rate": 4.5290329991903115e-05,
"loss": 2.295556831359863,
"step": 1750
},
{
"epoch": 0.22,
"grad_norm": 0.06255707144737244,
"learning_rate": 4.5231341750214256e-05,
"loss": 2.338383102416992,
"step": 1760
},
{
"epoch": 0.22125,
"grad_norm": 0.06699945777654648,
"learning_rate": 4.5172025311501004e-05,
"loss": 2.2839550018310546,
"step": 1770
},
{
"epoch": 0.2225,
"grad_norm": 0.06443452835083008,
"learning_rate": 4.511238163800692e-05,
"loss": 2.3070682525634765,
"step": 1780
},
{
"epoch": 0.22375,
"grad_norm": 0.06801515817642212,
"learning_rate": 4.505241169728407e-05,
"loss": 2.313974952697754,
"step": 1790
},
{
"epoch": 0.225,
"grad_norm": 0.06966620683670044,
"learning_rate": 4.499211646217727e-05,
"loss": 2.2580539703369142,
"step": 1800
}
],
"logging_steps": 10,
"max_steps": 8000,
"num_input_tokens_seen": 0,
"num_train_epochs": 9223372036854775807,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.1075411496665088e+17,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}