checkpoint-4382 / trainer_state.json
ChristophSchuhmann's picture
Initial upload of checkpoint-4382
288f8f3 verified
{
"best_global_step": 626,
"best_metric": 4.402504920959473,
"best_model_checkpoint": "/home/deployer/laion/Orpheus-3B-Continued-2E-V4-WithGen/checkpoint-626",
"epoch": 0.12624531117654175,
"eval_steps": 313,
"global_step": 4382,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00020166982616060984,
"grad_norm": 0.2099609375,
"learning_rate": 1.8e-05,
"loss": 4.9769,
"step": 10
},
{
"epoch": 0.0004033396523212197,
"grad_norm": 0.1552734375,
"learning_rate": 3.8e-05,
"loss": 4.8382,
"step": 20
},
{
"epoch": 0.0006050094784818296,
"grad_norm": 0.07666015625,
"learning_rate": 5.8e-05,
"loss": 4.6796,
"step": 30
},
{
"epoch": 0.0008066793046424393,
"grad_norm": 0.06787109375,
"learning_rate": 7.800000000000001e-05,
"loss": 4.6557,
"step": 40
},
{
"epoch": 0.0010083491308030493,
"grad_norm": 0.058837890625,
"learning_rate": 9.8e-05,
"loss": 4.6592,
"step": 50
},
{
"epoch": 0.0012100189569636591,
"grad_norm": 0.06982421875,
"learning_rate": 0.000118,
"loss": 4.612,
"step": 60
},
{
"epoch": 0.001411688783124269,
"grad_norm": 0.06396484375,
"learning_rate": 0.000138,
"loss": 4.5519,
"step": 70
},
{
"epoch": 0.0016133586092848787,
"grad_norm": 0.06787109375,
"learning_rate": 0.00015800000000000002,
"loss": 4.5649,
"step": 80
},
{
"epoch": 0.0018150284354454887,
"grad_norm": 0.07177734375,
"learning_rate": 0.00017800000000000002,
"loss": 4.5665,
"step": 90
},
{
"epoch": 0.0020166982616060987,
"grad_norm": 0.0615234375,
"learning_rate": 0.00019800000000000002,
"loss": 4.5854,
"step": 100
},
{
"epoch": 0.0022183680877667085,
"grad_norm": 0.06298828125,
"learning_rate": 0.00019999998367737306,
"loss": 4.6197,
"step": 110
},
{
"epoch": 0.0024200379139273183,
"grad_norm": 0.0693359375,
"learning_rate": 0.00019999992725348425,
"loss": 4.5047,
"step": 120
},
{
"epoch": 0.002621707740087928,
"grad_norm": 0.0732421875,
"learning_rate": 0.00019999983052684242,
"loss": 4.5378,
"step": 130
},
{
"epoch": 0.002823377566248538,
"grad_norm": 0.07763671875,
"learning_rate": 0.0001999996934974865,
"loss": 4.5702,
"step": 140
},
{
"epoch": 0.0030250473924091476,
"grad_norm": 0.08154296875,
"learning_rate": 0.00019999951616547182,
"loss": 4.5161,
"step": 150
},
{
"epoch": 0.0032267172185697574,
"grad_norm": 0.07177734375,
"learning_rate": 0.00019999929853086975,
"loss": 4.5427,
"step": 160
},
{
"epoch": 0.0034283870447303676,
"grad_norm": 0.064453125,
"learning_rate": 0.00019999904059376803,
"loss": 4.5205,
"step": 170
},
{
"epoch": 0.0036300568708909774,
"grad_norm": 0.06298828125,
"learning_rate": 0.00019999874235427067,
"loss": 4.5382,
"step": 180
},
{
"epoch": 0.003831726697051587,
"grad_norm": 0.0693359375,
"learning_rate": 0.0001999984038124978,
"loss": 4.5279,
"step": 190
},
{
"epoch": 0.004033396523212197,
"grad_norm": 0.06640625,
"learning_rate": 0.0001999980249685859,
"loss": 4.5047,
"step": 200
},
{
"epoch": 0.004235066349372807,
"grad_norm": 0.06494140625,
"learning_rate": 0.00019999760582268763,
"loss": 4.5041,
"step": 210
},
{
"epoch": 0.004436736175533417,
"grad_norm": 0.06689453125,
"learning_rate": 0.00019999714637497192,
"loss": 4.5513,
"step": 220
},
{
"epoch": 0.004638406001694026,
"grad_norm": 0.0654296875,
"learning_rate": 0.00019999664662562398,
"loss": 4.5115,
"step": 230
},
{
"epoch": 0.0048400758278546365,
"grad_norm": 0.06396484375,
"learning_rate": 0.0001999961065748452,
"loss": 4.5027,
"step": 240
},
{
"epoch": 0.005041745654015246,
"grad_norm": 0.07373046875,
"learning_rate": 0.00019999552622285317,
"loss": 4.448,
"step": 250
},
{
"epoch": 0.005243415480175856,
"grad_norm": 0.06689453125,
"learning_rate": 0.0001999949055698819,
"loss": 4.5337,
"step": 260
},
{
"epoch": 0.005445085306336466,
"grad_norm": 0.0654296875,
"learning_rate": 0.00019999424461618145,
"loss": 4.5358,
"step": 270
},
{
"epoch": 0.005646755132497076,
"grad_norm": 0.06103515625,
"learning_rate": 0.00019999354336201828,
"loss": 4.5168,
"step": 280
},
{
"epoch": 0.005848424958657686,
"grad_norm": 0.0634765625,
"learning_rate": 0.0001999928018076749,
"loss": 4.4455,
"step": 290
},
{
"epoch": 0.006050094784818295,
"grad_norm": 0.06591796875,
"learning_rate": 0.00019999201995345026,
"loss": 4.4747,
"step": 300
},
{
"epoch": 0.006251764610978905,
"grad_norm": 0.064453125,
"learning_rate": 0.00019999119779965947,
"loss": 4.4948,
"step": 310
},
{
"epoch": 0.006312265558827088,
"eval_loss": 4.493932723999023,
"eval_runtime": 8.8208,
"eval_samples_per_second": 22.674,
"eval_steps_per_second": 1.474,
"step": 313
},
{
"epoch": 0.00028233775662485377,
"grad_norm": 0.046875,
"learning_rate": 0.00019996118655688004,
"loss": 4.4607,
"step": 320
},
{
"epoch": 0.0006856774089460735,
"grad_norm": 0.04541015625,
"learning_rate": 0.00019995756127956854,
"loss": 4.4336,
"step": 330
},
{
"epoch": 0.0010890170612672932,
"grad_norm": 0.046142578125,
"learning_rate": 0.00019995377420631467,
"loss": 4.4016,
"step": 340
},
{
"epoch": 0.001492356713588513,
"grad_norm": 0.04541015625,
"learning_rate": 0.00019994982534324835,
"loss": 4.2976,
"step": 350
},
{
"epoch": 0.0018956963659097325,
"grad_norm": 0.05224609375,
"learning_rate": 0.00019994571469676142,
"loss": 4.2416,
"step": 360
},
{
"epoch": 0.0022990360182309523,
"grad_norm": 0.052734375,
"learning_rate": 0.00019994144227350756,
"loss": 4.1895,
"step": 370
},
{
"epoch": 0.002702375670552172,
"grad_norm": 0.058349609375,
"learning_rate": 0.00019993700808040233,
"loss": 4.1082,
"step": 380
},
{
"epoch": 0.003105715322873392,
"grad_norm": 0.057861328125,
"learning_rate": 0.0001999324121246231,
"loss": 4.148,
"step": 390
},
{
"epoch": 0.0035090549751946114,
"grad_norm": 0.06494140625,
"learning_rate": 0.00019992765441360905,
"loss": 4.1184,
"step": 400
},
{
"epoch": 0.003912394627515831,
"grad_norm": 0.068359375,
"learning_rate": 0.00019992273495506133,
"loss": 4.1018,
"step": 410
},
{
"epoch": 0.004315734279837051,
"grad_norm": 0.06005859375,
"learning_rate": 0.00019991765375694276,
"loss": 4.0995,
"step": 420
},
{
"epoch": 0.0047190739321582706,
"grad_norm": 0.0634765625,
"learning_rate": 0.00019991241082747795,
"loss": 4.1194,
"step": 430
},
{
"epoch": 0.00512241358447949,
"grad_norm": 0.0634765625,
"learning_rate": 0.00019990700617515344,
"loss": 4.0612,
"step": 440
},
{
"epoch": 0.00552575323680071,
"grad_norm": 0.06298828125,
"learning_rate": 0.00019990143980871738,
"loss": 4.1001,
"step": 450
},
{
"epoch": 0.005929092889121929,
"grad_norm": 0.0634765625,
"learning_rate": 0.00019989571173717975,
"loss": 4.075,
"step": 460
},
{
"epoch": 0.00633243254144315,
"grad_norm": 0.07275390625,
"learning_rate": 0.00019988982196981233,
"loss": 4.1117,
"step": 470
},
{
"epoch": 0.006735772193764369,
"grad_norm": 0.0517578125,
"learning_rate": 0.00019988377051614854,
"loss": 4.5104,
"step": 480
},
{
"epoch": 0.007139111846085589,
"grad_norm": 0.047607421875,
"learning_rate": 0.00019987755738598356,
"loss": 4.5086,
"step": 490
},
{
"epoch": 0.007542451498406808,
"grad_norm": 0.047607421875,
"learning_rate": 0.00019987118258937416,
"loss": 4.4517,
"step": 500
},
{
"epoch": 0.007945791150728028,
"grad_norm": 0.044677734375,
"learning_rate": 0.000199864646136639,
"loss": 4.419,
"step": 510
},
{
"epoch": 0.008349130803049248,
"grad_norm": 0.046142578125,
"learning_rate": 0.00019985794803835825,
"loss": 4.4749,
"step": 520
},
{
"epoch": 0.008752470455370467,
"grad_norm": 0.0478515625,
"learning_rate": 0.00019985108830537372,
"loss": 4.4646,
"step": 530
},
{
"epoch": 0.009155810107691688,
"grad_norm": 0.04345703125,
"learning_rate": 0.00019984406694878895,
"loss": 4.4207,
"step": 540
},
{
"epoch": 0.009559149760012906,
"grad_norm": 0.044677734375,
"learning_rate": 0.00019983688397996898,
"loss": 4.4308,
"step": 550
},
{
"epoch": 0.009962489412334127,
"grad_norm": 0.04345703125,
"learning_rate": 0.00019982953941054054,
"loss": 4.4311,
"step": 560
},
{
"epoch": 0.010365829064655347,
"grad_norm": 0.04541015625,
"learning_rate": 0.00019982203325239186,
"loss": 4.4623,
"step": 570
},
{
"epoch": 0.010769168716976566,
"grad_norm": 0.04833984375,
"learning_rate": 0.00019981436551767275,
"loss": 4.4461,
"step": 580
},
{
"epoch": 0.011172508369297786,
"grad_norm": 0.046630859375,
"learning_rate": 0.00019980653621879462,
"loss": 4.4197,
"step": 590
},
{
"epoch": 0.011575848021619005,
"grad_norm": 0.047119140625,
"learning_rate": 0.00019979854536843027,
"loss": 4.4135,
"step": 600
},
{
"epoch": 0.011979187673940225,
"grad_norm": 0.0498046875,
"learning_rate": 0.0001997903929795141,
"loss": 4.4105,
"step": 610
},
{
"epoch": 0.012382527326261444,
"grad_norm": 0.04638671875,
"learning_rate": 0.00019978207906524192,
"loss": 4.4107,
"step": 620
},
{
"epoch": 0.012624531117654176,
"eval_loss": 4.402504920959473,
"eval_runtime": 3.1085,
"eval_samples_per_second": 64.339,
"eval_steps_per_second": 8.042,
"step": 626
},
{
"epoch": 0.00016133586092848787,
"grad_norm": 0.047119140625,
"learning_rate": 0.000199773603639071,
"loss": 4.2779,
"step": 630
},
{
"epoch": 0.0005646755132497075,
"grad_norm": 0.0517578125,
"learning_rate": 0.0001997649667147201,
"loss": 4.2592,
"step": 640
},
{
"epoch": 0.0009680151655709273,
"grad_norm": 0.053466796875,
"learning_rate": 0.00019975616830616937,
"loss": 4.2228,
"step": 650
},
{
"epoch": 0.001371354817892147,
"grad_norm": 0.055908203125,
"learning_rate": 0.00019974720842766023,
"loss": 4.122,
"step": 660
},
{
"epoch": 0.0017746944702133668,
"grad_norm": 0.057373046875,
"learning_rate": 0.00019973808709369565,
"loss": 4.0099,
"step": 670
},
{
"epoch": 0.0021780341225345863,
"grad_norm": 0.0703125,
"learning_rate": 0.00019972880431903977,
"loss": 3.9065,
"step": 680
},
{
"epoch": 0.002581373774855806,
"grad_norm": 0.06982421875,
"learning_rate": 0.00019971936011871816,
"loss": 3.7953,
"step": 690
},
{
"epoch": 0.002984713427177026,
"grad_norm": 0.08349609375,
"learning_rate": 0.00019970975450801762,
"loss": 3.7788,
"step": 700
},
{
"epoch": 0.0033880530794982455,
"grad_norm": 0.08544921875,
"learning_rate": 0.00019969998750248626,
"loss": 3.7368,
"step": 710
},
{
"epoch": 0.003791392731819465,
"grad_norm": 0.09375,
"learning_rate": 0.0001996900591179334,
"loss": 3.7229,
"step": 720
},
{
"epoch": 0.004194732384140685,
"grad_norm": 0.095703125,
"learning_rate": 0.0001996799693704296,
"loss": 3.7192,
"step": 730
},
{
"epoch": 0.004598072036461905,
"grad_norm": 0.09228515625,
"learning_rate": 0.00019966971827630654,
"loss": 3.7457,
"step": 740
},
{
"epoch": 0.005001411688783124,
"grad_norm": 0.0986328125,
"learning_rate": 0.00019965930585215714,
"loss": 3.7031,
"step": 750
},
{
"epoch": 0.005404751341104344,
"grad_norm": 0.10595703125,
"learning_rate": 0.00019964873211483547,
"loss": 3.7335,
"step": 760
},
{
"epoch": 0.005808090993425563,
"grad_norm": 0.10986328125,
"learning_rate": 0.00019963799708145664,
"loss": 3.6902,
"step": 770
},
{
"epoch": 0.006211430645746784,
"grad_norm": 0.10400390625,
"learning_rate": 0.00019962710076939686,
"loss": 3.7408,
"step": 780
},
{
"epoch": 0.006614770298068003,
"grad_norm": 0.062255859375,
"learning_rate": 0.00019961604319629342,
"loss": 4.204,
"step": 790
},
{
"epoch": 0.007018109950389223,
"grad_norm": 0.053955078125,
"learning_rate": 0.00019960482438004462,
"loss": 4.364,
"step": 800
},
{
"epoch": 0.007421449602710442,
"grad_norm": 0.053955078125,
"learning_rate": 0.00019959344433880978,
"loss": 4.3305,
"step": 810
},
{
"epoch": 0.007824789255031663,
"grad_norm": 0.04833984375,
"learning_rate": 0.0001995819030910091,
"loss": 4.2992,
"step": 820
},
{
"epoch": 0.008228128907352882,
"grad_norm": 0.049072265625,
"learning_rate": 0.00019957020065532386,
"loss": 4.3347,
"step": 830
},
{
"epoch": 0.008631468559674102,
"grad_norm": 0.04931640625,
"learning_rate": 0.0001995583370506961,
"loss": 4.321,
"step": 840
},
{
"epoch": 0.00903480821199532,
"grad_norm": 0.0517578125,
"learning_rate": 0.00019954631229632884,
"loss": 4.3122,
"step": 850
},
{
"epoch": 0.009438147864316541,
"grad_norm": 0.05322265625,
"learning_rate": 0.00019953412641168588,
"loss": 4.2918,
"step": 860
},
{
"epoch": 0.00984148751663776,
"grad_norm": 0.052490234375,
"learning_rate": 0.00019952177941649185,
"loss": 4.2793,
"step": 870
},
{
"epoch": 0.01024482716895898,
"grad_norm": 0.048828125,
"learning_rate": 0.00019950927133073222,
"loss": 4.3363,
"step": 880
},
{
"epoch": 0.0106481668212802,
"grad_norm": 0.05126953125,
"learning_rate": 0.00019949660217465307,
"loss": 4.311,
"step": 890
},
{
"epoch": 0.01105150647360142,
"grad_norm": 0.05126953125,
"learning_rate": 0.00019948377196876138,
"loss": 4.2843,
"step": 900
},
{
"epoch": 0.01145484612592264,
"grad_norm": 0.053466796875,
"learning_rate": 0.00019947078073382466,
"loss": 4.2645,
"step": 910
},
{
"epoch": 0.011858185778243858,
"grad_norm": 0.05029296875,
"learning_rate": 0.00019945762849087113,
"loss": 4.2695,
"step": 920
},
{
"epoch": 0.012261525430565079,
"grad_norm": 0.054931640625,
"learning_rate": 0.00019944431526118964,
"loss": 4.279,
"step": 930
},
{
"epoch": 0.012624531117654176,
"eval_loss": 4.440088272094727,
"eval_runtime": 3.1398,
"eval_samples_per_second": 63.698,
"eval_steps_per_second": 7.962,
"step": 939
},
{
"epoch": 4.033396523212197e-05,
"grad_norm": 0.058837890625,
"learning_rate": 0.0,
"loss": 4.0975,
"step": 940
},
{
"epoch": 0.0004436736175533417,
"grad_norm": 0.07421875,
"learning_rate": 2e-05,
"loss": 4.1075,
"step": 950
},
{
"epoch": 0.0008470132698745614,
"grad_norm": 0.0703125,
"learning_rate": 4e-05,
"loss": 4.015,
"step": 960
},
{
"epoch": 0.001250352922195781,
"grad_norm": 0.07373046875,
"learning_rate": 6e-05,
"loss": 3.8564,
"step": 970
},
{
"epoch": 0.0016536925745170008,
"grad_norm": 0.1279296875,
"learning_rate": 8e-05,
"loss": 3.5457,
"step": 980
},
{
"epoch": 0.0020570322268382204,
"grad_norm": 0.1982421875,
"learning_rate": 0.0001,
"loss": 3.2341,
"step": 990
},
{
"epoch": 0.00246037187915944,
"grad_norm": 0.234375,
"learning_rate": 0.00012,
"loss": 3.0956,
"step": 1000
},
{
"epoch": 0.00286371153148066,
"grad_norm": 0.27734375,
"learning_rate": 0.00014,
"loss": 2.9331,
"step": 1010
},
{
"epoch": 0.0032670511838018795,
"grad_norm": 0.337890625,
"learning_rate": 0.00016,
"loss": 2.8938,
"step": 1020
},
{
"epoch": 0.003670390836123099,
"grad_norm": 0.328125,
"learning_rate": 0.00018,
"loss": 2.8145,
"step": 1030
},
{
"epoch": 0.004073730488444319,
"grad_norm": 0.296875,
"learning_rate": 0.0002,
"loss": 2.9022,
"step": 1040
},
{
"epoch": 0.004477070140765539,
"grad_norm": 0.279296875,
"learning_rate": 0.0001999999190676822,
"loss": 2.9572,
"step": 1050
},
{
"epoch": 0.004880409793086758,
"grad_norm": 0.33203125,
"learning_rate": 0.00019999967627085973,
"loss": 2.9381,
"step": 1060
},
{
"epoch": 0.005283749445407978,
"grad_norm": 0.2470703125,
"learning_rate": 0.00019999927160992563,
"loss": 2.9392,
"step": 1070
},
{
"epoch": 0.005687089097729197,
"grad_norm": 0.267578125,
"learning_rate": 0.00019999870508553488,
"loss": 2.8675,
"step": 1080
},
{
"epoch": 0.006090428750050418,
"grad_norm": 0.279296875,
"learning_rate": 0.00019999797669860455,
"loss": 2.9042,
"step": 1090
},
{
"epoch": 0.006493768402371637,
"grad_norm": 0.162109375,
"learning_rate": 0.00019999708645031353,
"loss": 3.4063,
"step": 1100
},
{
"epoch": 0.006897108054692857,
"grad_norm": 0.08349609375,
"learning_rate": 0.00019999603434210292,
"loss": 4.137,
"step": 1110
},
{
"epoch": 0.0073004477070140765,
"grad_norm": 0.0732421875,
"learning_rate": 0.00019999482037567565,
"loss": 4.1305,
"step": 1120
},
{
"epoch": 0.007703787359335296,
"grad_norm": 0.06591796875,
"learning_rate": 0.00019999344455299674,
"loss": 4.0303,
"step": 1130
},
{
"epoch": 0.008107127011656516,
"grad_norm": 0.0634765625,
"learning_rate": 0.0001999919068762931,
"loss": 4.0717,
"step": 1140
},
{
"epoch": 0.008510466663977735,
"grad_norm": 0.059326171875,
"learning_rate": 0.00019999020734805373,
"loss": 4.0664,
"step": 1150
},
{
"epoch": 0.008913806316298956,
"grad_norm": 0.060302734375,
"learning_rate": 0.0001999883459710296,
"loss": 4.0298,
"step": 1160
},
{
"epoch": 0.009317145968620174,
"grad_norm": 0.057861328125,
"learning_rate": 0.00019998632274823358,
"loss": 4.0348,
"step": 1170
},
{
"epoch": 0.009720485620941395,
"grad_norm": 0.06201171875,
"learning_rate": 0.00019998413768294052,
"loss": 4.0192,
"step": 1180
},
{
"epoch": 0.010123825273262615,
"grad_norm": 0.064453125,
"learning_rate": 0.0001999817907786873,
"loss": 4.033,
"step": 1190
},
{
"epoch": 0.010527164925583834,
"grad_norm": 0.0654296875,
"learning_rate": 0.00019997928203927275,
"loss": 4.0413,
"step": 1200
},
{
"epoch": 0.010930504577905054,
"grad_norm": 0.06884765625,
"learning_rate": 0.00019997661146875758,
"loss": 4.0011,
"step": 1210
},
{
"epoch": 0.011333844230226273,
"grad_norm": 0.068359375,
"learning_rate": 0.00019997377907146459,
"loss": 3.9817,
"step": 1220
},
{
"epoch": 0.011737183882547493,
"grad_norm": 0.06884765625,
"learning_rate": 0.0001999707848519783,
"loss": 4.0007,
"step": 1230
},
{
"epoch": 0.012140523534868712,
"grad_norm": 0.0654296875,
"learning_rate": 0.0001999676288151454,
"loss": 4.0265,
"step": 1240
},
{
"epoch": 0.012543863187189933,
"grad_norm": 0.060302734375,
"learning_rate": 0.00019996431096607438,
"loss": 4.172,
"step": 1250
},
{
"epoch": 0.012624531117654176,
"eval_loss": 4.49122428894043,
"eval_runtime": 3.1912,
"eval_samples_per_second": 62.672,
"eval_steps_per_second": 7.834,
"step": 1252
},
{
"epoch": 0.00032267172185697574,
"grad_norm": 0.08203125,
"learning_rate": 1.4000000000000001e-06,
"loss": 4.1169,
"step": 1260
},
{
"epoch": 0.0007260113741781954,
"grad_norm": 0.1220703125,
"learning_rate": 3.4000000000000005e-06,
"loss": 4.0418,
"step": 1270
},
{
"epoch": 0.001129351026499415,
"grad_norm": 0.1943359375,
"learning_rate": 5.400000000000001e-06,
"loss": 3.9211,
"step": 1280
},
{
"epoch": 0.0015326906788206349,
"grad_norm": 0.328125,
"learning_rate": 7.4e-06,
"loss": 3.6888,
"step": 1290
},
{
"epoch": 0.0019360303311418546,
"grad_norm": 0.375,
"learning_rate": 9.4e-06,
"loss": 3.4768,
"step": 1300
},
{
"epoch": 0.0023393699834630744,
"grad_norm": 0.408203125,
"learning_rate": 1.14e-05,
"loss": 3.242,
"step": 1310
},
{
"epoch": 0.002742709635784294,
"grad_norm": 0.3671875,
"learning_rate": 1.3400000000000002e-05,
"loss": 2.9401,
"step": 1320
},
{
"epoch": 0.0031460492881055136,
"grad_norm": 0.265625,
"learning_rate": 1.54e-05,
"loss": 2.7321,
"step": 1330
},
{
"epoch": 0.0035493889404267336,
"grad_norm": 0.23828125,
"learning_rate": 1.7400000000000003e-05,
"loss": 2.413,
"step": 1340
},
{
"epoch": 0.003952728592747953,
"grad_norm": 0.2353515625,
"learning_rate": 1.94e-05,
"loss": 2.2395,
"step": 1350
},
{
"epoch": 0.004356068245069173,
"grad_norm": 0.2353515625,
"learning_rate": 1.9999999012581816e-05,
"loss": 2.2062,
"step": 1360
},
{
"epoch": 0.004759407897390392,
"grad_norm": 0.232421875,
"learning_rate": 1.999999417624832e-05,
"loss": 2.1174,
"step": 1370
},
{
"epoch": 0.005162747549711612,
"grad_norm": 0.2265625,
"learning_rate": 1.999998530963894e-05,
"loss": 1.9498,
"step": 1380
},
{
"epoch": 0.005566087202032832,
"grad_norm": 0.23046875,
"learning_rate": 1.999997241275724e-05,
"loss": 1.8648,
"step": 1390
},
{
"epoch": 0.005969426854354052,
"grad_norm": 0.2392578125,
"learning_rate": 1.9999955485608426e-05,
"loss": 1.8192,
"step": 1400
},
{
"epoch": 0.006372766506675271,
"grad_norm": 0.296875,
"learning_rate": 1.999993452819932e-05,
"loss": 2.0548,
"step": 1410
},
{
"epoch": 0.006776106158996491,
"grad_norm": 0.34765625,
"learning_rate": 1.999990954053836e-05,
"loss": 3.5804,
"step": 1420
},
{
"epoch": 0.0071794458113177105,
"grad_norm": 0.21484375,
"learning_rate": 1.9999880522635625e-05,
"loss": 3.8265,
"step": 1430
},
{
"epoch": 0.00758278546363893,
"grad_norm": 0.130859375,
"learning_rate": 1.999984747450281e-05,
"loss": 3.7124,
"step": 1440
},
{
"epoch": 0.00798612511596015,
"grad_norm": 0.1142578125,
"learning_rate": 1.9999810396153232e-05,
"loss": 3.7061,
"step": 1450
},
{
"epoch": 0.00838946476828137,
"grad_norm": 0.103515625,
"learning_rate": 1.9999769287601834e-05,
"loss": 3.68,
"step": 1460
},
{
"epoch": 0.008792804420602589,
"grad_norm": 0.09814453125,
"learning_rate": 1.9999724148865183e-05,
"loss": 3.6099,
"step": 1470
},
{
"epoch": 0.00919614407292381,
"grad_norm": 0.09423828125,
"learning_rate": 1.9999674979961473e-05,
"loss": 3.613,
"step": 1480
},
{
"epoch": 0.00959948372524503,
"grad_norm": 0.0966796875,
"learning_rate": 1.999962178091052e-05,
"loss": 3.5696,
"step": 1490
},
{
"epoch": 0.010002823377566248,
"grad_norm": 0.10986328125,
"learning_rate": 1.9999564551733764e-05,
"loss": 3.5333,
"step": 1500
},
{
"epoch": 0.010406163029887469,
"grad_norm": 0.09375,
"learning_rate": 1.9999503292454275e-05,
"loss": 3.5119,
"step": 1510
},
{
"epoch": 0.010809502682208687,
"grad_norm": 0.103515625,
"learning_rate": 1.9999438003096733e-05,
"loss": 3.403,
"step": 1520
},
{
"epoch": 0.011212842334529908,
"grad_norm": 0.10546875,
"learning_rate": 1.9999368683687457e-05,
"loss": 3.3882,
"step": 1530
},
{
"epoch": 0.011616181986851127,
"grad_norm": 0.0986328125,
"learning_rate": 1.999929533425439e-05,
"loss": 3.3576,
"step": 1540
},
{
"epoch": 0.012019521639172347,
"grad_norm": 0.107421875,
"learning_rate": 1.999921795482708e-05,
"loss": 3.4566,
"step": 1550
},
{
"epoch": 0.012422861291493567,
"grad_norm": 0.158203125,
"learning_rate": 1.9999136545436727e-05,
"loss": 3.7767,
"step": 1560
},
{
"epoch": 0.012826200943814786,
"grad_norm": 0.2177734375,
"learning_rate": 1.999905110611613e-05,
"loss": 4.5098,
"step": 1570
},
{
"epoch": 0.013229540596136007,
"grad_norm": 0.1552734375,
"learning_rate": 1.9998961636899736e-05,
"loss": 4.6336,
"step": 1580
},
{
"epoch": 0.013632880248457225,
"grad_norm": 0.12255859375,
"learning_rate": 1.999886813782359e-05,
"loss": 4.5697,
"step": 1590
},
{
"epoch": 0.014036219900778446,
"grad_norm": 0.10498046875,
"learning_rate": 1.999877060892538e-05,
"loss": 4.5598,
"step": 1600
},
{
"epoch": 0.014439559553099664,
"grad_norm": 0.08984375,
"learning_rate": 1.9998669050244416e-05,
"loss": 4.5326,
"step": 1610
},
{
"epoch": 0.014842899205420885,
"grad_norm": 0.07958984375,
"learning_rate": 1.999856346182163e-05,
"loss": 4.5219,
"step": 1620
},
{
"epoch": 0.015246238857742105,
"grad_norm": 0.07177734375,
"learning_rate": 1.999845384369957e-05,
"loss": 4.5013,
"step": 1630
},
{
"epoch": 0.015649578510063326,
"grad_norm": 0.07080078125,
"learning_rate": 1.9998340195922418e-05,
"loss": 4.4859,
"step": 1640
},
{
"epoch": 0.016052918162384543,
"grad_norm": 0.061767578125,
"learning_rate": 1.999822251853598e-05,
"loss": 4.5328,
"step": 1650
},
{
"epoch": 0.016456257814705763,
"grad_norm": 0.05859375,
"learning_rate": 1.9998100811587686e-05,
"loss": 4.5024,
"step": 1660
},
{
"epoch": 0.016859597467026984,
"grad_norm": 0.06201171875,
"learning_rate": 1.9997975075126573e-05,
"loss": 4.5178,
"step": 1670
},
{
"epoch": 0.017262937119348204,
"grad_norm": 0.05908203125,
"learning_rate": 1.9997845309203333e-05,
"loss": 4.4892,
"step": 1680
},
{
"epoch": 0.017666276771669424,
"grad_norm": 0.055908203125,
"learning_rate": 1.9997711513870257e-05,
"loss": 4.4645,
"step": 1690
},
{
"epoch": 0.01806961642399064,
"grad_norm": 0.0546875,
"learning_rate": 1.9997573689181272e-05,
"loss": 4.4891,
"step": 1700
},
{
"epoch": 0.018472956076311862,
"grad_norm": 0.054443359375,
"learning_rate": 1.999743183519192e-05,
"loss": 4.4604,
"step": 1710
},
{
"epoch": 0.018876295728633082,
"grad_norm": 0.05224609375,
"learning_rate": 1.9997285951959372e-05,
"loss": 4.4935,
"step": 1720
},
{
"epoch": 0.019279635380954303,
"grad_norm": 0.052978515625,
"learning_rate": 1.999713603954243e-05,
"loss": 4.4723,
"step": 1730
},
{
"epoch": 0.01968297503327552,
"grad_norm": 0.052734375,
"learning_rate": 1.9996982098001508e-05,
"loss": 4.4923,
"step": 1740
},
{
"epoch": 0.02008631468559674,
"grad_norm": 0.050537109375,
"learning_rate": 1.9996824127398648e-05,
"loss": 4.4402,
"step": 1750
},
{
"epoch": 0.02048965433791796,
"grad_norm": 0.05078125,
"learning_rate": 1.999666212779752e-05,
"loss": 4.4716,
"step": 1760
},
{
"epoch": 0.02089299399023918,
"grad_norm": 0.0478515625,
"learning_rate": 1.999649609926341e-05,
"loss": 4.4952,
"step": 1770
},
{
"epoch": 0.0212963336425604,
"grad_norm": 0.052490234375,
"learning_rate": 1.9996326041863236e-05,
"loss": 4.4755,
"step": 1780
},
{
"epoch": 0.02169967329488162,
"grad_norm": 0.04931640625,
"learning_rate": 1.9996151955665535e-05,
"loss": 4.4645,
"step": 1790
},
{
"epoch": 0.02210301294720284,
"grad_norm": 0.05029296875,
"learning_rate": 1.9995973840740467e-05,
"loss": 4.474,
"step": 1800
},
{
"epoch": 0.02250635259952406,
"grad_norm": 0.05029296875,
"learning_rate": 1.999579169715982e-05,
"loss": 4.4614,
"step": 1810
},
{
"epoch": 0.02290969225184528,
"grad_norm": 0.049560546875,
"learning_rate": 1.9995605524996996e-05,
"loss": 4.4622,
"step": 1820
},
{
"epoch": 0.0233130319041665,
"grad_norm": 0.052490234375,
"learning_rate": 1.9995415324327038e-05,
"loss": 4.4785,
"step": 1830
},
{
"epoch": 0.023716371556487717,
"grad_norm": 0.048828125,
"learning_rate": 1.9995221095226597e-05,
"loss": 4.5104,
"step": 1840
},
{
"epoch": 0.024119711208808937,
"grad_norm": 0.048095703125,
"learning_rate": 1.999502283777395e-05,
"loss": 4.4759,
"step": 1850
},
{
"epoch": 0.024523050861130158,
"grad_norm": 0.0458984375,
"learning_rate": 1.9994820552049002e-05,
"loss": 4.4359,
"step": 1860
},
{
"epoch": 0.02492639051345138,
"grad_norm": 0.045166015625,
"learning_rate": 1.9994614238133282e-05,
"loss": 4.4895,
"step": 1870
},
{
"epoch": 0.0253297301657726,
"grad_norm": 0.04541015625,
"learning_rate": 1.9994403896109942e-05,
"loss": 4.4755,
"step": 1880
},
{
"epoch": 0.025733069818093816,
"grad_norm": 0.045654296875,
"learning_rate": 1.9994189526063746e-05,
"loss": 4.4661,
"step": 1890
},
{
"epoch": 0.026136409470415036,
"grad_norm": 0.046630859375,
"learning_rate": 1.99939711280811e-05,
"loss": 4.4775,
"step": 1900
},
{
"epoch": 0.026539749122736257,
"grad_norm": 0.04931640625,
"learning_rate": 1.999374870225003e-05,
"loss": 4.4954,
"step": 1910
},
{
"epoch": 0.026943088775057477,
"grad_norm": 0.046142578125,
"learning_rate": 1.9993522248660163e-05,
"loss": 4.4905,
"step": 1920
},
{
"epoch": 0.027346428427378694,
"grad_norm": 0.049072265625,
"learning_rate": 1.9993291767402776e-05,
"loss": 4.4545,
"step": 1930
},
{
"epoch": 0.027749768079699914,
"grad_norm": 0.045166015625,
"learning_rate": 1.9993057258570762e-05,
"loss": 4.4737,
"step": 1940
},
{
"epoch": 0.028153107732021135,
"grad_norm": 0.046630859375,
"learning_rate": 1.9992818722258626e-05,
"loss": 4.4908,
"step": 1950
},
{
"epoch": 0.028556447384342355,
"grad_norm": 0.047607421875,
"learning_rate": 1.9992576158562515e-05,
"loss": 4.4893,
"step": 1960
},
{
"epoch": 0.028959787036663576,
"grad_norm": 0.04736328125,
"learning_rate": 1.999232956758018e-05,
"loss": 4.4452,
"step": 1970
},
{
"epoch": 0.029363126688984793,
"grad_norm": 0.044677734375,
"learning_rate": 1.999207894941101e-05,
"loss": 4.4901,
"step": 1980
},
{
"epoch": 0.029766466341306013,
"grad_norm": 0.0478515625,
"learning_rate": 1.9991824304156006e-05,
"loss": 4.419,
"step": 1990
},
{
"epoch": 0.030169805993627234,
"grad_norm": 0.048828125,
"learning_rate": 1.99915656319178e-05,
"loss": 4.4521,
"step": 2000
},
{
"epoch": 0.030573145645948454,
"grad_norm": 0.046875,
"learning_rate": 1.999130293280065e-05,
"loss": 4.4561,
"step": 2010
},
{
"epoch": 0.030976485298269674,
"grad_norm": 0.0595703125,
"learning_rate": 1.9991036206910417e-05,
"loss": 4.4544,
"step": 2020
},
{
"epoch": 0.03137982495059089,
"grad_norm": 0.048095703125,
"learning_rate": 1.999076545435461e-05,
"loss": 4.4713,
"step": 2030
},
{
"epoch": 0.03178316460291211,
"grad_norm": 0.047119140625,
"learning_rate": 1.999049067524235e-05,
"loss": 4.4408,
"step": 2040
},
{
"epoch": 0.03218650425523333,
"grad_norm": 0.04541015625,
"learning_rate": 1.9990211869684374e-05,
"loss": 4.4686,
"step": 2050
},
{
"epoch": 0.03258984390755455,
"grad_norm": 0.047119140625,
"learning_rate": 1.998992903779305e-05,
"loss": 4.4673,
"step": 2060
},
{
"epoch": 0.03299318355987577,
"grad_norm": 0.046630859375,
"learning_rate": 1.9989642179682374e-05,
"loss": 4.4302,
"step": 2070
},
{
"epoch": 0.033396523212196993,
"grad_norm": 0.046875,
"learning_rate": 1.998935129546795e-05,
"loss": 4.4844,
"step": 2080
},
{
"epoch": 0.033799862864518214,
"grad_norm": 0.0439453125,
"learning_rate": 1.9989056385267015e-05,
"loss": 4.4739,
"step": 2090
},
{
"epoch": 0.03420320251683943,
"grad_norm": 0.047119140625,
"learning_rate": 1.9988757449198428e-05,
"loss": 4.499,
"step": 2100
},
{
"epoch": 0.03460654216916065,
"grad_norm": 0.046875,
"learning_rate": 1.9988454487382667e-05,
"loss": 4.4517,
"step": 2110
},
{
"epoch": 0.03500988182148187,
"grad_norm": 0.044677734375,
"learning_rate": 1.9988147499941832e-05,
"loss": 4.4608,
"step": 2120
},
{
"epoch": 0.03541322147380309,
"grad_norm": 0.0458984375,
"learning_rate": 1.998783648699965e-05,
"loss": 4.4515,
"step": 2130
},
{
"epoch": 0.03581656112612431,
"grad_norm": 0.04443359375,
"learning_rate": 1.9987521448681465e-05,
"loss": 4.4646,
"step": 2140
},
{
"epoch": 0.03621990077844553,
"grad_norm": 0.05908203125,
"learning_rate": 1.9987202385114252e-05,
"loss": 4.4586,
"step": 2150
},
{
"epoch": 0.03662324043076675,
"grad_norm": 0.04736328125,
"learning_rate": 1.99868792964266e-05,
"loss": 4.4699,
"step": 2160
},
{
"epoch": 0.03702658008308797,
"grad_norm": 0.044189453125,
"learning_rate": 1.9986552182748715e-05,
"loss": 4.4678,
"step": 2170
},
{
"epoch": 0.03742991973540919,
"grad_norm": 0.04541015625,
"learning_rate": 1.9986221044212442e-05,
"loss": 4.4581,
"step": 2180
},
{
"epoch": 0.037833259387730404,
"grad_norm": 0.045654296875,
"learning_rate": 1.998588588095124e-05,
"loss": 4.4273,
"step": 2190
},
{
"epoch": 0.038236599040051625,
"grad_norm": 0.044677734375,
"learning_rate": 1.9985546693100186e-05,
"loss": 4.4602,
"step": 2200
},
{
"epoch": 0.038639938692372845,
"grad_norm": 0.04931640625,
"learning_rate": 1.9985203480795977e-05,
"loss": 4.4766,
"step": 2210
},
{
"epoch": 0.039043278344694066,
"grad_norm": 0.045166015625,
"learning_rate": 1.9984856244176948e-05,
"loss": 4.4579,
"step": 2220
},
{
"epoch": 0.039446617997015286,
"grad_norm": 0.04638671875,
"learning_rate": 1.998450498338303e-05,
"loss": 4.4661,
"step": 2230
},
{
"epoch": 0.039849957649336507,
"grad_norm": 0.043701171875,
"learning_rate": 1.9984149698555808e-05,
"loss": 4.4415,
"step": 2240
},
{
"epoch": 0.04025329730165773,
"grad_norm": 0.046875,
"learning_rate": 1.998379038983846e-05,
"loss": 4.4461,
"step": 2250
},
{
"epoch": 0.04065663695397895,
"grad_norm": 0.04443359375,
"learning_rate": 1.9983427057375802e-05,
"loss": 4.4627,
"step": 2260
},
{
"epoch": 0.04105997660630017,
"grad_norm": 0.0458984375,
"learning_rate": 1.9983059701314267e-05,
"loss": 4.4399,
"step": 2270
},
{
"epoch": 0.04146331625862139,
"grad_norm": 0.045654296875,
"learning_rate": 1.9982688321801906e-05,
"loss": 4.4673,
"step": 2280
},
{
"epoch": 0.0418666559109426,
"grad_norm": 0.046142578125,
"learning_rate": 1.99823129189884e-05,
"loss": 4.478,
"step": 2290
},
{
"epoch": 0.04226999556326382,
"grad_norm": 0.045166015625,
"learning_rate": 1.9981933493025044e-05,
"loss": 4.4433,
"step": 2300
},
{
"epoch": 0.04267333521558504,
"grad_norm": 0.045166015625,
"learning_rate": 1.9981550044064756e-05,
"loss": 4.4535,
"step": 2310
},
{
"epoch": 0.04307667486790626,
"grad_norm": 0.046142578125,
"learning_rate": 1.998116257226208e-05,
"loss": 4.4665,
"step": 2320
},
{
"epoch": 0.043480014520227483,
"grad_norm": 0.045166015625,
"learning_rate": 1.9980771077773177e-05,
"loss": 4.4419,
"step": 2330
},
{
"epoch": 0.043883354172548704,
"grad_norm": 0.046875,
"learning_rate": 1.9980375560755833e-05,
"loss": 4.4724,
"step": 2340
},
{
"epoch": 0.044286693824869924,
"grad_norm": 0.044921875,
"learning_rate": 1.997997602136944e-05,
"loss": 4.4485,
"step": 2350
},
{
"epoch": 0.044690033477191145,
"grad_norm": 0.044921875,
"learning_rate": 1.997957245977504e-05,
"loss": 4.4484,
"step": 2360
},
{
"epoch": 0.045093373129512365,
"grad_norm": 0.044921875,
"learning_rate": 1.997916487613527e-05,
"loss": 4.4594,
"step": 2370
},
{
"epoch": 0.04549671278183358,
"grad_norm": 0.04638671875,
"learning_rate": 1.9978753270614403e-05,
"loss": 4.4836,
"step": 2380
},
{
"epoch": 0.0459000524341548,
"grad_norm": 0.044921875,
"learning_rate": 1.997833764337832e-05,
"loss": 4.4767,
"step": 2390
},
{
"epoch": 0.04630339208647602,
"grad_norm": 0.04736328125,
"learning_rate": 1.9977917994594537e-05,
"loss": 4.4358,
"step": 2400
},
{
"epoch": 0.04670673173879724,
"grad_norm": 0.04541015625,
"learning_rate": 1.997749432443218e-05,
"loss": 4.4552,
"step": 2410
},
{
"epoch": 0.04711007139111846,
"grad_norm": 0.044677734375,
"learning_rate": 1.9977066633062002e-05,
"loss": 4.4383,
"step": 2420
},
{
"epoch": 0.04751341104343968,
"grad_norm": 0.045654296875,
"learning_rate": 1.9976634920656374e-05,
"loss": 4.4605,
"step": 2430
},
{
"epoch": 0.0479167506957609,
"grad_norm": 0.042724609375,
"learning_rate": 1.9976199187389286e-05,
"loss": 4.4496,
"step": 2440
},
{
"epoch": 0.04832009034808212,
"grad_norm": 0.04638671875,
"learning_rate": 1.997575943343635e-05,
"loss": 4.454,
"step": 2450
},
{
"epoch": 0.04872343000040334,
"grad_norm": 0.044189453125,
"learning_rate": 1.997531565897481e-05,
"loss": 4.4212,
"step": 2460
},
{
"epoch": 0.04912676965272456,
"grad_norm": 0.043212890625,
"learning_rate": 1.9974867864183508e-05,
"loss": 4.4441,
"step": 2470
},
{
"epoch": 0.049530109305045776,
"grad_norm": 0.04541015625,
"learning_rate": 1.997441604924292e-05,
"loss": 4.4085,
"step": 2480
},
{
"epoch": 0.049933448957366997,
"grad_norm": 0.044189453125,
"learning_rate": 1.997396021433514e-05,
"loss": 4.4616,
"step": 2490
},
{
"epoch": 0.05033678860968822,
"grad_norm": 0.04833984375,
"learning_rate": 1.9973500359643885e-05,
"loss": 4.4544,
"step": 2500
},
{
"epoch": 0.05074012826200944,
"grad_norm": 0.044677734375,
"learning_rate": 1.9973036485354485e-05,
"loss": 4.453,
"step": 2510
},
{
"epoch": 0.05114346791433066,
"grad_norm": 0.04638671875,
"learning_rate": 1.99725685916539e-05,
"loss": 4.4722,
"step": 2520
},
{
"epoch": 0.05154680756665188,
"grad_norm": 0.04296875,
"learning_rate": 1.99720966787307e-05,
"loss": 4.4332,
"step": 2530
},
{
"epoch": 0.0519501472189731,
"grad_norm": 0.04638671875,
"learning_rate": 1.9971620746775077e-05,
"loss": 4.4757,
"step": 2540
},
{
"epoch": 0.05235348687129432,
"grad_norm": 0.04638671875,
"learning_rate": 1.997114079597885e-05,
"loss": 4.4264,
"step": 2550
},
{
"epoch": 0.05275682652361554,
"grad_norm": 0.04638671875,
"learning_rate": 1.997065682653545e-05,
"loss": 4.4372,
"step": 2560
},
{
"epoch": 0.05316016617593675,
"grad_norm": 0.045166015625,
"learning_rate": 1.997016883863993e-05,
"loss": 4.4566,
"step": 2570
},
{
"epoch": 0.05356350582825797,
"grad_norm": 0.04345703125,
"learning_rate": 1.9969676832488965e-05,
"loss": 4.4309,
"step": 2580
},
{
"epoch": 0.053966845480579194,
"grad_norm": 0.046142578125,
"learning_rate": 1.9969180808280845e-05,
"loss": 4.4621,
"step": 2590
},
{
"epoch": 0.054370185132900414,
"grad_norm": 0.04443359375,
"learning_rate": 1.9968680766215477e-05,
"loss": 4.4465,
"step": 2600
},
{
"epoch": 0.054773524785221635,
"grad_norm": 0.04541015625,
"learning_rate": 1.9968176706494403e-05,
"loss": 4.4239,
"step": 2610
},
{
"epoch": 0.055176864437542855,
"grad_norm": 0.042724609375,
"learning_rate": 1.996766862932076e-05,
"loss": 4.4567,
"step": 2620
},
{
"epoch": 0.055580204089864076,
"grad_norm": 0.046142578125,
"learning_rate": 1.996715653489933e-05,
"loss": 4.423,
"step": 2630
},
{
"epoch": 0.055983543742185296,
"grad_norm": 0.04248046875,
"learning_rate": 1.9966640423436492e-05,
"loss": 4.3849,
"step": 2640
},
{
"epoch": 0.056386883394506516,
"grad_norm": 0.0458984375,
"learning_rate": 1.9966120295140258e-05,
"loss": 4.4788,
"step": 2650
},
{
"epoch": 0.05679022304682774,
"grad_norm": 0.045166015625,
"learning_rate": 1.996559615022025e-05,
"loss": 4.4313,
"step": 2660
},
{
"epoch": 0.05719356269914895,
"grad_norm": 0.044921875,
"learning_rate": 1.996506798888772e-05,
"loss": 4.4493,
"step": 2670
},
{
"epoch": 0.05759690235147017,
"grad_norm": 0.047607421875,
"learning_rate": 1.9964535811355524e-05,
"loss": 4.4438,
"step": 2680
},
{
"epoch": 0.05800024200379139,
"grad_norm": 0.04296875,
"learning_rate": 1.996399961783815e-05,
"loss": 4.413,
"step": 2690
},
{
"epoch": 0.05840358165611261,
"grad_norm": 0.043701171875,
"learning_rate": 1.9963459408551693e-05,
"loss": 4.4538,
"step": 2700
},
{
"epoch": 0.05880692130843383,
"grad_norm": 0.0439453125,
"learning_rate": 1.996291518371388e-05,
"loss": 4.466,
"step": 2710
},
{
"epoch": 0.05921026096075505,
"grad_norm": 0.045654296875,
"learning_rate": 1.9962366943544045e-05,
"loss": 4.4963,
"step": 2720
},
{
"epoch": 0.05961360061307627,
"grad_norm": 0.04541015625,
"learning_rate": 1.9961814688263138e-05,
"loss": 4.46,
"step": 2730
},
{
"epoch": 0.06001694026539749,
"grad_norm": 0.04345703125,
"learning_rate": 1.9961258418093745e-05,
"loss": 4.4481,
"step": 2740
},
{
"epoch": 0.060420279917718714,
"grad_norm": 0.044677734375,
"learning_rate": 1.9960698133260053e-05,
"loss": 4.442,
"step": 2750
},
{
"epoch": 0.06082361957003993,
"grad_norm": 0.052490234375,
"learning_rate": 1.9960133833987866e-05,
"loss": 4.4473,
"step": 2760
},
{
"epoch": 0.06122695922236115,
"grad_norm": 0.07958984375,
"learning_rate": 1.9959565520504625e-05,
"loss": 4.4286,
"step": 2770
},
{
"epoch": 0.06163029887468237,
"grad_norm": 0.042724609375,
"learning_rate": 1.9958993193039365e-05,
"loss": 4.4616,
"step": 2780
},
{
"epoch": 0.06203363852700359,
"grad_norm": 0.04296875,
"learning_rate": 1.9958416851822755e-05,
"loss": 4.4409,
"step": 2790
},
{
"epoch": 0.06243697817932481,
"grad_norm": 0.042236328125,
"learning_rate": 1.9957836497087074e-05,
"loss": 4.4416,
"step": 2800
},
{
"epoch": 0.06284031783164602,
"grad_norm": 0.04541015625,
"learning_rate": 1.9957252129066227e-05,
"loss": 4.4552,
"step": 2810
},
{
"epoch": 0.06324365748396725,
"grad_norm": 0.046875,
"learning_rate": 1.9956663747995724e-05,
"loss": 4.4536,
"step": 2820
},
{
"epoch": 0.06364699713628846,
"grad_norm": 0.045654296875,
"learning_rate": 1.99560713541127e-05,
"loss": 4.4544,
"step": 2830
},
{
"epoch": 0.06405033678860969,
"grad_norm": 0.0419921875,
"learning_rate": 1.9955474947655912e-05,
"loss": 4.4288,
"step": 2840
},
{
"epoch": 0.0644536764409309,
"grad_norm": 0.041748046875,
"learning_rate": 1.995487452886572e-05,
"loss": 4.4397,
"step": 2850
},
{
"epoch": 0.06485701609325213,
"grad_norm": 0.044677734375,
"learning_rate": 1.995427009798411e-05,
"loss": 4.4909,
"step": 2860
},
{
"epoch": 0.06526035574557335,
"grad_norm": 0.043701171875,
"learning_rate": 1.9953661655254695e-05,
"loss": 4.4528,
"step": 2870
},
{
"epoch": 0.06566369539789457,
"grad_norm": 0.045654296875,
"learning_rate": 1.9953049200922684e-05,
"loss": 4.4308,
"step": 2880
},
{
"epoch": 0.06606703505021579,
"grad_norm": 0.04541015625,
"learning_rate": 1.9952432735234918e-05,
"loss": 4.4585,
"step": 2890
},
{
"epoch": 0.066470374702537,
"grad_norm": 0.046630859375,
"learning_rate": 1.9951812258439846e-05,
"loss": 4.4663,
"step": 2900
},
{
"epoch": 0.06687371435485823,
"grad_norm": 0.0439453125,
"learning_rate": 1.995118777078754e-05,
"loss": 4.4549,
"step": 2910
},
{
"epoch": 0.06727705400717944,
"grad_norm": 0.04931640625,
"learning_rate": 1.9950559272529686e-05,
"loss": 4.4434,
"step": 2920
},
{
"epoch": 0.06768039365950067,
"grad_norm": 0.04443359375,
"learning_rate": 1.9949926763919586e-05,
"loss": 4.4086,
"step": 2930
},
{
"epoch": 0.06808373331182188,
"grad_norm": 0.045166015625,
"learning_rate": 1.9949290245212157e-05,
"loss": 4.4456,
"step": 2940
},
{
"epoch": 0.06848707296414311,
"grad_norm": 0.042724609375,
"learning_rate": 1.9948649716663936e-05,
"loss": 4.4395,
"step": 2950
},
{
"epoch": 0.06889041261646432,
"grad_norm": 0.048095703125,
"learning_rate": 1.994800517853307e-05,
"loss": 4.4357,
"step": 2960
},
{
"epoch": 0.06929375226878555,
"grad_norm": 0.041015625,
"learning_rate": 1.9947356631079337e-05,
"loss": 4.4129,
"step": 2970
},
{
"epoch": 0.06969709192110676,
"grad_norm": 0.042724609375,
"learning_rate": 1.9946704074564105e-05,
"loss": 4.4587,
"step": 2980
},
{
"epoch": 0.07010043157342799,
"grad_norm": 0.04248046875,
"learning_rate": 1.994604750925038e-05,
"loss": 4.4844,
"step": 2990
},
{
"epoch": 0.0705037712257492,
"grad_norm": 0.044921875,
"learning_rate": 1.9945386935402775e-05,
"loss": 4.4578,
"step": 3000
},
{
"epoch": 0.07090711087807042,
"grad_norm": 0.04443359375,
"learning_rate": 1.9944722353287518e-05,
"loss": 4.4306,
"step": 3010
},
{
"epoch": 0.07131045053039164,
"grad_norm": 0.04248046875,
"learning_rate": 1.994405376317246e-05,
"loss": 4.4542,
"step": 3020
},
{
"epoch": 0.07171379018271286,
"grad_norm": 0.044921875,
"learning_rate": 1.9943381165327053e-05,
"loss": 4.4166,
"step": 3030
},
{
"epoch": 0.07211712983503409,
"grad_norm": 0.045166015625,
"learning_rate": 1.9942704560022378e-05,
"loss": 4.4745,
"step": 3040
},
{
"epoch": 0.0725204694873553,
"grad_norm": 0.045166015625,
"learning_rate": 1.9942023947531122e-05,
"loss": 4.4555,
"step": 3050
},
{
"epoch": 0.07292380913967653,
"grad_norm": 0.047119140625,
"learning_rate": 1.99413393281276e-05,
"loss": 4.4133,
"step": 3060
},
{
"epoch": 0.07332714879199774,
"grad_norm": 0.04638671875,
"learning_rate": 1.9940650702087718e-05,
"loss": 4.4555,
"step": 3070
},
{
"epoch": 0.07373048844431897,
"grad_norm": 0.044677734375,
"learning_rate": 1.9939958069689026e-05,
"loss": 4.4011,
"step": 3080
},
{
"epoch": 0.07413382809664018,
"grad_norm": 0.04541015625,
"learning_rate": 1.9939261431210664e-05,
"loss": 4.4595,
"step": 3090
},
{
"epoch": 0.0745371677489614,
"grad_norm": 0.04296875,
"learning_rate": 1.9938560786933398e-05,
"loss": 4.452,
"step": 3100
},
{
"epoch": 0.07494050740128262,
"grad_norm": 0.04345703125,
"learning_rate": 1.9937856137139612e-05,
"loss": 4.4497,
"step": 3110
},
{
"epoch": 0.07534384705360384,
"grad_norm": 0.044677734375,
"learning_rate": 1.9937147482113296e-05,
"loss": 4.4514,
"step": 3120
},
{
"epoch": 0.07574718670592506,
"grad_norm": 0.04541015625,
"learning_rate": 1.993643482214006e-05,
"loss": 4.4674,
"step": 3130
},
{
"epoch": 0.07615052635824628,
"grad_norm": 0.0419921875,
"learning_rate": 1.9935718157507124e-05,
"loss": 4.4503,
"step": 3140
},
{
"epoch": 0.0765538660105675,
"grad_norm": 0.044677734375,
"learning_rate": 1.9934997488503325e-05,
"loss": 4.4512,
"step": 3150
},
{
"epoch": 0.07695720566288872,
"grad_norm": 0.044677734375,
"learning_rate": 1.993427281541911e-05,
"loss": 4.4441,
"step": 3160
},
{
"epoch": 0.07736054531520994,
"grad_norm": 0.044189453125,
"learning_rate": 1.9933544138546542e-05,
"loss": 4.4542,
"step": 3170
},
{
"epoch": 0.07776388496753116,
"grad_norm": 0.0439453125,
"learning_rate": 1.9932811458179305e-05,
"loss": 4.4436,
"step": 3180
},
{
"epoch": 0.07816722461985237,
"grad_norm": 0.047607421875,
"learning_rate": 1.993207477461268e-05,
"loss": 4.4158,
"step": 3190
},
{
"epoch": 0.0785705642721736,
"grad_norm": 0.04443359375,
"learning_rate": 1.993133408814358e-05,
"loss": 4.4518,
"step": 3200
},
{
"epoch": 0.07897390392449481,
"grad_norm": 0.044921875,
"learning_rate": 1.9930589399070515e-05,
"loss": 4.4289,
"step": 3210
},
{
"epoch": 0.07937724357681604,
"grad_norm": 0.045654296875,
"learning_rate": 1.9929840707693618e-05,
"loss": 4.4452,
"step": 3220
},
{
"epoch": 0.07978058322913725,
"grad_norm": 0.042724609375,
"learning_rate": 1.9929088014314636e-05,
"loss": 4.462,
"step": 3230
},
{
"epoch": 0.08018392288145848,
"grad_norm": 0.04345703125,
"learning_rate": 1.992833131923692e-05,
"loss": 4.4282,
"step": 3240
},
{
"epoch": 0.0805872625337797,
"grad_norm": 0.0439453125,
"learning_rate": 1.9927570622765443e-05,
"loss": 4.4584,
"step": 3250
},
{
"epoch": 0.08099060218610092,
"grad_norm": 0.045166015625,
"learning_rate": 1.9926805925206784e-05,
"loss": 4.455,
"step": 3260
},
{
"epoch": 0.08139394183842213,
"grad_norm": 0.043212890625,
"learning_rate": 1.992603722686914e-05,
"loss": 4.4233,
"step": 3270
},
{
"epoch": 0.08179728149074335,
"grad_norm": 0.043701171875,
"learning_rate": 1.9925264528062317e-05,
"loss": 4.4435,
"step": 3280
},
{
"epoch": 0.08220062114306458,
"grad_norm": 0.04541015625,
"learning_rate": 1.9924487829097733e-05,
"loss": 4.4385,
"step": 3290
},
{
"epoch": 0.08260396079538579,
"grad_norm": 0.04248046875,
"learning_rate": 1.9923707130288415e-05,
"loss": 4.4204,
"step": 3300
},
{
"epoch": 0.08300730044770702,
"grad_norm": 0.043701171875,
"learning_rate": 1.9922922431949017e-05,
"loss": 4.4202,
"step": 3310
},
{
"epoch": 0.08341064010002823,
"grad_norm": 0.04296875,
"learning_rate": 1.9922133734395787e-05,
"loss": 4.394,
"step": 3320
},
{
"epoch": 0.08381397975234946,
"grad_norm": 0.04443359375,
"learning_rate": 1.9921341037946592e-05,
"loss": 4.4216,
"step": 3330
},
{
"epoch": 0.08421731940467067,
"grad_norm": 0.044677734375,
"learning_rate": 1.9920544342920913e-05,
"loss": 4.4231,
"step": 3340
},
{
"epoch": 0.0846206590569919,
"grad_norm": 0.0458984375,
"learning_rate": 1.991974364963984e-05,
"loss": 4.4847,
"step": 3350
},
{
"epoch": 0.08502399870931311,
"grad_norm": 0.04248046875,
"learning_rate": 1.9918938958426075e-05,
"loss": 4.4063,
"step": 3360
},
{
"epoch": 0.08542733836163434,
"grad_norm": 0.0458984375,
"learning_rate": 1.9918130269603926e-05,
"loss": 4.4668,
"step": 3370
},
{
"epoch": 0.08583067801395555,
"grad_norm": 0.0458984375,
"learning_rate": 1.991731758349933e-05,
"loss": 4.4454,
"step": 3380
},
{
"epoch": 0.08623401766627677,
"grad_norm": 0.04541015625,
"learning_rate": 1.9916500900439806e-05,
"loss": 4.4527,
"step": 3390
},
{
"epoch": 0.086637357318598,
"grad_norm": 0.042724609375,
"learning_rate": 1.991568022075451e-05,
"loss": 4.4057,
"step": 3400
},
{
"epoch": 0.0870406969709192,
"grad_norm": 0.044189453125,
"learning_rate": 1.9914855544774195e-05,
"loss": 4.4503,
"step": 3410
},
{
"epoch": 0.08744403662324043,
"grad_norm": 0.043212890625,
"learning_rate": 1.991402687283123e-05,
"loss": 4.4968,
"step": 3420
},
{
"epoch": 0.08784737627556165,
"grad_norm": 0.045166015625,
"learning_rate": 1.9913194205259595e-05,
"loss": 4.4642,
"step": 3430
},
{
"epoch": 0.08825071592788288,
"grad_norm": 0.0439453125,
"learning_rate": 1.9912357542394873e-05,
"loss": 4.4283,
"step": 3440
},
{
"epoch": 0.08865405558020409,
"grad_norm": 0.0458984375,
"learning_rate": 1.9911516884574262e-05,
"loss": 4.4776,
"step": 3450
},
{
"epoch": 0.08905739523252532,
"grad_norm": 0.045166015625,
"learning_rate": 1.9910672232136578e-05,
"loss": 4.4578,
"step": 3460
},
{
"epoch": 0.08946073488484653,
"grad_norm": 0.04443359375,
"learning_rate": 1.990982358542223e-05,
"loss": 4.4718,
"step": 3470
},
{
"epoch": 0.08986407453716774,
"grad_norm": 0.043212890625,
"learning_rate": 1.9908970944773255e-05,
"loss": 4.4575,
"step": 3480
},
{
"epoch": 0.09026741418948897,
"grad_norm": 0.04296875,
"learning_rate": 1.9908114310533285e-05,
"loss": 4.4147,
"step": 3490
},
{
"epoch": 0.09067075384181018,
"grad_norm": 0.048095703125,
"learning_rate": 1.990725368304757e-05,
"loss": 4.4363,
"step": 3500
},
{
"epoch": 0.09107409349413141,
"grad_norm": 0.0439453125,
"learning_rate": 1.990638906266297e-05,
"loss": 4.4575,
"step": 3510
},
{
"epoch": 0.09147743314645262,
"grad_norm": 0.043212890625,
"learning_rate": 1.990552044972794e-05,
"loss": 4.4545,
"step": 3520
},
{
"epoch": 0.09188077279877385,
"grad_norm": 0.044189453125,
"learning_rate": 1.9904647844592572e-05,
"loss": 4.4439,
"step": 3530
},
{
"epoch": 0.09228411245109507,
"grad_norm": 0.043701171875,
"learning_rate": 1.9903771247608535e-05,
"loss": 4.4192,
"step": 3540
},
{
"epoch": 0.09268745210341629,
"grad_norm": 0.044677734375,
"learning_rate": 1.9902890659129125e-05,
"loss": 4.4251,
"step": 3550
},
{
"epoch": 0.0930907917557375,
"grad_norm": 0.041259765625,
"learning_rate": 1.990200607950925e-05,
"loss": 4.4339,
"step": 3560
},
{
"epoch": 0.09349413140805872,
"grad_norm": 0.04541015625,
"learning_rate": 1.9901117509105417e-05,
"loss": 4.4635,
"step": 3570
},
{
"epoch": 0.09389747106037995,
"grad_norm": 0.044921875,
"learning_rate": 1.990022494827574e-05,
"loss": 4.4255,
"step": 3580
},
{
"epoch": 0.09430081071270116,
"grad_norm": 0.044677734375,
"learning_rate": 1.9899328397379955e-05,
"loss": 4.4412,
"step": 3590
},
{
"epoch": 0.09470415036502239,
"grad_norm": 0.042724609375,
"learning_rate": 1.989842785677939e-05,
"loss": 4.437,
"step": 3600
},
{
"epoch": 0.0951074900173436,
"grad_norm": 0.04541015625,
"learning_rate": 1.9897523326836987e-05,
"loss": 4.4136,
"step": 3610
},
{
"epoch": 0.09551082966966483,
"grad_norm": 0.047607421875,
"learning_rate": 1.98966148079173e-05,
"loss": 4.4353,
"step": 3620
},
{
"epoch": 0.09591416932198604,
"grad_norm": 0.044677734375,
"learning_rate": 1.989570230038649e-05,
"loss": 4.4331,
"step": 3630
},
{
"epoch": 0.09631750897430727,
"grad_norm": 0.043212890625,
"learning_rate": 1.989478580461232e-05,
"loss": 4.4311,
"step": 3640
},
{
"epoch": 0.09672084862662848,
"grad_norm": 0.04443359375,
"learning_rate": 1.9893865320964162e-05,
"loss": 4.3961,
"step": 3650
},
{
"epoch": 0.0971241882789497,
"grad_norm": 0.044921875,
"learning_rate": 1.9892940849812997e-05,
"loss": 4.4645,
"step": 3660
},
{
"epoch": 0.09752752793127092,
"grad_norm": 0.046142578125,
"learning_rate": 1.9892012391531413e-05,
"loss": 4.495,
"step": 3670
},
{
"epoch": 0.09793086758359214,
"grad_norm": 0.044677734375,
"learning_rate": 1.989107994649361e-05,
"loss": 4.4074,
"step": 3680
},
{
"epoch": 0.09833420723591337,
"grad_norm": 0.04638671875,
"learning_rate": 1.989014351507538e-05,
"loss": 4.441,
"step": 3690
},
{
"epoch": 0.09873754688823458,
"grad_norm": 0.045654296875,
"learning_rate": 1.988920309765413e-05,
"loss": 4.446,
"step": 3700
},
{
"epoch": 0.0991408865405558,
"grad_norm": 0.045166015625,
"learning_rate": 1.9888258694608886e-05,
"loss": 4.4378,
"step": 3710
},
{
"epoch": 0.09954422619287702,
"grad_norm": 0.044189453125,
"learning_rate": 1.988731030632026e-05,
"loss": 4.4068,
"step": 3720
},
{
"epoch": 0.09994756584519825,
"grad_norm": 0.0458984375,
"learning_rate": 1.988635793317048e-05,
"loss": 4.4587,
"step": 3730
},
{
"epoch": 0.10035090549751946,
"grad_norm": 0.042236328125,
"learning_rate": 1.9885401575543384e-05,
"loss": 4.4586,
"step": 3740
},
{
"epoch": 0.10075424514984069,
"grad_norm": 0.044677734375,
"learning_rate": 1.98844412338244e-05,
"loss": 4.4342,
"step": 3750
},
{
"epoch": 0.1011575848021619,
"grad_norm": 0.04638671875,
"learning_rate": 1.9883476908400587e-05,
"loss": 4.4647,
"step": 3760
},
{
"epoch": 0.10156092445448311,
"grad_norm": 0.04345703125,
"learning_rate": 1.9882508599660583e-05,
"loss": 4.4362,
"step": 3770
},
{
"epoch": 0.10196426410680434,
"grad_norm": 0.04248046875,
"learning_rate": 1.9881536307994645e-05,
"loss": 4.4402,
"step": 3780
},
{
"epoch": 0.10236760375912556,
"grad_norm": 0.0439453125,
"learning_rate": 1.9880560033794637e-05,
"loss": 4.4238,
"step": 3790
},
{
"epoch": 0.10277094341144678,
"grad_norm": 0.04443359375,
"learning_rate": 1.9879579777454027e-05,
"loss": 4.4556,
"step": 3800
},
{
"epoch": 0.103174283063768,
"grad_norm": 0.049560546875,
"learning_rate": 1.987859553936788e-05,
"loss": 4.4287,
"step": 3810
},
{
"epoch": 0.10357762271608922,
"grad_norm": 0.045654296875,
"learning_rate": 1.9877607319932872e-05,
"loss": 4.421,
"step": 3820
},
{
"epoch": 0.10398096236841044,
"grad_norm": 0.04638671875,
"learning_rate": 1.9876615119547286e-05,
"loss": 4.4607,
"step": 3830
},
{
"epoch": 0.10438430202073166,
"grad_norm": 0.04638671875,
"learning_rate": 1.9875618938611008e-05,
"loss": 4.4433,
"step": 3840
},
{
"epoch": 0.10478764167305288,
"grad_norm": 0.04443359375,
"learning_rate": 1.987461877752552e-05,
"loss": 4.4713,
"step": 3850
},
{
"epoch": 0.10519098132537409,
"grad_norm": 0.043701171875,
"learning_rate": 1.9873614636693918e-05,
"loss": 4.4631,
"step": 3860
},
{
"epoch": 0.10559432097769532,
"grad_norm": 0.045166015625,
"learning_rate": 1.9872606516520898e-05,
"loss": 4.3911,
"step": 3870
},
{
"epoch": 0.10599766063001653,
"grad_norm": 0.0458984375,
"learning_rate": 1.9871594417412763e-05,
"loss": 4.451,
"step": 3880
},
{
"epoch": 0.10640100028233776,
"grad_norm": 0.04541015625,
"learning_rate": 1.9870578339777416e-05,
"loss": 4.4256,
"step": 3890
},
{
"epoch": 0.10680433993465897,
"grad_norm": 0.04541015625,
"learning_rate": 1.9869558284024363e-05,
"loss": 4.4336,
"step": 3900
},
{
"epoch": 0.1072076795869802,
"grad_norm": 0.044921875,
"learning_rate": 1.9868534250564713e-05,
"loss": 4.464,
"step": 3910
},
{
"epoch": 0.10761101923930141,
"grad_norm": 0.042236328125,
"learning_rate": 1.9867506239811188e-05,
"loss": 4.4258,
"step": 3920
},
{
"epoch": 0.10801435889162264,
"grad_norm": 0.04736328125,
"learning_rate": 1.9866474252178096e-05,
"loss": 4.4037,
"step": 3930
},
{
"epoch": 0.10841769854394386,
"grad_norm": 0.04345703125,
"learning_rate": 1.9865438288081366e-05,
"loss": 4.383,
"step": 3940
},
{
"epoch": 0.10882103819626507,
"grad_norm": 0.0439453125,
"learning_rate": 1.986439834793851e-05,
"loss": 4.4667,
"step": 3950
},
{
"epoch": 0.1092243778485863,
"grad_norm": 0.044677734375,
"learning_rate": 1.986335443216866e-05,
"loss": 4.4176,
"step": 3960
},
{
"epoch": 0.10962771750090751,
"grad_norm": 0.044189453125,
"learning_rate": 1.9862306541192536e-05,
"loss": 4.4293,
"step": 3970
},
{
"epoch": 0.11003105715322874,
"grad_norm": 0.047119140625,
"learning_rate": 1.9861254675432478e-05,
"loss": 4.4302,
"step": 3980
},
{
"epoch": 0.11043439680554995,
"grad_norm": 0.04541015625,
"learning_rate": 1.9860198835312408e-05,
"loss": 4.4271,
"step": 3990
},
{
"epoch": 0.11083773645787118,
"grad_norm": 0.044677734375,
"learning_rate": 1.985913902125786e-05,
"loss": 4.4409,
"step": 4000
},
{
"epoch": 0.11124107611019239,
"grad_norm": 0.04541015625,
"learning_rate": 1.9858075233695974e-05,
"loss": 4.4272,
"step": 4010
},
{
"epoch": 0.11164441576251362,
"grad_norm": 0.0439453125,
"learning_rate": 1.9857007473055482e-05,
"loss": 4.426,
"step": 4020
},
{
"epoch": 0.11204775541483483,
"grad_norm": 0.0439453125,
"learning_rate": 1.9855935739766724e-05,
"loss": 4.4782,
"step": 4030
},
{
"epoch": 0.11245109506715605,
"grad_norm": 0.047119140625,
"learning_rate": 1.9854860034261635e-05,
"loss": 4.4529,
"step": 4040
},
{
"epoch": 0.11285443471947727,
"grad_norm": 0.044677734375,
"learning_rate": 1.9853780356973757e-05,
"loss": 4.413,
"step": 4050
},
{
"epoch": 0.11325777437179849,
"grad_norm": 0.045166015625,
"learning_rate": 1.9852696708338224e-05,
"loss": 4.4179,
"step": 4060
},
{
"epoch": 0.11366111402411971,
"grad_norm": 0.04443359375,
"learning_rate": 1.9851609088791783e-05,
"loss": 4.4679,
"step": 4070
},
{
"epoch": 0.11406445367644093,
"grad_norm": 0.04345703125,
"learning_rate": 1.9850517498772775e-05,
"loss": 4.4556,
"step": 4080
},
{
"epoch": 0.11446779332876215,
"grad_norm": 0.043701171875,
"learning_rate": 1.9849421938721137e-05,
"loss": 4.4282,
"step": 4090
},
{
"epoch": 0.11487113298108337,
"grad_norm": 0.0439453125,
"learning_rate": 1.9848322409078412e-05,
"loss": 4.4264,
"step": 4100
},
{
"epoch": 0.1152744726334046,
"grad_norm": 0.045166015625,
"learning_rate": 1.9847218910287743e-05,
"loss": 4.4565,
"step": 4110
},
{
"epoch": 0.11567781228572581,
"grad_norm": 0.04736328125,
"learning_rate": 1.9846111442793866e-05,
"loss": 4.4284,
"step": 4120
},
{
"epoch": 0.11608115193804702,
"grad_norm": 0.045166015625,
"learning_rate": 1.984500000704313e-05,
"loss": 4.4254,
"step": 4130
},
{
"epoch": 0.11648449159036825,
"grad_norm": 0.0458984375,
"learning_rate": 1.9843884603483464e-05,
"loss": 4.4437,
"step": 4140
},
{
"epoch": 0.11688783124268946,
"grad_norm": 0.0439453125,
"learning_rate": 1.9842765232564415e-05,
"loss": 4.4201,
"step": 4150
},
{
"epoch": 0.11729117089501069,
"grad_norm": 0.044921875,
"learning_rate": 1.9841641894737113e-05,
"loss": 4.442,
"step": 4160
},
{
"epoch": 0.1176945105473319,
"grad_norm": 0.044189453125,
"learning_rate": 1.98405145904543e-05,
"loss": 4.4023,
"step": 4170
},
{
"epoch": 0.11809785019965313,
"grad_norm": 0.044921875,
"learning_rate": 1.9839383320170308e-05,
"loss": 4.4158,
"step": 4180
},
{
"epoch": 0.11850118985197435,
"grad_norm": 0.0439453125,
"learning_rate": 1.9838248084341077e-05,
"loss": 4.4407,
"step": 4190
},
{
"epoch": 0.11890452950429557,
"grad_norm": 0.044189453125,
"learning_rate": 1.9837108883424128e-05,
"loss": 4.4053,
"step": 4200
},
{
"epoch": 0.11930786915661679,
"grad_norm": 0.044189453125,
"learning_rate": 1.98359657178786e-05,
"loss": 4.453,
"step": 4210
},
{
"epoch": 0.11971120880893801,
"grad_norm": 0.0439453125,
"learning_rate": 1.9834818588165216e-05,
"loss": 4.428,
"step": 4220
},
{
"epoch": 0.12011454846125923,
"grad_norm": 0.046875,
"learning_rate": 1.98336674947463e-05,
"loss": 4.4208,
"step": 4230
},
{
"epoch": 0.12051788811358044,
"grad_norm": 0.0439453125,
"learning_rate": 1.9832512438085776e-05,
"loss": 4.4668,
"step": 4240
},
{
"epoch": 0.12092122776590167,
"grad_norm": 0.04345703125,
"learning_rate": 1.9831353418649168e-05,
"loss": 4.4151,
"step": 4250
},
{
"epoch": 0.12132456741822288,
"grad_norm": 0.0478515625,
"learning_rate": 1.9830190436903587e-05,
"loss": 4.4326,
"step": 4260
},
{
"epoch": 0.12172790707054411,
"grad_norm": 0.044677734375,
"learning_rate": 1.982902349331775e-05,
"loss": 4.4254,
"step": 4270
},
{
"epoch": 0.12213124672286532,
"grad_norm": 0.043212890625,
"learning_rate": 1.9827852588361966e-05,
"loss": 4.46,
"step": 4280
},
{
"epoch": 0.12253458637518655,
"grad_norm": 0.04541015625,
"learning_rate": 1.982667772250815e-05,
"loss": 4.4317,
"step": 4290
},
{
"epoch": 0.12293792602750776,
"grad_norm": 0.044677734375,
"learning_rate": 1.9825498896229793e-05,
"loss": 4.4348,
"step": 4300
},
{
"epoch": 0.12334126567982899,
"grad_norm": 0.043701171875,
"learning_rate": 1.9824316110002e-05,
"loss": 4.4246,
"step": 4310
},
{
"epoch": 0.1237446053321502,
"grad_norm": 0.0439453125,
"learning_rate": 1.9823129364301474e-05,
"loss": 4.4576,
"step": 4320
},
{
"epoch": 0.12414794498447142,
"grad_norm": 0.04541015625,
"learning_rate": 1.9821938659606496e-05,
"loss": 4.4288,
"step": 4330
},
{
"epoch": 0.12455128463679264,
"grad_norm": 0.0439453125,
"learning_rate": 1.9820743996396957e-05,
"loss": 4.4153,
"step": 4340
},
{
"epoch": 0.12495462428911386,
"grad_norm": 0.046142578125,
"learning_rate": 1.981954537515434e-05,
"loss": 4.4412,
"step": 4350
},
{
"epoch": 0.12535796394143509,
"grad_norm": 0.0458984375,
"learning_rate": 1.9818342796361723e-05,
"loss": 4.4165,
"step": 4360
},
{
"epoch": 0.1257613035937563,
"grad_norm": 0.046630859375,
"learning_rate": 1.981713626050378e-05,
"loss": 4.4294,
"step": 4370
},
{
"epoch": 0.1261646432460775,
"grad_norm": 0.045166015625,
"learning_rate": 1.9815925768066776e-05,
"loss": 4.4297,
"step": 4380
}
],
"logging_steps": 10,
"max_steps": 49586,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 313,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.014857922095612e+19,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}