llama_ko / trainer_state.json
Gdragon's picture
Upload 12 files
5f8da84 verified
{
"best_global_step": 7210,
"best_metric": 2.6293044090270996,
"best_model_checkpoint": "/workspace/AI/Tunning/FFT/_cpu_save/Llama-3.2-1B-ins-korean_best2/checkpoint-7210",
"epoch": 8.0,
"eval_steps": 103,
"global_step": 8280,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00966533768273529,
"grad_norm": 22.400630950927734,
"learning_rate": 2.1739130434782606e-08,
"loss": 3.6412,
"step": 10
},
{
"epoch": 0.01933067536547058,
"grad_norm": 21.54340934753418,
"learning_rate": 4.589371980676329e-08,
"loss": 3.6599,
"step": 20
},
{
"epoch": 0.02899601304820587,
"grad_norm": 17.44480323791504,
"learning_rate": 7.004830917874397e-08,
"loss": 3.6637,
"step": 30
},
{
"epoch": 0.03866135073094116,
"grad_norm": 14.80634593963623,
"learning_rate": 9.420289855072464e-08,
"loss": 3.6108,
"step": 40
},
{
"epoch": 0.048326688413676454,
"grad_norm": 11.775601387023926,
"learning_rate": 1.1835748792270531e-07,
"loss": 3.6227,
"step": 50
},
{
"epoch": 0.05799202609641174,
"grad_norm": 17.34461212158203,
"learning_rate": 1.4251207729468598e-07,
"loss": 3.611,
"step": 60
},
{
"epoch": 0.06765736377914704,
"grad_norm": 9.619295120239258,
"learning_rate": 1.6666666666666665e-07,
"loss": 3.5769,
"step": 70
},
{
"epoch": 0.07732270146188232,
"grad_norm": 8.995287895202637,
"learning_rate": 1.9082125603864732e-07,
"loss": 3.5653,
"step": 80
},
{
"epoch": 0.08698803914461761,
"grad_norm": 7.953859329223633,
"learning_rate": 2.1497584541062802e-07,
"loss": 3.5337,
"step": 90
},
{
"epoch": 0.09665337682735291,
"grad_norm": 7.673056125640869,
"learning_rate": 2.391304347826087e-07,
"loss": 3.5483,
"step": 100
},
{
"epoch": 0.0995529781321735,
"eval_loss": 3.5090503692626953,
"eval_runtime": 217.1473,
"eval_samples_per_second": 101.636,
"eval_steps_per_second": 4.237,
"step": 103
},
{
"epoch": 0.1063187145100882,
"grad_norm": 7.639675617218018,
"learning_rate": 2.6328502415458936e-07,
"loss": 3.5037,
"step": 110
},
{
"epoch": 0.11598405219282348,
"grad_norm": 7.536130428314209,
"learning_rate": 2.8743961352657006e-07,
"loss": 3.5268,
"step": 120
},
{
"epoch": 0.12564938987555876,
"grad_norm": 7.575876235961914,
"learning_rate": 3.115942028985507e-07,
"loss": 3.4664,
"step": 130
},
{
"epoch": 0.13531472755829407,
"grad_norm": 7.081434726715088,
"learning_rate": 3.357487922705314e-07,
"loss": 3.4784,
"step": 140
},
{
"epoch": 0.14498006524102935,
"grad_norm": 7.422059059143066,
"learning_rate": 3.5990338164251205e-07,
"loss": 3.475,
"step": 150
},
{
"epoch": 0.15464540292376464,
"grad_norm": 7.345850467681885,
"learning_rate": 3.8405797101449274e-07,
"loss": 3.4595,
"step": 160
},
{
"epoch": 0.16431074060649994,
"grad_norm": 7.430078029632568,
"learning_rate": 4.082125603864734e-07,
"loss": 3.4618,
"step": 170
},
{
"epoch": 0.17397607828923523,
"grad_norm": 7.277656078338623,
"learning_rate": 4.323671497584541e-07,
"loss": 3.4182,
"step": 180
},
{
"epoch": 0.18364141597197053,
"grad_norm": 7.314350605010986,
"learning_rate": 4.5652173913043473e-07,
"loss": 3.4357,
"step": 190
},
{
"epoch": 0.19330675365470582,
"grad_norm": 7.0931267738342285,
"learning_rate": 4.806763285024155e-07,
"loss": 3.359,
"step": 200
},
{
"epoch": 0.199105956264347,
"eval_loss": 3.377324342727661,
"eval_runtime": 216.8839,
"eval_samples_per_second": 101.759,
"eval_steps_per_second": 4.242,
"step": 206
},
{
"epoch": 0.2029720913374411,
"grad_norm": 7.167511940002441,
"learning_rate": 5.048309178743962e-07,
"loss": 3.3954,
"step": 210
},
{
"epoch": 0.2126374290201764,
"grad_norm": 6.995325565338135,
"learning_rate": 5.289855072463768e-07,
"loss": 3.3675,
"step": 220
},
{
"epoch": 0.22230276670291169,
"grad_norm": 7.191945552825928,
"learning_rate": 5.531400966183575e-07,
"loss": 3.333,
"step": 230
},
{
"epoch": 0.23196810438564697,
"grad_norm": 7.308806419372559,
"learning_rate": 5.772946859903382e-07,
"loss": 3.3752,
"step": 240
},
{
"epoch": 0.24163344206838228,
"grad_norm": 7.029453277587891,
"learning_rate": 6.014492753623189e-07,
"loss": 3.3537,
"step": 250
},
{
"epoch": 0.25129877975111753,
"grad_norm": 7.261192798614502,
"learning_rate": 6.256038647342995e-07,
"loss": 3.3393,
"step": 260
},
{
"epoch": 0.26096411743385284,
"grad_norm": 7.451327800750732,
"learning_rate": 6.497584541062802e-07,
"loss": 3.3266,
"step": 270
},
{
"epoch": 0.27062945511658815,
"grad_norm": 7.12129020690918,
"learning_rate": 6.739130434782609e-07,
"loss": 3.293,
"step": 280
},
{
"epoch": 0.2802947927993234,
"grad_norm": 7.322017192840576,
"learning_rate": 6.980676328502416e-07,
"loss": 3.3214,
"step": 290
},
{
"epoch": 0.2899601304820587,
"grad_norm": 7.200961589813232,
"learning_rate": 7.222222222222221e-07,
"loss": 3.298,
"step": 300
},
{
"epoch": 0.29865893439652047,
"eval_loss": 3.266552448272705,
"eval_runtime": 217.0856,
"eval_samples_per_second": 101.665,
"eval_steps_per_second": 4.238,
"step": 309
},
{
"epoch": 0.299625468164794,
"grad_norm": 7.080708026885986,
"learning_rate": 7.463768115942028e-07,
"loss": 3.2466,
"step": 310
},
{
"epoch": 0.30929080584752927,
"grad_norm": 7.21260929107666,
"learning_rate": 7.705314009661835e-07,
"loss": 3.2607,
"step": 320
},
{
"epoch": 0.3189561435302646,
"grad_norm": 7.206089973449707,
"learning_rate": 7.946859903381642e-07,
"loss": 3.2278,
"step": 330
},
{
"epoch": 0.3286214812129999,
"grad_norm": 7.031430244445801,
"learning_rate": 8.188405797101448e-07,
"loss": 3.2724,
"step": 340
},
{
"epoch": 0.3382868188957352,
"grad_norm": 7.544980525970459,
"learning_rate": 8.429951690821255e-07,
"loss": 3.2514,
"step": 350
},
{
"epoch": 0.34795215657847045,
"grad_norm": 7.031979084014893,
"learning_rate": 8.671497584541062e-07,
"loss": 3.223,
"step": 360
},
{
"epoch": 0.35761749426120576,
"grad_norm": 7.519832611083984,
"learning_rate": 8.913043478260869e-07,
"loss": 3.2113,
"step": 370
},
{
"epoch": 0.36728283194394107,
"grad_norm": 7.309943675994873,
"learning_rate": 9.154589371980675e-07,
"loss": 3.2434,
"step": 380
},
{
"epoch": 0.3769481696266763,
"grad_norm": 7.484668254852295,
"learning_rate": 9.396135265700482e-07,
"loss": 3.1784,
"step": 390
},
{
"epoch": 0.38661350730941163,
"grad_norm": 7.425693988800049,
"learning_rate": 9.637681159420288e-07,
"loss": 3.224,
"step": 400
},
{
"epoch": 0.39627884499214694,
"grad_norm": 7.163578987121582,
"learning_rate": 9.879227053140095e-07,
"loss": 3.1703,
"step": 410
},
{
"epoch": 0.398211912528694,
"eval_loss": 3.1767160892486572,
"eval_runtime": 217.0763,
"eval_samples_per_second": 101.669,
"eval_steps_per_second": 4.238,
"step": 412
},
{
"epoch": 0.4059441826748822,
"grad_norm": 7.424842357635498,
"learning_rate": 1.0120772946859904e-06,
"loss": 3.1604,
"step": 420
},
{
"epoch": 0.4156095203576175,
"grad_norm": 7.412405967712402,
"learning_rate": 1.036231884057971e-06,
"loss": 3.1459,
"step": 430
},
{
"epoch": 0.4252748580403528,
"grad_norm": 7.094534397125244,
"learning_rate": 1.0603864734299516e-06,
"loss": 3.1701,
"step": 440
},
{
"epoch": 0.43494019572308806,
"grad_norm": 7.796970844268799,
"learning_rate": 1.0845410628019323e-06,
"loss": 3.1437,
"step": 450
},
{
"epoch": 0.44460553340582337,
"grad_norm": 7.447244644165039,
"learning_rate": 1.108695652173913e-06,
"loss": 3.1425,
"step": 460
},
{
"epoch": 0.4542708710885587,
"grad_norm": 7.39740514755249,
"learning_rate": 1.1328502415458937e-06,
"loss": 3.1273,
"step": 470
},
{
"epoch": 0.46393620877129393,
"grad_norm": 7.5719428062438965,
"learning_rate": 1.1570048309178744e-06,
"loss": 3.1535,
"step": 480
},
{
"epoch": 0.47360154645402924,
"grad_norm": 7.861938953399658,
"learning_rate": 1.1811594202898549e-06,
"loss": 3.1265,
"step": 490
},
{
"epoch": 0.48326688413676455,
"grad_norm": 7.544646263122559,
"learning_rate": 1.2053140096618358e-06,
"loss": 3.1087,
"step": 500
},
{
"epoch": 0.4929322218194998,
"grad_norm": 7.783000469207764,
"learning_rate": 1.2294685990338163e-06,
"loss": 3.1148,
"step": 510
},
{
"epoch": 0.4977648906608675,
"eval_loss": 3.1075122356414795,
"eval_runtime": 216.8998,
"eval_samples_per_second": 101.752,
"eval_steps_per_second": 4.242,
"step": 515
},
{
"epoch": 0.5025975595022351,
"grad_norm": 7.453518390655518,
"learning_rate": 1.253623188405797e-06,
"loss": 3.1099,
"step": 520
},
{
"epoch": 0.5122628971849704,
"grad_norm": 7.3546600341796875,
"learning_rate": 1.2777777777777777e-06,
"loss": 3.0977,
"step": 530
},
{
"epoch": 0.5219282348677057,
"grad_norm": 7.336857318878174,
"learning_rate": 1.3019323671497584e-06,
"loss": 3.1053,
"step": 540
},
{
"epoch": 0.5315935725504409,
"grad_norm": 7.137737274169922,
"learning_rate": 1.326086956521739e-06,
"loss": 3.1258,
"step": 550
},
{
"epoch": 0.5412589102331763,
"grad_norm": 7.388270854949951,
"learning_rate": 1.3502415458937198e-06,
"loss": 3.0867,
"step": 560
},
{
"epoch": 0.5509242479159115,
"grad_norm": 7.798702239990234,
"learning_rate": 1.3743961352657005e-06,
"loss": 3.0653,
"step": 570
},
{
"epoch": 0.5605895855986468,
"grad_norm": 7.419099807739258,
"learning_rate": 1.3985507246376811e-06,
"loss": 3.0953,
"step": 580
},
{
"epoch": 0.5702549232813822,
"grad_norm": 7.186514377593994,
"learning_rate": 1.4227053140096618e-06,
"loss": 3.0377,
"step": 590
},
{
"epoch": 0.5799202609641174,
"grad_norm": 7.876040935516357,
"learning_rate": 1.4468599033816423e-06,
"loss": 3.0786,
"step": 600
},
{
"epoch": 0.5895855986468527,
"grad_norm": 7.537592887878418,
"learning_rate": 1.4710144927536232e-06,
"loss": 3.0611,
"step": 610
},
{
"epoch": 0.5973178687930409,
"eval_loss": 3.0519590377807617,
"eval_runtime": 217.0226,
"eval_samples_per_second": 101.694,
"eval_steps_per_second": 4.239,
"step": 618
},
{
"epoch": 0.599250936329588,
"grad_norm": 7.553813457489014,
"learning_rate": 1.4951690821256037e-06,
"loss": 3.0642,
"step": 620
},
{
"epoch": 0.6089162740123233,
"grad_norm": 7.489668369293213,
"learning_rate": 1.5193236714975846e-06,
"loss": 3.0466,
"step": 630
},
{
"epoch": 0.6185816116950585,
"grad_norm": 7.333913803100586,
"learning_rate": 1.5434782608695651e-06,
"loss": 3.0133,
"step": 640
},
{
"epoch": 0.6282469493777939,
"grad_norm": 7.396905422210693,
"learning_rate": 1.5676328502415458e-06,
"loss": 3.0207,
"step": 650
},
{
"epoch": 0.6379122870605292,
"grad_norm": 7.6018266677856445,
"learning_rate": 1.5917874396135265e-06,
"loss": 3.0175,
"step": 660
},
{
"epoch": 0.6475776247432644,
"grad_norm": 7.45857048034668,
"learning_rate": 1.6159420289855072e-06,
"loss": 3.0618,
"step": 670
},
{
"epoch": 0.6572429624259998,
"grad_norm": 7.37148904800415,
"learning_rate": 1.6400966183574877e-06,
"loss": 3.0365,
"step": 680
},
{
"epoch": 0.666908300108735,
"grad_norm": 7.367425918579102,
"learning_rate": 1.6642512077294686e-06,
"loss": 3.0245,
"step": 690
},
{
"epoch": 0.6765736377914704,
"grad_norm": 7.418240547180176,
"learning_rate": 1.688405797101449e-06,
"loss": 3.0269,
"step": 700
},
{
"epoch": 0.6862389754742056,
"grad_norm": 7.5704426765441895,
"learning_rate": 1.71256038647343e-06,
"loss": 3.0419,
"step": 710
},
{
"epoch": 0.6959043131569409,
"grad_norm": 7.671561241149902,
"learning_rate": 1.7367149758454105e-06,
"loss": 3.0498,
"step": 720
},
{
"epoch": 0.6968708469252145,
"eval_loss": 3.005807876586914,
"eval_runtime": 216.8392,
"eval_samples_per_second": 101.781,
"eval_steps_per_second": 4.243,
"step": 721
},
{
"epoch": 0.7055696508396763,
"grad_norm": 7.509618759155273,
"learning_rate": 1.7608695652173912e-06,
"loss": 3.0009,
"step": 730
},
{
"epoch": 0.7152349885224115,
"grad_norm": 7.517166614532471,
"learning_rate": 1.7850241545893719e-06,
"loss": 2.992,
"step": 740
},
{
"epoch": 0.7249003262051468,
"grad_norm": 7.742889881134033,
"learning_rate": 1.8091787439613526e-06,
"loss": 2.9995,
"step": 750
},
{
"epoch": 0.7345656638878821,
"grad_norm": 7.825319290161133,
"learning_rate": 1.833333333333333e-06,
"loss": 3.021,
"step": 760
},
{
"epoch": 0.7442310015706174,
"grad_norm": 7.782094478607178,
"learning_rate": 1.857487922705314e-06,
"loss": 3.0349,
"step": 770
},
{
"epoch": 0.7538963392533526,
"grad_norm": 7.751500606536865,
"learning_rate": 1.8816425120772945e-06,
"loss": 2.9727,
"step": 780
},
{
"epoch": 0.763561676936088,
"grad_norm": 7.506228923797607,
"learning_rate": 1.9057971014492754e-06,
"loss": 2.9381,
"step": 790
},
{
"epoch": 0.7732270146188233,
"grad_norm": 7.2402849197387695,
"learning_rate": 1.9299516908212557e-06,
"loss": 2.9831,
"step": 800
},
{
"epoch": 0.7828923523015585,
"grad_norm": 7.474483013153076,
"learning_rate": 1.9541062801932366e-06,
"loss": 3.0057,
"step": 810
},
{
"epoch": 0.7925576899842939,
"grad_norm": 7.3539934158325195,
"learning_rate": 1.978260869565217e-06,
"loss": 2.969,
"step": 820
},
{
"epoch": 0.796423825057388,
"eval_loss": 2.966066360473633,
"eval_runtime": 216.579,
"eval_samples_per_second": 101.903,
"eval_steps_per_second": 4.248,
"step": 824
},
{
"epoch": 0.8022230276670291,
"grad_norm": 7.149211406707764,
"learning_rate": 1.99999991113637e-06,
"loss": 2.974,
"step": 830
},
{
"epoch": 0.8118883653497644,
"grad_norm": 7.528937339782715,
"learning_rate": 1.9999892475198546e-06,
"loss": 2.9405,
"step": 840
},
{
"epoch": 0.8215537030324997,
"grad_norm": 7.3311614990234375,
"learning_rate": 1.9999608113944566e-06,
"loss": 2.9328,
"step": 850
},
{
"epoch": 0.831219040715235,
"grad_norm": 7.876272678375244,
"learning_rate": 1.9999146032655633e-06,
"loss": 2.9703,
"step": 860
},
{
"epoch": 0.8408843783979703,
"grad_norm": 7.37180757522583,
"learning_rate": 1.999850623954417e-06,
"loss": 2.9538,
"step": 870
},
{
"epoch": 0.8505497160807056,
"grad_norm": 7.159989833831787,
"learning_rate": 1.999768874598104e-06,
"loss": 2.934,
"step": 880
},
{
"epoch": 0.8602150537634409,
"grad_norm": 7.683533191680908,
"learning_rate": 1.9996693566495293e-06,
"loss": 2.9945,
"step": 890
},
{
"epoch": 0.8698803914461761,
"grad_norm": 7.5594353675842285,
"learning_rate": 1.999552071877397e-06,
"loss": 2.9649,
"step": 900
},
{
"epoch": 0.8795457291289115,
"grad_norm": 7.358302116394043,
"learning_rate": 1.999417022366174e-06,
"loss": 2.9244,
"step": 910
},
{
"epoch": 0.8892110668116467,
"grad_norm": 7.427350997924805,
"learning_rate": 1.9992642105160544e-06,
"loss": 2.9356,
"step": 920
},
{
"epoch": 0.8959768031895614,
"eval_loss": 2.9331393241882324,
"eval_runtime": 216.9209,
"eval_samples_per_second": 101.742,
"eval_steps_per_second": 4.241,
"step": 927
},
{
"epoch": 0.898876404494382,
"grad_norm": 7.108737945556641,
"learning_rate": 1.9990936390429174e-06,
"loss": 2.8991,
"step": 930
},
{
"epoch": 0.9085417421771174,
"grad_norm": 7.447180271148682,
"learning_rate": 1.9989053109782786e-06,
"loss": 2.9311,
"step": 940
},
{
"epoch": 0.9182070798598526,
"grad_norm": 7.615334987640381,
"learning_rate": 1.998699229669236e-06,
"loss": 2.9204,
"step": 950
},
{
"epoch": 0.9278724175425879,
"grad_norm": 7.724397659301758,
"learning_rate": 1.9984753987784118e-06,
"loss": 2.9048,
"step": 960
},
{
"epoch": 0.9375377552253232,
"grad_norm": 7.22221565246582,
"learning_rate": 1.9982338222838842e-06,
"loss": 2.9372,
"step": 970
},
{
"epoch": 0.9472030929080585,
"grad_norm": 7.188429832458496,
"learning_rate": 1.9979745044791204e-06,
"loss": 2.9471,
"step": 980
},
{
"epoch": 0.9568684305907937,
"grad_norm": 7.243603706359863,
"learning_rate": 1.9976974499728977e-06,
"loss": 2.9325,
"step": 990
},
{
"epoch": 0.9665337682735291,
"grad_norm": 7.491036891937256,
"learning_rate": 1.9974026636892224e-06,
"loss": 2.9511,
"step": 1000
},
{
"epoch": 0.9761991059562644,
"grad_norm": 7.160167217254639,
"learning_rate": 1.9970901508672435e-06,
"loss": 2.9265,
"step": 1010
},
{
"epoch": 0.9858644436389996,
"grad_norm": 7.02191686630249,
"learning_rate": 1.9967599170611565e-06,
"loss": 2.9277,
"step": 1020
},
{
"epoch": 0.995529781321735,
"grad_norm": 7.492205619812012,
"learning_rate": 1.996411968140109e-06,
"loss": 2.9185,
"step": 1030
},
{
"epoch": 0.995529781321735,
"eval_loss": 2.905233383178711,
"eval_runtime": 216.909,
"eval_samples_per_second": 101.748,
"eval_steps_per_second": 4.241,
"step": 1030
},
{
"epoch": 1.0048326688413676,
"grad_norm": 7.374047756195068,
"learning_rate": 1.996046310288092e-06,
"loss": 2.9196,
"step": 1040
},
{
"epoch": 1.014498006524103,
"grad_norm": 7.255711555480957,
"learning_rate": 1.995662950003833e-06,
"loss": 2.8683,
"step": 1050
},
{
"epoch": 1.0241633442068383,
"grad_norm": 7.150922775268555,
"learning_rate": 1.9952618941006794e-06,
"loss": 2.8806,
"step": 1060
},
{
"epoch": 1.0338286818895734,
"grad_norm": 7.3063883781433105,
"learning_rate": 1.9948431497064772e-06,
"loss": 2.8827,
"step": 1070
},
{
"epoch": 1.0434940195723088,
"grad_norm": 7.305731773376465,
"learning_rate": 1.994406724263445e-06,
"loss": 2.8813,
"step": 1080
},
{
"epoch": 1.0531593572550442,
"grad_norm": 7.360948085784912,
"learning_rate": 1.9939526255280415e-06,
"loss": 2.8888,
"step": 1090
},
{
"epoch": 1.0628246949377793,
"grad_norm": 7.436374187469482,
"learning_rate": 1.993480861570827e-06,
"loss": 2.8621,
"step": 1100
},
{
"epoch": 1.0724900326205147,
"grad_norm": 7.136722564697266,
"learning_rate": 1.9929914407763206e-06,
"loss": 2.8604,
"step": 1110
},
{
"epoch": 1.08215537030325,
"grad_norm": 7.82803201675415,
"learning_rate": 1.9924843718428517e-06,
"loss": 2.8565,
"step": 1120
},
{
"epoch": 1.0918207079859852,
"grad_norm": 7.341372489929199,
"learning_rate": 1.9919596637824044e-06,
"loss": 2.8914,
"step": 1130
},
{
"epoch": 1.0947203092908058,
"eval_loss": 2.883575201034546,
"eval_runtime": 216.9173,
"eval_samples_per_second": 101.744,
"eval_steps_per_second": 4.241,
"step": 1133
},
{
"epoch": 1.1014860456687205,
"grad_norm": 7.483231544494629,
"learning_rate": 1.991417325920457e-06,
"loss": 2.8729,
"step": 1140
},
{
"epoch": 1.111151383351456,
"grad_norm": 7.345571041107178,
"learning_rate": 1.990857367895818e-06,
"loss": 2.8265,
"step": 1150
},
{
"epoch": 1.120816721034191,
"grad_norm": 7.247069835662842,
"learning_rate": 1.9902797996604535e-06,
"loss": 2.8562,
"step": 1160
},
{
"epoch": 1.1304820587169264,
"grad_norm": 7.213609218597412,
"learning_rate": 1.9896846314793106e-06,
"loss": 2.8768,
"step": 1170
},
{
"epoch": 1.1401473963996618,
"grad_norm": 7.366117477416992,
"learning_rate": 1.989071873930134e-06,
"loss": 2.8697,
"step": 1180
},
{
"epoch": 1.149812734082397,
"grad_norm": 7.163483619689941,
"learning_rate": 1.9884415379032807e-06,
"loss": 2.8694,
"step": 1190
},
{
"epoch": 1.1594780717651323,
"grad_norm": 7.064028739929199,
"learning_rate": 1.9877936346015232e-06,
"loss": 2.8417,
"step": 1200
},
{
"epoch": 1.1691434094478677,
"grad_norm": 7.20824670791626,
"learning_rate": 1.987128175539853e-06,
"loss": 2.8613,
"step": 1210
},
{
"epoch": 1.1788087471306028,
"grad_norm": 7.681613922119141,
"learning_rate": 1.9864451725452738e-06,
"loss": 2.8694,
"step": 1220
},
{
"epoch": 1.1884740848133382,
"grad_norm": 7.184121131896973,
"learning_rate": 1.985744637756593e-06,
"loss": 2.8297,
"step": 1230
},
{
"epoch": 1.1942732874229793,
"eval_loss": 2.863403081893921,
"eval_runtime": 216.6561,
"eval_samples_per_second": 101.867,
"eval_steps_per_second": 4.246,
"step": 1236
},
{
"epoch": 1.1981394224960735,
"grad_norm": 7.362229347229004,
"learning_rate": 1.985026583624206e-06,
"loss": 2.876,
"step": 1240
},
{
"epoch": 1.2078047601788087,
"grad_norm": 6.909084320068359,
"learning_rate": 1.9842910229098727e-06,
"loss": 2.8253,
"step": 1250
},
{
"epoch": 1.217470097861544,
"grad_norm": 7.297941207885742,
"learning_rate": 1.983537968686493e-06,
"loss": 2.832,
"step": 1260
},
{
"epoch": 1.2271354355442794,
"grad_norm": 7.0146164894104,
"learning_rate": 1.982767434337874e-06,
"loss": 2.8104,
"step": 1270
},
{
"epoch": 1.2368007732270145,
"grad_norm": 7.416932582855225,
"learning_rate": 1.9819794335584913e-06,
"loss": 2.8184,
"step": 1280
},
{
"epoch": 1.24646611090975,
"grad_norm": 7.231618881225586,
"learning_rate": 1.981173980353246e-06,
"loss": 2.8503,
"step": 1290
},
{
"epoch": 1.2561314485924853,
"grad_norm": 7.189469337463379,
"learning_rate": 1.9803510890372174e-06,
"loss": 2.8199,
"step": 1300
},
{
"epoch": 1.2657967862752204,
"grad_norm": 7.181421279907227,
"learning_rate": 1.9795107742354046e-06,
"loss": 2.8475,
"step": 1310
},
{
"epoch": 1.2754621239579558,
"grad_norm": 7.317245960235596,
"learning_rate": 1.9786530508824715e-06,
"loss": 2.8784,
"step": 1320
},
{
"epoch": 1.2851274616406911,
"grad_norm": 7.532864570617676,
"learning_rate": 1.9777779342224776e-06,
"loss": 2.8484,
"step": 1330
},
{
"epoch": 1.293826265555153,
"eval_loss": 2.8460726737976074,
"eval_runtime": 216.904,
"eval_samples_per_second": 101.75,
"eval_steps_per_second": 4.242,
"step": 1339
},
{
"epoch": 1.2947927993234263,
"grad_norm": 6.849309921264648,
"learning_rate": 1.9768854398086084e-06,
"loss": 2.8427,
"step": 1340
},
{
"epoch": 1.3044581370061616,
"grad_norm": 7.047804355621338,
"learning_rate": 1.9759755835029e-06,
"loss": 2.8141,
"step": 1350
},
{
"epoch": 1.314123474688897,
"grad_norm": 7.175754547119141,
"learning_rate": 1.975048381475955e-06,
"loss": 2.8576,
"step": 1360
},
{
"epoch": 1.3237888123716322,
"grad_norm": 7.228835105895996,
"learning_rate": 1.9741038502066566e-06,
"loss": 2.8288,
"step": 1370
},
{
"epoch": 1.3334541500543675,
"grad_norm": 7.117552757263184,
"learning_rate": 1.9731420064818765e-06,
"loss": 2.8667,
"step": 1380
},
{
"epoch": 1.3431194877371029,
"grad_norm": 7.141602039337158,
"learning_rate": 1.9721628673961743e-06,
"loss": 2.8021,
"step": 1390
},
{
"epoch": 1.352784825419838,
"grad_norm": 7.065165042877197,
"learning_rate": 1.971166450351495e-06,
"loss": 2.8376,
"step": 1400
},
{
"epoch": 1.3624501631025734,
"grad_norm": 7.072872638702393,
"learning_rate": 1.97015277305686e-06,
"loss": 2.8373,
"step": 1410
},
{
"epoch": 1.3721155007853088,
"grad_norm": 6.841920375823975,
"learning_rate": 1.9691218535280517e-06,
"loss": 2.8458,
"step": 1420
},
{
"epoch": 1.381780838468044,
"grad_norm": 7.044722557067871,
"learning_rate": 1.968073710087294e-06,
"loss": 2.8373,
"step": 1430
},
{
"epoch": 1.3914461761507793,
"grad_norm": 7.605813980102539,
"learning_rate": 1.9670083613629245e-06,
"loss": 2.8032,
"step": 1440
},
{
"epoch": 1.3933792436873262,
"eval_loss": 2.8305492401123047,
"eval_runtime": 216.7183,
"eval_samples_per_second": 101.837,
"eval_steps_per_second": 4.245,
"step": 1442
},
{
"epoch": 1.4011115138335146,
"grad_norm": 7.031517505645752,
"learning_rate": 1.965925826289068e-06,
"loss": 2.8395,
"step": 1450
},
{
"epoch": 1.4107768515162498,
"grad_norm": 7.169219970703125,
"learning_rate": 1.9648261241052952e-06,
"loss": 2.8244,
"step": 1460
},
{
"epoch": 1.4204421891989851,
"grad_norm": 7.192759037017822,
"learning_rate": 1.9637092743562823e-06,
"loss": 2.8123,
"step": 1470
},
{
"epoch": 1.4301075268817205,
"grad_norm": 7.061282634735107,
"learning_rate": 1.9625752968914656e-06,
"loss": 2.7968,
"step": 1480
},
{
"epoch": 1.4397728645644556,
"grad_norm": 7.095070838928223,
"learning_rate": 1.9614242118646858e-06,
"loss": 2.8005,
"step": 1490
},
{
"epoch": 1.449438202247191,
"grad_norm": 6.964037895202637,
"learning_rate": 1.9602560397338325e-06,
"loss": 2.7837,
"step": 1500
},
{
"epoch": 1.4591035399299264,
"grad_norm": 6.997467517852783,
"learning_rate": 1.959070801260478e-06,
"loss": 2.8355,
"step": 1510
},
{
"epoch": 1.4687688776126615,
"grad_norm": 6.82883882522583,
"learning_rate": 1.9578685175095098e-06,
"loss": 2.817,
"step": 1520
},
{
"epoch": 1.4784342152953969,
"grad_norm": 6.964229583740234,
"learning_rate": 1.9566492098487572e-06,
"loss": 2.7913,
"step": 1530
},
{
"epoch": 1.4880995529781322,
"grad_norm": 6.858983516693115,
"learning_rate": 1.9554128999486085e-06,
"loss": 2.7997,
"step": 1540
},
{
"epoch": 1.4929322218194998,
"eval_loss": 2.816601037979126,
"eval_runtime": 217.0118,
"eval_samples_per_second": 101.7,
"eval_steps_per_second": 4.239,
"step": 1545
},
{
"epoch": 1.4977648906608674,
"grad_norm": 6.959181308746338,
"learning_rate": 1.954159609781629e-06,
"loss": 2.7839,
"step": 1550
},
{
"epoch": 1.5074302283436027,
"grad_norm": 6.717589855194092,
"learning_rate": 1.952889361622169e-06,
"loss": 2.7969,
"step": 1560
},
{
"epoch": 1.517095566026338,
"grad_norm": 6.919014930725098,
"learning_rate": 1.9516021780459678e-06,
"loss": 2.7675,
"step": 1570
},
{
"epoch": 1.5267609037090732,
"grad_norm": 7.024904251098633,
"learning_rate": 1.950298081929753e-06,
"loss": 2.7854,
"step": 1580
},
{
"epoch": 1.5364262413918086,
"grad_norm": 7.037530422210693,
"learning_rate": 1.9489770964508335e-06,
"loss": 2.7911,
"step": 1590
},
{
"epoch": 1.546091579074544,
"grad_norm": 6.958497524261475,
"learning_rate": 1.947639245086688e-06,
"loss": 2.801,
"step": 1600
},
{
"epoch": 1.5557569167572791,
"grad_norm": 6.978152751922607,
"learning_rate": 1.9462845516145473e-06,
"loss": 2.7855,
"step": 1610
},
{
"epoch": 1.5654222544400145,
"grad_norm": 6.826775550842285,
"learning_rate": 1.944913040110972e-06,
"loss": 2.7614,
"step": 1620
},
{
"epoch": 1.5750875921227498,
"grad_norm": 6.996805191040039,
"learning_rate": 1.943524734951425e-06,
"loss": 2.8176,
"step": 1630
},
{
"epoch": 1.584752929805485,
"grad_norm": 7.008688449859619,
"learning_rate": 1.942119660809837e-06,
"loss": 2.7718,
"step": 1640
},
{
"epoch": 1.5924851999516734,
"eval_loss": 2.8026342391967773,
"eval_runtime": 216.9019,
"eval_samples_per_second": 101.751,
"eval_steps_per_second": 4.242,
"step": 1648
},
{
"epoch": 1.5944182674882204,
"grad_norm": 7.19225549697876,
"learning_rate": 1.940697842658169e-06,
"loss": 2.7797,
"step": 1650
},
{
"epoch": 1.6040836051709557,
"grad_norm": 7.009806156158447,
"learning_rate": 1.939259305765969e-06,
"loss": 2.7773,
"step": 1660
},
{
"epoch": 1.6137489428536909,
"grad_norm": 6.812699317932129,
"learning_rate": 1.937804075699921e-06,
"loss": 2.7684,
"step": 1670
},
{
"epoch": 1.6234142805364262,
"grad_norm": 7.124710559844971,
"learning_rate": 1.936332178323393e-06,
"loss": 2.789,
"step": 1680
},
{
"epoch": 1.6330796182191616,
"grad_norm": 7.376682758331299,
"learning_rate": 1.934843639795975e-06,
"loss": 2.801,
"step": 1690
},
{
"epoch": 1.6427449559018967,
"grad_norm": 6.929679870605469,
"learning_rate": 1.933338486573016e-06,
"loss": 2.8004,
"step": 1700
},
{
"epoch": 1.652410293584632,
"grad_norm": 7.01241397857666,
"learning_rate": 1.9318167454051523e-06,
"loss": 2.7923,
"step": 1710
},
{
"epoch": 1.6620756312673675,
"grad_norm": 6.667963981628418,
"learning_rate": 1.930278443337833e-06,
"loss": 2.7892,
"step": 1720
},
{
"epoch": 1.6717409689501026,
"grad_norm": 6.875312805175781,
"learning_rate": 1.92872360771084e-06,
"loss": 2.7754,
"step": 1730
},
{
"epoch": 1.681406306632838,
"grad_norm": 6.666006565093994,
"learning_rate": 1.9271522661577995e-06,
"loss": 2.7922,
"step": 1740
},
{
"epoch": 1.6910716443155733,
"grad_norm": 7.197988986968994,
"learning_rate": 1.925564446605694e-06,
"loss": 2.7951,
"step": 1750
},
{
"epoch": 1.692038178083847,
"eval_loss": 2.791018009185791,
"eval_runtime": 216.9287,
"eval_samples_per_second": 101.739,
"eval_steps_per_second": 4.241,
"step": 1751
},
{
"epoch": 1.7007369819983085,
"grad_norm": 6.819860935211182,
"learning_rate": 1.9239601772743645e-06,
"loss": 2.7526,
"step": 1760
},
{
"epoch": 1.7104023196810438,
"grad_norm": 6.862004280090332,
"learning_rate": 1.9223394866760073e-06,
"loss": 2.7449,
"step": 1770
},
{
"epoch": 1.7200676573637792,
"grad_norm": 6.870265007019043,
"learning_rate": 1.920702403614671e-06,
"loss": 2.7771,
"step": 1780
},
{
"epoch": 1.7297329950465143,
"grad_norm": 7.154959678649902,
"learning_rate": 1.9190489571857423e-06,
"loss": 2.8067,
"step": 1790
},
{
"epoch": 1.7393983327292497,
"grad_norm": 6.809077739715576,
"learning_rate": 1.9173791767754272e-06,
"loss": 2.7733,
"step": 1800
},
{
"epoch": 1.749063670411985,
"grad_norm": 6.7708282470703125,
"learning_rate": 1.915693092060232e-06,
"loss": 2.7649,
"step": 1810
},
{
"epoch": 1.7587290080947202,
"grad_norm": 7.031125068664551,
"learning_rate": 1.913990733006436e-06,
"loss": 2.7686,
"step": 1820
},
{
"epoch": 1.7683943457774556,
"grad_norm": 6.831878662109375,
"learning_rate": 1.912272129869554e-06,
"loss": 2.7496,
"step": 1830
},
{
"epoch": 1.778059683460191,
"grad_norm": 6.8769097328186035,
"learning_rate": 1.9105373131938057e-06,
"loss": 2.7351,
"step": 1840
},
{
"epoch": 1.787725021142926,
"grad_norm": 6.5916266441345215,
"learning_rate": 1.9087863138115665e-06,
"loss": 2.7503,
"step": 1850
},
{
"epoch": 1.7915911562160203,
"eval_loss": 2.7797155380249023,
"eval_runtime": 216.9502,
"eval_samples_per_second": 101.728,
"eval_steps_per_second": 4.241,
"step": 1854
},
{
"epoch": 1.7973903588256614,
"grad_norm": 6.822347164154053,
"learning_rate": 1.907019162842823e-06,
"loss": 2.7729,
"step": 1860
},
{
"epoch": 1.8070556965083968,
"grad_norm": 7.012885093688965,
"learning_rate": 1.9052358916946192e-06,
"loss": 2.7429,
"step": 1870
},
{
"epoch": 1.816721034191132,
"grad_norm": 6.920743942260742,
"learning_rate": 1.9034365320604987e-06,
"loss": 2.7963,
"step": 1880
},
{
"epoch": 1.8263863718738673,
"grad_norm": 7.041555404663086,
"learning_rate": 1.90162111591994e-06,
"loss": 2.752,
"step": 1890
},
{
"epoch": 1.8360517095566027,
"grad_norm": 6.854340076446533,
"learning_rate": 1.8997896755377898e-06,
"loss": 2.786,
"step": 1900
},
{
"epoch": 1.8457170472393378,
"grad_norm": 7.011345386505127,
"learning_rate": 1.897942243463688e-06,
"loss": 2.7627,
"step": 1910
},
{
"epoch": 1.8553823849220732,
"grad_norm": 6.391646385192871,
"learning_rate": 1.8960788525314912e-06,
"loss": 2.7706,
"step": 1920
},
{
"epoch": 1.8650477226048086,
"grad_norm": 6.750019550323486,
"learning_rate": 1.8941995358586874e-06,
"loss": 2.7245,
"step": 1930
},
{
"epoch": 1.8747130602875437,
"grad_norm": 6.734500408172607,
"learning_rate": 1.8923043268458074e-06,
"loss": 2.7494,
"step": 1940
},
{
"epoch": 1.884378397970279,
"grad_norm": 6.839148998260498,
"learning_rate": 1.8903932591758328e-06,
"loss": 2.7551,
"step": 1950
},
{
"epoch": 1.8911441343481938,
"eval_loss": 2.7689898014068604,
"eval_runtime": 216.9508,
"eval_samples_per_second": 101.728,
"eval_steps_per_second": 4.241,
"step": 1957
},
{
"epoch": 1.8940437356530144,
"grad_norm": 6.925171852111816,
"learning_rate": 1.888466366813596e-06,
"loss": 2.7679,
"step": 1960
},
{
"epoch": 1.9037090733357496,
"grad_norm": 7.017655849456787,
"learning_rate": 1.8865236840051769e-06,
"loss": 2.7386,
"step": 1970
},
{
"epoch": 1.913374411018485,
"grad_norm": 6.900379180908203,
"learning_rate": 1.8845652452772938e-06,
"loss": 2.7741,
"step": 1980
},
{
"epoch": 1.9230397487012203,
"grad_norm": 6.903608798980713,
"learning_rate": 1.8825910854366912e-06,
"loss": 2.7476,
"step": 1990
},
{
"epoch": 1.9327050863839554,
"grad_norm": 6.967295169830322,
"learning_rate": 1.8806012395695193e-06,
"loss": 2.7653,
"step": 2000
},
{
"epoch": 1.9423704240666908,
"grad_norm": 6.658292770385742,
"learning_rate": 1.8785957430407116e-06,
"loss": 2.7692,
"step": 2010
},
{
"epoch": 1.9520357617494262,
"grad_norm": 6.8208909034729,
"learning_rate": 1.8765746314933564e-06,
"loss": 2.7523,
"step": 2020
},
{
"epoch": 1.9617010994321613,
"grad_norm": 6.817293643951416,
"learning_rate": 1.8745379408480631e-06,
"loss": 2.7417,
"step": 2030
},
{
"epoch": 1.9713664371148967,
"grad_norm": 6.492961406707764,
"learning_rate": 1.8724857073023232e-06,
"loss": 2.7474,
"step": 2040
},
{
"epoch": 1.981031774797632,
"grad_norm": 6.4757585525512695,
"learning_rate": 1.8704179673298678e-06,
"loss": 2.7271,
"step": 2050
},
{
"epoch": 1.9906971124803672,
"grad_norm": 6.683932781219482,
"learning_rate": 1.8683347576800194e-06,
"loss": 2.763,
"step": 2060
},
{
"epoch": 1.9906971124803672,
"eval_loss": 2.759366750717163,
"eval_runtime": 216.6345,
"eval_samples_per_second": 101.877,
"eval_steps_per_second": 4.247,
"step": 2060
},
{
"epoch": 2.0,
"grad_norm": 9.310201644897461,
"learning_rate": 1.8662361153770377e-06,
"loss": 2.7428,
"step": 2070
},
{
"epoch": 2.009665337682735,
"grad_norm": 6.855327129364014,
"learning_rate": 1.8641220777194628e-06,
"loss": 2.6795,
"step": 2080
},
{
"epoch": 2.0193306753654707,
"grad_norm": 6.812604904174805,
"learning_rate": 1.8619926822794519e-06,
"loss": 2.6903,
"step": 2090
},
{
"epoch": 2.028996013048206,
"grad_norm": 6.613392353057861,
"learning_rate": 1.8598479669021114e-06,
"loss": 2.7386,
"step": 2100
},
{
"epoch": 2.038661350730941,
"grad_norm": 6.695821762084961,
"learning_rate": 1.857687969704824e-06,
"loss": 2.6984,
"step": 2110
},
{
"epoch": 2.0483266884136766,
"grad_norm": 6.728259563446045,
"learning_rate": 1.8555127290765717e-06,
"loss": 2.7247,
"step": 2120
},
{
"epoch": 2.0579920260964117,
"grad_norm": 6.941239833831787,
"learning_rate": 1.853322283677254e-06,
"loss": 2.6862,
"step": 2130
},
{
"epoch": 2.067657363779147,
"grad_norm": 6.898434162139893,
"learning_rate": 1.8511166724369996e-06,
"loss": 2.6707,
"step": 2140
},
{
"epoch": 2.0773227014618825,
"grad_norm": 6.764381408691406,
"learning_rate": 1.8488959345554754e-06,
"loss": 2.6758,
"step": 2150
},
{
"epoch": 2.0869880391446176,
"grad_norm": 6.8859639167785645,
"learning_rate": 1.8466601095011893e-06,
"loss": 2.7344,
"step": 2160
},
{
"epoch": 2.0898876404494384,
"eval_loss": 2.7534408569335938,
"eval_runtime": 216.6942,
"eval_samples_per_second": 101.849,
"eval_steps_per_second": 4.246,
"step": 2163
},
{
"epoch": 2.0966533768273528,
"grad_norm": 6.86688756942749,
"learning_rate": 1.8444092370107897e-06,
"loss": 2.7162,
"step": 2170
},
{
"epoch": 2.1063187145100883,
"grad_norm": 6.992837905883789,
"learning_rate": 1.8421433570883576e-06,
"loss": 2.7357,
"step": 2180
},
{
"epoch": 2.1159840521928235,
"grad_norm": 6.954036712646484,
"learning_rate": 1.8398625100046971e-06,
"loss": 2.7045,
"step": 2190
},
{
"epoch": 2.1256493898755586,
"grad_norm": 6.828469753265381,
"learning_rate": 1.8375667362966193e-06,
"loss": 2.6887,
"step": 2200
},
{
"epoch": 2.135314727558294,
"grad_norm": 6.695276260375977,
"learning_rate": 1.8352560767662217e-06,
"loss": 2.6672,
"step": 2210
},
{
"epoch": 2.1449800652410294,
"grad_norm": 6.662326335906982,
"learning_rate": 1.8329305724801624e-06,
"loss": 2.6651,
"step": 2220
},
{
"epoch": 2.1546454029237645,
"grad_norm": 6.725667953491211,
"learning_rate": 1.830590264768932e-06,
"loss": 2.6712,
"step": 2230
},
{
"epoch": 2.1643107406065,
"grad_norm": 6.9403395652771,
"learning_rate": 1.8282351952261167e-06,
"loss": 2.6914,
"step": 2240
},
{
"epoch": 2.1739760782892352,
"grad_norm": 6.7621636390686035,
"learning_rate": 1.8258654057076614e-06,
"loss": 2.7119,
"step": 2250
},
{
"epoch": 2.1836414159719704,
"grad_norm": 6.9445719718933105,
"learning_rate": 1.823480938331124e-06,
"loss": 2.7022,
"step": 2260
},
{
"epoch": 2.1894406185816115,
"eval_loss": 2.7453184127807617,
"eval_runtime": 216.9137,
"eval_samples_per_second": 101.746,
"eval_steps_per_second": 4.241,
"step": 2266
},
{
"epoch": 2.193306753654706,
"grad_norm": 6.655174255371094,
"learning_rate": 1.8210818354749279e-06,
"loss": 2.7272,
"step": 2270
},
{
"epoch": 2.202972091337441,
"grad_norm": 6.8438262939453125,
"learning_rate": 1.818668139777608e-06,
"loss": 2.707,
"step": 2280
},
{
"epoch": 2.2126374290201762,
"grad_norm": 6.986091613769531,
"learning_rate": 1.8162398941370545e-06,
"loss": 2.7106,
"step": 2290
},
{
"epoch": 2.222302766702912,
"grad_norm": 6.6938347816467285,
"learning_rate": 1.813797141709748e-06,
"loss": 2.7015,
"step": 2300
},
{
"epoch": 2.231968104385647,
"grad_norm": 6.803264617919922,
"learning_rate": 1.811339925909995e-06,
"loss": 2.7209,
"step": 2310
},
{
"epoch": 2.241633442068382,
"grad_norm": 6.701619625091553,
"learning_rate": 1.8088682904091543e-06,
"loss": 2.6589,
"step": 2320
},
{
"epoch": 2.2512987797511177,
"grad_norm": 6.809988498687744,
"learning_rate": 1.8063822791348624e-06,
"loss": 2.7053,
"step": 2330
},
{
"epoch": 2.260964117433853,
"grad_norm": 6.671308517456055,
"learning_rate": 1.803881936270252e-06,
"loss": 2.7065,
"step": 2340
},
{
"epoch": 2.270629455116588,
"grad_norm": 6.887976169586182,
"learning_rate": 1.8013673062531663e-06,
"loss": 2.712,
"step": 2350
},
{
"epoch": 2.2802947927993236,
"grad_norm": 6.573399543762207,
"learning_rate": 1.7988384337753702e-06,
"loss": 2.6855,
"step": 2360
},
{
"epoch": 2.288993596713785,
"eval_loss": 2.737840175628662,
"eval_runtime": 217.0075,
"eval_samples_per_second": 101.702,
"eval_steps_per_second": 4.239,
"step": 2369
},
{
"epoch": 2.2899601304820587,
"grad_norm": 6.888051509857178,
"learning_rate": 1.7962953637817556e-06,
"loss": 2.6826,
"step": 2370
},
{
"epoch": 2.299625468164794,
"grad_norm": 6.976341724395752,
"learning_rate": 1.7937381414695428e-06,
"loss": 2.688,
"step": 2380
},
{
"epoch": 2.3092908058475294,
"grad_norm": 6.785758018493652,
"learning_rate": 1.7911668122874764e-06,
"loss": 2.6971,
"step": 2390
},
{
"epoch": 2.3189561435302646,
"grad_norm": 6.791547775268555,
"learning_rate": 1.7885814219350187e-06,
"loss": 2.7279,
"step": 2400
},
{
"epoch": 2.3286214812129997,
"grad_norm": 6.7897820472717285,
"learning_rate": 1.7859820163615367e-06,
"loss": 2.697,
"step": 2410
},
{
"epoch": 2.3382868188957353,
"grad_norm": 6.9951395988464355,
"learning_rate": 1.7833686417654857e-06,
"loss": 2.6994,
"step": 2420
},
{
"epoch": 2.3479521565784705,
"grad_norm": 6.830650806427002,
"learning_rate": 1.7807413445935886e-06,
"loss": 2.6833,
"step": 2430
},
{
"epoch": 2.3576174942612056,
"grad_norm": 6.746628761291504,
"learning_rate": 1.7781001715400094e-06,
"loss": 2.699,
"step": 2440
},
{
"epoch": 2.367282831943941,
"grad_norm": 6.582501411437988,
"learning_rate": 1.7754451695455248e-06,
"loss": 2.6798,
"step": 2450
},
{
"epoch": 2.3769481696266763,
"grad_norm": 6.743264675140381,
"learning_rate": 1.7727763857966887e-06,
"loss": 2.6569,
"step": 2460
},
{
"epoch": 2.3866135073094115,
"grad_norm": 6.615793228149414,
"learning_rate": 1.7700938677249934e-06,
"loss": 2.6972,
"step": 2470
},
{
"epoch": 2.3885465748459587,
"eval_loss": 2.7308976650238037,
"eval_runtime": 217.0492,
"eval_samples_per_second": 101.682,
"eval_steps_per_second": 4.239,
"step": 2472
},
{
"epoch": 2.396278844992147,
"grad_norm": 6.7577948570251465,
"learning_rate": 1.7673976630060287e-06,
"loss": 2.6751,
"step": 2480
},
{
"epoch": 2.405944182674882,
"grad_norm": 6.334952354431152,
"learning_rate": 1.7646878195586311e-06,
"loss": 2.679,
"step": 2490
},
{
"epoch": 2.4156095203576173,
"grad_norm": 7.024070739746094,
"learning_rate": 1.7619643855440362e-06,
"loss": 2.6967,
"step": 2500
},
{
"epoch": 2.425274858040353,
"grad_norm": 6.562467575073242,
"learning_rate": 1.7592274093650191e-06,
"loss": 2.6722,
"step": 2510
},
{
"epoch": 2.434940195723088,
"grad_norm": 6.788969993591309,
"learning_rate": 1.7564769396650366e-06,
"loss": 2.6808,
"step": 2520
},
{
"epoch": 2.444605533405823,
"grad_norm": 6.727795124053955,
"learning_rate": 1.753713025327361e-06,
"loss": 2.6603,
"step": 2530
},
{
"epoch": 2.454270871088559,
"grad_norm": 6.815580368041992,
"learning_rate": 1.7509357154742132e-06,
"loss": 2.662,
"step": 2540
},
{
"epoch": 2.463936208771294,
"grad_norm": 6.863372802734375,
"learning_rate": 1.7481450594658873e-06,
"loss": 2.678,
"step": 2550
},
{
"epoch": 2.473601546454029,
"grad_norm": 6.5276641845703125,
"learning_rate": 1.7453411068998762e-06,
"loss": 2.6925,
"step": 2560
},
{
"epoch": 2.4832668841367647,
"grad_norm": 6.627063274383545,
"learning_rate": 1.7425239076099866e-06,
"loss": 2.6452,
"step": 2570
},
{
"epoch": 2.4880995529781322,
"eval_loss": 2.7234959602355957,
"eval_runtime": 216.9128,
"eval_samples_per_second": 101.746,
"eval_steps_per_second": 4.241,
"step": 2575
},
{
"epoch": 2.4929322218195,
"grad_norm": 6.625070095062256,
"learning_rate": 1.739693511665457e-06,
"loss": 2.6579,
"step": 2580
},
{
"epoch": 2.502597559502235,
"grad_norm": 6.858110427856445,
"learning_rate": 1.7368499693700652e-06,
"loss": 2.6886,
"step": 2590
},
{
"epoch": 2.5122628971849705,
"grad_norm": 6.678401470184326,
"learning_rate": 1.7339933312612352e-06,
"loss": 2.6669,
"step": 2600
},
{
"epoch": 2.5219282348677057,
"grad_norm": 6.686957359313965,
"learning_rate": 1.731123648109139e-06,
"loss": 2.7024,
"step": 2610
},
{
"epoch": 2.531593572550441,
"grad_norm": 6.644192695617676,
"learning_rate": 1.7282409709157947e-06,
"loss": 2.6749,
"step": 2620
},
{
"epoch": 2.5412589102331764,
"grad_norm": 6.746331214904785,
"learning_rate": 1.7253453509141586e-06,
"loss": 2.6588,
"step": 2630
},
{
"epoch": 2.5509242479159115,
"grad_norm": 6.613391876220703,
"learning_rate": 1.7224368395672166e-06,
"loss": 2.6554,
"step": 2640
},
{
"epoch": 2.5605895855986467,
"grad_norm": 6.989151477813721,
"learning_rate": 1.7195154885670683e-06,
"loss": 2.6336,
"step": 2650
},
{
"epoch": 2.5702549232813823,
"grad_norm": 6.685927867889404,
"learning_rate": 1.7165813498340088e-06,
"loss": 2.6931,
"step": 2660
},
{
"epoch": 2.5799202609641174,
"grad_norm": 6.835086345672607,
"learning_rate": 1.7136344755156048e-06,
"loss": 2.6702,
"step": 2670
},
{
"epoch": 2.587652531110306,
"eval_loss": 2.716982126235962,
"eval_runtime": 216.725,
"eval_samples_per_second": 101.834,
"eval_steps_per_second": 4.245,
"step": 2678
},
{
"epoch": 2.5895855986468526,
"grad_norm": 6.729923248291016,
"learning_rate": 1.7106749179857701e-06,
"loss": 2.6708,
"step": 2680
},
{
"epoch": 2.599250936329588,
"grad_norm": 6.472827911376953,
"learning_rate": 1.7077027298438327e-06,
"loss": 2.6505,
"step": 2690
},
{
"epoch": 2.6089162740123233,
"grad_norm": 6.623499393463135,
"learning_rate": 1.7047179639136e-06,
"loss": 2.666,
"step": 2700
},
{
"epoch": 2.6185816116950584,
"grad_norm": 6.723398208618164,
"learning_rate": 1.7017206732424226e-06,
"loss": 2.6724,
"step": 2710
},
{
"epoch": 2.628246949377794,
"grad_norm": 6.559345245361328,
"learning_rate": 1.6987109111002474e-06,
"loss": 2.663,
"step": 2720
},
{
"epoch": 2.637912287060529,
"grad_norm": 6.853271484375,
"learning_rate": 1.6956887309786743e-06,
"loss": 2.6874,
"step": 2730
},
{
"epoch": 2.6475776247432643,
"grad_norm": 7.0866312980651855,
"learning_rate": 1.692654186590004e-06,
"loss": 2.6809,
"step": 2740
},
{
"epoch": 2.657242962426,
"grad_norm": 6.323083877563477,
"learning_rate": 1.6896073318662834e-06,
"loss": 2.6679,
"step": 2750
},
{
"epoch": 2.666908300108735,
"grad_norm": 6.670552730560303,
"learning_rate": 1.6865482209583473e-06,
"loss": 2.6977,
"step": 2760
},
{
"epoch": 2.6765736377914706,
"grad_norm": 6.876490592956543,
"learning_rate": 1.6834769082348563e-06,
"loss": 2.6711,
"step": 2770
},
{
"epoch": 2.6862389754742058,
"grad_norm": 6.621648788452148,
"learning_rate": 1.6803934482813297e-06,
"loss": 2.6526,
"step": 2780
},
{
"epoch": 2.687205509242479,
"eval_loss": 2.711458444595337,
"eval_runtime": 216.9564,
"eval_samples_per_second": 101.725,
"eval_steps_per_second": 4.24,
"step": 2781
},
{
"epoch": 2.695904313156941,
"grad_norm": 6.591291904449463,
"learning_rate": 1.6772978958991766e-06,
"loss": 2.6639,
"step": 2790
},
{
"epoch": 2.705569650839676,
"grad_norm": 6.6821608543396,
"learning_rate": 1.6741903061047201e-06,
"loss": 2.6498,
"step": 2800
},
{
"epoch": 2.7152349885224116,
"grad_norm": 6.35032320022583,
"learning_rate": 1.6710707341282223e-06,
"loss": 2.6378,
"step": 2810
},
{
"epoch": 2.7249003262051468,
"grad_norm": 6.479367733001709,
"learning_rate": 1.667939235412899e-06,
"loss": 2.6458,
"step": 2820
},
{
"epoch": 2.7345656638878824,
"grad_norm": 6.6215009689331055,
"learning_rate": 1.6647958656139376e-06,
"loss": 2.6879,
"step": 2830
},
{
"epoch": 2.7442310015706175,
"grad_norm": 6.726338863372803,
"learning_rate": 1.6616406805975069e-06,
"loss": 2.6433,
"step": 2840
},
{
"epoch": 2.7538963392533526,
"grad_norm": 6.45868444442749,
"learning_rate": 1.6584737364397637e-06,
"loss": 2.6362,
"step": 2850
},
{
"epoch": 2.763561676936088,
"grad_norm": 6.664327144622803,
"learning_rate": 1.6552950894258555e-06,
"loss": 2.6657,
"step": 2860
},
{
"epoch": 2.7732270146188234,
"grad_norm": 6.683444976806641,
"learning_rate": 1.6521047960489228e-06,
"loss": 2.6648,
"step": 2870
},
{
"epoch": 2.7828923523015585,
"grad_norm": 6.655669212341309,
"learning_rate": 1.6489029130090921e-06,
"loss": 2.6701,
"step": 2880
},
{
"epoch": 2.7867584873746525,
"eval_loss": 2.705122470855713,
"eval_runtime": 217.1999,
"eval_samples_per_second": 101.611,
"eval_steps_per_second": 4.236,
"step": 2884
},
{
"epoch": 2.792557689984294,
"grad_norm": 6.554884910583496,
"learning_rate": 1.6456894972124707e-06,
"loss": 2.646,
"step": 2890
},
{
"epoch": 2.8022230276670292,
"grad_norm": 6.624351978302002,
"learning_rate": 1.6424646057701325e-06,
"loss": 2.6591,
"step": 2900
},
{
"epoch": 2.8118883653497644,
"grad_norm": 6.485014915466309,
"learning_rate": 1.6392282959971068e-06,
"loss": 2.6668,
"step": 2910
},
{
"epoch": 2.8215537030324995,
"grad_norm": 6.348880767822266,
"learning_rate": 1.6359806254113554e-06,
"loss": 2.6663,
"step": 2920
},
{
"epoch": 2.831219040715235,
"grad_norm": 6.690762519836426,
"learning_rate": 1.6327216517327535e-06,
"loss": 2.6598,
"step": 2930
},
{
"epoch": 2.8408843783979703,
"grad_norm": 6.5298309326171875,
"learning_rate": 1.6294514328820629e-06,
"loss": 2.6988,
"step": 2940
},
{
"epoch": 2.850549716080706,
"grad_norm": 6.622351169586182,
"learning_rate": 1.6261700269799014e-06,
"loss": 2.6194,
"step": 2950
},
{
"epoch": 2.860215053763441,
"grad_norm": 6.835710048675537,
"learning_rate": 1.622877492345712e-06,
"loss": 2.6718,
"step": 2960
},
{
"epoch": 2.869880391446176,
"grad_norm": 6.526007175445557,
"learning_rate": 1.6195738874967249e-06,
"loss": 2.6562,
"step": 2970
},
{
"epoch": 2.8795457291289113,
"grad_norm": 6.33076286315918,
"learning_rate": 1.6162592711469181e-06,
"loss": 2.6509,
"step": 2980
},
{
"epoch": 2.886311465506826,
"eval_loss": 2.699232578277588,
"eval_runtime": 217.0251,
"eval_samples_per_second": 101.693,
"eval_steps_per_second": 4.239,
"step": 2987
},
{
"epoch": 2.889211066811647,
"grad_norm": 6.612087249755859,
"learning_rate": 1.6129337022059733e-06,
"loss": 2.6695,
"step": 2990
},
{
"epoch": 2.898876404494382,
"grad_norm": 6.748770236968994,
"learning_rate": 1.6095972397782298e-06,
"loss": 2.6739,
"step": 3000
},
{
"epoch": 2.9085417421771176,
"grad_norm": 6.782710552215576,
"learning_rate": 1.6062499431616331e-06,
"loss": 2.6703,
"step": 3010
},
{
"epoch": 2.9182070798598527,
"grad_norm": 6.288000106811523,
"learning_rate": 1.602891871846682e-06,
"loss": 2.6443,
"step": 3020
},
{
"epoch": 2.927872417542588,
"grad_norm": 6.792652130126953,
"learning_rate": 1.59952308551537e-06,
"loss": 2.6951,
"step": 3030
},
{
"epoch": 2.937537755225323,
"grad_norm": 6.761292934417725,
"learning_rate": 1.5961436440401267e-06,
"loss": 2.6623,
"step": 3040
},
{
"epoch": 2.9472030929080586,
"grad_norm": 6.930576324462891,
"learning_rate": 1.5927536074827503e-06,
"loss": 2.6476,
"step": 3050
},
{
"epoch": 2.9568684305907937,
"grad_norm": 6.642486095428467,
"learning_rate": 1.5893530360933448e-06,
"loss": 2.6679,
"step": 3060
},
{
"epoch": 2.9665337682735293,
"grad_norm": 6.4895501136779785,
"learning_rate": 1.5859419903092445e-06,
"loss": 2.6638,
"step": 3070
},
{
"epoch": 2.9761991059562645,
"grad_norm": 6.47406005859375,
"learning_rate": 1.582520530753943e-06,
"loss": 2.6515,
"step": 3080
},
{
"epoch": 2.9858644436389996,
"grad_norm": 6.589229106903076,
"learning_rate": 1.5790887182360143e-06,
"loss": 2.671,
"step": 3090
},
{
"epoch": 2.9858644436389996,
"eval_loss": 2.694453239440918,
"eval_runtime": 216.9988,
"eval_samples_per_second": 101.706,
"eval_steps_per_second": 4.24,
"step": 3090
},
{
"epoch": 2.9955297813217348,
"grad_norm": 6.626533508300781,
"learning_rate": 1.575646613748033e-06,
"loss": 2.6769,
"step": 3100
},
{
"epoch": 3.0048326688413676,
"grad_norm": 6.644468307495117,
"learning_rate": 1.5721942784654892e-06,
"loss": 2.6425,
"step": 3110
},
{
"epoch": 3.0144980065241027,
"grad_norm": 6.786801338195801,
"learning_rate": 1.5687317737457023e-06,
"loss": 2.6216,
"step": 3120
},
{
"epoch": 3.0241633442068383,
"grad_norm": 6.773380756378174,
"learning_rate": 1.5652591611267293e-06,
"loss": 2.6165,
"step": 3130
},
{
"epoch": 3.0338286818895734,
"grad_norm": 6.535521507263184,
"learning_rate": 1.5617765023262734e-06,
"loss": 2.6002,
"step": 3140
},
{
"epoch": 3.043494019572309,
"grad_norm": 6.627861499786377,
"learning_rate": 1.5582838592405838e-06,
"loss": 2.6218,
"step": 3150
},
{
"epoch": 3.053159357255044,
"grad_norm": 6.7703938484191895,
"learning_rate": 1.5547812939433587e-06,
"loss": 2.6333,
"step": 3160
},
{
"epoch": 3.0628246949377793,
"grad_norm": 6.72239875793457,
"learning_rate": 1.5512688686846402e-06,
"loss": 2.6356,
"step": 3170
},
{
"epoch": 3.0724900326205145,
"grad_norm": 6.52492618560791,
"learning_rate": 1.5477466458897083e-06,
"loss": 2.6048,
"step": 3180
},
{
"epoch": 3.08215537030325,
"grad_norm": 6.695172309875488,
"learning_rate": 1.5442146881579726e-06,
"loss": 2.5957,
"step": 3190
},
{
"epoch": 3.0850549716080704,
"eval_loss": 2.6920766830444336,
"eval_runtime": 216.947,
"eval_samples_per_second": 101.73,
"eval_steps_per_second": 4.241,
"step": 3193
},
{
"epoch": 3.091820707985985,
"grad_norm": 6.52951717376709,
"learning_rate": 1.5406730582618575e-06,
"loss": 2.5731,
"step": 3200
},
{
"epoch": 3.1014860456687208,
"grad_norm": 6.658172130584717,
"learning_rate": 1.5371218191456892e-06,
"loss": 2.599,
"step": 3210
},
{
"epoch": 3.111151383351456,
"grad_norm": 6.837924003601074,
"learning_rate": 1.5335610339245748e-06,
"loss": 2.6061,
"step": 3220
},
{
"epoch": 3.120816721034191,
"grad_norm": 6.788025856018066,
"learning_rate": 1.5299907658832817e-06,
"loss": 2.593,
"step": 3230
},
{
"epoch": 3.130482058716926,
"grad_norm": 6.745009899139404,
"learning_rate": 1.526411078475113e-06,
"loss": 2.6085,
"step": 3240
},
{
"epoch": 3.140147396399662,
"grad_norm": 6.781652927398682,
"learning_rate": 1.5228220353207784e-06,
"loss": 2.6148,
"step": 3250
},
{
"epoch": 3.149812734082397,
"grad_norm": 6.939257621765137,
"learning_rate": 1.5192237002072656e-06,
"loss": 2.6116,
"step": 3260
},
{
"epoch": 3.1594780717651325,
"grad_norm": 6.560710430145264,
"learning_rate": 1.5156161370867052e-06,
"loss": 2.6266,
"step": 3270
},
{
"epoch": 3.1691434094478677,
"grad_norm": 6.522158145904541,
"learning_rate": 1.5119994100752337e-06,
"loss": 2.6049,
"step": 3280
},
{
"epoch": 3.178808747130603,
"grad_norm": 6.933811187744141,
"learning_rate": 1.5083735834518555e-06,
"loss": 2.6168,
"step": 3290
},
{
"epoch": 3.184607949740244,
"eval_loss": 2.68808650970459,
"eval_runtime": 216.8361,
"eval_samples_per_second": 101.782,
"eval_steps_per_second": 4.243,
"step": 3296
},
{
"epoch": 3.188474084813338,
"grad_norm": 6.851550579071045,
"learning_rate": 1.5047387216572994e-06,
"loss": 2.6326,
"step": 3300
},
{
"epoch": 3.1981394224960735,
"grad_norm": 6.676671504974365,
"learning_rate": 1.5010948892928732e-06,
"loss": 2.5766,
"step": 3310
},
{
"epoch": 3.2078047601788087,
"grad_norm": 6.791378974914551,
"learning_rate": 1.4974421511193164e-06,
"loss": 2.5921,
"step": 3320
},
{
"epoch": 3.2174700978615443,
"grad_norm": 6.412371635437012,
"learning_rate": 1.493780572055649e-06,
"loss": 2.6231,
"step": 3330
},
{
"epoch": 3.2271354355442794,
"grad_norm": 6.870903968811035,
"learning_rate": 1.4901102171780174e-06,
"loss": 2.6049,
"step": 3340
},
{
"epoch": 3.2368007732270145,
"grad_norm": 6.820573806762695,
"learning_rate": 1.4864311517185369e-06,
"loss": 2.592,
"step": 3350
},
{
"epoch": 3.2464661109097497,
"grad_norm": 6.859145641326904,
"learning_rate": 1.482743441064134e-06,
"loss": 2.6487,
"step": 3360
},
{
"epoch": 3.2561314485924853,
"grad_norm": 6.515082359313965,
"learning_rate": 1.4790471507553847e-06,
"loss": 2.6272,
"step": 3370
},
{
"epoch": 3.2657967862752204,
"grad_norm": 6.639008045196533,
"learning_rate": 1.4753423464853473e-06,
"loss": 2.597,
"step": 3380
},
{
"epoch": 3.275462123957956,
"grad_norm": 6.76662540435791,
"learning_rate": 1.4716290940983965e-06,
"loss": 2.6137,
"step": 3390
},
{
"epoch": 3.2841609278724175,
"eval_loss": 2.6844358444213867,
"eval_runtime": 216.9567,
"eval_samples_per_second": 101.725,
"eval_steps_per_second": 4.24,
"step": 3399
},
{
"epoch": 3.285127461640691,
"grad_norm": 6.6699347496032715,
"learning_rate": 1.4679074595890532e-06,
"loss": 2.5737,
"step": 3400
},
{
"epoch": 3.2947927993234263,
"grad_norm": 6.792168617248535,
"learning_rate": 1.4641775091008106e-06,
"loss": 2.6094,
"step": 3410
},
{
"epoch": 3.3044581370061614,
"grad_norm": 6.676168441772461,
"learning_rate": 1.4604393089249599e-06,
"loss": 2.6151,
"step": 3420
},
{
"epoch": 3.314123474688897,
"grad_norm": 6.540125846862793,
"learning_rate": 1.456692925499411e-06,
"loss": 2.626,
"step": 3430
},
{
"epoch": 3.323788812371632,
"grad_norm": 6.471897602081299,
"learning_rate": 1.4529384254075133e-06,
"loss": 2.6007,
"step": 3440
},
{
"epoch": 3.3334541500543677,
"grad_norm": 6.628961086273193,
"learning_rate": 1.4491758753768694e-06,
"loss": 2.6175,
"step": 3450
},
{
"epoch": 3.343119487737103,
"grad_norm": 6.617523670196533,
"learning_rate": 1.4454053422781532e-06,
"loss": 2.6116,
"step": 3460
},
{
"epoch": 3.352784825419838,
"grad_norm": 6.7695722579956055,
"learning_rate": 1.4416268931239177e-06,
"loss": 2.5806,
"step": 3470
},
{
"epoch": 3.362450163102573,
"grad_norm": 6.675383567810059,
"learning_rate": 1.4378405950674065e-06,
"loss": 2.5972,
"step": 3480
},
{
"epoch": 3.3721155007853088,
"grad_norm": 6.70187520980835,
"learning_rate": 1.4340465154013585e-06,
"loss": 2.5809,
"step": 3490
},
{
"epoch": 3.381780838468044,
"grad_norm": 6.520016670227051,
"learning_rate": 1.4302447215568136e-06,
"loss": 2.6,
"step": 3500
},
{
"epoch": 3.383713906004591,
"eval_loss": 2.6802666187286377,
"eval_runtime": 216.9945,
"eval_samples_per_second": 101.708,
"eval_steps_per_second": 4.24,
"step": 3502
},
{
"epoch": 3.3914461761507795,
"grad_norm": 6.802120208740234,
"learning_rate": 1.426435281101913e-06,
"loss": 2.6175,
"step": 3510
},
{
"epoch": 3.4011115138335146,
"grad_norm": 6.581682205200195,
"learning_rate": 1.4226182617406994e-06,
"loss": 2.6121,
"step": 3520
},
{
"epoch": 3.4107768515162498,
"grad_norm": 6.738428115844727,
"learning_rate": 1.4187937313119124e-06,
"loss": 2.6015,
"step": 3530
},
{
"epoch": 3.420442189198985,
"grad_norm": 6.7162675857543945,
"learning_rate": 1.4149617577877841e-06,
"loss": 2.6161,
"step": 3540
},
{
"epoch": 3.4301075268817205,
"grad_norm": 6.830008029937744,
"learning_rate": 1.41112240927283e-06,
"loss": 2.6125,
"step": 3550
},
{
"epoch": 3.4397728645644556,
"grad_norm": 6.8902435302734375,
"learning_rate": 1.4072757540026396e-06,
"loss": 2.6181,
"step": 3560
},
{
"epoch": 3.449438202247191,
"grad_norm": 6.644567012786865,
"learning_rate": 1.403421860342663e-06,
"loss": 2.5965,
"step": 3570
},
{
"epoch": 3.4591035399299264,
"grad_norm": 6.674381732940674,
"learning_rate": 1.3995607967869963e-06,
"loss": 2.6186,
"step": 3580
},
{
"epoch": 3.4687688776126615,
"grad_norm": 6.683381080627441,
"learning_rate": 1.3956926319571628e-06,
"loss": 2.6271,
"step": 3590
},
{
"epoch": 3.4784342152953966,
"grad_norm": 6.642363548278809,
"learning_rate": 1.3918174346008963e-06,
"loss": 2.6047,
"step": 3600
},
{
"epoch": 3.4832668841367647,
"eval_loss": 2.6766762733459473,
"eval_runtime": 216.9538,
"eval_samples_per_second": 101.727,
"eval_steps_per_second": 4.241,
"step": 3605
},
{
"epoch": 3.4880995529781322,
"grad_norm": 6.7301106452941895,
"learning_rate": 1.3879352735909163e-06,
"loss": 2.6016,
"step": 3610
},
{
"epoch": 3.4977648906608674,
"grad_norm": 6.565543174743652,
"learning_rate": 1.3840462179237058e-06,
"loss": 2.6413,
"step": 3620
},
{
"epoch": 3.507430228343603,
"grad_norm": 6.726136207580566,
"learning_rate": 1.3801503367182846e-06,
"loss": 2.5709,
"step": 3630
},
{
"epoch": 3.517095566026338,
"grad_norm": 6.778688907623291,
"learning_rate": 1.3762476992149803e-06,
"loss": 2.5772,
"step": 3640
},
{
"epoch": 3.5267609037090732,
"grad_norm": 6.823915004730225,
"learning_rate": 1.3723383747741988e-06,
"loss": 2.6417,
"step": 3650
},
{
"epoch": 3.5364262413918084,
"grad_norm": 6.821535110473633,
"learning_rate": 1.3684224328751904e-06,
"loss": 2.6092,
"step": 3660
},
{
"epoch": 3.546091579074544,
"grad_norm": 6.66444730758667,
"learning_rate": 1.364499943114815e-06,
"loss": 2.5744,
"step": 3670
},
{
"epoch": 3.555756916757279,
"grad_norm": 6.7145161628723145,
"learning_rate": 1.3605709752063072e-06,
"loss": 2.6028,
"step": 3680
},
{
"epoch": 3.5654222544400147,
"grad_norm": 6.7199931144714355,
"learning_rate": 1.356635598978034e-06,
"loss": 2.5938,
"step": 3690
},
{
"epoch": 3.57508759212275,
"grad_norm": 6.810650825500488,
"learning_rate": 1.352693884372257e-06,
"loss": 2.5833,
"step": 3700
},
{
"epoch": 3.582819862268938,
"eval_loss": 2.67258882522583,
"eval_runtime": 216.9499,
"eval_samples_per_second": 101.729,
"eval_steps_per_second": 4.241,
"step": 3708
},
{
"epoch": 3.584752929805485,
"grad_norm": 6.88958740234375,
"learning_rate": 1.3487459014438875e-06,
"loss": 2.6103,
"step": 3710
},
{
"epoch": 3.59441826748822,
"grad_norm": 6.527899742126465,
"learning_rate": 1.344791720359241e-06,
"loss": 2.5896,
"step": 3720
},
{
"epoch": 3.6040836051709557,
"grad_norm": 6.705896854400635,
"learning_rate": 1.340831411394792e-06,
"loss": 2.5731,
"step": 3730
},
{
"epoch": 3.613748942853691,
"grad_norm": 6.65834903717041,
"learning_rate": 1.3368650449359237e-06,
"loss": 2.5992,
"step": 3740
},
{
"epoch": 3.6234142805364264,
"grad_norm": 6.513383865356445,
"learning_rate": 1.3328926914756764e-06,
"loss": 2.59,
"step": 3750
},
{
"epoch": 3.6330796182191616,
"grad_norm": 6.5526652336120605,
"learning_rate": 1.3289144216134973e-06,
"loss": 2.5497,
"step": 3760
},
{
"epoch": 3.6427449559018967,
"grad_norm": 6.795760631561279,
"learning_rate": 1.324930306053983e-06,
"loss": 2.6445,
"step": 3770
},
{
"epoch": 3.652410293584632,
"grad_norm": 6.859256744384766,
"learning_rate": 1.3209404156056234e-06,
"loss": 2.5714,
"step": 3780
},
{
"epoch": 3.6620756312673675,
"grad_norm": 6.862630367279053,
"learning_rate": 1.316944821179545e-06,
"loss": 2.5888,
"step": 3790
},
{
"epoch": 3.6717409689501026,
"grad_norm": 6.7294511795043945,
"learning_rate": 1.3129435937882485e-06,
"loss": 2.5981,
"step": 3800
},
{
"epoch": 3.681406306632838,
"grad_norm": 6.428034782409668,
"learning_rate": 1.3089368045443478e-06,
"loss": 2.6094,
"step": 3810
},
{
"epoch": 3.6823728404011113,
"eval_loss": 2.6688151359558105,
"eval_runtime": 216.9738,
"eval_samples_per_second": 101.717,
"eval_steps_per_second": 4.24,
"step": 3811
},
{
"epoch": 3.6910716443155733,
"grad_norm": 6.683334827423096,
"learning_rate": 1.3049245246593063e-06,
"loss": 2.5955,
"step": 3820
},
{
"epoch": 3.7007369819983085,
"grad_norm": 6.5523481369018555,
"learning_rate": 1.3009068254421707e-06,
"loss": 2.5874,
"step": 3830
},
{
"epoch": 3.7104023196810436,
"grad_norm": 6.8417253494262695,
"learning_rate": 1.2968837782983032e-06,
"loss": 2.5853,
"step": 3840
},
{
"epoch": 3.720067657363779,
"grad_norm": 6.959075450897217,
"learning_rate": 1.292855454728115e-06,
"loss": 2.5751,
"step": 3850
},
{
"epoch": 3.7297329950465143,
"grad_norm": 6.557989120483398,
"learning_rate": 1.288821926325791e-06,
"loss": 2.6214,
"step": 3860
},
{
"epoch": 3.73939833272925,
"grad_norm": 6.578183650970459,
"learning_rate": 1.2847832647780218e-06,
"loss": 2.5918,
"step": 3870
},
{
"epoch": 3.749063670411985,
"grad_norm": 6.700976371765137,
"learning_rate": 1.2807395418627277e-06,
"loss": 2.5793,
"step": 3880
},
{
"epoch": 3.75872900809472,
"grad_norm": 6.564661502838135,
"learning_rate": 1.2766908294477826e-06,
"loss": 2.5966,
"step": 3890
},
{
"epoch": 3.7683943457774554,
"grad_norm": 6.6865410804748535,
"learning_rate": 1.2726371994897376e-06,
"loss": 2.5844,
"step": 3900
},
{
"epoch": 3.778059683460191,
"grad_norm": 6.696436882019043,
"learning_rate": 1.2685787240325417e-06,
"loss": 2.5761,
"step": 3910
},
{
"epoch": 3.781925818533285,
"eval_loss": 2.665092706680298,
"eval_runtime": 216.965,
"eval_samples_per_second": 101.721,
"eval_steps_per_second": 4.24,
"step": 3914
},
{
"epoch": 3.787725021142926,
"grad_norm": 6.7038702964782715,
"learning_rate": 1.264515475206262e-06,
"loss": 2.5867,
"step": 3920
},
{
"epoch": 3.7973903588256617,
"grad_norm": 6.5760064125061035,
"learning_rate": 1.2604475252258005e-06,
"loss": 2.5927,
"step": 3930
},
{
"epoch": 3.807055696508397,
"grad_norm": 6.533830642700195,
"learning_rate": 1.2563749463896116e-06,
"loss": 2.5828,
"step": 3940
},
{
"epoch": 3.816721034191132,
"grad_norm": 6.66879415512085,
"learning_rate": 1.2522978110784177e-06,
"loss": 2.5703,
"step": 3950
},
{
"epoch": 3.826386371873867,
"grad_norm": 6.830549716949463,
"learning_rate": 1.2482161917539209e-06,
"loss": 2.6076,
"step": 3960
},
{
"epoch": 3.8360517095566027,
"grad_norm": 6.542690753936768,
"learning_rate": 1.2441301609575172e-06,
"loss": 2.5879,
"step": 3970
},
{
"epoch": 3.845717047239338,
"grad_norm": 6.628766059875488,
"learning_rate": 1.2400397913090059e-06,
"loss": 2.6014,
"step": 3980
},
{
"epoch": 3.8553823849220734,
"grad_norm": 6.663356781005859,
"learning_rate": 1.2359451555052997e-06,
"loss": 2.5364,
"step": 3990
},
{
"epoch": 3.8650477226048086,
"grad_norm": 6.802213191986084,
"learning_rate": 1.2318463263191323e-06,
"loss": 2.5678,
"step": 4000
},
{
"epoch": 3.8747130602875437,
"grad_norm": 6.468794345855713,
"learning_rate": 1.227743376597765e-06,
"loss": 2.5762,
"step": 4010
},
{
"epoch": 3.8814787966654585,
"eval_loss": 2.6623401641845703,
"eval_runtime": 217.0567,
"eval_samples_per_second": 101.678,
"eval_steps_per_second": 4.239,
"step": 4017
},
{
"epoch": 3.884378397970279,
"grad_norm": 6.709383964538574,
"learning_rate": 1.2236363792616923e-06,
"loss": 2.6092,
"step": 4020
},
{
"epoch": 3.8940437356530144,
"grad_norm": 6.665591239929199,
"learning_rate": 1.2195254073033455e-06,
"loss": 2.5627,
"step": 4030
},
{
"epoch": 3.9037090733357496,
"grad_norm": 6.547743797302246,
"learning_rate": 1.2154105337857963e-06,
"loss": 2.6063,
"step": 4040
},
{
"epoch": 3.913374411018485,
"grad_norm": 6.582630634307861,
"learning_rate": 1.2112918318414572e-06,
"loss": 2.5997,
"step": 4050
},
{
"epoch": 3.9230397487012203,
"grad_norm": 6.448917865753174,
"learning_rate": 1.2071693746707812e-06,
"loss": 2.5644,
"step": 4060
},
{
"epoch": 3.9327050863839554,
"grad_norm": 6.690165042877197,
"learning_rate": 1.203043235540964e-06,
"loss": 2.6261,
"step": 4070
},
{
"epoch": 3.9423704240666906,
"grad_norm": 6.645413398742676,
"learning_rate": 1.1989134877846377e-06,
"loss": 2.6075,
"step": 4080
},
{
"epoch": 3.952035761749426,
"grad_norm": 6.471231460571289,
"learning_rate": 1.1947802047985701e-06,
"loss": 2.5694,
"step": 4090
},
{
"epoch": 3.9617010994321613,
"grad_norm": 6.488655090332031,
"learning_rate": 1.1906434600423605e-06,
"loss": 2.5728,
"step": 4100
},
{
"epoch": 3.971366437114897,
"grad_norm": 6.397697448730469,
"learning_rate": 1.1865033270371317e-06,
"loss": 2.6184,
"step": 4110
},
{
"epoch": 3.981031774797632,
"grad_norm": 6.57921838760376,
"learning_rate": 1.1823598793642256e-06,
"loss": 2.5828,
"step": 4120
},
{
"epoch": 3.981031774797632,
"eval_loss": 2.658400774002075,
"eval_runtime": 216.9957,
"eval_samples_per_second": 101.707,
"eval_steps_per_second": 4.24,
"step": 4120
},
{
"epoch": 3.990697112480367,
"grad_norm": 6.8846306800842285,
"learning_rate": 1.1782131906638947e-06,
"loss": 2.5556,
"step": 4130
},
{
"epoch": 4.0,
"grad_norm": 8.866212844848633,
"learning_rate": 1.174063334633993e-06,
"loss": 2.5868,
"step": 4140
},
{
"epoch": 4.009665337682735,
"grad_norm": 6.461270332336426,
"learning_rate": 1.1699103850286667e-06,
"loss": 2.5343,
"step": 4150
},
{
"epoch": 4.01933067536547,
"grad_norm": 6.403049945831299,
"learning_rate": 1.1657544156570433e-06,
"loss": 2.5403,
"step": 4160
},
{
"epoch": 4.028996013048205,
"grad_norm": 6.411298751831055,
"learning_rate": 1.1615955003819195e-06,
"loss": 2.5484,
"step": 4170
},
{
"epoch": 4.0386613507309415,
"grad_norm": 6.594933032989502,
"learning_rate": 1.157433713118449e-06,
"loss": 2.5664,
"step": 4180
},
{
"epoch": 4.048326688413677,
"grad_norm": 6.8718342781066895,
"learning_rate": 1.1532691278328282e-06,
"loss": 2.5654,
"step": 4190
},
{
"epoch": 4.057992026096412,
"grad_norm": 6.683941841125488,
"learning_rate": 1.1491018185409813e-06,
"loss": 2.532,
"step": 4200
},
{
"epoch": 4.067657363779147,
"grad_norm": 6.692328929901123,
"learning_rate": 1.1449318593072465e-06,
"loss": 2.5486,
"step": 4210
},
{
"epoch": 4.077322701461882,
"grad_norm": 6.605168342590332,
"learning_rate": 1.1407593242430586e-06,
"loss": 2.5821,
"step": 4220
},
{
"epoch": 4.080222302766703,
"eval_loss": 2.6606152057647705,
"eval_runtime": 217.0244,
"eval_samples_per_second": 101.694,
"eval_steps_per_second": 4.239,
"step": 4223
},
{
"epoch": 4.086988039144618,
"grad_norm": 6.7313337326049805,
"learning_rate": 1.1365842875056311e-06,
"loss": 2.554,
"step": 4230
},
{
"epoch": 4.096653376827353,
"grad_norm": 6.845920085906982,
"learning_rate": 1.1324068232966392e-06,
"loss": 2.5234,
"step": 4240
},
{
"epoch": 4.106318714510088,
"grad_norm": 6.765261650085449,
"learning_rate": 1.1282270058609013e-06,
"loss": 2.547,
"step": 4250
},
{
"epoch": 4.1159840521928235,
"grad_norm": 6.755435943603516,
"learning_rate": 1.1240449094850584e-06,
"loss": 2.5661,
"step": 4260
},
{
"epoch": 4.125649389875559,
"grad_norm": 6.7944655418396,
"learning_rate": 1.1198606084962547e-06,
"loss": 2.5423,
"step": 4270
},
{
"epoch": 4.135314727558294,
"grad_norm": 6.660946369171143,
"learning_rate": 1.1156741772608165e-06,
"loss": 2.5314,
"step": 4280
},
{
"epoch": 4.144980065241029,
"grad_norm": 6.597288608551025,
"learning_rate": 1.11148569018293e-06,
"loss": 2.5404,
"step": 4290
},
{
"epoch": 4.154645402923765,
"grad_norm": 6.629675388336182,
"learning_rate": 1.1072952217033195e-06,
"loss": 2.5392,
"step": 4300
},
{
"epoch": 4.1643107406065,
"grad_norm": 6.65186882019043,
"learning_rate": 1.103102846297924e-06,
"loss": 2.6007,
"step": 4310
},
{
"epoch": 4.173976078289235,
"grad_norm": 6.881357669830322,
"learning_rate": 1.0989086384765737e-06,
"loss": 2.5721,
"step": 4320
},
{
"epoch": 4.179775280898877,
"eval_loss": 2.6572110652923584,
"eval_runtime": 216.941,
"eval_samples_per_second": 101.733,
"eval_steps_per_second": 4.241,
"step": 4326
},
{
"epoch": 4.18364141597197,
"grad_norm": 6.678983688354492,
"learning_rate": 1.0947126727816665e-06,
"loss": 2.5719,
"step": 4330
},
{
"epoch": 4.1933067536547055,
"grad_norm": 6.7589945793151855,
"learning_rate": 1.090515023786841e-06,
"loss": 2.5502,
"step": 4340
},
{
"epoch": 4.2029720913374415,
"grad_norm": 6.886190414428711,
"learning_rate": 1.0863157660956538e-06,
"loss": 2.5236,
"step": 4350
},
{
"epoch": 4.212637429020177,
"grad_norm": 6.677691459655762,
"learning_rate": 1.082114974340252e-06,
"loss": 2.5518,
"step": 4360
},
{
"epoch": 4.222302766702912,
"grad_norm": 6.715696811676025,
"learning_rate": 1.0779127231800474e-06,
"loss": 2.5375,
"step": 4370
},
{
"epoch": 4.231968104385647,
"grad_norm": 6.802489757537842,
"learning_rate": 1.0737090873003884e-06,
"loss": 2.5263,
"step": 4380
},
{
"epoch": 4.241633442068382,
"grad_norm": 7.024548053741455,
"learning_rate": 1.069504141411235e-06,
"loss": 2.5676,
"step": 4390
},
{
"epoch": 4.251298779751117,
"grad_norm": 6.528923034667969,
"learning_rate": 1.0652979602458287e-06,
"loss": 2.5676,
"step": 4400
},
{
"epoch": 4.260964117433852,
"grad_norm": 6.7343244552612305,
"learning_rate": 1.0610906185593653e-06,
"loss": 2.5323,
"step": 4410
},
{
"epoch": 4.270629455116588,
"grad_norm": 6.739745140075684,
"learning_rate": 1.0568821911276668e-06,
"loss": 2.5333,
"step": 4420
},
{
"epoch": 4.2793282590310495,
"eval_loss": 2.6547646522521973,
"eval_runtime": 217.1095,
"eval_samples_per_second": 101.654,
"eval_steps_per_second": 4.237,
"step": 4429
},
{
"epoch": 4.280294792799324,
"grad_norm": 6.441568851470947,
"learning_rate": 1.0526727527458508e-06,
"loss": 2.561,
"step": 4430
},
{
"epoch": 4.289960130482059,
"grad_norm": 6.605863571166992,
"learning_rate": 1.048462378227003e-06,
"loss": 2.5611,
"step": 4440
},
{
"epoch": 4.299625468164794,
"grad_norm": 6.771341323852539,
"learning_rate": 1.0442511424008464e-06,
"loss": 2.5735,
"step": 4450
},
{
"epoch": 4.309290805847529,
"grad_norm": 6.989775657653809,
"learning_rate": 1.040039120112412e-06,
"loss": 2.5473,
"step": 4460
},
{
"epoch": 4.318956143530265,
"grad_norm": 7.052975654602051,
"learning_rate": 1.0358263862207083e-06,
"loss": 2.5443,
"step": 4470
},
{
"epoch": 4.328621481213,
"grad_norm": 6.5822858810424805,
"learning_rate": 1.031613015597391e-06,
"loss": 2.5585,
"step": 4480
},
{
"epoch": 4.338286818895735,
"grad_norm": 6.736408710479736,
"learning_rate": 1.0273990831254319e-06,
"loss": 2.5436,
"step": 4490
},
{
"epoch": 4.3479521565784705,
"grad_norm": 6.6356120109558105,
"learning_rate": 1.0231846636977882e-06,
"loss": 2.5533,
"step": 4500
},
{
"epoch": 4.357617494261206,
"grad_norm": 6.671817779541016,
"learning_rate": 1.0189698322160732e-06,
"loss": 2.5469,
"step": 4510
},
{
"epoch": 4.367282831943941,
"grad_norm": 7.152904987335205,
"learning_rate": 1.0147546635892209e-06,
"loss": 2.537,
"step": 4520
},
{
"epoch": 4.376948169626676,
"grad_norm": 6.494625568389893,
"learning_rate": 1.010539232732159e-06,
"loss": 2.5045,
"step": 4530
},
{
"epoch": 4.378881237163223,
"eval_loss": 2.653103828430176,
"eval_runtime": 217.106,
"eval_samples_per_second": 101.655,
"eval_steps_per_second": 4.238,
"step": 4532
},
{
"epoch": 4.386613507309412,
"grad_norm": 6.714130401611328,
"learning_rate": 1.0063236145644762e-06,
"loss": 2.5592,
"step": 4540
},
{
"epoch": 4.396278844992147,
"grad_norm": 6.6748528480529785,
"learning_rate": 1.0021078840090886e-06,
"loss": 2.5473,
"step": 4550
},
{
"epoch": 4.405944182674882,
"grad_norm": 6.548516273498535,
"learning_rate": 9.978921159909113e-07,
"loss": 2.5382,
"step": 4560
},
{
"epoch": 4.415609520357617,
"grad_norm": 6.959436416625977,
"learning_rate": 9.93676385435524e-07,
"loss": 2.537,
"step": 4570
},
{
"epoch": 4.4252748580403525,
"grad_norm": 6.954132556915283,
"learning_rate": 9.894607672678408e-07,
"loss": 2.5511,
"step": 4580
},
{
"epoch": 4.4349401957230885,
"grad_norm": 6.731428146362305,
"learning_rate": 9.852453364107792e-07,
"loss": 2.522,
"step": 4590
},
{
"epoch": 4.444605533405824,
"grad_norm": 6.622025489807129,
"learning_rate": 9.81030167783927e-07,
"loss": 2.5363,
"step": 4600
},
{
"epoch": 4.454270871088559,
"grad_norm": 6.871068954467773,
"learning_rate": 9.768153363022115e-07,
"loss": 2.5361,
"step": 4610
},
{
"epoch": 4.463936208771294,
"grad_norm": 6.612332820892334,
"learning_rate": 9.72600916874568e-07,
"loss": 2.5385,
"step": 4620
},
{
"epoch": 4.473601546454029,
"grad_norm": 6.64678955078125,
"learning_rate": 9.683869844026089e-07,
"loss": 2.5374,
"step": 4630
},
{
"epoch": 4.478434215295397,
"eval_loss": 2.650099515914917,
"eval_runtime": 216.8756,
"eval_samples_per_second": 101.763,
"eval_steps_per_second": 4.242,
"step": 4635
},
{
"epoch": 4.483266884136764,
"grad_norm": 6.953402996063232,
"learning_rate": 9.641736137792914e-07,
"loss": 2.5646,
"step": 4640
},
{
"epoch": 4.492932221819499,
"grad_norm": 6.753168106079102,
"learning_rate": 9.59960879887588e-07,
"loss": 2.545,
"step": 4650
},
{
"epoch": 4.502597559502235,
"grad_norm": 6.863531589508057,
"learning_rate": 9.557488575991537e-07,
"loss": 2.5303,
"step": 4660
},
{
"epoch": 4.5122628971849705,
"grad_norm": 6.904908657073975,
"learning_rate": 9.515376217729971e-07,
"loss": 2.5381,
"step": 4670
},
{
"epoch": 4.521928234867706,
"grad_norm": 6.753827095031738,
"learning_rate": 9.473272472541492e-07,
"loss": 2.5243,
"step": 4680
},
{
"epoch": 4.531593572550441,
"grad_norm": 6.280510902404785,
"learning_rate": 9.431178088723333e-07,
"loss": 2.5566,
"step": 4690
},
{
"epoch": 4.541258910233176,
"grad_norm": 6.862277030944824,
"learning_rate": 9.389093814406347e-07,
"loss": 2.5286,
"step": 4700
},
{
"epoch": 4.550924247915912,
"grad_norm": 6.67894172668457,
"learning_rate": 9.347020397541714e-07,
"loss": 2.5297,
"step": 4710
},
{
"epoch": 4.560589585598647,
"grad_norm": 6.786202907562256,
"learning_rate": 9.30495858588765e-07,
"loss": 2.565,
"step": 4720
},
{
"epoch": 4.570254923281382,
"grad_norm": 6.705028533935547,
"learning_rate": 9.262909126996116e-07,
"loss": 2.5141,
"step": 4730
},
{
"epoch": 4.57798719342757,
"eval_loss": 2.6482667922973633,
"eval_runtime": 216.9673,
"eval_samples_per_second": 101.72,
"eval_steps_per_second": 4.24,
"step": 4738
},
{
"epoch": 4.579920260964117,
"grad_norm": 6.745487213134766,
"learning_rate": 9.220872768199527e-07,
"loss": 2.5456,
"step": 4740
},
{
"epoch": 4.589585598646853,
"grad_norm": 6.562654495239258,
"learning_rate": 9.178850256597481e-07,
"loss": 2.5743,
"step": 4750
},
{
"epoch": 4.599250936329588,
"grad_norm": 6.653303623199463,
"learning_rate": 9.136842339043463e-07,
"loss": 2.522,
"step": 4760
},
{
"epoch": 4.608916274012323,
"grad_norm": 7.028505325317383,
"learning_rate": 9.09484976213159e-07,
"loss": 2.5192,
"step": 4770
},
{
"epoch": 4.618581611695059,
"grad_norm": 6.768604278564453,
"learning_rate": 9.052873272183335e-07,
"loss": 2.5391,
"step": 4780
},
{
"epoch": 4.628246949377794,
"grad_norm": 7.153252601623535,
"learning_rate": 9.01091361523426e-07,
"loss": 2.5597,
"step": 4790
},
{
"epoch": 4.637912287060529,
"grad_norm": 6.854784965515137,
"learning_rate": 8.968971537020757e-07,
"loss": 2.553,
"step": 4800
},
{
"epoch": 4.647577624743264,
"grad_norm": 6.819455146789551,
"learning_rate": 8.927047782966806e-07,
"loss": 2.5752,
"step": 4810
},
{
"epoch": 4.657242962425999,
"grad_norm": 6.677375793457031,
"learning_rate": 8.8851430981707e-07,
"loss": 2.562,
"step": 4820
},
{
"epoch": 4.6669083001087355,
"grad_norm": 6.6845245361328125,
"learning_rate": 8.843258227391834e-07,
"loss": 2.5613,
"step": 4830
},
{
"epoch": 4.676573637791471,
"grad_norm": 6.7105607986450195,
"learning_rate": 8.801393915037456e-07,
"loss": 2.558,
"step": 4840
},
{
"epoch": 4.677540171559744,
"eval_loss": 2.6455490589141846,
"eval_runtime": 217.0811,
"eval_samples_per_second": 101.667,
"eval_steps_per_second": 4.238,
"step": 4841
},
{
"epoch": 4.686238975474206,
"grad_norm": 6.720954418182373,
"learning_rate": 8.759550905149419e-07,
"loss": 2.5752,
"step": 4850
},
{
"epoch": 4.695904313156941,
"grad_norm": 6.768190383911133,
"learning_rate": 8.717729941390988e-07,
"loss": 2.5586,
"step": 4860
},
{
"epoch": 4.705569650839676,
"grad_norm": 6.576201915740967,
"learning_rate": 8.675931767033609e-07,
"loss": 2.5396,
"step": 4870
},
{
"epoch": 4.715234988522411,
"grad_norm": 6.709794521331787,
"learning_rate": 8.63415712494369e-07,
"loss": 2.5761,
"step": 4880
},
{
"epoch": 4.724900326205146,
"grad_norm": 6.836709022521973,
"learning_rate": 8.592406757569415e-07,
"loss": 2.5602,
"step": 4890
},
{
"epoch": 4.734565663887882,
"grad_norm": 6.63275146484375,
"learning_rate": 8.550681406927533e-07,
"loss": 2.5477,
"step": 4900
},
{
"epoch": 4.7442310015706175,
"grad_norm": 6.481536865234375,
"learning_rate": 8.508981814590188e-07,
"loss": 2.5443,
"step": 4910
},
{
"epoch": 4.753896339253353,
"grad_norm": 6.737277507781982,
"learning_rate": 8.46730872167172e-07,
"loss": 2.5516,
"step": 4920
},
{
"epoch": 4.763561676936088,
"grad_norm": 6.8111138343811035,
"learning_rate": 8.425662868815509e-07,
"loss": 2.5538,
"step": 4930
},
{
"epoch": 4.773227014618823,
"grad_norm": 6.9275078773498535,
"learning_rate": 8.384044996180805e-07,
"loss": 2.4995,
"step": 4940
},
{
"epoch": 4.777093149691917,
"eval_loss": 2.6439967155456543,
"eval_runtime": 217.1401,
"eval_samples_per_second": 101.639,
"eval_steps_per_second": 4.237,
"step": 4944
},
{
"epoch": 4.782892352301559,
"grad_norm": 6.729687690734863,
"learning_rate": 8.342455843429568e-07,
"loss": 2.5391,
"step": 4950
},
{
"epoch": 4.792557689984294,
"grad_norm": 6.650339126586914,
"learning_rate": 8.300896149713334e-07,
"loss": 2.5402,
"step": 4960
},
{
"epoch": 4.802223027667029,
"grad_norm": 6.950220108032227,
"learning_rate": 8.259366653660071e-07,
"loss": 2.5354,
"step": 4970
},
{
"epoch": 4.811888365349764,
"grad_norm": 6.6570305824279785,
"learning_rate": 8.217868093361053e-07,
"loss": 2.5529,
"step": 4980
},
{
"epoch": 4.8215537030324995,
"grad_norm": 6.657363414764404,
"learning_rate": 8.176401206357742e-07,
"loss": 2.5434,
"step": 4990
},
{
"epoch": 4.831219040715235,
"grad_norm": 6.5188398361206055,
"learning_rate": 8.134966729628683e-07,
"loss": 2.5375,
"step": 5000
},
{
"epoch": 4.84088437839797,
"grad_norm": 6.585175037384033,
"learning_rate": 8.093565399576394e-07,
"loss": 2.537,
"step": 5010
},
{
"epoch": 4.850549716080706,
"grad_norm": 6.604196548461914,
"learning_rate": 8.052197952014296e-07,
"loss": 2.5458,
"step": 5020
},
{
"epoch": 4.860215053763441,
"grad_norm": 6.6913909912109375,
"learning_rate": 8.010865122153627e-07,
"loss": 2.5668,
"step": 5030
},
{
"epoch": 4.869880391446176,
"grad_norm": 6.809241771697998,
"learning_rate": 7.969567644590365e-07,
"loss": 2.5216,
"step": 5040
},
{
"epoch": 4.876646127824091,
"eval_loss": 2.641507863998413,
"eval_runtime": 217.12,
"eval_samples_per_second": 101.649,
"eval_steps_per_second": 4.237,
"step": 5047
},
{
"epoch": 4.879545729128911,
"grad_norm": 6.640186309814453,
"learning_rate": 7.928306253292189e-07,
"loss": 2.556,
"step": 5050
},
{
"epoch": 4.889211066811646,
"grad_norm": 6.924322605133057,
"learning_rate": 7.887081681585432e-07,
"loss": 2.5384,
"step": 5060
},
{
"epoch": 4.898876404494382,
"grad_norm": 6.675876140594482,
"learning_rate": 7.845894662142037e-07,
"loss": 2.4959,
"step": 5070
},
{
"epoch": 4.908541742177118,
"grad_norm": 6.73758602142334,
"learning_rate": 7.804745926966546e-07,
"loss": 2.5323,
"step": 5080
},
{
"epoch": 4.918207079859853,
"grad_norm": 6.658973217010498,
"learning_rate": 7.763636207383079e-07,
"loss": 2.5538,
"step": 5090
},
{
"epoch": 4.927872417542588,
"grad_norm": 6.829054355621338,
"learning_rate": 7.722566234022351e-07,
"loss": 2.539,
"step": 5100
},
{
"epoch": 4.937537755225323,
"grad_norm": 6.643087863922119,
"learning_rate": 7.681536736808678e-07,
"loss": 2.5454,
"step": 5110
},
{
"epoch": 4.947203092908058,
"grad_norm": 6.710460186004639,
"learning_rate": 7.640548444947003e-07,
"loss": 2.5398,
"step": 5120
},
{
"epoch": 4.956868430590793,
"grad_norm": 7.006237030029297,
"learning_rate": 7.599602086909942e-07,
"loss": 2.5618,
"step": 5130
},
{
"epoch": 4.966533768273529,
"grad_norm": 6.898321628570557,
"learning_rate": 7.558698390424829e-07,
"loss": 2.5254,
"step": 5140
},
{
"epoch": 4.9761991059562645,
"grad_norm": 6.923776149749756,
"learning_rate": 7.51783808246079e-07,
"loss": 2.5391,
"step": 5150
},
{
"epoch": 4.9761991059562645,
"eval_loss": 2.640099048614502,
"eval_runtime": 216.9681,
"eval_samples_per_second": 101.72,
"eval_steps_per_second": 4.24,
"step": 5150
},
{
"epoch": 4.985864443639,
"grad_norm": 6.776469707489014,
"learning_rate": 7.477021889215822e-07,
"loss": 2.5117,
"step": 5160
},
{
"epoch": 4.995529781321735,
"grad_norm": 6.805541515350342,
"learning_rate": 7.43625053610388e-07,
"loss": 2.5361,
"step": 5170
},
{
"epoch": 5.004832668841368,
"grad_norm": 6.649566173553467,
"learning_rate": 7.395524747741995e-07,
"loss": 2.5263,
"step": 5180
},
{
"epoch": 5.014498006524103,
"grad_norm": 6.7946977615356445,
"learning_rate": 7.35484524793738e-07,
"loss": 2.5104,
"step": 5190
},
{
"epoch": 5.024163344206838,
"grad_norm": 6.8785624504089355,
"learning_rate": 7.314212759674581e-07,
"loss": 2.5303,
"step": 5200
},
{
"epoch": 5.033828681889574,
"grad_norm": 6.662807464599609,
"learning_rate": 7.273628005102628e-07,
"loss": 2.5041,
"step": 5210
},
{
"epoch": 5.043494019572309,
"grad_norm": 6.747055530548096,
"learning_rate": 7.233091705522179e-07,
"loss": 2.5294,
"step": 5220
},
{
"epoch": 5.053159357255044,
"grad_norm": 6.868143558502197,
"learning_rate": 7.192604581372726e-07,
"loss": 2.4964,
"step": 5230
},
{
"epoch": 5.062824694937779,
"grad_norm": 6.930590629577637,
"learning_rate": 7.152167352219783e-07,
"loss": 2.5184,
"step": 5240
},
{
"epoch": 5.0724900326205145,
"grad_norm": 6.698469638824463,
"learning_rate": 7.111780736742093e-07,
"loss": 2.4956,
"step": 5250
},
{
"epoch": 5.075389633925335,
"eval_loss": 2.642467498779297,
"eval_runtime": 217.038,
"eval_samples_per_second": 101.687,
"eval_steps_per_second": 4.239,
"step": 5253
},
{
"epoch": 5.08215537030325,
"grad_norm": 6.79097318649292,
"learning_rate": 7.071445452718852e-07,
"loss": 2.5243,
"step": 5260
},
{
"epoch": 5.091820707985986,
"grad_norm": 6.664259910583496,
"learning_rate": 7.031162217016966e-07,
"loss": 2.5117,
"step": 5270
},
{
"epoch": 5.101486045668721,
"grad_norm": 6.928717613220215,
"learning_rate": 6.990931745578295e-07,
"loss": 2.5242,
"step": 5280
},
{
"epoch": 5.111151383351456,
"grad_norm": 6.610539436340332,
"learning_rate": 6.950754753406937e-07,
"loss": 2.474,
"step": 5290
},
{
"epoch": 5.120816721034191,
"grad_norm": 6.7854790687561035,
"learning_rate": 6.910631954556522e-07,
"loss": 2.4971,
"step": 5300
},
{
"epoch": 5.130482058716926,
"grad_norm": 6.860823154449463,
"learning_rate": 6.870564062117517e-07,
"loss": 2.5279,
"step": 5310
},
{
"epoch": 5.140147396399661,
"grad_norm": 6.848371982574463,
"learning_rate": 6.830551788204551e-07,
"loss": 2.5232,
"step": 5320
},
{
"epoch": 5.149812734082397,
"grad_norm": 6.649808883666992,
"learning_rate": 6.790595843943768e-07,
"loss": 2.4639,
"step": 5330
},
{
"epoch": 5.1594780717651325,
"grad_norm": 6.732810974121094,
"learning_rate": 6.750696939460172e-07,
"loss": 2.5131,
"step": 5340
},
{
"epoch": 5.169143409447868,
"grad_norm": 6.754150867462158,
"learning_rate": 6.710855783865025e-07,
"loss": 2.5343,
"step": 5350
},
{
"epoch": 5.174942612057508,
"eval_loss": 2.640747547149658,
"eval_runtime": 216.8589,
"eval_samples_per_second": 101.771,
"eval_steps_per_second": 4.242,
"step": 5356
},
{
"epoch": 5.178808747130603,
"grad_norm": 6.860896587371826,
"learning_rate": 6.671073085243233e-07,
"loss": 2.5313,
"step": 5360
},
{
"epoch": 5.188474084813338,
"grad_norm": 6.547597408294678,
"learning_rate": 6.631349550640764e-07,
"loss": 2.5394,
"step": 5370
},
{
"epoch": 5.198139422496073,
"grad_norm": 6.75359582901001,
"learning_rate": 6.591685886052079e-07,
"loss": 2.5012,
"step": 5380
},
{
"epoch": 5.207804760178809,
"grad_norm": 6.7049455642700195,
"learning_rate": 6.552082796407589e-07,
"loss": 2.5162,
"step": 5390
},
{
"epoch": 5.217470097861544,
"grad_norm": 6.600377082824707,
"learning_rate": 6.512540985561125e-07,
"loss": 2.5347,
"step": 5400
},
{
"epoch": 5.227135435544279,
"grad_norm": 6.767351150512695,
"learning_rate": 6.473061156277431e-07,
"loss": 2.4873,
"step": 5410
},
{
"epoch": 5.2368007732270145,
"grad_norm": 6.60963773727417,
"learning_rate": 6.433644010219661e-07,
"loss": 2.4962,
"step": 5420
},
{
"epoch": 5.24646611090975,
"grad_norm": 6.64891242980957,
"learning_rate": 6.394290247936931e-07,
"loss": 2.5101,
"step": 5430
},
{
"epoch": 5.256131448592485,
"grad_norm": 6.72605037689209,
"learning_rate": 6.355000568851849e-07,
"loss": 2.4919,
"step": 5440
},
{
"epoch": 5.265796786275221,
"grad_norm": 7.019696235656738,
"learning_rate": 6.315775671248098e-07,
"loss": 2.4964,
"step": 5450
},
{
"epoch": 5.274495590189682,
"eval_loss": 2.639216184616089,
"eval_runtime": 216.9692,
"eval_samples_per_second": 101.719,
"eval_steps_per_second": 4.24,
"step": 5459
},
{
"epoch": 5.275462123957956,
"grad_norm": 6.810911178588867,
"learning_rate": 6.276616252258014e-07,
"loss": 2.499,
"step": 5460
},
{
"epoch": 5.285127461640691,
"grad_norm": 6.777184009552002,
"learning_rate": 6.237523007850196e-07,
"loss": 2.5246,
"step": 5470
},
{
"epoch": 5.294792799323426,
"grad_norm": 6.932974338531494,
"learning_rate": 6.198496632817154e-07,
"loss": 2.5384,
"step": 5480
},
{
"epoch": 5.304458137006161,
"grad_norm": 6.819568634033203,
"learning_rate": 6.159537820762942e-07,
"loss": 2.5205,
"step": 5490
},
{
"epoch": 5.314123474688897,
"grad_norm": 6.879855632781982,
"learning_rate": 6.120647264090838e-07,
"loss": 2.5193,
"step": 5500
},
{
"epoch": 5.323788812371633,
"grad_norm": 6.756879806518555,
"learning_rate": 6.081825653991037e-07,
"loss": 2.4839,
"step": 5510
},
{
"epoch": 5.333454150054368,
"grad_norm": 6.868306636810303,
"learning_rate": 6.043073680428372e-07,
"loss": 2.5402,
"step": 5520
},
{
"epoch": 5.343119487737103,
"grad_norm": 6.817827224731445,
"learning_rate": 6.004392032130039e-07,
"loss": 2.5098,
"step": 5530
},
{
"epoch": 5.352784825419838,
"grad_norm": 6.814731597900391,
"learning_rate": 5.965781396573367e-07,
"loss": 2.5142,
"step": 5540
},
{
"epoch": 5.362450163102573,
"grad_norm": 7.052011013031006,
"learning_rate": 5.9272424599736e-07,
"loss": 2.5233,
"step": 5550
},
{
"epoch": 5.372115500785308,
"grad_norm": 6.802794456481934,
"learning_rate": 5.888775907271696e-07,
"loss": 2.5157,
"step": 5560
},
{
"epoch": 5.3740485683218555,
"eval_loss": 2.6385231018066406,
"eval_runtime": 216.8661,
"eval_samples_per_second": 101.768,
"eval_steps_per_second": 4.242,
"step": 5562
},
{
"epoch": 5.381780838468044,
"grad_norm": 6.800307273864746,
"learning_rate": 5.850382422122159e-07,
"loss": 2.5085,
"step": 5570
},
{
"epoch": 5.3914461761507795,
"grad_norm": 6.743267059326172,
"learning_rate": 5.812062686880879e-07,
"loss": 2.5179,
"step": 5580
},
{
"epoch": 5.401111513833515,
"grad_norm": 6.773606300354004,
"learning_rate": 5.773817382593007e-07,
"loss": 2.5203,
"step": 5590
},
{
"epoch": 5.41077685151625,
"grad_norm": 7.061532497406006,
"learning_rate": 5.735647188980871e-07,
"loss": 2.5013,
"step": 5600
},
{
"epoch": 5.420442189198985,
"grad_norm": 6.810518264770508,
"learning_rate": 5.697552784431865e-07,
"loss": 2.4812,
"step": 5610
},
{
"epoch": 5.43010752688172,
"grad_norm": 6.675147533416748,
"learning_rate": 5.659534845986417e-07,
"loss": 2.5223,
"step": 5620
},
{
"epoch": 5.439772864564456,
"grad_norm": 6.812685489654541,
"learning_rate": 5.621594049325939e-07,
"loss": 2.4554,
"step": 5630
},
{
"epoch": 5.449438202247191,
"grad_norm": 6.6521992683410645,
"learning_rate": 5.583731068760822e-07,
"loss": 2.5003,
"step": 5640
},
{
"epoch": 5.459103539929926,
"grad_norm": 6.802224159240723,
"learning_rate": 5.545946577218469e-07,
"loss": 2.509,
"step": 5650
},
{
"epoch": 5.4687688776126615,
"grad_norm": 6.921814441680908,
"learning_rate": 5.508241246231303e-07,
"loss": 2.5157,
"step": 5660
},
{
"epoch": 5.473601546454029,
"eval_loss": 2.6369357109069824,
"eval_runtime": 217.0718,
"eval_samples_per_second": 101.671,
"eval_steps_per_second": 4.238,
"step": 5665
},
{
"epoch": 5.478434215295397,
"grad_norm": 6.57248592376709,
"learning_rate": 5.470615745924869e-07,
"loss": 2.5003,
"step": 5670
},
{
"epoch": 5.488099552978132,
"grad_norm": 6.756587028503418,
"learning_rate": 5.433070745005889e-07,
"loss": 2.4746,
"step": 5680
},
{
"epoch": 5.497764890660868,
"grad_norm": 6.835014343261719,
"learning_rate": 5.395606910750401e-07,
"loss": 2.4871,
"step": 5690
},
{
"epoch": 5.507430228343603,
"grad_norm": 7.088858604431152,
"learning_rate": 5.358224908991895e-07,
"loss": 2.4799,
"step": 5700
},
{
"epoch": 5.517095566026338,
"grad_norm": 7.005571365356445,
"learning_rate": 5.320925404109466e-07,
"loss": 2.5439,
"step": 5710
},
{
"epoch": 5.526760903709073,
"grad_norm": 6.776179313659668,
"learning_rate": 5.283709059016033e-07,
"loss": 2.5226,
"step": 5720
},
{
"epoch": 5.536426241391808,
"grad_norm": 6.528167247772217,
"learning_rate": 5.246576535146523e-07,
"loss": 2.516,
"step": 5730
},
{
"epoch": 5.5460915790745435,
"grad_norm": 6.705376625061035,
"learning_rate": 5.20952849244615e-07,
"loss": 2.5408,
"step": 5740
},
{
"epoch": 5.55575691675728,
"grad_norm": 6.841287136077881,
"learning_rate": 5.172565589358658e-07,
"loss": 2.4921,
"step": 5750
},
{
"epoch": 5.565422254440015,
"grad_norm": 6.934755802154541,
"learning_rate": 5.13568848281463e-07,
"loss": 2.5019,
"step": 5760
},
{
"epoch": 5.573154524586203,
"eval_loss": 2.6358206272125244,
"eval_runtime": 216.9617,
"eval_samples_per_second": 101.723,
"eval_steps_per_second": 4.24,
"step": 5768
},
{
"epoch": 5.57508759212275,
"grad_norm": 6.7719621658325195,
"learning_rate": 5.09889782821983e-07,
"loss": 2.5053,
"step": 5770
},
{
"epoch": 5.584752929805485,
"grad_norm": 6.658035755157471,
"learning_rate": 5.062194279443508e-07,
"loss": 2.5161,
"step": 5780
},
{
"epoch": 5.59441826748822,
"grad_norm": 7.038604736328125,
"learning_rate": 5.025578488806836e-07,
"loss": 2.5208,
"step": 5790
},
{
"epoch": 5.604083605170955,
"grad_norm": 6.811711311340332,
"learning_rate": 4.989051107071268e-07,
"loss": 2.5461,
"step": 5800
},
{
"epoch": 5.613748942853691,
"grad_norm": 7.0844526290893555,
"learning_rate": 4.952612783427008e-07,
"loss": 2.5076,
"step": 5810
},
{
"epoch": 5.6234142805364264,
"grad_norm": 6.744603157043457,
"learning_rate": 4.916264165481448e-07,
"loss": 2.4963,
"step": 5820
},
{
"epoch": 5.633079618219162,
"grad_norm": 6.718221187591553,
"learning_rate": 4.880005899247663e-07,
"loss": 2.4961,
"step": 5830
},
{
"epoch": 5.642744955901897,
"grad_norm": 7.078195095062256,
"learning_rate": 4.843838629132949e-07,
"loss": 2.5171,
"step": 5840
},
{
"epoch": 5.652410293584632,
"grad_norm": 6.604617118835449,
"learning_rate": 4.80776299792734e-07,
"loss": 2.5146,
"step": 5850
},
{
"epoch": 5.662075631267367,
"grad_norm": 6.939282417297363,
"learning_rate": 4.771779646792216e-07,
"loss": 2.4872,
"step": 5860
},
{
"epoch": 5.671740968950103,
"grad_norm": 7.01817512512207,
"learning_rate": 4.7358892152488726e-07,
"loss": 2.487,
"step": 5870
},
{
"epoch": 5.672707502718376,
"eval_loss": 2.6341195106506348,
"eval_runtime": 216.8979,
"eval_samples_per_second": 101.753,
"eval_steps_per_second": 4.242,
"step": 5871
},
{
"epoch": 5.681406306632838,
"grad_norm": 7.151918888092041,
"learning_rate": 4.700092341167182e-07,
"loss": 2.5233,
"step": 5880
},
{
"epoch": 5.691071644315573,
"grad_norm": 6.588758945465088,
"learning_rate": 4.664389660754253e-07,
"loss": 2.5248,
"step": 5890
},
{
"epoch": 5.7007369819983085,
"grad_norm": 6.761035442352295,
"learning_rate": 4.6287818085431064e-07,
"loss": 2.4915,
"step": 5900
},
{
"epoch": 5.710402319681044,
"grad_norm": 6.56200647354126,
"learning_rate": 4.5932694173814246e-07,
"loss": 2.5467,
"step": 5910
},
{
"epoch": 5.72006765736378,
"grad_norm": 6.924496650695801,
"learning_rate": 4.5578531184202726e-07,
"loss": 2.5266,
"step": 5920
},
{
"epoch": 5.729732995046515,
"grad_norm": 6.611005783081055,
"learning_rate": 4.522533541102914e-07,
"loss": 2.5247,
"step": 5930
},
{
"epoch": 5.73939833272925,
"grad_norm": 6.712198734283447,
"learning_rate": 4.487311313153598e-07,
"loss": 2.4902,
"step": 5940
},
{
"epoch": 5.749063670411985,
"grad_norm": 6.897911071777344,
"learning_rate": 4.452187060566409e-07,
"loss": 2.4757,
"step": 5950
},
{
"epoch": 5.75872900809472,
"grad_norm": 6.599294185638428,
"learning_rate": 4.417161407594163e-07,
"loss": 2.5007,
"step": 5960
},
{
"epoch": 5.768394345777455,
"grad_norm": 6.836195945739746,
"learning_rate": 4.3822349767372667e-07,
"loss": 2.511,
"step": 5970
},
{
"epoch": 5.77226048085055,
"eval_loss": 2.633843421936035,
"eval_runtime": 216.737,
"eval_samples_per_second": 101.828,
"eval_steps_per_second": 4.245,
"step": 5974
},
{
"epoch": 5.7780596834601905,
"grad_norm": 6.8379974365234375,
"learning_rate": 4.3474083887327076e-07,
"loss": 2.5229,
"step": 5980
},
{
"epoch": 5.7877250211429265,
"grad_norm": 6.931849002838135,
"learning_rate": 4.312682262542978e-07,
"loss": 2.4955,
"step": 5990
},
{
"epoch": 5.797390358825662,
"grad_norm": 6.655295372009277,
"learning_rate": 4.278057215345109e-07,
"loss": 2.5219,
"step": 6000
},
{
"epoch": 5.807055696508397,
"grad_norm": 6.884836673736572,
"learning_rate": 4.2435338625196727e-07,
"loss": 2.5298,
"step": 6010
},
{
"epoch": 5.816721034191132,
"grad_norm": 6.96956205368042,
"learning_rate": 4.209112817639856e-07,
"loss": 2.5129,
"step": 6020
},
{
"epoch": 5.826386371873867,
"grad_norm": 6.961551666259766,
"learning_rate": 4.174794692460571e-07,
"loss": 2.4823,
"step": 6030
},
{
"epoch": 5.836051709556603,
"grad_norm": 6.782859802246094,
"learning_rate": 4.1405800969075534e-07,
"loss": 2.5208,
"step": 6040
},
{
"epoch": 5.845717047239338,
"grad_norm": 6.812822341918945,
"learning_rate": 4.106469639066552e-07,
"loss": 2.4742,
"step": 6050
},
{
"epoch": 5.855382384922073,
"grad_norm": 6.870484352111816,
"learning_rate": 4.072463925172497e-07,
"loss": 2.5023,
"step": 6060
},
{
"epoch": 5.8650477226048086,
"grad_norm": 6.649935722351074,
"learning_rate": 4.0385635595987344e-07,
"loss": 2.5102,
"step": 6070
},
{
"epoch": 5.871813458982723,
"eval_loss": 2.6324048042297363,
"eval_runtime": 216.9994,
"eval_samples_per_second": 101.705,
"eval_steps_per_second": 4.24,
"step": 6077
},
{
"epoch": 5.874713060287544,
"grad_norm": 6.996982097625732,
"learning_rate": 4.004769144846299e-07,
"loss": 2.523,
"step": 6080
},
{
"epoch": 5.884378397970279,
"grad_norm": 6.676599979400635,
"learning_rate": 3.9710812815331797e-07,
"loss": 2.4997,
"step": 6090
},
{
"epoch": 5.894043735653014,
"grad_norm": 6.77655029296875,
"learning_rate": 3.9375005683836683e-07,
"loss": 2.498,
"step": 6100
},
{
"epoch": 5.90370907333575,
"grad_norm": 6.703573703765869,
"learning_rate": 3.9040276022176996e-07,
"loss": 2.5095,
"step": 6110
},
{
"epoch": 5.913374411018485,
"grad_norm": 6.955379486083984,
"learning_rate": 3.870662977940264e-07,
"loss": 2.5226,
"step": 6120
},
{
"epoch": 5.92303974870122,
"grad_norm": 6.614869117736816,
"learning_rate": 3.8374072885308184e-07,
"loss": 2.5237,
"step": 6130
},
{
"epoch": 5.932705086383955,
"grad_norm": 6.875811576843262,
"learning_rate": 3.8042611250327516e-07,
"loss": 2.551,
"step": 6140
},
{
"epoch": 5.942370424066691,
"grad_norm": 6.832152843475342,
"learning_rate": 3.7712250765428824e-07,
"loss": 2.4825,
"step": 6150
},
{
"epoch": 5.952035761749427,
"grad_norm": 6.9608683586120605,
"learning_rate": 3.738299730200987e-07,
"loss": 2.4832,
"step": 6160
},
{
"epoch": 5.961701099432162,
"grad_norm": 6.698019504547119,
"learning_rate": 3.7054856711793736e-07,
"loss": 2.5244,
"step": 6170
},
{
"epoch": 5.971366437114897,
"grad_norm": 7.041402816772461,
"learning_rate": 3.6727834826724634e-07,
"loss": 2.5294,
"step": 6180
},
{
"epoch": 5.971366437114897,
"eval_loss": 2.631476879119873,
"eval_runtime": 216.9835,
"eval_samples_per_second": 101.713,
"eval_steps_per_second": 4.24,
"step": 6180
},
{
"epoch": 5.981031774797632,
"grad_norm": 6.974484443664551,
"learning_rate": 3.640193745886446e-07,
"loss": 2.5173,
"step": 6190
},
{
"epoch": 5.990697112480367,
"grad_norm": 6.792074680328369,
"learning_rate": 3.6077170400289337e-07,
"loss": 2.5276,
"step": 6200
},
{
"epoch": 6.0,
"grad_norm": 9.531786918640137,
"learning_rate": 3.5753539422986725e-07,
"loss": 2.5002,
"step": 6210
},
{
"epoch": 6.009665337682735,
"grad_norm": 6.820946216583252,
"learning_rate": 3.543105027875296e-07,
"loss": 2.5161,
"step": 6220
},
{
"epoch": 6.01933067536547,
"grad_norm": 6.973489284515381,
"learning_rate": 3.5109708699090777e-07,
"loss": 2.5012,
"step": 6230
},
{
"epoch": 6.028996013048205,
"grad_norm": 7.010828495025635,
"learning_rate": 3.478952039510774e-07,
"loss": 2.4784,
"step": 6240
},
{
"epoch": 6.0386613507309415,
"grad_norm": 6.972498416900635,
"learning_rate": 3.4470491057414475e-07,
"loss": 2.5262,
"step": 6250
},
{
"epoch": 6.048326688413677,
"grad_norm": 6.622044563293457,
"learning_rate": 3.415262635602364e-07,
"loss": 2.4681,
"step": 6260
},
{
"epoch": 6.057992026096412,
"grad_norm": 7.308681011199951,
"learning_rate": 3.3835931940249294e-07,
"loss": 2.5143,
"step": 6270
},
{
"epoch": 6.067657363779147,
"grad_norm": 6.736526966094971,
"learning_rate": 3.352041343860621e-07,
"loss": 2.4933,
"step": 6280
},
{
"epoch": 6.070556965083967,
"eval_loss": 2.6330554485321045,
"eval_runtime": 216.8848,
"eval_samples_per_second": 101.759,
"eval_steps_per_second": 4.242,
"step": 6283
},
{
"epoch": 6.077322701461882,
"grad_norm": 6.952606201171875,
"learning_rate": 3.320607645871011e-07,
"loss": 2.4915,
"step": 6290
},
{
"epoch": 6.086988039144618,
"grad_norm": 6.871399402618408,
"learning_rate": 3.289292658717776e-07,
"loss": 2.4658,
"step": 6300
},
{
"epoch": 6.096653376827353,
"grad_norm": 6.710361480712891,
"learning_rate": 3.2580969389527955e-07,
"loss": 2.4651,
"step": 6310
},
{
"epoch": 6.106318714510088,
"grad_norm": 6.776742458343506,
"learning_rate": 3.2270210410082345e-07,
"loss": 2.5001,
"step": 6320
},
{
"epoch": 6.1159840521928235,
"grad_norm": 6.880600452423096,
"learning_rate": 3.1960655171867037e-07,
"loss": 2.467,
"step": 6330
},
{
"epoch": 6.125649389875559,
"grad_norm": 6.831501483917236,
"learning_rate": 3.1652309176514405e-07,
"loss": 2.4905,
"step": 6340
},
{
"epoch": 6.135314727558294,
"grad_norm": 6.918708324432373,
"learning_rate": 3.134517790416528e-07,
"loss": 2.4513,
"step": 6350
},
{
"epoch": 6.144980065241029,
"grad_norm": 6.8542561531066895,
"learning_rate": 3.103926681337168e-07,
"loss": 2.5004,
"step": 6360
},
{
"epoch": 6.154645402923765,
"grad_norm": 6.689946174621582,
"learning_rate": 3.07345813409996e-07,
"loss": 2.4817,
"step": 6370
},
{
"epoch": 6.1643107406065,
"grad_norm": 6.745342254638672,
"learning_rate": 3.0431126902132575e-07,
"loss": 2.492,
"step": 6380
},
{
"epoch": 6.170109943216141,
"eval_loss": 2.632904291152954,
"eval_runtime": 216.9357,
"eval_samples_per_second": 101.735,
"eval_steps_per_second": 4.241,
"step": 6386
},
{
"epoch": 6.173976078289235,
"grad_norm": 6.870169639587402,
"learning_rate": 3.012890888997528e-07,
"loss": 2.4701,
"step": 6390
},
{
"epoch": 6.18364141597197,
"grad_norm": 6.764638900756836,
"learning_rate": 2.982793267575775e-07,
"loss": 2.4718,
"step": 6400
},
{
"epoch": 6.1933067536547055,
"grad_norm": 7.040791988372803,
"learning_rate": 2.952820360863999e-07,
"loss": 2.4883,
"step": 6410
},
{
"epoch": 6.2029720913374415,
"grad_norm": 6.948042392730713,
"learning_rate": 2.9229727015616735e-07,
"loss": 2.4951,
"step": 6420
},
{
"epoch": 6.212637429020177,
"grad_norm": 6.7030463218688965,
"learning_rate": 2.893250820142299e-07,
"loss": 2.4897,
"step": 6430
},
{
"epoch": 6.222302766702912,
"grad_norm": 6.751480579376221,
"learning_rate": 2.863655244843949e-07,
"loss": 2.5016,
"step": 6440
},
{
"epoch": 6.231968104385647,
"grad_norm": 6.898858070373535,
"learning_rate": 2.834186501659912e-07,
"loss": 2.5046,
"step": 6450
},
{
"epoch": 6.241633442068382,
"grad_norm": 6.5952301025390625,
"learning_rate": 2.804845114329316e-07,
"loss": 2.5032,
"step": 6460
},
{
"epoch": 6.251298779751117,
"grad_norm": 6.756124019622803,
"learning_rate": 2.7756316043278315e-07,
"loss": 2.5008,
"step": 6470
},
{
"epoch": 6.260964117433852,
"grad_norm": 6.867387771606445,
"learning_rate": 2.7465464908584135e-07,
"loss": 2.5096,
"step": 6480
},
{
"epoch": 6.269662921348314,
"eval_loss": 2.6323742866516113,
"eval_runtime": 216.8606,
"eval_samples_per_second": 101.77,
"eval_steps_per_second": 4.242,
"step": 6489
},
{
"epoch": 6.270629455116588,
"grad_norm": 6.6687703132629395,
"learning_rate": 2.717590290842051e-07,
"loss": 2.5012,
"step": 6490
},
{
"epoch": 6.280294792799324,
"grad_norm": 7.008709907531738,
"learning_rate": 2.6887635189086077e-07,
"loss": 2.4931,
"step": 6500
},
{
"epoch": 6.289960130482059,
"grad_norm": 6.839788436889648,
"learning_rate": 2.6600666873876474e-07,
"loss": 2.4933,
"step": 6510
},
{
"epoch": 6.299625468164794,
"grad_norm": 6.8228654861450195,
"learning_rate": 2.631500306299349e-07,
"loss": 2.4999,
"step": 6520
},
{
"epoch": 6.309290805847529,
"grad_norm": 6.8285393714904785,
"learning_rate": 2.6030648833454307e-07,
"loss": 2.4577,
"step": 6530
},
{
"epoch": 6.318956143530265,
"grad_norm": 6.882188320159912,
"learning_rate": 2.574760923900133e-07,
"loss": 2.4612,
"step": 6540
},
{
"epoch": 6.328621481213,
"grad_norm": 6.814681053161621,
"learning_rate": 2.54658893100124e-07,
"loss": 2.4773,
"step": 6550
},
{
"epoch": 6.338286818895735,
"grad_norm": 6.943098068237305,
"learning_rate": 2.518549405341125e-07,
"loss": 2.5,
"step": 6560
},
{
"epoch": 6.3479521565784705,
"grad_norm": 6.890428066253662,
"learning_rate": 2.49064284525787e-07,
"loss": 2.5145,
"step": 6570
},
{
"epoch": 6.357617494261206,
"grad_norm": 6.88997220993042,
"learning_rate": 2.4628697467263913e-07,
"loss": 2.486,
"step": 6580
},
{
"epoch": 6.367282831943941,
"grad_norm": 6.790536403656006,
"learning_rate": 2.435230603349635e-07,
"loss": 2.4943,
"step": 6590
},
{
"epoch": 6.369215899480488,
"eval_loss": 2.6316308975219727,
"eval_runtime": 216.8902,
"eval_samples_per_second": 101.757,
"eval_steps_per_second": 4.242,
"step": 6592
},
{
"epoch": 6.376948169626676,
"grad_norm": 6.717286109924316,
"learning_rate": 2.4077259063498087e-07,
"loss": 2.4916,
"step": 6600
},
{
"epoch": 6.386613507309412,
"grad_norm": 6.9318647384643555,
"learning_rate": 2.3803561445596366e-07,
"loss": 2.5037,
"step": 6610
},
{
"epoch": 6.396278844992147,
"grad_norm": 7.279837608337402,
"learning_rate": 2.353121804413687e-07,
"loss": 2.4524,
"step": 6620
},
{
"epoch": 6.405944182674882,
"grad_norm": 6.827029705047607,
"learning_rate": 2.3260233699397126e-07,
"loss": 2.4705,
"step": 6630
},
{
"epoch": 6.415609520357617,
"grad_norm": 6.849607467651367,
"learning_rate": 2.2990613227500645e-07,
"loss": 2.4501,
"step": 6640
},
{
"epoch": 6.4252748580403525,
"grad_norm": 6.832201957702637,
"learning_rate": 2.272236142033115e-07,
"loss": 2.4941,
"step": 6650
},
{
"epoch": 6.4349401957230885,
"grad_norm": 7.062440395355225,
"learning_rate": 2.2455483045447498e-07,
"loss": 2.4879,
"step": 6660
},
{
"epoch": 6.444605533405824,
"grad_norm": 7.106632709503174,
"learning_rate": 2.2189982845999057e-07,
"loss": 2.5048,
"step": 6670
},
{
"epoch": 6.454270871088559,
"grad_norm": 6.965383052825928,
"learning_rate": 2.1925865540641132e-07,
"loss": 2.5375,
"step": 6680
},
{
"epoch": 6.463936208771294,
"grad_norm": 6.780632972717285,
"learning_rate": 2.1663135823451418e-07,
"loss": 2.4805,
"step": 6690
},
{
"epoch": 6.4687688776126615,
"eval_loss": 2.63108229637146,
"eval_runtime": 216.9071,
"eval_samples_per_second": 101.749,
"eval_steps_per_second": 4.241,
"step": 6695
},
{
"epoch": 6.473601546454029,
"grad_norm": 6.964837074279785,
"learning_rate": 2.1401798363846336e-07,
"loss": 2.5012,
"step": 6700
},
{
"epoch": 6.483266884136764,
"grad_norm": 6.848756313323975,
"learning_rate": 2.1141857806498143e-07,
"loss": 2.4931,
"step": 6710
},
{
"epoch": 6.492932221819499,
"grad_norm": 6.881086349487305,
"learning_rate": 2.088331877125238e-07,
"loss": 2.5012,
"step": 6720
},
{
"epoch": 6.502597559502235,
"grad_norm": 6.669610500335693,
"learning_rate": 2.062618585304573e-07,
"loss": 2.488,
"step": 6730
},
{
"epoch": 6.5122628971849705,
"grad_norm": 6.833005905151367,
"learning_rate": 2.037046362182444e-07,
"loss": 2.4838,
"step": 6740
},
{
"epoch": 6.521928234867706,
"grad_norm": 6.958085060119629,
"learning_rate": 2.0116156622462977e-07,
"loss": 2.4601,
"step": 6750
},
{
"epoch": 6.531593572550441,
"grad_norm": 6.800379276275635,
"learning_rate": 1.986326937468339e-07,
"loss": 2.4848,
"step": 6760
},
{
"epoch": 6.541258910233176,
"grad_norm": 6.915721416473389,
"learning_rate": 1.9611806372974816e-07,
"loss": 2.4784,
"step": 6770
},
{
"epoch": 6.550924247915912,
"grad_norm": 6.845888614654541,
"learning_rate": 1.936177208651374e-07,
"loss": 2.5028,
"step": 6780
},
{
"epoch": 6.560589585598647,
"grad_norm": 6.784244060516357,
"learning_rate": 1.9113170959084569e-07,
"loss": 2.4622,
"step": 6790
},
{
"epoch": 6.568321855744835,
"eval_loss": 2.6306145191192627,
"eval_runtime": 216.9083,
"eval_samples_per_second": 101.748,
"eval_steps_per_second": 4.241,
"step": 6798
},
{
"epoch": 6.570254923281382,
"grad_norm": 6.741705894470215,
"learning_rate": 1.8866007409000495e-07,
"loss": 2.4956,
"step": 6800
},
{
"epoch": 6.579920260964117,
"grad_norm": 6.939857006072998,
"learning_rate": 1.8620285829025196e-07,
"loss": 2.4728,
"step": 6810
},
{
"epoch": 6.589585598646853,
"grad_norm": 6.82913064956665,
"learning_rate": 1.8376010586294542e-07,
"loss": 2.457,
"step": 6820
},
{
"epoch": 6.599250936329588,
"grad_norm": 6.889610290527344,
"learning_rate": 1.8133186022239188e-07,
"loss": 2.4735,
"step": 6830
},
{
"epoch": 6.608916274012323,
"grad_norm": 6.8231120109558105,
"learning_rate": 1.7891816452507236e-07,
"loss": 2.4836,
"step": 6840
},
{
"epoch": 6.618581611695059,
"grad_norm": 6.723885536193848,
"learning_rate": 1.7651906166887598e-07,
"loss": 2.4752,
"step": 6850
},
{
"epoch": 6.628246949377794,
"grad_norm": 6.732489585876465,
"learning_rate": 1.7413459429233857e-07,
"loss": 2.4853,
"step": 6860
},
{
"epoch": 6.637912287060529,
"grad_norm": 6.842257022857666,
"learning_rate": 1.71764804773883e-07,
"loss": 2.5025,
"step": 6870
},
{
"epoch": 6.647577624743264,
"grad_norm": 6.585004806518555,
"learning_rate": 1.6940973523106794e-07,
"loss": 2.4823,
"step": 6880
},
{
"epoch": 6.657242962425999,
"grad_norm": 6.8570966720581055,
"learning_rate": 1.6706942751983745e-07,
"loss": 2.5025,
"step": 6890
},
{
"epoch": 6.6669083001087355,
"grad_norm": 6.777892112731934,
"learning_rate": 1.6474392323377828e-07,
"loss": 2.4994,
"step": 6900
},
{
"epoch": 6.667874833877009,
"eval_loss": 2.6304829120635986,
"eval_runtime": 216.9988,
"eval_samples_per_second": 101.706,
"eval_steps_per_second": 4.24,
"step": 6901
},
{
"epoch": 6.676573637791471,
"grad_norm": 7.168168544769287,
"learning_rate": 1.6243326370338062e-07,
"loss": 2.5063,
"step": 6910
},
{
"epoch": 6.686238975474206,
"grad_norm": 6.96233606338501,
"learning_rate": 1.6013748999530276e-07,
"loss": 2.4905,
"step": 6920
},
{
"epoch": 6.695904313156941,
"grad_norm": 6.856019496917725,
"learning_rate": 1.5785664291164246e-07,
"loss": 2.4624,
"step": 6930
},
{
"epoch": 6.705569650839676,
"grad_norm": 6.714591979980469,
"learning_rate": 1.5559076298921025e-07,
"loss": 2.4877,
"step": 6940
},
{
"epoch": 6.715234988522411,
"grad_norm": 6.764096736907959,
"learning_rate": 1.5333989049881058e-07,
"loss": 2.4733,
"step": 6950
},
{
"epoch": 6.724900326205146,
"grad_norm": 6.856327056884766,
"learning_rate": 1.511040654445247e-07,
"loss": 2.4986,
"step": 6960
},
{
"epoch": 6.734565663887882,
"grad_norm": 6.840289115905762,
"learning_rate": 1.4888332756300027e-07,
"loss": 2.4862,
"step": 6970
},
{
"epoch": 6.7442310015706175,
"grad_norm": 7.013392448425293,
"learning_rate": 1.4667771632274596e-07,
"loss": 2.5031,
"step": 6980
},
{
"epoch": 6.753896339253353,
"grad_norm": 6.860467433929443,
"learning_rate": 1.4448727092342816e-07,
"loss": 2.511,
"step": 6990
},
{
"epoch": 6.763561676936088,
"grad_norm": 6.86760139465332,
"learning_rate": 1.4231203029517615e-07,
"loss": 2.5187,
"step": 7000
},
{
"epoch": 6.767427812009182,
"eval_loss": 2.6297874450683594,
"eval_runtime": 216.6476,
"eval_samples_per_second": 101.871,
"eval_steps_per_second": 4.247,
"step": 7004
},
{
"epoch": 6.773227014618823,
"grad_norm": 6.677191734313965,
"learning_rate": 1.4015203309788848e-07,
"loss": 2.4913,
"step": 7010
},
{
"epoch": 6.782892352301559,
"grad_norm": 6.7386155128479,
"learning_rate": 1.3800731772054796e-07,
"loss": 2.4563,
"step": 7020
},
{
"epoch": 6.792557689984294,
"grad_norm": 6.725878715515137,
"learning_rate": 1.3587792228053718e-07,
"loss": 2.4936,
"step": 7030
},
{
"epoch": 6.802223027667029,
"grad_norm": 6.603059768676758,
"learning_rate": 1.3376388462296217e-07,
"loss": 2.4484,
"step": 7040
},
{
"epoch": 6.811888365349764,
"grad_norm": 6.6211042404174805,
"learning_rate": 1.3166524231998055e-07,
"loss": 2.4887,
"step": 7050
},
{
"epoch": 6.8215537030324995,
"grad_norm": 6.715404510498047,
"learning_rate": 1.295820326701319e-07,
"loss": 2.5034,
"step": 7060
},
{
"epoch": 6.831219040715235,
"grad_norm": 6.930505275726318,
"learning_rate": 1.2751429269767667e-07,
"loss": 2.4997,
"step": 7070
},
{
"epoch": 6.84088437839797,
"grad_norm": 6.84429931640625,
"learning_rate": 1.2546205915193687e-07,
"loss": 2.5128,
"step": 7080
},
{
"epoch": 6.850549716080706,
"grad_norm": 7.0082478523254395,
"learning_rate": 1.2342536850664354e-07,
"loss": 2.4893,
"step": 7090
},
{
"epoch": 6.860215053763441,
"grad_norm": 6.885166645050049,
"learning_rate": 1.2140425695928858e-07,
"loss": 2.5144,
"step": 7100
},
{
"epoch": 6.866980790141356,
"eval_loss": 2.629671335220337,
"eval_runtime": 216.8264,
"eval_samples_per_second": 101.786,
"eval_steps_per_second": 4.243,
"step": 7107
},
{
"epoch": 6.869880391446176,
"grad_norm": 7.1056599617004395,
"learning_rate": 1.193987604304809e-07,
"loss": 2.4867,
"step": 7110
},
{
"epoch": 6.879545729128911,
"grad_norm": 6.84068489074707,
"learning_rate": 1.1740891456330892e-07,
"loss": 2.4736,
"step": 7120
},
{
"epoch": 6.889211066811646,
"grad_norm": 6.737675189971924,
"learning_rate": 1.1543475472270613e-07,
"loss": 2.4826,
"step": 7130
},
{
"epoch": 6.898876404494382,
"grad_norm": 6.845552444458008,
"learning_rate": 1.1347631599482321e-07,
"loss": 2.5346,
"step": 7140
},
{
"epoch": 6.908541742177118,
"grad_norm": 6.8293776512146,
"learning_rate": 1.1153363318640396e-07,
"loss": 2.4788,
"step": 7150
},
{
"epoch": 6.918207079859853,
"grad_norm": 6.818983554840088,
"learning_rate": 1.096067408241671e-07,
"loss": 2.5074,
"step": 7160
},
{
"epoch": 6.927872417542588,
"grad_norm": 6.84577751159668,
"learning_rate": 1.076956731541927e-07,
"loss": 2.494,
"step": 7170
},
{
"epoch": 6.937537755225323,
"grad_norm": 6.472375869750977,
"learning_rate": 1.0580046414131261e-07,
"loss": 2.5109,
"step": 7180
},
{
"epoch": 6.947203092908058,
"grad_norm": 7.070675373077393,
"learning_rate": 1.0392114746850867e-07,
"loss": 2.4984,
"step": 7190
},
{
"epoch": 6.956868430590793,
"grad_norm": 7.004439353942871,
"learning_rate": 1.0205775653631176e-07,
"loss": 2.4856,
"step": 7200
},
{
"epoch": 6.966533768273529,
"grad_norm": 6.824222564697266,
"learning_rate": 1.0021032446221023e-07,
"loss": 2.507,
"step": 7210
},
{
"epoch": 6.966533768273529,
"eval_loss": 2.6293044090270996,
"eval_runtime": 216.6368,
"eval_samples_per_second": 101.876,
"eval_steps_per_second": 4.247,
"step": 7210
},
{
"epoch": 6.9761991059562645,
"grad_norm": 6.812296390533447,
"learning_rate": 9.837888408006e-08,
"loss": 2.5143,
"step": 7220
},
{
"epoch": 6.985864443639,
"grad_norm": 6.687375068664551,
"learning_rate": 9.65634679395011e-08,
"loss": 2.4881,
"step": 7230
},
{
"epoch": 6.995529781321735,
"grad_norm": 6.915345668792725,
"learning_rate": 9.476410830538063e-08,
"loss": 2.4649,
"step": 7240
},
{
"epoch": 7.004832668841368,
"grad_norm": 7.076071739196777,
"learning_rate": 9.298083715717686e-08,
"loss": 2.4633,
"step": 7250
},
{
"epoch": 7.014498006524103,
"grad_norm": 6.80445671081543,
"learning_rate": 9.121368618843361e-08,
"loss": 2.4815,
"step": 7260
},
{
"epoch": 7.024163344206838,
"grad_norm": 6.910451412200928,
"learning_rate": 8.946268680619407e-08,
"loss": 2.5297,
"step": 7270
},
{
"epoch": 7.033828681889574,
"grad_norm": 6.937068462371826,
"learning_rate": 8.772787013044558e-08,
"loss": 2.4856,
"step": 7280
},
{
"epoch": 7.043494019572309,
"grad_norm": 6.77278995513916,
"learning_rate": 8.600926699356414e-08,
"loss": 2.4626,
"step": 7290
},
{
"epoch": 7.053159357255044,
"grad_norm": 6.759533882141113,
"learning_rate": 8.430690793976758e-08,
"loss": 2.4845,
"step": 7300
},
{
"epoch": 7.062824694937779,
"grad_norm": 6.756450653076172,
"learning_rate": 8.262082322457297e-08,
"loss": 2.4992,
"step": 7310
},
{
"epoch": 7.0657242962426,
"eval_loss": 2.6299219131469727,
"eval_runtime": 216.8528,
"eval_samples_per_second": 101.774,
"eval_steps_per_second": 4.243,
"step": 7313
},
{
"epoch": 7.0724900326205145,
"grad_norm": 7.033835411071777,
"learning_rate": 8.09510428142578e-08,
"loss": 2.4857,
"step": 7320
},
{
"epoch": 7.08215537030325,
"grad_norm": 6.858207702636719,
"learning_rate": 7.929759638532851e-08,
"loss": 2.4614,
"step": 7330
},
{
"epoch": 7.091820707985986,
"grad_norm": 6.827333927154541,
"learning_rate": 7.766051332399226e-08,
"loss": 2.4479,
"step": 7340
},
{
"epoch": 7.101486045668721,
"grad_norm": 7.060365676879883,
"learning_rate": 7.603982272563547e-08,
"loss": 2.5128,
"step": 7350
},
{
"epoch": 7.111151383351456,
"grad_norm": 6.903797626495361,
"learning_rate": 7.443555339430573e-08,
"loss": 2.4585,
"step": 7360
},
{
"epoch": 7.120816721034191,
"grad_norm": 6.946906566619873,
"learning_rate": 7.284773384220034e-08,
"loss": 2.4955,
"step": 7370
},
{
"epoch": 7.130482058716926,
"grad_norm": 7.100877285003662,
"learning_rate": 7.127639228916004e-08,
"loss": 2.4804,
"step": 7380
},
{
"epoch": 7.140147396399661,
"grad_norm": 6.856292724609375,
"learning_rate": 6.972155666216684e-08,
"loss": 2.4788,
"step": 7390
},
{
"epoch": 7.149812734082397,
"grad_norm": 7.044023036956787,
"learning_rate": 6.818325459484786e-08,
"loss": 2.5149,
"step": 7400
},
{
"epoch": 7.1594780717651325,
"grad_norm": 6.6180572509765625,
"learning_rate": 6.666151342698412e-08,
"loss": 2.4784,
"step": 7410
},
{
"epoch": 7.165277274374773,
"eval_loss": 2.6300928592681885,
"eval_runtime": 216.7445,
"eval_samples_per_second": 101.825,
"eval_steps_per_second": 4.245,
"step": 7416
},
{
"epoch": 7.169143409447868,
"grad_norm": 6.7274932861328125,
"learning_rate": 6.515636020402481e-08,
"loss": 2.4571,
"step": 7420
},
{
"epoch": 7.178808747130603,
"grad_norm": 6.896090507507324,
"learning_rate": 6.36678216766069e-08,
"loss": 2.4801,
"step": 7430
},
{
"epoch": 7.188474084813338,
"grad_norm": 6.6245622634887695,
"learning_rate": 6.219592430007869e-08,
"loss": 2.4743,
"step": 7440
},
{
"epoch": 7.198139422496073,
"grad_norm": 6.647139549255371,
"learning_rate": 6.074069423403105e-08,
"loss": 2.4691,
"step": 7450
},
{
"epoch": 7.207804760178809,
"grad_norm": 6.740390777587891,
"learning_rate": 5.9302157341830864e-08,
"loss": 2.4892,
"step": 7460
},
{
"epoch": 7.217470097861544,
"grad_norm": 6.634116172790527,
"learning_rate": 5.788033919016311e-08,
"loss": 2.4693,
"step": 7470
},
{
"epoch": 7.227135435544279,
"grad_norm": 6.560644149780273,
"learning_rate": 5.647526504857514e-08,
"loss": 2.4722,
"step": 7480
},
{
"epoch": 7.2368007732270145,
"grad_norm": 6.656121730804443,
"learning_rate": 5.5086959889027894e-08,
"loss": 2.4589,
"step": 7490
},
{
"epoch": 7.24646611090975,
"grad_norm": 6.90827751159668,
"learning_rate": 5.371544838545283e-08,
"loss": 2.4813,
"step": 7500
},
{
"epoch": 7.256131448592485,
"grad_norm": 6.8788323402404785,
"learning_rate": 5.236075491331205e-08,
"loss": 2.4721,
"step": 7510
},
{
"epoch": 7.264830252506947,
"eval_loss": 2.629812479019165,
"eval_runtime": 216.6204,
"eval_samples_per_second": 101.883,
"eval_steps_per_second": 4.247,
"step": 7519
},
{
"epoch": 7.265796786275221,
"grad_norm": 6.811087131500244,
"learning_rate": 5.10229035491665e-08,
"loss": 2.4821,
"step": 7520
},
{
"epoch": 7.275462123957956,
"grad_norm": 6.634406089782715,
"learning_rate": 4.970191807024693e-08,
"loss": 2.501,
"step": 7530
},
{
"epoch": 7.285127461640691,
"grad_norm": 6.741237640380859,
"learning_rate": 4.8397821954032194e-08,
"loss": 2.4787,
"step": 7540
},
{
"epoch": 7.294792799323426,
"grad_norm": 6.966506481170654,
"learning_rate": 4.7110638377831025e-08,
"loss": 2.4442,
"step": 7550
},
{
"epoch": 7.304458137006161,
"grad_norm": 6.796621799468994,
"learning_rate": 4.584039021837094e-08,
"loss": 2.4723,
"step": 7560
},
{
"epoch": 7.314123474688897,
"grad_norm": 6.76340913772583,
"learning_rate": 4.4587100051391547e-08,
"loss": 2.4815,
"step": 7570
},
{
"epoch": 7.323788812371633,
"grad_norm": 6.656253337860107,
"learning_rate": 4.3350790151242876e-08,
"loss": 2.5071,
"step": 7580
},
{
"epoch": 7.333454150054368,
"grad_norm": 6.851405620574951,
"learning_rate": 4.2131482490490035e-08,
"loss": 2.4624,
"step": 7590
},
{
"epoch": 7.343119487737103,
"grad_norm": 6.677999973297119,
"learning_rate": 4.092919873952205e-08,
"loss": 2.4766,
"step": 7600
},
{
"epoch": 7.352784825419838,
"grad_norm": 6.938327312469482,
"learning_rate": 3.9743960266167334e-08,
"loss": 2.4837,
"step": 7610
},
{
"epoch": 7.362450163102573,
"grad_norm": 6.797394275665283,
"learning_rate": 3.857578813531392e-08,
"loss": 2.4926,
"step": 7620
},
{
"epoch": 7.36438323063912,
"eval_loss": 2.6299386024475098,
"eval_runtime": 216.5913,
"eval_samples_per_second": 101.897,
"eval_steps_per_second": 4.248,
"step": 7622
},
{
"epoch": 7.372115500785308,
"grad_norm": 6.87264347076416,
"learning_rate": 3.742470310853441e-08,
"loss": 2.4897,
"step": 7630
},
{
"epoch": 7.381780838468044,
"grad_norm": 7.126553535461426,
"learning_rate": 3.6290725643717715e-08,
"loss": 2.4782,
"step": 7640
},
{
"epoch": 7.3914461761507795,
"grad_norm": 6.758020401000977,
"learning_rate": 3.5173875894704886e-08,
"loss": 2.4454,
"step": 7650
},
{
"epoch": 7.401111513833515,
"grad_norm": 7.062627792358398,
"learning_rate": 3.4074173710931796e-08,
"loss": 2.4876,
"step": 7660
},
{
"epoch": 7.41077685151625,
"grad_norm": 6.85789680480957,
"learning_rate": 3.299163863707522e-08,
"loss": 2.4866,
"step": 7670
},
{
"epoch": 7.420442189198985,
"grad_norm": 6.748703956604004,
"learning_rate": 3.1926289912706185e-08,
"loss": 2.5081,
"step": 7680
},
{
"epoch": 7.43010752688172,
"grad_norm": 6.617964744567871,
"learning_rate": 3.08781464719482e-08,
"loss": 2.4899,
"step": 7690
},
{
"epoch": 7.439772864564456,
"grad_norm": 6.949032306671143,
"learning_rate": 2.98472269431399e-08,
"loss": 2.4634,
"step": 7700
},
{
"epoch": 7.449438202247191,
"grad_norm": 6.690010070800781,
"learning_rate": 2.883354964850493e-08,
"loss": 2.489,
"step": 7710
},
{
"epoch": 7.459103539929926,
"grad_norm": 6.775691509246826,
"learning_rate": 2.7837132603825696e-08,
"loss": 2.4834,
"step": 7720
},
{
"epoch": 7.463936208771294,
"eval_loss": 2.629732370376587,
"eval_runtime": 216.725,
"eval_samples_per_second": 101.834,
"eval_steps_per_second": 4.245,
"step": 7725
},
{
"epoch": 7.4687688776126615,
"grad_norm": 7.173056602478027,
"learning_rate": 2.6857993518123455e-08,
"loss": 2.4805,
"step": 7730
},
{
"epoch": 7.478434215295397,
"grad_norm": 6.82067346572876,
"learning_rate": 2.5896149793343423e-08,
"loss": 2.4992,
"step": 7740
},
{
"epoch": 7.488099552978132,
"grad_norm": 6.860231876373291,
"learning_rate": 2.495161852404526e-08,
"loss": 2.4786,
"step": 7750
},
{
"epoch": 7.497764890660868,
"grad_norm": 6.8974432945251465,
"learning_rate": 2.4024416497100298e-08,
"loss": 2.5135,
"step": 7760
},
{
"epoch": 7.507430228343603,
"grad_norm": 6.890160083770752,
"learning_rate": 2.3114560191391575e-08,
"loss": 2.4968,
"step": 7770
},
{
"epoch": 7.517095566026338,
"grad_norm": 6.756450176239014,
"learning_rate": 2.22220657775225e-08,
"loss": 2.5055,
"step": 7780
},
{
"epoch": 7.526760903709073,
"grad_norm": 6.810698986053467,
"learning_rate": 2.1346949117528435e-08,
"loss": 2.4775,
"step": 7790
},
{
"epoch": 7.536426241391808,
"grad_norm": 6.732553482055664,
"learning_rate": 2.048922576459522e-08,
"loss": 2.4679,
"step": 7800
},
{
"epoch": 7.5460915790745435,
"grad_norm": 7.055318355560303,
"learning_rate": 1.964891096278276e-08,
"loss": 2.4829,
"step": 7810
},
{
"epoch": 7.55575691675728,
"grad_norm": 6.638191223144531,
"learning_rate": 1.882601964675379e-08,
"loss": 2.4568,
"step": 7820
},
{
"epoch": 7.5634891869034675,
"eval_loss": 2.6296987533569336,
"eval_runtime": 216.739,
"eval_samples_per_second": 101.828,
"eval_steps_per_second": 4.245,
"step": 7828
},
{
"epoch": 7.565422254440015,
"grad_norm": 6.956434726715088,
"learning_rate": 1.8020566441508843e-08,
"loss": 2.4692,
"step": 7830
},
{
"epoch": 7.57508759212275,
"grad_norm": 6.698721885681152,
"learning_rate": 1.7232565662126164e-08,
"loss": 2.4487,
"step": 7840
},
{
"epoch": 7.584752929805485,
"grad_norm": 6.7048726081848145,
"learning_rate": 1.6462031313507096e-08,
"loss": 2.4678,
"step": 7850
},
{
"epoch": 7.59441826748822,
"grad_norm": 6.834412097930908,
"learning_rate": 1.5708977090127417e-08,
"loss": 2.4789,
"step": 7860
},
{
"epoch": 7.604083605170955,
"grad_norm": 6.895366191864014,
"learning_rate": 1.4973416375793967e-08,
"loss": 2.4942,
"step": 7870
},
{
"epoch": 7.613748942853691,
"grad_norm": 7.1303324699401855,
"learning_rate": 1.4255362243406621e-08,
"loss": 2.5303,
"step": 7880
},
{
"epoch": 7.6234142805364264,
"grad_norm": 6.808503150939941,
"learning_rate": 1.3554827454726136e-08,
"loss": 2.4786,
"step": 7890
},
{
"epoch": 7.633079618219162,
"grad_norm": 6.925174713134766,
"learning_rate": 1.2871824460147007e-08,
"loss": 2.4547,
"step": 7900
},
{
"epoch": 7.642744955901897,
"grad_norm": 6.883964538574219,
"learning_rate": 1.2206365398476637e-08,
"loss": 2.506,
"step": 7910
},
{
"epoch": 7.652410293584632,
"grad_norm": 6.8656086921691895,
"learning_rate": 1.155846209671918e-08,
"loss": 2.5117,
"step": 7920
},
{
"epoch": 7.662075631267367,
"grad_norm": 6.914687633514404,
"learning_rate": 1.0928126069865818e-08,
"loss": 2.4916,
"step": 7930
},
{
"epoch": 7.663042165035641,
"eval_loss": 2.629610061645508,
"eval_runtime": 216.9358,
"eval_samples_per_second": 101.735,
"eval_steps_per_second": 4.241,
"step": 7931
},
{
"epoch": 7.671740968950103,
"grad_norm": 6.8651323318481445,
"learning_rate": 1.0315368520689372e-08,
"loss": 2.4809,
"step": 7940
},
{
"epoch": 7.681406306632838,
"grad_norm": 6.866594314575195,
"learning_rate": 9.720200339546236e-09,
"loss": 2.4871,
"step": 7950
},
{
"epoch": 7.691071644315573,
"grad_norm": 6.732547283172607,
"learning_rate": 9.142632104181648e-09,
"loss": 2.4786,
"step": 7960
},
{
"epoch": 7.7007369819983085,
"grad_norm": 6.74143123626709,
"learning_rate": 8.58267407954283e-09,
"loss": 2.4774,
"step": 7970
},
{
"epoch": 7.710402319681044,
"grad_norm": 6.904804706573486,
"learning_rate": 8.040336217595588e-09,
"loss": 2.4869,
"step": 7980
},
{
"epoch": 7.72006765736378,
"grad_norm": 6.923266410827637,
"learning_rate": 7.515628157148012e-09,
"loss": 2.4933,
"step": 7990
},
{
"epoch": 7.729732995046515,
"grad_norm": 7.009883880615234,
"learning_rate": 7.008559223679156e-09,
"loss": 2.4392,
"step": 8000
},
{
"epoch": 7.73939833272925,
"grad_norm": 6.632232189178467,
"learning_rate": 6.519138429172954e-09,
"loss": 2.4594,
"step": 8010
},
{
"epoch": 7.749063670411985,
"grad_norm": 6.914027214050293,
"learning_rate": 6.047374471958466e-09,
"loss": 2.4873,
"step": 8020
},
{
"epoch": 7.75872900809472,
"grad_norm": 7.000319957733154,
"learning_rate": 5.59327573655477e-09,
"loss": 2.5232,
"step": 8030
},
{
"epoch": 7.762595143167815,
"eval_loss": 2.6296308040618896,
"eval_runtime": 216.9179,
"eval_samples_per_second": 101.744,
"eval_steps_per_second": 4.241,
"step": 8034
},
{
"epoch": 7.768394345777455,
"grad_norm": 6.82133674621582,
"learning_rate": 5.156850293522752e-09,
"loss": 2.4846,
"step": 8040
},
{
"epoch": 7.7780596834601905,
"grad_norm": 6.938842296600342,
"learning_rate": 4.7381058993205545e-09,
"loss": 2.474,
"step": 8050
},
{
"epoch": 7.7877250211429265,
"grad_norm": 6.969207286834717,
"learning_rate": 4.3370499961667975e-09,
"loss": 2.4701,
"step": 8060
},
{
"epoch": 7.797390358825662,
"grad_norm": 6.929446220397949,
"learning_rate": 3.953689711907792e-09,
"loss": 2.4944,
"step": 8070
},
{
"epoch": 7.807055696508397,
"grad_norm": 6.946097373962402,
"learning_rate": 3.588031859890761e-09,
"loss": 2.4888,
"step": 8080
},
{
"epoch": 7.816721034191132,
"grad_norm": 6.968674182891846,
"learning_rate": 3.240082938843147e-09,
"loss": 2.4995,
"step": 8090
},
{
"epoch": 7.826386371873867,
"grad_norm": 6.90951681137085,
"learning_rate": 2.9098491327564924e-09,
"loss": 2.4757,
"step": 8100
},
{
"epoch": 7.836051709556603,
"grad_norm": 6.8854756355285645,
"learning_rate": 2.5973363107774095e-09,
"loss": 2.4781,
"step": 8110
},
{
"epoch": 7.845717047239338,
"grad_norm": 7.083747386932373,
"learning_rate": 2.3025500271023346e-09,
"loss": 2.4496,
"step": 8120
},
{
"epoch": 7.855382384922073,
"grad_norm": 6.7878851890563965,
"learning_rate": 2.0254955208794944e-09,
"loss": 2.4695,
"step": 8130
},
{
"epoch": 7.862148121299988,
"eval_loss": 2.629627227783203,
"eval_runtime": 216.9855,
"eval_samples_per_second": 101.712,
"eval_steps_per_second": 4.24,
"step": 8137
},
{
"epoch": 7.8650477226048086,
"grad_norm": 6.706069469451904,
"learning_rate": 1.7661777161156467e-09,
"loss": 2.4856,
"step": 8140
},
{
"epoch": 7.874713060287544,
"grad_norm": 6.6752705574035645,
"learning_rate": 1.524601221588151e-09,
"loss": 2.4755,
"step": 8150
},
{
"epoch": 7.884378397970279,
"grad_norm": 6.708074569702148,
"learning_rate": 1.3007703307635897e-09,
"loss": 2.4643,
"step": 8160
},
{
"epoch": 7.894043735653014,
"grad_norm": 7.023872375488281,
"learning_rate": 1.0946890217213844e-09,
"loss": 2.499,
"step": 8170
},
{
"epoch": 7.90370907333575,
"grad_norm": 6.72566032409668,
"learning_rate": 9.063609570826302e-10,
"loss": 2.4384,
"step": 8180
},
{
"epoch": 7.913374411018485,
"grad_norm": 6.935588836669922,
"learning_rate": 7.357894839457035e-10,
"loss": 2.4758,
"step": 8190
},
{
"epoch": 7.92303974870122,
"grad_norm": 6.855502128601074,
"learning_rate": 5.829776338259762e-10,
"loss": 2.4676,
"step": 8200
},
{
"epoch": 7.932705086383955,
"grad_norm": 6.738182544708252,
"learning_rate": 4.479281226028586e-10,
"loss": 2.4874,
"step": 8210
},
{
"epoch": 7.942370424066691,
"grad_norm": 6.814088821411133,
"learning_rate": 3.3064335047061633e-10,
"loss": 2.5022,
"step": 8220
},
{
"epoch": 7.952035761749427,
"grad_norm": 6.909735202789307,
"learning_rate": 2.3112540189640372e-10,
"loss": 2.4729,
"step": 8230
},
{
"epoch": 7.961701099432162,
"grad_norm": 6.548181533813477,
"learning_rate": 1.4937604558284967e-10,
"loss": 2.4591,
"step": 8240
},
{
"epoch": 7.961701099432162,
"eval_loss": 2.629610061645508,
"eval_runtime": 217.2357,
"eval_samples_per_second": 101.595,
"eval_steps_per_second": 4.235,
"step": 8240
},
{
"epoch": 7.971366437114897,
"grad_norm": 6.895495891571045,
"learning_rate": 8.539673443686002e-11,
"loss": 2.4801,
"step": 8250
},
{
"epoch": 7.981031774797632,
"grad_norm": 6.714371204376221,
"learning_rate": 3.91886055433055e-11,
"loss": 2.4589,
"step": 8260
},
{
"epoch": 7.990697112480367,
"grad_norm": 6.664721965789795,
"learning_rate": 1.0752480145481691e-11,
"loss": 2.511,
"step": 8270
},
{
"epoch": 8.0,
"grad_norm": 9.770947456359863,
"learning_rate": 8.886363023208332e-14,
"loss": 2.4974,
"step": 8280
}
],
"logging_steps": 10,
"max_steps": 8280,
"num_input_tokens_seen": 0,
"num_train_epochs": 8,
"save_steps": 103,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.118170148932321e+18,
"train_batch_size": 24,
"trial_name": null,
"trial_params": null
}