LogicCoder-8B / trainer_state.json
azzzacs's picture
Upload folder using huggingface_hub
2e3409d verified
raw
history blame
75.3 kB
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 8.0,
"eval_steps": 500,
"global_step": 4576,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.017497812773403325,
"grad_norm": 15.0,
"learning_rate": 1.0465116279069768e-06,
"loss": 1.1604,
"step": 10
},
{
"epoch": 0.03499562554680665,
"grad_norm": 9.6875,
"learning_rate": 2.2093023255813954e-06,
"loss": 1.1391,
"step": 20
},
{
"epoch": 0.05249343832020997,
"grad_norm": 4.09375,
"learning_rate": 3.372093023255814e-06,
"loss": 1.0272,
"step": 30
},
{
"epoch": 0.0699912510936133,
"grad_norm": 2.3125,
"learning_rate": 4.5348837209302326e-06,
"loss": 0.8968,
"step": 40
},
{
"epoch": 0.08748906386701662,
"grad_norm": 1.515625,
"learning_rate": 5.697674418604652e-06,
"loss": 0.8108,
"step": 50
},
{
"epoch": 0.10498687664041995,
"grad_norm": 1.421875,
"learning_rate": 6.86046511627907e-06,
"loss": 0.8071,
"step": 60
},
{
"epoch": 0.12248468941382328,
"grad_norm": 1.1953125,
"learning_rate": 8.023255813953488e-06,
"loss": 0.759,
"step": 70
},
{
"epoch": 0.1399825021872266,
"grad_norm": 1.25,
"learning_rate": 9.186046511627908e-06,
"loss": 0.767,
"step": 80
},
{
"epoch": 0.15748031496062992,
"grad_norm": 1.3046875,
"learning_rate": 1.0348837209302327e-05,
"loss": 0.7569,
"step": 90
},
{
"epoch": 0.17497812773403323,
"grad_norm": 1.1171875,
"learning_rate": 1.1511627906976746e-05,
"loss": 0.7603,
"step": 100
},
{
"epoch": 0.19247594050743658,
"grad_norm": 1.046875,
"learning_rate": 1.2674418604651164e-05,
"loss": 0.7424,
"step": 110
},
{
"epoch": 0.2099737532808399,
"grad_norm": 1.3125,
"learning_rate": 1.3837209302325583e-05,
"loss": 0.7358,
"step": 120
},
{
"epoch": 0.2274715660542432,
"grad_norm": 1.1015625,
"learning_rate": 1.5000000000000002e-05,
"loss": 0.7119,
"step": 130
},
{
"epoch": 0.24496937882764655,
"grad_norm": 1.140625,
"learning_rate": 1.616279069767442e-05,
"loss": 0.7219,
"step": 140
},
{
"epoch": 0.26246719160104987,
"grad_norm": 1.0859375,
"learning_rate": 1.7325581395348837e-05,
"loss": 0.7073,
"step": 150
},
{
"epoch": 0.2799650043744532,
"grad_norm": 1.078125,
"learning_rate": 1.8488372093023256e-05,
"loss": 0.7135,
"step": 160
},
{
"epoch": 0.2974628171478565,
"grad_norm": 1.046875,
"learning_rate": 1.9651162790697676e-05,
"loss": 0.72,
"step": 170
},
{
"epoch": 0.31496062992125984,
"grad_norm": 0.96484375,
"learning_rate": 1.9999929041918377e-05,
"loss": 0.7111,
"step": 180
},
{
"epoch": 0.3324584426946632,
"grad_norm": 1.0859375,
"learning_rate": 1.999958149482438e-05,
"loss": 0.7308,
"step": 190
},
{
"epoch": 0.34995625546806647,
"grad_norm": 1.09375,
"learning_rate": 1.9998944336771236e-05,
"loss": 0.711,
"step": 200
},
{
"epoch": 0.3674540682414698,
"grad_norm": 1.03125,
"learning_rate": 1.9998017588263007e-05,
"loss": 0.7096,
"step": 210
},
{
"epoch": 0.38495188101487315,
"grad_norm": 1.15625,
"learning_rate": 1.9996801279122917e-05,
"loss": 0.7193,
"step": 220
},
{
"epoch": 0.40244969378827644,
"grad_norm": 0.97265625,
"learning_rate": 1.9995295448492383e-05,
"loss": 0.7187,
"step": 230
},
{
"epoch": 0.4199475065616798,
"grad_norm": 1.0546875,
"learning_rate": 1.9993500144829784e-05,
"loss": 0.7197,
"step": 240
},
{
"epoch": 0.4374453193350831,
"grad_norm": 1.046875,
"learning_rate": 1.9991415425908868e-05,
"loss": 0.7117,
"step": 250
},
{
"epoch": 0.4549431321084864,
"grad_norm": 1.1328125,
"learning_rate": 1.9989041358816926e-05,
"loss": 0.6958,
"step": 260
},
{
"epoch": 0.47244094488188976,
"grad_norm": 1.0625,
"learning_rate": 1.9986378019952595e-05,
"loss": 0.7127,
"step": 270
},
{
"epoch": 0.4899387576552931,
"grad_norm": 1.1640625,
"learning_rate": 1.998342549502343e-05,
"loss": 0.7068,
"step": 280
},
{
"epoch": 0.5074365704286964,
"grad_norm": 1.0546875,
"learning_rate": 1.998018387904314e-05,
"loss": 0.6994,
"step": 290
},
{
"epoch": 0.5249343832020997,
"grad_norm": 1.078125,
"learning_rate": 1.997665327632852e-05,
"loss": 0.713,
"step": 300
},
{
"epoch": 0.5424321959755031,
"grad_norm": 1.2109375,
"learning_rate": 1.9972833800496107e-05,
"loss": 0.7119,
"step": 310
},
{
"epoch": 0.5599300087489064,
"grad_norm": 1.0703125,
"learning_rate": 1.99687255744585e-05,
"loss": 0.6993,
"step": 320
},
{
"epoch": 0.5774278215223098,
"grad_norm": 0.97265625,
"learning_rate": 1.9964328730420445e-05,
"loss": 0.706,
"step": 330
},
{
"epoch": 0.594925634295713,
"grad_norm": 1.1484375,
"learning_rate": 1.995964340987454e-05,
"loss": 0.7167,
"step": 340
},
{
"epoch": 0.6124234470691163,
"grad_norm": 1.109375,
"learning_rate": 1.9954669763596708e-05,
"loss": 0.7206,
"step": 350
},
{
"epoch": 0.6299212598425197,
"grad_norm": 0.97265625,
"learning_rate": 1.9949407951641325e-05,
"loss": 0.6875,
"step": 360
},
{
"epoch": 0.647419072615923,
"grad_norm": 1.0625,
"learning_rate": 1.99438581433361e-05,
"loss": 0.7256,
"step": 370
},
{
"epoch": 0.6649168853893264,
"grad_norm": 1.1015625,
"learning_rate": 1.9938020517276583e-05,
"loss": 0.689,
"step": 380
},
{
"epoch": 0.6824146981627297,
"grad_norm": 0.98828125,
"learning_rate": 1.9931895261320463e-05,
"loss": 0.7029,
"step": 390
},
{
"epoch": 0.6999125109361329,
"grad_norm": 1.0078125,
"learning_rate": 1.9925482572581477e-05,
"loss": 0.7189,
"step": 400
},
{
"epoch": 0.7174103237095363,
"grad_norm": 1.015625,
"learning_rate": 1.991878265742311e-05,
"loss": 0.7103,
"step": 410
},
{
"epoch": 0.7349081364829396,
"grad_norm": 0.98046875,
"learning_rate": 1.9911795731451928e-05,
"loss": 0.7045,
"step": 420
},
{
"epoch": 0.752405949256343,
"grad_norm": 1.1171875,
"learning_rate": 1.9904522019510647e-05,
"loss": 0.7112,
"step": 430
},
{
"epoch": 0.7699037620297463,
"grad_norm": 1.03125,
"learning_rate": 1.989696175567089e-05,
"loss": 0.711,
"step": 440
},
{
"epoch": 0.7874015748031497,
"grad_norm": 1.0546875,
"learning_rate": 1.988911518322566e-05,
"loss": 0.7174,
"step": 450
},
{
"epoch": 0.8048993875765529,
"grad_norm": 1.078125,
"learning_rate": 1.9880982554681534e-05,
"loss": 0.7023,
"step": 460
},
{
"epoch": 0.8223972003499562,
"grad_norm": 1.09375,
"learning_rate": 1.9872564131750488e-05,
"loss": 0.6956,
"step": 470
},
{
"epoch": 0.8398950131233596,
"grad_norm": 1.0859375,
"learning_rate": 1.9863860185341514e-05,
"loss": 0.7003,
"step": 480
},
{
"epoch": 0.8573928258967629,
"grad_norm": 1.03125,
"learning_rate": 1.985487099555189e-05,
"loss": 0.6849,
"step": 490
},
{
"epoch": 0.8748906386701663,
"grad_norm": 0.9140625,
"learning_rate": 1.9845596851658168e-05,
"loss": 0.7017,
"step": 500
},
{
"epoch": 0.8923884514435696,
"grad_norm": 0.9609375,
"learning_rate": 1.9836038052106853e-05,
"loss": 0.6809,
"step": 510
},
{
"epoch": 0.9098862642169728,
"grad_norm": 0.96484375,
"learning_rate": 1.9826194904504824e-05,
"loss": 0.6961,
"step": 520
},
{
"epoch": 0.9273840769903762,
"grad_norm": 0.97265625,
"learning_rate": 1.9816067725609403e-05,
"loss": 0.698,
"step": 530
},
{
"epoch": 0.9448818897637795,
"grad_norm": 1.125,
"learning_rate": 1.9805656841318198e-05,
"loss": 0.7022,
"step": 540
},
{
"epoch": 0.9623797025371829,
"grad_norm": 0.97265625,
"learning_rate": 1.9794962586658582e-05,
"loss": 0.6896,
"step": 550
},
{
"epoch": 0.9798775153105862,
"grad_norm": 1.0390625,
"learning_rate": 1.978398530577693e-05,
"loss": 0.6915,
"step": 560
},
{
"epoch": 0.9973753280839895,
"grad_norm": 0.99609375,
"learning_rate": 1.977272535192754e-05,
"loss": 0.6996,
"step": 570
},
{
"epoch": 1.0139982502187226,
"grad_norm": 1.0390625,
"learning_rate": 1.9761183087461264e-05,
"loss": 0.6237,
"step": 580
},
{
"epoch": 1.031496062992126,
"grad_norm": 1.1328125,
"learning_rate": 1.974935888381385e-05,
"loss": 0.5954,
"step": 590
},
{
"epoch": 1.0489938757655293,
"grad_norm": 1.1484375,
"learning_rate": 1.9737253121493994e-05,
"loss": 0.6124,
"step": 600
},
{
"epoch": 1.0664916885389326,
"grad_norm": 1.1640625,
"learning_rate": 1.9724866190071075e-05,
"loss": 0.6023,
"step": 610
},
{
"epoch": 1.083989501312336,
"grad_norm": 1.1875,
"learning_rate": 1.971219848816264e-05,
"loss": 0.6128,
"step": 620
},
{
"epoch": 1.1014873140857393,
"grad_norm": 1.2109375,
"learning_rate": 1.9699250423421568e-05,
"loss": 0.6167,
"step": 630
},
{
"epoch": 1.1189851268591426,
"grad_norm": 1.203125,
"learning_rate": 1.9686022412522953e-05,
"loss": 0.6322,
"step": 640
},
{
"epoch": 1.136482939632546,
"grad_norm": 0.94921875,
"learning_rate": 1.967251488115069e-05,
"loss": 0.6,
"step": 650
},
{
"epoch": 1.1539807524059493,
"grad_norm": 1.0546875,
"learning_rate": 1.9658728263983783e-05,
"loss": 0.6204,
"step": 660
},
{
"epoch": 1.1714785651793527,
"grad_norm": 1.140625,
"learning_rate": 1.964466300468235e-05,
"loss": 0.627,
"step": 670
},
{
"epoch": 1.188976377952756,
"grad_norm": 1.0,
"learning_rate": 1.9630319555873364e-05,
"loss": 0.5988,
"step": 680
},
{
"epoch": 1.2064741907261591,
"grad_norm": 1.1015625,
"learning_rate": 1.961569837913605e-05,
"loss": 0.6141,
"step": 690
},
{
"epoch": 1.2239720034995625,
"grad_norm": 1.03125,
"learning_rate": 1.9600799944987077e-05,
"loss": 0.6171,
"step": 700
},
{
"epoch": 1.2414698162729658,
"grad_norm": 1.1328125,
"learning_rate": 1.958562473286538e-05,
"loss": 0.613,
"step": 710
},
{
"epoch": 1.2589676290463692,
"grad_norm": 1.109375,
"learning_rate": 1.957017323111675e-05,
"loss": 0.6169,
"step": 720
},
{
"epoch": 1.2764654418197725,
"grad_norm": 1.1015625,
"learning_rate": 1.955444593697811e-05,
"loss": 0.6312,
"step": 730
},
{
"epoch": 1.2939632545931758,
"grad_norm": 1.2109375,
"learning_rate": 1.9538443356561528e-05,
"loss": 0.6144,
"step": 740
},
{
"epoch": 1.3114610673665792,
"grad_norm": 1.140625,
"learning_rate": 1.9522166004837908e-05,
"loss": 0.6096,
"step": 750
},
{
"epoch": 1.3289588801399825,
"grad_norm": 1.1953125,
"learning_rate": 1.9505614405620436e-05,
"loss": 0.6225,
"step": 760
},
{
"epoch": 1.3464566929133859,
"grad_norm": 1.1953125,
"learning_rate": 1.9488789091547716e-05,
"loss": 0.6156,
"step": 770
},
{
"epoch": 1.3639545056867892,
"grad_norm": 1.078125,
"learning_rate": 1.9471690604066633e-05,
"loss": 0.644,
"step": 780
},
{
"epoch": 1.3814523184601923,
"grad_norm": 1.140625,
"learning_rate": 1.945431949341492e-05,
"loss": 0.6221,
"step": 790
},
{
"epoch": 1.3989501312335957,
"grad_norm": 1.046875,
"learning_rate": 1.9436676318603465e-05,
"loss": 0.6164,
"step": 800
},
{
"epoch": 1.416447944006999,
"grad_norm": 1.0546875,
"learning_rate": 1.941876164739831e-05,
"loss": 0.6187,
"step": 810
},
{
"epoch": 1.4339457567804024,
"grad_norm": 1.1953125,
"learning_rate": 1.940057605630239e-05,
"loss": 0.6041,
"step": 820
},
{
"epoch": 1.4514435695538057,
"grad_norm": 1.0078125,
"learning_rate": 1.938212013053697e-05,
"loss": 0.6172,
"step": 830
},
{
"epoch": 1.468941382327209,
"grad_norm": 1.125,
"learning_rate": 1.9363394464022814e-05,
"loss": 0.6107,
"step": 840
},
{
"epoch": 1.4864391951006124,
"grad_norm": 1.1484375,
"learning_rate": 1.934439965936109e-05,
"loss": 0.6109,
"step": 850
},
{
"epoch": 1.5039370078740157,
"grad_norm": 1.15625,
"learning_rate": 1.9325136327813945e-05,
"loss": 0.606,
"step": 860
},
{
"epoch": 1.521434820647419,
"grad_norm": 1.1171875,
"learning_rate": 1.9305605089284873e-05,
"loss": 0.6264,
"step": 870
},
{
"epoch": 1.5389326334208224,
"grad_norm": 1.140625,
"learning_rate": 1.9285806572298726e-05,
"loss": 0.6185,
"step": 880
},
{
"epoch": 1.5564304461942258,
"grad_norm": 1.1953125,
"learning_rate": 1.926574141398153e-05,
"loss": 0.6243,
"step": 890
},
{
"epoch": 1.5739282589676291,
"grad_norm": 1.0703125,
"learning_rate": 1.924541026003994e-05,
"loss": 0.6035,
"step": 900
},
{
"epoch": 1.5914260717410325,
"grad_norm": 1.203125,
"learning_rate": 1.9224813764740496e-05,
"loss": 0.6297,
"step": 910
},
{
"epoch": 1.6089238845144358,
"grad_norm": 1.078125,
"learning_rate": 1.9203952590888547e-05,
"loss": 0.6138,
"step": 920
},
{
"epoch": 1.6264216972878391,
"grad_norm": 1.1640625,
"learning_rate": 1.918282740980693e-05,
"loss": 0.6211,
"step": 930
},
{
"epoch": 1.6439195100612425,
"grad_norm": 1.140625,
"learning_rate": 1.9161438901314352e-05,
"loss": 0.6082,
"step": 940
},
{
"epoch": 1.6614173228346458,
"grad_norm": 1.09375,
"learning_rate": 1.9139787753703554e-05,
"loss": 0.6023,
"step": 950
},
{
"epoch": 1.678915135608049,
"grad_norm": 1.0859375,
"learning_rate": 1.91178746637191e-05,
"loss": 0.6152,
"step": 960
},
{
"epoch": 1.6964129483814523,
"grad_norm": 1.296875,
"learning_rate": 1.9095700336535017e-05,
"loss": 0.6076,
"step": 970
},
{
"epoch": 1.7139107611548556,
"grad_norm": 1.078125,
"learning_rate": 1.907326548573205e-05,
"loss": 0.6257,
"step": 980
},
{
"epoch": 1.731408573928259,
"grad_norm": 1.21875,
"learning_rate": 1.905057083327474e-05,
"loss": 0.616,
"step": 990
},
{
"epoch": 1.7489063867016623,
"grad_norm": 1.3203125,
"learning_rate": 1.9027617109488164e-05,
"loss": 0.6239,
"step": 1000
},
{
"epoch": 1.7664041994750657,
"grad_norm": 1.125,
"learning_rate": 1.9004405053034445e-05,
"loss": 0.5979,
"step": 1010
},
{
"epoch": 1.7839020122484688,
"grad_norm": 1.03125,
"learning_rate": 1.8980935410888973e-05,
"loss": 0.6177,
"step": 1020
},
{
"epoch": 1.8013998250218721,
"grad_norm": 1.0859375,
"learning_rate": 1.895720893831638e-05,
"loss": 0.5923,
"step": 1030
},
{
"epoch": 1.8188976377952755,
"grad_norm": 1.078125,
"learning_rate": 1.893322639884622e-05,
"loss": 0.6029,
"step": 1040
},
{
"epoch": 1.8363954505686788,
"grad_norm": 1.1328125,
"learning_rate": 1.8908988564248415e-05,
"loss": 0.6273,
"step": 1050
},
{
"epoch": 1.8538932633420822,
"grad_norm": 1.1953125,
"learning_rate": 1.8884496214508398e-05,
"loss": 0.6358,
"step": 1060
},
{
"epoch": 1.8713910761154855,
"grad_norm": 1.09375,
"learning_rate": 1.8859750137802037e-05,
"loss": 0.6321,
"step": 1070
},
{
"epoch": 1.8888888888888888,
"grad_norm": 1.1953125,
"learning_rate": 1.8834751130470247e-05,
"loss": 0.6118,
"step": 1080
},
{
"epoch": 1.9063867016622922,
"grad_norm": 1.140625,
"learning_rate": 1.8809499996993385e-05,
"loss": 0.6095,
"step": 1090
},
{
"epoch": 1.9238845144356955,
"grad_norm": 1.09375,
"learning_rate": 1.8783997549965344e-05,
"loss": 0.622,
"step": 1100
},
{
"epoch": 1.9413823272090989,
"grad_norm": 1.2421875,
"learning_rate": 1.875824461006741e-05,
"loss": 0.636,
"step": 1110
},
{
"epoch": 1.9588801399825022,
"grad_norm": 1.125,
"learning_rate": 1.8732242006041864e-05,
"loss": 0.6089,
"step": 1120
},
{
"epoch": 1.9763779527559056,
"grad_norm": 1.1328125,
"learning_rate": 1.8705990574665288e-05,
"loss": 0.6169,
"step": 1130
},
{
"epoch": 1.993875765529309,
"grad_norm": 1.1640625,
"learning_rate": 1.8679491160721654e-05,
"loss": 0.6174,
"step": 1140
},
{
"epoch": 2.010498687664042,
"grad_norm": 1.640625,
"learning_rate": 1.8652744616975142e-05,
"loss": 0.566,
"step": 1150
},
{
"epoch": 2.027996500437445,
"grad_norm": 1.6328125,
"learning_rate": 1.8625751804142685e-05,
"loss": 0.5181,
"step": 1160
},
{
"epoch": 2.0454943132108485,
"grad_norm": 1.2890625,
"learning_rate": 1.8598513590866278e-05,
"loss": 0.5257,
"step": 1170
},
{
"epoch": 2.062992125984252,
"grad_norm": 1.2890625,
"learning_rate": 1.8571030853685016e-05,
"loss": 0.5193,
"step": 1180
},
{
"epoch": 2.080489938757655,
"grad_norm": 1.3125,
"learning_rate": 1.8543304477006906e-05,
"loss": 0.5252,
"step": 1190
},
{
"epoch": 2.0979877515310585,
"grad_norm": 1.390625,
"learning_rate": 1.851533535308038e-05,
"loss": 0.5194,
"step": 1200
},
{
"epoch": 2.115485564304462,
"grad_norm": 1.1875,
"learning_rate": 1.848712438196561e-05,
"loss": 0.5126,
"step": 1210
},
{
"epoch": 2.1329833770778652,
"grad_norm": 1.2890625,
"learning_rate": 1.8458672471505522e-05,
"loss": 0.5007,
"step": 1220
},
{
"epoch": 2.1504811898512686,
"grad_norm": 1.2421875,
"learning_rate": 1.8429980537296587e-05,
"loss": 0.5205,
"step": 1230
},
{
"epoch": 2.167979002624672,
"grad_norm": 1.25,
"learning_rate": 1.840104950265936e-05,
"loss": 0.5018,
"step": 1240
},
{
"epoch": 2.1854768153980753,
"grad_norm": 1.359375,
"learning_rate": 1.8371880298608768e-05,
"loss": 0.5216,
"step": 1250
},
{
"epoch": 2.2029746281714786,
"grad_norm": 1.4921875,
"learning_rate": 1.834247386382414e-05,
"loss": 0.5109,
"step": 1260
},
{
"epoch": 2.220472440944882,
"grad_norm": 1.359375,
"learning_rate": 1.8312831144619024e-05,
"loss": 0.512,
"step": 1270
},
{
"epoch": 2.2379702537182853,
"grad_norm": 1.390625,
"learning_rate": 1.828295309491069e-05,
"loss": 0.5139,
"step": 1280
},
{
"epoch": 2.2554680664916886,
"grad_norm": 1.3203125,
"learning_rate": 1.8252840676189484e-05,
"loss": 0.5277,
"step": 1290
},
{
"epoch": 2.272965879265092,
"grad_norm": 1.4375,
"learning_rate": 1.8222494857487845e-05,
"loss": 0.5175,
"step": 1300
},
{
"epoch": 2.2904636920384953,
"grad_norm": 1.4609375,
"learning_rate": 1.819191661534915e-05,
"loss": 0.5247,
"step": 1310
},
{
"epoch": 2.3079615048118987,
"grad_norm": 1.3046875,
"learning_rate": 1.8161106933796268e-05,
"loss": 0.5173,
"step": 1320
},
{
"epoch": 2.325459317585302,
"grad_norm": 2.0,
"learning_rate": 1.813006680429991e-05,
"loss": 0.5262,
"step": 1330
},
{
"epoch": 2.3429571303587053,
"grad_norm": 1.328125,
"learning_rate": 1.8098797225746703e-05,
"loss": 0.5167,
"step": 1340
},
{
"epoch": 2.3604549431321082,
"grad_norm": 1.2890625,
"learning_rate": 1.8067299204407075e-05,
"loss": 0.5254,
"step": 1350
},
{
"epoch": 2.377952755905512,
"grad_norm": 1.40625,
"learning_rate": 1.803557375390284e-05,
"loss": 0.5234,
"step": 1360
},
{
"epoch": 2.395450568678915,
"grad_norm": 1.4375,
"learning_rate": 1.8003621895174603e-05,
"loss": 0.5113,
"step": 1370
},
{
"epoch": 2.4129483814523183,
"grad_norm": 1.296875,
"learning_rate": 1.797144465644889e-05,
"loss": 0.5177,
"step": 1380
},
{
"epoch": 2.4304461942257216,
"grad_norm": 1.34375,
"learning_rate": 1.7939043073205076e-05,
"loss": 0.5289,
"step": 1390
},
{
"epoch": 2.447944006999125,
"grad_norm": 1.3671875,
"learning_rate": 1.7906418188142045e-05,
"loss": 0.524,
"step": 1400
},
{
"epoch": 2.4654418197725283,
"grad_norm": 1.34375,
"learning_rate": 1.7873571051144644e-05,
"loss": 0.5193,
"step": 1410
},
{
"epoch": 2.4829396325459316,
"grad_norm": 1.421875,
"learning_rate": 1.7840502719249898e-05,
"loss": 0.5117,
"step": 1420
},
{
"epoch": 2.500437445319335,
"grad_norm": 1.375,
"learning_rate": 1.7807214256612995e-05,
"loss": 0.5436,
"step": 1430
},
{
"epoch": 2.5179352580927383,
"grad_norm": 1.3203125,
"learning_rate": 1.777370673447303e-05,
"loss": 0.5229,
"step": 1440
},
{
"epoch": 2.5354330708661417,
"grad_norm": 1.375,
"learning_rate": 1.7739981231118558e-05,
"loss": 0.5243,
"step": 1450
},
{
"epoch": 2.552930883639545,
"grad_norm": 1.3046875,
"learning_rate": 1.770603883185286e-05,
"loss": 0.5244,
"step": 1460
},
{
"epoch": 2.5704286964129484,
"grad_norm": 1.375,
"learning_rate": 1.7671880628959034e-05,
"loss": 0.5284,
"step": 1470
},
{
"epoch": 2.5879265091863517,
"grad_norm": 1.234375,
"learning_rate": 1.763750772166486e-05,
"loss": 0.5243,
"step": 1480
},
{
"epoch": 2.605424321959755,
"grad_norm": 1.4375,
"learning_rate": 1.7602921216107397e-05,
"loss": 0.5299,
"step": 1490
},
{
"epoch": 2.6229221347331584,
"grad_norm": 1.3046875,
"learning_rate": 1.7568122225297407e-05,
"loss": 0.5129,
"step": 1500
},
{
"epoch": 2.6404199475065617,
"grad_norm": 1.40625,
"learning_rate": 1.753311186908353e-05,
"loss": 0.5222,
"step": 1510
},
{
"epoch": 2.657917760279965,
"grad_norm": 1.3515625,
"learning_rate": 1.7497891274116246e-05,
"loss": 0.5314,
"step": 1520
},
{
"epoch": 2.6754155730533684,
"grad_norm": 1.3203125,
"learning_rate": 1.7462461573811632e-05,
"loss": 0.5145,
"step": 1530
},
{
"epoch": 2.6929133858267718,
"grad_norm": 1.3203125,
"learning_rate": 1.742682390831487e-05,
"loss": 0.5252,
"step": 1540
},
{
"epoch": 2.710411198600175,
"grad_norm": 1.515625,
"learning_rate": 1.739097942446356e-05,
"loss": 0.5354,
"step": 1550
},
{
"epoch": 2.7279090113735784,
"grad_norm": 1.328125,
"learning_rate": 1.7354929275750832e-05,
"loss": 0.5128,
"step": 1560
},
{
"epoch": 2.745406824146982,
"grad_norm": 1.2890625,
"learning_rate": 1.731867462228819e-05,
"loss": 0.5337,
"step": 1570
},
{
"epoch": 2.7629046369203847,
"grad_norm": 1.2890625,
"learning_rate": 1.7282216630768223e-05,
"loss": 0.51,
"step": 1580
},
{
"epoch": 2.7804024496937885,
"grad_norm": 1.296875,
"learning_rate": 1.7245556474427036e-05,
"loss": 0.5161,
"step": 1590
},
{
"epoch": 2.7979002624671914,
"grad_norm": 1.4765625,
"learning_rate": 1.7208695333006483e-05,
"loss": 0.5197,
"step": 1600
},
{
"epoch": 2.815398075240595,
"grad_norm": 1.3046875,
"learning_rate": 1.717163439271623e-05,
"loss": 0.5267,
"step": 1610
},
{
"epoch": 2.832895888013998,
"grad_norm": 1.3828125,
"learning_rate": 1.7134374846195566e-05,
"loss": 0.5144,
"step": 1620
},
{
"epoch": 2.850393700787402,
"grad_norm": 1.3046875,
"learning_rate": 1.7096917892475024e-05,
"loss": 0.5255,
"step": 1630
},
{
"epoch": 2.8678915135608047,
"grad_norm": 1.546875,
"learning_rate": 1.7059264736937796e-05,
"loss": 0.5197,
"step": 1640
},
{
"epoch": 2.885389326334208,
"grad_norm": 1.3828125,
"learning_rate": 1.702141659128095e-05,
"loss": 0.5165,
"step": 1650
},
{
"epoch": 2.9028871391076114,
"grad_norm": 1.296875,
"learning_rate": 1.6983374673476428e-05,
"loss": 0.5202,
"step": 1660
},
{
"epoch": 2.9203849518810148,
"grad_norm": 1.328125,
"learning_rate": 1.694514020773186e-05,
"loss": 0.5227,
"step": 1670
},
{
"epoch": 2.937882764654418,
"grad_norm": 1.328125,
"learning_rate": 1.6906714424451154e-05,
"loss": 0.5223,
"step": 1680
},
{
"epoch": 2.9553805774278215,
"grad_norm": 1.3359375,
"learning_rate": 1.6868098560194926e-05,
"loss": 0.528,
"step": 1690
},
{
"epoch": 2.972878390201225,
"grad_norm": 1.3046875,
"learning_rate": 1.6829293857640675e-05,
"loss": 0.518,
"step": 1700
},
{
"epoch": 2.990376202974628,
"grad_norm": 1.3671875,
"learning_rate": 1.679030156554282e-05,
"loss": 0.5345,
"step": 1710
},
{
"epoch": 3.0069991251093615,
"grad_norm": 1.6015625,
"learning_rate": 1.6751122938692493e-05,
"loss": 0.4821,
"step": 1720
},
{
"epoch": 3.024496937882765,
"grad_norm": 1.6796875,
"learning_rate": 1.6711759237877188e-05,
"loss": 0.4391,
"step": 1730
},
{
"epoch": 3.041994750656168,
"grad_norm": 1.890625,
"learning_rate": 1.667221172984015e-05,
"loss": 0.4258,
"step": 1740
},
{
"epoch": 3.059492563429571,
"grad_norm": 1.6796875,
"learning_rate": 1.663248168723964e-05,
"loss": 0.4333,
"step": 1750
},
{
"epoch": 3.0769903762029744,
"grad_norm": 1.7890625,
"learning_rate": 1.6592570388607978e-05,
"loss": 0.4318,
"step": 1760
},
{
"epoch": 3.094488188976378,
"grad_norm": 1.703125,
"learning_rate": 1.6552479118310378e-05,
"loss": 0.4146,
"step": 1770
},
{
"epoch": 3.111986001749781,
"grad_norm": 1.5078125,
"learning_rate": 1.651220916650364e-05,
"loss": 0.4209,
"step": 1780
},
{
"epoch": 3.1294838145231845,
"grad_norm": 1.6171875,
"learning_rate": 1.6471761829094623e-05,
"loss": 0.4331,
"step": 1790
},
{
"epoch": 3.146981627296588,
"grad_norm": 1.578125,
"learning_rate": 1.643113840769854e-05,
"loss": 0.428,
"step": 1800
},
{
"epoch": 3.164479440069991,
"grad_norm": 1.9609375,
"learning_rate": 1.6390340209597074e-05,
"loss": 0.4195,
"step": 1810
},
{
"epoch": 3.1819772528433945,
"grad_norm": 1.7109375,
"learning_rate": 1.6349368547696297e-05,
"loss": 0.4232,
"step": 1820
},
{
"epoch": 3.199475065616798,
"grad_norm": 1.6171875,
"learning_rate": 1.6308224740484456e-05,
"loss": 0.4258,
"step": 1830
},
{
"epoch": 3.216972878390201,
"grad_norm": 1.640625,
"learning_rate": 1.6266910111989498e-05,
"loss": 0.424,
"step": 1840
},
{
"epoch": 3.2344706911636045,
"grad_norm": 1.5703125,
"learning_rate": 1.622542599173649e-05,
"loss": 0.4326,
"step": 1850
},
{
"epoch": 3.251968503937008,
"grad_norm": 1.4921875,
"learning_rate": 1.6183773714704824e-05,
"loss": 0.436,
"step": 1860
},
{
"epoch": 3.269466316710411,
"grad_norm": 1.671875,
"learning_rate": 1.6141954621285267e-05,
"loss": 0.4423,
"step": 1870
},
{
"epoch": 3.2869641294838146,
"grad_norm": 1.71875,
"learning_rate": 1.6099970057236808e-05,
"loss": 0.4348,
"step": 1880
},
{
"epoch": 3.304461942257218,
"grad_norm": 1.8125,
"learning_rate": 1.605782137364338e-05,
"loss": 0.4306,
"step": 1890
},
{
"epoch": 3.3219597550306212,
"grad_norm": 1.609375,
"learning_rate": 1.6015509926870343e-05,
"loss": 0.4321,
"step": 1900
},
{
"epoch": 3.3394575678040246,
"grad_norm": 1.8203125,
"learning_rate": 1.597303707852087e-05,
"loss": 0.446,
"step": 1910
},
{
"epoch": 3.356955380577428,
"grad_norm": 1.6484375,
"learning_rate": 1.5930404195392114e-05,
"loss": 0.4246,
"step": 1920
},
{
"epoch": 3.3744531933508313,
"grad_norm": 1.7734375,
"learning_rate": 1.588761264943122e-05,
"loss": 0.4184,
"step": 1930
},
{
"epoch": 3.3919510061242346,
"grad_norm": 1.8828125,
"learning_rate": 1.5844663817691187e-05,
"loss": 0.4267,
"step": 1940
},
{
"epoch": 3.409448818897638,
"grad_norm": 2.0,
"learning_rate": 1.5801559082286546e-05,
"loss": 0.4358,
"step": 1950
},
{
"epoch": 3.4269466316710413,
"grad_norm": 1.671875,
"learning_rate": 1.5758299830348884e-05,
"loss": 0.4438,
"step": 1960
},
{
"epoch": 3.4444444444444446,
"grad_norm": 1.578125,
"learning_rate": 1.5714887453982204e-05,
"loss": 0.4413,
"step": 1970
},
{
"epoch": 3.4619422572178475,
"grad_norm": 1.7109375,
"learning_rate": 1.5671323350218135e-05,
"loss": 0.4425,
"step": 1980
},
{
"epoch": 3.4794400699912513,
"grad_norm": 1.640625,
"learning_rate": 1.562760892097096e-05,
"loss": 0.4326,
"step": 1990
},
{
"epoch": 3.4969378827646542,
"grad_norm": 1.59375,
"learning_rate": 1.5583745572992518e-05,
"loss": 0.4335,
"step": 2000
},
{
"epoch": 3.514435695538058,
"grad_norm": 1.734375,
"learning_rate": 1.553973471782692e-05,
"loss": 0.4405,
"step": 2010
},
{
"epoch": 3.531933508311461,
"grad_norm": 1.7265625,
"learning_rate": 1.5495577771765134e-05,
"loss": 0.4259,
"step": 2020
},
{
"epoch": 3.5494313210848643,
"grad_norm": 1.5546875,
"learning_rate": 1.5451276155799405e-05,
"loss": 0.4218,
"step": 2030
},
{
"epoch": 3.5669291338582676,
"grad_norm": 1.65625,
"learning_rate": 1.540683129557752e-05,
"loss": 0.4248,
"step": 2040
},
{
"epoch": 3.584426946631671,
"grad_norm": 1.71875,
"learning_rate": 1.5362244621356946e-05,
"loss": 0.4276,
"step": 2050
},
{
"epoch": 3.6019247594050743,
"grad_norm": 1.6640625,
"learning_rate": 1.531751756795879e-05,
"loss": 0.4408,
"step": 2060
},
{
"epoch": 3.6194225721784776,
"grad_norm": 1.59375,
"learning_rate": 1.5272651574721632e-05,
"loss": 0.4264,
"step": 2070
},
{
"epoch": 3.636920384951881,
"grad_norm": 1.7109375,
"learning_rate": 1.5227648085455202e-05,
"loss": 0.4295,
"step": 2080
},
{
"epoch": 3.6544181977252843,
"grad_norm": 1.7734375,
"learning_rate": 1.5182508548393917e-05,
"loss": 0.427,
"step": 2090
},
{
"epoch": 3.6719160104986877,
"grad_norm": 1.6953125,
"learning_rate": 1.5137234416150288e-05,
"loss": 0.4398,
"step": 2100
},
{
"epoch": 3.689413823272091,
"grad_norm": 1.546875,
"learning_rate": 1.5091827145668156e-05,
"loss": 0.4411,
"step": 2110
},
{
"epoch": 3.7069116360454943,
"grad_norm": 1.734375,
"learning_rate": 1.504628819817582e-05,
"loss": 0.4181,
"step": 2120
},
{
"epoch": 3.7244094488188977,
"grad_norm": 1.6484375,
"learning_rate": 1.5000619039139011e-05,
"loss": 0.4384,
"step": 2130
},
{
"epoch": 3.741907261592301,
"grad_norm": 1.7109375,
"learning_rate": 1.495482113821373e-05,
"loss": 0.4293,
"step": 2140
},
{
"epoch": 3.7594050743657044,
"grad_norm": 1.7734375,
"learning_rate": 1.4908895969198951e-05,
"loss": 0.4298,
"step": 2150
},
{
"epoch": 3.7769028871391077,
"grad_norm": 1.625,
"learning_rate": 1.4862845009989211e-05,
"loss": 0.4403,
"step": 2160
},
{
"epoch": 3.794400699912511,
"grad_norm": 1.6640625,
"learning_rate": 1.4816669742527018e-05,
"loss": 0.4373,
"step": 2170
},
{
"epoch": 3.8118985126859144,
"grad_norm": 1.546875,
"learning_rate": 1.4770371652755203e-05,
"loss": 0.4268,
"step": 2180
},
{
"epoch": 3.8293963254593177,
"grad_norm": 1.8203125,
"learning_rate": 1.4723952230569057e-05,
"loss": 0.4274,
"step": 2190
},
{
"epoch": 3.846894138232721,
"grad_norm": 1.6953125,
"learning_rate": 1.4677412969768427e-05,
"loss": 0.4313,
"step": 2200
},
{
"epoch": 3.864391951006124,
"grad_norm": 1.6484375,
"learning_rate": 1.4630755368009611e-05,
"loss": 0.4465,
"step": 2210
},
{
"epoch": 3.8818897637795278,
"grad_norm": 1.546875,
"learning_rate": 1.4583980926757184e-05,
"loss": 0.4454,
"step": 2220
},
{
"epoch": 3.8993875765529307,
"grad_norm": 1.921875,
"learning_rate": 1.4537091151235673e-05,
"loss": 0.4412,
"step": 2230
},
{
"epoch": 3.9168853893263345,
"grad_norm": 1.796875,
"learning_rate": 1.449008755038112e-05,
"loss": 0.4289,
"step": 2240
},
{
"epoch": 3.9343832020997374,
"grad_norm": 1.8515625,
"learning_rate": 1.4442971636792518e-05,
"loss": 0.4311,
"step": 2250
},
{
"epoch": 3.9518810148731407,
"grad_norm": 1.6640625,
"learning_rate": 1.4395744926683142e-05,
"loss": 0.4263,
"step": 2260
},
{
"epoch": 3.969378827646544,
"grad_norm": 1.875,
"learning_rate": 1.4348408939831758e-05,
"loss": 0.4435,
"step": 2270
},
{
"epoch": 3.9868766404199474,
"grad_norm": 1.65625,
"learning_rate": 1.4300965199533696e-05,
"loss": 0.433,
"step": 2280
},
{
"epoch": 4.003499562554681,
"grad_norm": 2.015625,
"learning_rate": 1.4253415232551861e-05,
"loss": 0.4271,
"step": 2290
},
{
"epoch": 4.020997375328084,
"grad_norm": 2.28125,
"learning_rate": 1.4205760569067577e-05,
"loss": 0.3527,
"step": 2300
},
{
"epoch": 4.038495188101487,
"grad_norm": 1.7265625,
"learning_rate": 1.4158002742631359e-05,
"loss": 0.3734,
"step": 2310
},
{
"epoch": 4.05599300087489,
"grad_norm": 2.296875,
"learning_rate": 1.4110143290113546e-05,
"loss": 0.3634,
"step": 2320
},
{
"epoch": 4.073490813648294,
"grad_norm": 1.9921875,
"learning_rate": 1.4062183751654868e-05,
"loss": 0.3462,
"step": 2330
},
{
"epoch": 4.090988626421697,
"grad_norm": 1.96875,
"learning_rate": 1.4014125670616856e-05,
"loss": 0.3487,
"step": 2340
},
{
"epoch": 4.108486439195101,
"grad_norm": 1.9921875,
"learning_rate": 1.3965970593532201e-05,
"loss": 0.3794,
"step": 2350
},
{
"epoch": 4.125984251968504,
"grad_norm": 1.890625,
"learning_rate": 1.3917720070054965e-05,
"loss": 0.3634,
"step": 2360
},
{
"epoch": 4.1434820647419075,
"grad_norm": 1.9609375,
"learning_rate": 1.386937565291073e-05,
"loss": 0.3601,
"step": 2370
},
{
"epoch": 4.16097987751531,
"grad_norm": 1.9453125,
"learning_rate": 1.382093889784662e-05,
"loss": 0.3654,
"step": 2380
},
{
"epoch": 4.178477690288714,
"grad_norm": 2.078125,
"learning_rate": 1.3772411363581238e-05,
"loss": 0.3492,
"step": 2390
},
{
"epoch": 4.195975503062117,
"grad_norm": 1.96875,
"learning_rate": 1.3723794611754502e-05,
"loss": 0.3683,
"step": 2400
},
{
"epoch": 4.213473315835521,
"grad_norm": 1.859375,
"learning_rate": 1.3675090206877399e-05,
"loss": 0.3634,
"step": 2410
},
{
"epoch": 4.230971128608924,
"grad_norm": 2.1875,
"learning_rate": 1.3626299716281639e-05,
"loss": 0.3623,
"step": 2420
},
{
"epoch": 4.248468941382328,
"grad_norm": 1.890625,
"learning_rate": 1.3577424710069202e-05,
"loss": 0.3596,
"step": 2430
},
{
"epoch": 4.2659667541557305,
"grad_norm": 2.25,
"learning_rate": 1.3528466761061839e-05,
"loss": 0.3623,
"step": 2440
},
{
"epoch": 4.283464566929134,
"grad_norm": 1.9453125,
"learning_rate": 1.3479427444750415e-05,
"loss": 0.3535,
"step": 2450
},
{
"epoch": 4.300962379702537,
"grad_norm": 2.09375,
"learning_rate": 1.343030833924426e-05,
"loss": 0.3592,
"step": 2460
},
{
"epoch": 4.318460192475941,
"grad_norm": 1.9765625,
"learning_rate": 1.338111102522035e-05,
"loss": 0.3644,
"step": 2470
},
{
"epoch": 4.335958005249344,
"grad_norm": 2.25,
"learning_rate": 1.3331837085872444e-05,
"loss": 0.3565,
"step": 2480
},
{
"epoch": 4.353455818022747,
"grad_norm": 2.09375,
"learning_rate": 1.3282488106860146e-05,
"loss": 0.3613,
"step": 2490
},
{
"epoch": 4.3709536307961505,
"grad_norm": 1.984375,
"learning_rate": 1.323306567625788e-05,
"loss": 0.3574,
"step": 2500
},
{
"epoch": 4.388451443569553,
"grad_norm": 2.484375,
"learning_rate": 1.3183571384503767e-05,
"loss": 0.3507,
"step": 2510
},
{
"epoch": 4.405949256342957,
"grad_norm": 1.8671875,
"learning_rate": 1.3134006824348464e-05,
"loss": 0.3583,
"step": 2520
},
{
"epoch": 4.42344706911636,
"grad_norm": 2.0,
"learning_rate": 1.3084373590803898e-05,
"loss": 0.3567,
"step": 2530
},
{
"epoch": 4.440944881889764,
"grad_norm": 1.96875,
"learning_rate": 1.303467328109193e-05,
"loss": 0.3538,
"step": 2540
},
{
"epoch": 4.458442694663167,
"grad_norm": 2.046875,
"learning_rate": 1.2984907494592983e-05,
"loss": 0.3623,
"step": 2550
},
{
"epoch": 4.475940507436571,
"grad_norm": 1.9375,
"learning_rate": 1.2935077832794533e-05,
"loss": 0.3649,
"step": 2560
},
{
"epoch": 4.4934383202099735,
"grad_norm": 1.984375,
"learning_rate": 1.2885185899239617e-05,
"loss": 0.3551,
"step": 2570
},
{
"epoch": 4.510936132983377,
"grad_norm": 2.140625,
"learning_rate": 1.2835233299475192e-05,
"loss": 0.354,
"step": 2580
},
{
"epoch": 4.52843394575678,
"grad_norm": 2.109375,
"learning_rate": 1.2785221641000487e-05,
"loss": 0.3652,
"step": 2590
},
{
"epoch": 4.545931758530184,
"grad_norm": 2.0625,
"learning_rate": 1.2735152533215275e-05,
"loss": 0.3597,
"step": 2600
},
{
"epoch": 4.563429571303587,
"grad_norm": 1.9375,
"learning_rate": 1.2685027587368072e-05,
"loss": 0.3567,
"step": 2610
},
{
"epoch": 4.580927384076991,
"grad_norm": 1.75,
"learning_rate": 1.2634848416504289e-05,
"loss": 0.3577,
"step": 2620
},
{
"epoch": 4.5984251968503935,
"grad_norm": 1.9453125,
"learning_rate": 1.2584616635414325e-05,
"loss": 0.3576,
"step": 2630
},
{
"epoch": 4.615923009623797,
"grad_norm": 1.9609375,
"learning_rate": 1.2534333860581607e-05,
"loss": 0.3609,
"step": 2640
},
{
"epoch": 4.6334208223972,
"grad_norm": 2.109375,
"learning_rate": 1.248400171013056e-05,
"loss": 0.3525,
"step": 2650
},
{
"epoch": 4.650918635170604,
"grad_norm": 1.7734375,
"learning_rate": 1.2433621803774542e-05,
"loss": 0.3567,
"step": 2660
},
{
"epoch": 4.668416447944007,
"grad_norm": 1.921875,
"learning_rate": 1.2383195762763718e-05,
"loss": 0.3623,
"step": 2670
},
{
"epoch": 4.685914260717411,
"grad_norm": 1.8046875,
"learning_rate": 1.2332725209832892e-05,
"loss": 0.3643,
"step": 2680
},
{
"epoch": 4.703412073490814,
"grad_norm": 1.8828125,
"learning_rate": 1.2282211769149283e-05,
"loss": 0.3642,
"step": 2690
},
{
"epoch": 4.7209098862642165,
"grad_norm": 1.96875,
"learning_rate": 1.2231657066260261e-05,
"loss": 0.348,
"step": 2700
},
{
"epoch": 4.73840769903762,
"grad_norm": 2.03125,
"learning_rate": 1.2181062728041029e-05,
"loss": 0.3545,
"step": 2710
},
{
"epoch": 4.755905511811024,
"grad_norm": 2.5625,
"learning_rate": 1.2130430382642276e-05,
"loss": 0.3624,
"step": 2720
},
{
"epoch": 4.773403324584427,
"grad_norm": 1.984375,
"learning_rate": 1.2079761659437781e-05,
"loss": 0.3719,
"step": 2730
},
{
"epoch": 4.79090113735783,
"grad_norm": 2.140625,
"learning_rate": 1.202905818897198e-05,
"loss": 0.362,
"step": 2740
},
{
"epoch": 4.808398950131234,
"grad_norm": 2.15625,
"learning_rate": 1.1978321602907497e-05,
"loss": 0.3691,
"step": 2750
},
{
"epoch": 4.8258967629046365,
"grad_norm": 1.953125,
"learning_rate": 1.192755353397262e-05,
"loss": 0.3611,
"step": 2760
},
{
"epoch": 4.84339457567804,
"grad_norm": 2.21875,
"learning_rate": 1.187675561590878e-05,
"loss": 0.356,
"step": 2770
},
{
"epoch": 4.860892388451443,
"grad_norm": 2.0,
"learning_rate": 1.1825929483417975e-05,
"loss": 0.349,
"step": 2780
},
{
"epoch": 4.878390201224847,
"grad_norm": 2.265625,
"learning_rate": 1.1775076772110136e-05,
"loss": 0.3643,
"step": 2790
},
{
"epoch": 4.89588801399825,
"grad_norm": 1.953125,
"learning_rate": 1.1724199118450537e-05,
"loss": 0.358,
"step": 2800
},
{
"epoch": 4.913385826771654,
"grad_norm": 2.359375,
"learning_rate": 1.1673298159707086e-05,
"loss": 0.358,
"step": 2810
},
{
"epoch": 4.930883639545057,
"grad_norm": 1.875,
"learning_rate": 1.1622375533897683e-05,
"loss": 0.3705,
"step": 2820
},
{
"epoch": 4.94838145231846,
"grad_norm": 2.171875,
"learning_rate": 1.157143287973747e-05,
"loss": 0.3638,
"step": 2830
},
{
"epoch": 4.965879265091863,
"grad_norm": 2.46875,
"learning_rate": 1.1520471836586115e-05,
"loss": 0.3417,
"step": 2840
},
{
"epoch": 4.983377077865267,
"grad_norm": 2.21875,
"learning_rate": 1.1469494044395055e-05,
"loss": 0.3732,
"step": 2850
},
{
"epoch": 5.0,
"grad_norm": 2.796875,
"learning_rate": 1.1418501143654713e-05,
"loss": 0.376,
"step": 2860
},
{
"epoch": 5.017497812773403,
"grad_norm": 2.6875,
"learning_rate": 1.1367494775341716e-05,
"loss": 0.3109,
"step": 2870
},
{
"epoch": 5.034995625546807,
"grad_norm": 2.265625,
"learning_rate": 1.1316476580866091e-05,
"loss": 0.3136,
"step": 2880
},
{
"epoch": 5.05249343832021,
"grad_norm": 2.78125,
"learning_rate": 1.1265448202018432e-05,
"loss": 0.3137,
"step": 2890
},
{
"epoch": 5.069991251093613,
"grad_norm": 2.71875,
"learning_rate": 1.121441128091706e-05,
"loss": 0.3082,
"step": 2900
},
{
"epoch": 5.087489063867016,
"grad_norm": 2.109375,
"learning_rate": 1.1163367459955208e-05,
"loss": 0.3102,
"step": 2910
},
{
"epoch": 5.10498687664042,
"grad_norm": 2.4375,
"learning_rate": 1.1112318381748135e-05,
"loss": 0.3092,
"step": 2920
},
{
"epoch": 5.122484689413823,
"grad_norm": 2.953125,
"learning_rate": 1.1061265689080296e-05,
"loss": 0.3054,
"step": 2930
},
{
"epoch": 5.139982502187227,
"grad_norm": 2.0625,
"learning_rate": 1.1010211024852445e-05,
"loss": 0.3165,
"step": 2940
},
{
"epoch": 5.15748031496063,
"grad_norm": 2.078125,
"learning_rate": 1.0959156032028789e-05,
"loss": 0.3081,
"step": 2950
},
{
"epoch": 5.174978127734033,
"grad_norm": 2.015625,
"learning_rate": 1.0908102353584117e-05,
"loss": 0.3047,
"step": 2960
},
{
"epoch": 5.192475940507436,
"grad_norm": 2.0,
"learning_rate": 1.0857051632450905e-05,
"loss": 0.3016,
"step": 2970
},
{
"epoch": 5.20997375328084,
"grad_norm": 2.1875,
"learning_rate": 1.0806005511466478e-05,
"loss": 0.306,
"step": 2980
},
{
"epoch": 5.227471566054243,
"grad_norm": 2.5,
"learning_rate": 1.075496563332011e-05,
"loss": 0.3036,
"step": 2990
},
{
"epoch": 5.244969378827647,
"grad_norm": 2.203125,
"learning_rate": 1.0703933640500183e-05,
"loss": 0.3041,
"step": 3000
},
{
"epoch": 5.26246719160105,
"grad_norm": 2.515625,
"learning_rate": 1.0652911175241336e-05,
"loss": 0.3183,
"step": 3010
},
{
"epoch": 5.2799650043744535,
"grad_norm": 2.515625,
"learning_rate": 1.0601899879471593e-05,
"loss": 0.2964,
"step": 3020
},
{
"epoch": 5.297462817147856,
"grad_norm": 2.453125,
"learning_rate": 1.0550901394759543e-05,
"loss": 0.3006,
"step": 3030
},
{
"epoch": 5.31496062992126,
"grad_norm": 2.25,
"learning_rate": 1.0499917362261505e-05,
"loss": 0.3146,
"step": 3040
},
{
"epoch": 5.332458442694663,
"grad_norm": 2.421875,
"learning_rate": 1.0448949422668729e-05,
"loss": 0.3153,
"step": 3050
},
{
"epoch": 5.349956255468067,
"grad_norm": 2.234375,
"learning_rate": 1.039799921615458e-05,
"loss": 0.3166,
"step": 3060
},
{
"epoch": 5.36745406824147,
"grad_norm": 2.234375,
"learning_rate": 1.0347068382321758e-05,
"loss": 0.3069,
"step": 3070
},
{
"epoch": 5.3849518810148735,
"grad_norm": 2.1875,
"learning_rate": 1.0296158560149554e-05,
"loss": 0.3031,
"step": 3080
},
{
"epoch": 5.402449693788276,
"grad_norm": 2.75,
"learning_rate": 1.0245271387941085e-05,
"loss": 0.308,
"step": 3090
},
{
"epoch": 5.41994750656168,
"grad_norm": 2.09375,
"learning_rate": 1.0194408503270582e-05,
"loss": 0.3081,
"step": 3100
},
{
"epoch": 5.437445319335083,
"grad_norm": 2.25,
"learning_rate": 1.0143571542930695e-05,
"loss": 0.3052,
"step": 3110
},
{
"epoch": 5.454943132108486,
"grad_norm": 2.140625,
"learning_rate": 1.0092762142879803e-05,
"loss": 0.3156,
"step": 3120
},
{
"epoch": 5.47244094488189,
"grad_norm": 2.109375,
"learning_rate": 1.0041981938189401e-05,
"loss": 0.3076,
"step": 3130
},
{
"epoch": 5.489938757655293,
"grad_norm": 2.171875,
"learning_rate": 9.99123256299145e-06,
"loss": 0.3164,
"step": 3140
},
{
"epoch": 5.5074365704286965,
"grad_norm": 2.234375,
"learning_rate": 9.94051565042581e-06,
"loss": 0.3215,
"step": 3150
},
{
"epoch": 5.524934383202099,
"grad_norm": 2.3125,
"learning_rate": 9.889832832587667e-06,
"loss": 0.3076,
"step": 3160
},
{
"epoch": 5.542432195975503,
"grad_norm": 2.296875,
"learning_rate": 9.83918574047503e-06,
"loss": 0.288,
"step": 3170
},
{
"epoch": 5.559930008748906,
"grad_norm": 2.25,
"learning_rate": 9.788576003936239e-06,
"loss": 0.3116,
"step": 3180
},
{
"epoch": 5.57742782152231,
"grad_norm": 2.203125,
"learning_rate": 9.738005251617508e-06,
"loss": 0.3083,
"step": 3190
},
{
"epoch": 5.594925634295713,
"grad_norm": 2.234375,
"learning_rate": 9.68747511091052e-06,
"loss": 0.3078,
"step": 3200
},
{
"epoch": 5.6124234470691166,
"grad_norm": 2.421875,
"learning_rate": 9.636987207900056e-06,
"loss": 0.3131,
"step": 3210
},
{
"epoch": 5.6299212598425195,
"grad_norm": 2.203125,
"learning_rate": 9.58654316731167e-06,
"loss": 0.3147,
"step": 3220
},
{
"epoch": 5.647419072615923,
"grad_norm": 2.109375,
"learning_rate": 9.536144612459399e-06,
"loss": 0.3131,
"step": 3230
},
{
"epoch": 5.664916885389326,
"grad_norm": 1.953125,
"learning_rate": 9.485793165193531e-06,
"loss": 0.3147,
"step": 3240
},
{
"epoch": 5.68241469816273,
"grad_norm": 2.234375,
"learning_rate": 9.435490445848403e-06,
"loss": 0.3124,
"step": 3250
},
{
"epoch": 5.699912510936133,
"grad_norm": 2.125,
"learning_rate": 9.385238073190264e-06,
"loss": 0.3091,
"step": 3260
},
{
"epoch": 5.717410323709537,
"grad_norm": 2.59375,
"learning_rate": 9.335037664365185e-06,
"loss": 0.2988,
"step": 3270
},
{
"epoch": 5.7349081364829395,
"grad_norm": 2.3125,
"learning_rate": 9.284890834847022e-06,
"loss": 0.306,
"step": 3280
},
{
"epoch": 5.752405949256343,
"grad_norm": 2.515625,
"learning_rate": 9.234799198385402e-06,
"loss": 0.3214,
"step": 3290
},
{
"epoch": 5.769903762029746,
"grad_norm": 2.046875,
"learning_rate": 9.184764366953831e-06,
"loss": 0.3178,
"step": 3300
},
{
"epoch": 5.78740157480315,
"grad_norm": 2.046875,
"learning_rate": 9.134787950697798e-06,
"loss": 0.3061,
"step": 3310
},
{
"epoch": 5.804899387576553,
"grad_norm": 2.25,
"learning_rate": 9.084871557882957e-06,
"loss": 0.3096,
"step": 3320
},
{
"epoch": 5.822397200349956,
"grad_norm": 2.28125,
"learning_rate": 9.035016794843383e-06,
"loss": 0.3131,
"step": 3330
},
{
"epoch": 5.83989501312336,
"grad_norm": 2.09375,
"learning_rate": 8.985225265929872e-06,
"loss": 0.307,
"step": 3340
},
{
"epoch": 5.857392825896763,
"grad_norm": 2.390625,
"learning_rate": 8.93549857345831e-06,
"loss": 0.3107,
"step": 3350
},
{
"epoch": 5.874890638670166,
"grad_norm": 2.09375,
"learning_rate": 8.885838317658126e-06,
"loss": 0.3041,
"step": 3360
},
{
"epoch": 5.892388451443569,
"grad_norm": 1.9765625,
"learning_rate": 8.83624609662078e-06,
"loss": 0.3123,
"step": 3370
},
{
"epoch": 5.909886264216973,
"grad_norm": 2.234375,
"learning_rate": 8.786723506248334e-06,
"loss": 0.316,
"step": 3380
},
{
"epoch": 5.927384076990376,
"grad_norm": 2.625,
"learning_rate": 8.737272140202105e-06,
"loss": 0.3015,
"step": 3390
},
{
"epoch": 5.94488188976378,
"grad_norm": 2.1875,
"learning_rate": 8.687893589851375e-06,
"loss": 0.3216,
"step": 3400
},
{
"epoch": 5.9623797025371825,
"grad_norm": 2.21875,
"learning_rate": 8.638589444222191e-06,
"loss": 0.3143,
"step": 3410
},
{
"epoch": 5.979877515310586,
"grad_norm": 2.40625,
"learning_rate": 8.5893612899462e-06,
"loss": 0.3089,
"step": 3420
},
{
"epoch": 5.997375328083989,
"grad_norm": 2.171875,
"learning_rate": 8.540210711209626e-06,
"loss": 0.3137,
"step": 3430
},
{
"epoch": 6.013998250218723,
"grad_norm": 2.40625,
"learning_rate": 8.491139289702262e-06,
"loss": 0.2933,
"step": 3440
},
{
"epoch": 6.031496062992126,
"grad_norm": 2.28125,
"learning_rate": 8.442148604566598e-06,
"loss": 0.2787,
"step": 3450
},
{
"epoch": 6.04899387576553,
"grad_norm": 2.46875,
"learning_rate": 8.393240232346976e-06,
"loss": 0.2763,
"step": 3460
},
{
"epoch": 6.066491688538933,
"grad_norm": 2.328125,
"learning_rate": 8.34441574693887e-06,
"loss": 0.2706,
"step": 3470
},
{
"epoch": 6.083989501312336,
"grad_norm": 2.21875,
"learning_rate": 8.295676719538231e-06,
"loss": 0.2732,
"step": 3480
},
{
"epoch": 6.101487314085739,
"grad_norm": 2.25,
"learning_rate": 8.247024718590941e-06,
"loss": 0.2772,
"step": 3490
},
{
"epoch": 6.118985126859142,
"grad_norm": 2.484375,
"learning_rate": 8.19846130974232e-06,
"loss": 0.2791,
"step": 3500
},
{
"epoch": 6.136482939632546,
"grad_norm": 2.484375,
"learning_rate": 8.14998805578675e-06,
"loss": 0.2808,
"step": 3510
},
{
"epoch": 6.153980752405949,
"grad_norm": 2.46875,
"learning_rate": 8.101606516617384e-06,
"loss": 0.2908,
"step": 3520
},
{
"epoch": 6.171478565179353,
"grad_norm": 2.359375,
"learning_rate": 8.053318249175955e-06,
"loss": 0.2809,
"step": 3530
},
{
"epoch": 6.188976377952756,
"grad_norm": 2.296875,
"learning_rate": 8.005124807402657e-06,
"loss": 0.2708,
"step": 3540
},
{
"epoch": 6.206474190726159,
"grad_norm": 2.265625,
"learning_rate": 7.957027742186153e-06,
"loss": 0.2789,
"step": 3550
},
{
"epoch": 6.223972003499562,
"grad_norm": 2.234375,
"learning_rate": 7.909028601313658e-06,
"loss": 0.2752,
"step": 3560
},
{
"epoch": 6.241469816272966,
"grad_norm": 2.234375,
"learning_rate": 7.861128929421133e-06,
"loss": 0.2778,
"step": 3570
},
{
"epoch": 6.258967629046369,
"grad_norm": 2.640625,
"learning_rate": 7.813330267943586e-06,
"loss": 0.2779,
"step": 3580
},
{
"epoch": 6.276465441819773,
"grad_norm": 2.015625,
"learning_rate": 7.765634155065451e-06,
"loss": 0.2696,
"step": 3590
},
{
"epoch": 6.293963254593176,
"grad_norm": 2.171875,
"learning_rate": 7.718042125671102e-06,
"loss": 0.2813,
"step": 3600
},
{
"epoch": 6.311461067366579,
"grad_norm": 2.21875,
"learning_rate": 7.670555711295446e-06,
"loss": 0.2847,
"step": 3610
},
{
"epoch": 6.328958880139982,
"grad_norm": 2.515625,
"learning_rate": 7.623176440074667e-06,
"loss": 0.2761,
"step": 3620
},
{
"epoch": 6.346456692913386,
"grad_norm": 2.546875,
"learning_rate": 7.575905836697008e-06,
"loss": 0.2801,
"step": 3630
},
{
"epoch": 6.363954505686789,
"grad_norm": 2.109375,
"learning_rate": 7.528745422353745e-06,
"loss": 0.2848,
"step": 3640
},
{
"epoch": 6.381452318460193,
"grad_norm": 2.75,
"learning_rate": 7.481696714690203e-06,
"loss": 0.2834,
"step": 3650
},
{
"epoch": 6.398950131233596,
"grad_norm": 2.140625,
"learning_rate": 7.4347612277569345e-06,
"loss": 0.2743,
"step": 3660
},
{
"epoch": 6.4164479440069995,
"grad_norm": 2.015625,
"learning_rate": 7.387940471961001e-06,
"loss": 0.2794,
"step": 3670
},
{
"epoch": 6.433945756780402,
"grad_norm": 2.234375,
"learning_rate": 7.341235954017351e-06,
"loss": 0.277,
"step": 3680
},
{
"epoch": 6.451443569553806,
"grad_norm": 2.15625,
"learning_rate": 7.294649176900344e-06,
"loss": 0.2817,
"step": 3690
},
{
"epoch": 6.468941382327209,
"grad_norm": 2.265625,
"learning_rate": 7.248181639795384e-06,
"loss": 0.2816,
"step": 3700
},
{
"epoch": 6.486439195100612,
"grad_norm": 2.328125,
"learning_rate": 7.201834838050668e-06,
"loss": 0.2765,
"step": 3710
},
{
"epoch": 6.503937007874016,
"grad_norm": 2.34375,
"learning_rate": 7.155610263129082e-06,
"loss": 0.2778,
"step": 3720
},
{
"epoch": 6.5214348206474195,
"grad_norm": 2.328125,
"learning_rate": 7.109509402560171e-06,
"loss": 0.27,
"step": 3730
},
{
"epoch": 6.538932633420822,
"grad_norm": 2.359375,
"learning_rate": 7.063533739892312e-06,
"loss": 0.2713,
"step": 3740
},
{
"epoch": 6.556430446194225,
"grad_norm": 2.375,
"learning_rate": 7.017684754644938e-06,
"loss": 0.2801,
"step": 3750
},
{
"epoch": 6.573928258967629,
"grad_norm": 2.234375,
"learning_rate": 6.971963922260954e-06,
"loss": 0.2893,
"step": 3760
},
{
"epoch": 6.591426071741032,
"grad_norm": 2.71875,
"learning_rate": 6.926372714059227e-06,
"loss": 0.2785,
"step": 3770
},
{
"epoch": 6.608923884514436,
"grad_norm": 2.28125,
"learning_rate": 6.880912597187266e-06,
"loss": 0.2762,
"step": 3780
},
{
"epoch": 6.626421697287839,
"grad_norm": 2.421875,
"learning_rate": 6.835585034573993e-06,
"loss": 0.2791,
"step": 3790
},
{
"epoch": 6.6439195100612425,
"grad_norm": 2.796875,
"learning_rate": 6.790391484882665e-06,
"loss": 0.2795,
"step": 3800
},
{
"epoch": 6.661417322834645,
"grad_norm": 2.078125,
"learning_rate": 6.745333402463944e-06,
"loss": 0.2845,
"step": 3810
},
{
"epoch": 6.678915135608049,
"grad_norm": 2.46875,
"learning_rate": 6.700412237309084e-06,
"loss": 0.2829,
"step": 3820
},
{
"epoch": 6.696412948381452,
"grad_norm": 2.421875,
"learning_rate": 6.655629435003274e-06,
"loss": 0.2814,
"step": 3830
},
{
"epoch": 6.713910761154856,
"grad_norm": 2.515625,
"learning_rate": 6.6109864366791225e-06,
"loss": 0.2703,
"step": 3840
},
{
"epoch": 6.731408573928259,
"grad_norm": 2.1875,
"learning_rate": 6.56648467897027e-06,
"loss": 0.2856,
"step": 3850
},
{
"epoch": 6.7489063867016625,
"grad_norm": 2.109375,
"learning_rate": 6.522125593965171e-06,
"loss": 0.2766,
"step": 3860
},
{
"epoch": 6.766404199475065,
"grad_norm": 2.546875,
"learning_rate": 6.477910609161004e-06,
"loss": 0.2792,
"step": 3870
},
{
"epoch": 6.783902012248469,
"grad_norm": 2.5,
"learning_rate": 6.433841147417717e-06,
"loss": 0.28,
"step": 3880
},
{
"epoch": 6.801399825021872,
"grad_norm": 2.125,
"learning_rate": 6.389918626912277e-06,
"loss": 0.2773,
"step": 3890
},
{
"epoch": 6.818897637795276,
"grad_norm": 2.3125,
"learning_rate": 6.346144461092991e-06,
"loss": 0.2761,
"step": 3900
},
{
"epoch": 6.836395450568679,
"grad_norm": 2.484375,
"learning_rate": 6.302520058634057e-06,
"loss": 0.2765,
"step": 3910
},
{
"epoch": 6.853893263342083,
"grad_norm": 2.125,
"learning_rate": 6.259046823390184e-06,
"loss": 0.2746,
"step": 3920
},
{
"epoch": 6.8713910761154855,
"grad_norm": 2.140625,
"learning_rate": 6.2157261543514825e-06,
"loss": 0.2832,
"step": 3930
},
{
"epoch": 6.888888888888889,
"grad_norm": 2.171875,
"learning_rate": 6.172559445598385e-06,
"loss": 0.2831,
"step": 3940
},
{
"epoch": 6.906386701662292,
"grad_norm": 2.09375,
"learning_rate": 6.129548086256816e-06,
"loss": 0.2821,
"step": 3950
},
{
"epoch": 6.923884514435695,
"grad_norm": 2.421875,
"learning_rate": 6.086693460453466e-06,
"loss": 0.2905,
"step": 3960
},
{
"epoch": 6.941382327209099,
"grad_norm": 2.046875,
"learning_rate": 6.0439969472712734e-06,
"loss": 0.2772,
"step": 3970
},
{
"epoch": 6.958880139982503,
"grad_norm": 2.28125,
"learning_rate": 6.001459920705046e-06,
"loss": 0.2814,
"step": 3980
},
{
"epoch": 6.9763779527559056,
"grad_norm": 2.59375,
"learning_rate": 5.95908374961721e-06,
"loss": 0.2674,
"step": 3990
},
{
"epoch": 6.9938757655293085,
"grad_norm": 2.234375,
"learning_rate": 5.916869797693794e-06,
"loss": 0.2746,
"step": 4000
},
{
"epoch": 7.010498687664042,
"grad_norm": 2.4375,
"learning_rate": 5.874819423400538e-06,
"loss": 0.2766,
"step": 4010
},
{
"epoch": 7.027996500437445,
"grad_norm": 2.1875,
"learning_rate": 5.832933979939177e-06,
"loss": 0.257,
"step": 4020
},
{
"epoch": 7.045494313210849,
"grad_norm": 2.328125,
"learning_rate": 5.791214815203874e-06,
"loss": 0.2575,
"step": 4030
},
{
"epoch": 7.062992125984252,
"grad_norm": 2.03125,
"learning_rate": 5.749663271737873e-06,
"loss": 0.2589,
"step": 4040
},
{
"epoch": 7.080489938757656,
"grad_norm": 2.046875,
"learning_rate": 5.708280686690284e-06,
"loss": 0.2585,
"step": 4050
},
{
"epoch": 7.0979877515310585,
"grad_norm": 2.109375,
"learning_rate": 5.667068391773045e-06,
"loss": 0.2696,
"step": 4060
},
{
"epoch": 7.115485564304462,
"grad_norm": 2.703125,
"learning_rate": 5.62602771321808e-06,
"loss": 0.263,
"step": 4070
},
{
"epoch": 7.132983377077865,
"grad_norm": 2.453125,
"learning_rate": 5.5851599717346106e-06,
"loss": 0.2598,
"step": 4080
},
{
"epoch": 7.150481189851269,
"grad_norm": 2.421875,
"learning_rate": 5.5444664824666594e-06,
"loss": 0.2683,
"step": 4090
},
{
"epoch": 7.167979002624672,
"grad_norm": 2.3125,
"learning_rate": 5.503948554950727e-06,
"loss": 0.2669,
"step": 4100
},
{
"epoch": 7.185476815398075,
"grad_norm": 2.328125,
"learning_rate": 5.4636074930736525e-06,
"loss": 0.2546,
"step": 4110
},
{
"epoch": 7.202974628171479,
"grad_norm": 2.328125,
"learning_rate": 5.423444595030648e-06,
"loss": 0.262,
"step": 4120
},
{
"epoch": 7.2204724409448815,
"grad_norm": 2.21875,
"learning_rate": 5.383461153283529e-06,
"loss": 0.2772,
"step": 4130
},
{
"epoch": 7.237970253718285,
"grad_norm": 2.203125,
"learning_rate": 5.343658454519113e-06,
"loss": 0.2584,
"step": 4140
},
{
"epoch": 7.255468066491688,
"grad_norm": 2.3125,
"learning_rate": 5.3040377796078295e-06,
"loss": 0.2626,
"step": 4150
},
{
"epoch": 7.272965879265092,
"grad_norm": 2.09375,
"learning_rate": 5.264600403562482e-06,
"loss": 0.2643,
"step": 4160
},
{
"epoch": 7.290463692038495,
"grad_norm": 2.1875,
"learning_rate": 5.225347595497234e-06,
"loss": 0.2598,
"step": 4170
},
{
"epoch": 7.307961504811899,
"grad_norm": 2.09375,
"learning_rate": 5.186280618586752e-06,
"loss": 0.2584,
"step": 4180
},
{
"epoch": 7.3254593175853016,
"grad_norm": 2.421875,
"learning_rate": 5.147400730025567e-06,
"loss": 0.2551,
"step": 4190
},
{
"epoch": 7.342957130358705,
"grad_norm": 2.03125,
"learning_rate": 5.108709180987623e-06,
"loss": 0.2573,
"step": 4200
},
{
"epoch": 7.360454943132108,
"grad_norm": 2.09375,
"learning_rate": 5.070207216586e-06,
"loss": 0.2546,
"step": 4210
},
{
"epoch": 7.377952755905512,
"grad_norm": 2.203125,
"learning_rate": 5.031896075832846e-06,
"loss": 0.2503,
"step": 4220
},
{
"epoch": 7.395450568678915,
"grad_norm": 2.171875,
"learning_rate": 4.993776991599511e-06,
"loss": 0.2612,
"step": 4230
},
{
"epoch": 7.412948381452319,
"grad_norm": 2.40625,
"learning_rate": 4.955851190576886e-06,
"loss": 0.2624,
"step": 4240
},
{
"epoch": 7.430446194225722,
"grad_norm": 2.328125,
"learning_rate": 4.918119893235894e-06,
"loss": 0.2627,
"step": 4250
},
{
"epoch": 7.447944006999125,
"grad_norm": 2.390625,
"learning_rate": 4.880584313788245e-06,
"loss": 0.2505,
"step": 4260
},
{
"epoch": 7.465441819772528,
"grad_norm": 2.546875,
"learning_rate": 4.843245660147346e-06,
"loss": 0.2545,
"step": 4270
},
{
"epoch": 7.482939632545932,
"grad_norm": 2.3125,
"learning_rate": 4.806105133889444e-06,
"loss": 0.265,
"step": 4280
},
{
"epoch": 7.500437445319335,
"grad_norm": 2.046875,
"learning_rate": 4.7691639302149365e-06,
"loss": 0.2661,
"step": 4290
},
{
"epoch": 7.517935258092739,
"grad_norm": 2.203125,
"learning_rate": 4.732423237909929e-06,
"loss": 0.2628,
"step": 4300
},
{
"epoch": 7.535433070866142,
"grad_norm": 2.140625,
"learning_rate": 4.695884239307972e-06,
"loss": 0.2691,
"step": 4310
},
{
"epoch": 7.5529308836395455,
"grad_norm": 2.296875,
"learning_rate": 4.659548110252012e-06,
"loss": 0.2632,
"step": 4320
},
{
"epoch": 7.570428696412948,
"grad_norm": 2.71875,
"learning_rate": 4.623416020056556e-06,
"loss": 0.2625,
"step": 4330
},
{
"epoch": 7.587926509186351,
"grad_norm": 2.296875,
"learning_rate": 4.587489131470039e-06,
"loss": 0.2514,
"step": 4340
},
{
"epoch": 7.605424321959755,
"grad_norm": 2.21875,
"learning_rate": 4.551768600637407e-06,
"loss": 0.2686,
"step": 4350
},
{
"epoch": 7.622922134733159,
"grad_norm": 2.515625,
"learning_rate": 4.516255577062913e-06,
"loss": 0.2701,
"step": 4360
},
{
"epoch": 7.640419947506562,
"grad_norm": 2.265625,
"learning_rate": 4.48095120357312e-06,
"loss": 0.2617,
"step": 4370
},
{
"epoch": 7.657917760279965,
"grad_norm": 2.09375,
"learning_rate": 4.445856616280136e-06,
"loss": 0.2597,
"step": 4380
},
{
"epoch": 7.675415573053368,
"grad_norm": 2.140625,
"learning_rate": 4.410972944545041e-06,
"loss": 0.2695,
"step": 4390
},
{
"epoch": 7.692913385826771,
"grad_norm": 2.21875,
"learning_rate": 4.376301310941552e-06,
"loss": 0.2632,
"step": 4400
},
{
"epoch": 7.710411198600175,
"grad_norm": 2.484375,
"learning_rate": 4.3418428312198835e-06,
"loss": 0.2629,
"step": 4410
},
{
"epoch": 7.727909011373578,
"grad_norm": 2.203125,
"learning_rate": 4.307598614270871e-06,
"loss": 0.2505,
"step": 4420
},
{
"epoch": 7.745406824146982,
"grad_norm": 2.171875,
"learning_rate": 4.273569762090261e-06,
"loss": 0.2695,
"step": 4430
},
{
"epoch": 7.762904636920385,
"grad_norm": 2.296875,
"learning_rate": 4.239757369743248e-06,
"loss": 0.2636,
"step": 4440
},
{
"epoch": 7.7804024496937885,
"grad_norm": 2.5,
"learning_rate": 4.2061625253292484e-06,
"loss": 0.257,
"step": 4450
},
{
"epoch": 7.797900262467191,
"grad_norm": 2.265625,
"learning_rate": 4.172786309946885e-06,
"loss": 0.2672,
"step": 4460
},
{
"epoch": 7.815398075240595,
"grad_norm": 2.40625,
"learning_rate": 4.139629797659188e-06,
"loss": 0.2641,
"step": 4470
},
{
"epoch": 7.832895888013998,
"grad_norm": 2.1875,
"learning_rate": 4.106694055459023e-06,
"loss": 0.2654,
"step": 4480
},
{
"epoch": 7.850393700787402,
"grad_norm": 2.671875,
"learning_rate": 4.073980143234777e-06,
"loss": 0.2606,
"step": 4490
},
{
"epoch": 7.867891513560805,
"grad_norm": 2.421875,
"learning_rate": 4.041489113736244e-06,
"loss": 0.2669,
"step": 4500
},
{
"epoch": 7.8853893263342085,
"grad_norm": 2.125,
"learning_rate": 4.009222012540725e-06,
"loss": 0.274,
"step": 4510
},
{
"epoch": 7.902887139107611,
"grad_norm": 2.125,
"learning_rate": 3.977179878019412e-06,
"loss": 0.259,
"step": 4520
},
{
"epoch": 7.920384951881015,
"grad_norm": 2.53125,
"learning_rate": 3.9453637413039536e-06,
"loss": 0.2611,
"step": 4530
},
{
"epoch": 7.937882764654418,
"grad_norm": 2.40625,
"learning_rate": 3.913774626253279e-06,
"loss": 0.2541,
"step": 4540
},
{
"epoch": 7.955380577427822,
"grad_norm": 2.171875,
"learning_rate": 3.882413549420649e-06,
"loss": 0.2657,
"step": 4550
},
{
"epoch": 7.972878390201225,
"grad_norm": 2.3125,
"learning_rate": 3.851281520020941e-06,
"loss": 0.2671,
"step": 4560
},
{
"epoch": 7.990376202974629,
"grad_norm": 1.9921875,
"learning_rate": 3.820379539898173e-06,
"loss": 0.271,
"step": 4570
}
],
"logging_steps": 10,
"max_steps": 5710,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.2912807922064032e+19,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}