LogicCoder-7B / trainer_state.json
azzzacs's picture
Upload folder using huggingface_hub
5f45462 verified
raw
history blame
47.8 kB
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 5.0,
"eval_steps": 500,
"global_step": 2860,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.017497812773403325,
"grad_norm": 4.1875,
"learning_rate": 2.0930232558139536e-06,
"loss": 1.0581,
"step": 10
},
{
"epoch": 0.03499562554680665,
"grad_norm": 3.328125,
"learning_rate": 4.418604651162791e-06,
"loss": 1.0503,
"step": 20
},
{
"epoch": 0.05249343832020997,
"grad_norm": 2.71875,
"learning_rate": 6.744186046511628e-06,
"loss": 0.9804,
"step": 30
},
{
"epoch": 0.0699912510936133,
"grad_norm": 1.4296875,
"learning_rate": 9.069767441860465e-06,
"loss": 0.8786,
"step": 40
},
{
"epoch": 0.08748906386701662,
"grad_norm": 1.0,
"learning_rate": 1.1395348837209304e-05,
"loss": 0.8081,
"step": 50
},
{
"epoch": 0.10498687664041995,
"grad_norm": 0.8203125,
"learning_rate": 1.372093023255814e-05,
"loss": 0.8047,
"step": 60
},
{
"epoch": 0.12248468941382328,
"grad_norm": 0.765625,
"learning_rate": 1.6046511627906977e-05,
"loss": 0.7453,
"step": 70
},
{
"epoch": 0.1399825021872266,
"grad_norm": 0.8125,
"learning_rate": 1.8372093023255815e-05,
"loss": 0.7581,
"step": 80
},
{
"epoch": 0.15748031496062992,
"grad_norm": 0.734375,
"learning_rate": 2.0697674418604654e-05,
"loss": 0.7399,
"step": 90
},
{
"epoch": 0.17497812773403323,
"grad_norm": 0.71875,
"learning_rate": 2.3023255813953492e-05,
"loss": 0.7476,
"step": 100
},
{
"epoch": 0.19247594050743658,
"grad_norm": 0.7109375,
"learning_rate": 2.5348837209302327e-05,
"loss": 0.7247,
"step": 110
},
{
"epoch": 0.2099737532808399,
"grad_norm": 0.81640625,
"learning_rate": 2.7674418604651166e-05,
"loss": 0.7179,
"step": 120
},
{
"epoch": 0.2274715660542432,
"grad_norm": 0.77734375,
"learning_rate": 3.0000000000000004e-05,
"loss": 0.6949,
"step": 130
},
{
"epoch": 0.24496937882764655,
"grad_norm": 0.82421875,
"learning_rate": 3.232558139534884e-05,
"loss": 0.7043,
"step": 140
},
{
"epoch": 0.26246719160104987,
"grad_norm": 0.69921875,
"learning_rate": 3.4651162790697674e-05,
"loss": 0.6934,
"step": 150
},
{
"epoch": 0.2799650043744532,
"grad_norm": 0.6796875,
"learning_rate": 3.697674418604651e-05,
"loss": 0.6993,
"step": 160
},
{
"epoch": 0.2974628171478565,
"grad_norm": 0.66015625,
"learning_rate": 3.930232558139535e-05,
"loss": 0.7027,
"step": 170
},
{
"epoch": 0.31496062992125984,
"grad_norm": 0.66015625,
"learning_rate": 3.9999858083836754e-05,
"loss": 0.6935,
"step": 180
},
{
"epoch": 0.3324584426946632,
"grad_norm": 0.6875,
"learning_rate": 3.999916298964876e-05,
"loss": 0.7137,
"step": 190
},
{
"epoch": 0.34995625546806647,
"grad_norm": 0.7578125,
"learning_rate": 3.999788867354247e-05,
"loss": 0.6916,
"step": 200
},
{
"epoch": 0.3674540682414698,
"grad_norm": 0.69921875,
"learning_rate": 3.9996035176526014e-05,
"loss": 0.6947,
"step": 210
},
{
"epoch": 0.38495188101487315,
"grad_norm": 0.77734375,
"learning_rate": 3.9993602558245834e-05,
"loss": 0.7045,
"step": 220
},
{
"epoch": 0.40244969378827644,
"grad_norm": 0.6484375,
"learning_rate": 3.9990590896984766e-05,
"loss": 0.7016,
"step": 230
},
{
"epoch": 0.4199475065616798,
"grad_norm": 0.6328125,
"learning_rate": 3.998700028965957e-05,
"loss": 0.7072,
"step": 240
},
{
"epoch": 0.4374453193350831,
"grad_norm": 0.65234375,
"learning_rate": 3.9982830851817736e-05,
"loss": 0.6986,
"step": 250
},
{
"epoch": 0.4549431321084864,
"grad_norm": 0.66015625,
"learning_rate": 3.997808271763385e-05,
"loss": 0.6804,
"step": 260
},
{
"epoch": 0.47244094488188976,
"grad_norm": 0.6640625,
"learning_rate": 3.997275603990519e-05,
"loss": 0.7016,
"step": 270
},
{
"epoch": 0.4899387576552931,
"grad_norm": 0.703125,
"learning_rate": 3.996685099004686e-05,
"loss": 0.6921,
"step": 280
},
{
"epoch": 0.5074365704286964,
"grad_norm": 0.6875,
"learning_rate": 3.996036775808628e-05,
"loss": 0.6857,
"step": 290
},
{
"epoch": 0.5249343832020997,
"grad_norm": 0.6875,
"learning_rate": 3.995330655265704e-05,
"loss": 0.6967,
"step": 300
},
{
"epoch": 0.5424321959755031,
"grad_norm": 0.72265625,
"learning_rate": 3.994566760099221e-05,
"loss": 0.6969,
"step": 310
},
{
"epoch": 0.5599300087489064,
"grad_norm": 0.59765625,
"learning_rate": 3.9937451148917e-05,
"loss": 0.6862,
"step": 320
},
{
"epoch": 0.5774278215223098,
"grad_norm": 0.625,
"learning_rate": 3.992865746084089e-05,
"loss": 0.695,
"step": 330
},
{
"epoch": 0.594925634295713,
"grad_norm": 0.62109375,
"learning_rate": 3.991928681974908e-05,
"loss": 0.7036,
"step": 340
},
{
"epoch": 0.6124234470691163,
"grad_norm": 0.734375,
"learning_rate": 3.9909339527193416e-05,
"loss": 0.7064,
"step": 350
},
{
"epoch": 0.6299212598425197,
"grad_norm": 0.6171875,
"learning_rate": 3.989881590328265e-05,
"loss": 0.6768,
"step": 360
},
{
"epoch": 0.647419072615923,
"grad_norm": 0.6640625,
"learning_rate": 3.98877162866722e-05,
"loss": 0.7161,
"step": 370
},
{
"epoch": 0.6649168853893264,
"grad_norm": 0.625,
"learning_rate": 3.9876041034553165e-05,
"loss": 0.6731,
"step": 380
},
{
"epoch": 0.6824146981627297,
"grad_norm": 0.65625,
"learning_rate": 3.9863790522640926e-05,
"loss": 0.6905,
"step": 390
},
{
"epoch": 0.6999125109361329,
"grad_norm": 0.66796875,
"learning_rate": 3.9850965145162954e-05,
"loss": 0.7067,
"step": 400
},
{
"epoch": 0.7174103237095363,
"grad_norm": 0.6015625,
"learning_rate": 3.983756531484622e-05,
"loss": 0.6982,
"step": 410
},
{
"epoch": 0.7349081364829396,
"grad_norm": 0.59375,
"learning_rate": 3.9823591462903856e-05,
"loss": 0.6932,
"step": 420
},
{
"epoch": 0.752405949256343,
"grad_norm": 0.6953125,
"learning_rate": 3.9809044039021295e-05,
"loss": 0.7004,
"step": 430
},
{
"epoch": 0.7699037620297463,
"grad_norm": 0.63671875,
"learning_rate": 3.979392351134178e-05,
"loss": 0.7002,
"step": 440
},
{
"epoch": 0.7874015748031497,
"grad_norm": 0.63671875,
"learning_rate": 3.977823036645132e-05,
"loss": 0.7065,
"step": 450
},
{
"epoch": 0.8048993875765529,
"grad_norm": 0.62109375,
"learning_rate": 3.976196510936307e-05,
"loss": 0.692,
"step": 460
},
{
"epoch": 0.8223972003499562,
"grad_norm": 0.58984375,
"learning_rate": 3.9745128263500976e-05,
"loss": 0.6847,
"step": 470
},
{
"epoch": 0.8398950131233596,
"grad_norm": 0.75,
"learning_rate": 3.972772037068303e-05,
"loss": 0.6868,
"step": 480
},
{
"epoch": 0.8573928258967629,
"grad_norm": 0.69140625,
"learning_rate": 3.970974199110378e-05,
"loss": 0.6727,
"step": 490
},
{
"epoch": 0.8748906386701663,
"grad_norm": 0.54296875,
"learning_rate": 3.9691193703316336e-05,
"loss": 0.69,
"step": 500
},
{
"epoch": 0.8923884514435696,
"grad_norm": 0.578125,
"learning_rate": 3.9672076104213706e-05,
"loss": 0.6696,
"step": 510
},
{
"epoch": 0.9098862642169728,
"grad_norm": 0.61328125,
"learning_rate": 3.965238980900965e-05,
"loss": 0.6852,
"step": 520
},
{
"epoch": 0.9273840769903762,
"grad_norm": 0.6328125,
"learning_rate": 3.963213545121881e-05,
"loss": 0.6852,
"step": 530
},
{
"epoch": 0.9448818897637795,
"grad_norm": 0.63671875,
"learning_rate": 3.9611313682636395e-05,
"loss": 0.6907,
"step": 540
},
{
"epoch": 0.9623797025371829,
"grad_norm": 0.59375,
"learning_rate": 3.9589925173317165e-05,
"loss": 0.6753,
"step": 550
},
{
"epoch": 0.9798775153105862,
"grad_norm": 0.63671875,
"learning_rate": 3.956797061155386e-05,
"loss": 0.6796,
"step": 560
},
{
"epoch": 0.9973753280839895,
"grad_norm": 0.640625,
"learning_rate": 3.954545070385508e-05,
"loss": 0.6875,
"step": 570
},
{
"epoch": 1.0139982502187226,
"grad_norm": 0.59765625,
"learning_rate": 3.952236617492253e-05,
"loss": 0.6204,
"step": 580
},
{
"epoch": 1.031496062992126,
"grad_norm": 0.59765625,
"learning_rate": 3.94987177676277e-05,
"loss": 0.5968,
"step": 590
},
{
"epoch": 1.0489938757655293,
"grad_norm": 0.671875,
"learning_rate": 3.947450624298799e-05,
"loss": 0.6147,
"step": 600
},
{
"epoch": 1.0664916885389326,
"grad_norm": 0.65234375,
"learning_rate": 3.944973238014215e-05,
"loss": 0.6052,
"step": 610
},
{
"epoch": 1.083989501312336,
"grad_norm": 0.6640625,
"learning_rate": 3.942439697632528e-05,
"loss": 0.6152,
"step": 620
},
{
"epoch": 1.1014873140857393,
"grad_norm": 0.62109375,
"learning_rate": 3.9398500846843136e-05,
"loss": 0.622,
"step": 630
},
{
"epoch": 1.1189851268591426,
"grad_norm": 0.578125,
"learning_rate": 3.9372044825045905e-05,
"loss": 0.6361,
"step": 640
},
{
"epoch": 1.136482939632546,
"grad_norm": 0.5703125,
"learning_rate": 3.934502976230138e-05,
"loss": 0.6015,
"step": 650
},
{
"epoch": 1.1539807524059493,
"grad_norm": 0.6171875,
"learning_rate": 3.9317456527967566e-05,
"loss": 0.6231,
"step": 660
},
{
"epoch": 1.1714785651793527,
"grad_norm": 0.640625,
"learning_rate": 3.92893260093647e-05,
"loss": 0.6325,
"step": 670
},
{
"epoch": 1.188976377952756,
"grad_norm": 0.58203125,
"learning_rate": 3.926063911174673e-05,
"loss": 0.6,
"step": 680
},
{
"epoch": 1.2064741907261591,
"grad_norm": 0.58984375,
"learning_rate": 3.92313967582721e-05,
"loss": 0.6158,
"step": 690
},
{
"epoch": 1.2239720034995625,
"grad_norm": 0.62109375,
"learning_rate": 3.9201599889974155e-05,
"loss": 0.6184,
"step": 700
},
{
"epoch": 1.2414698162729658,
"grad_norm": 0.6015625,
"learning_rate": 3.917124946573076e-05,
"loss": 0.6159,
"step": 710
},
{
"epoch": 1.2589676290463692,
"grad_norm": 0.57421875,
"learning_rate": 3.91403464622335e-05,
"loss": 0.6221,
"step": 720
},
{
"epoch": 1.2764654418197725,
"grad_norm": 0.640625,
"learning_rate": 3.910889187395622e-05,
"loss": 0.6393,
"step": 730
},
{
"epoch": 1.2939632545931758,
"grad_norm": 0.6484375,
"learning_rate": 3.9076886713123056e-05,
"loss": 0.6168,
"step": 740
},
{
"epoch": 1.3114610673665792,
"grad_norm": 0.578125,
"learning_rate": 3.9044332009675816e-05,
"loss": 0.611,
"step": 750
},
{
"epoch": 1.3289588801399825,
"grad_norm": 0.66796875,
"learning_rate": 3.901122881124087e-05,
"loss": 0.625,
"step": 760
},
{
"epoch": 1.3464566929133859,
"grad_norm": 0.6640625,
"learning_rate": 3.897757818309543e-05,
"loss": 0.6196,
"step": 770
},
{
"epoch": 1.3639545056867892,
"grad_norm": 0.62109375,
"learning_rate": 3.8943381208133266e-05,
"loss": 0.6451,
"step": 780
},
{
"epoch": 1.3814523184601923,
"grad_norm": 0.62109375,
"learning_rate": 3.890863898682984e-05,
"loss": 0.6255,
"step": 790
},
{
"epoch": 1.3989501312335957,
"grad_norm": 0.6171875,
"learning_rate": 3.887335263720693e-05,
"loss": 0.6182,
"step": 800
},
{
"epoch": 1.416447944006999,
"grad_norm": 0.57421875,
"learning_rate": 3.883752329479662e-05,
"loss": 0.6206,
"step": 810
},
{
"epoch": 1.4339457567804024,
"grad_norm": 0.65625,
"learning_rate": 3.880115211260478e-05,
"loss": 0.6068,
"step": 820
},
{
"epoch": 1.4514435695538057,
"grad_norm": 0.52734375,
"learning_rate": 3.876424026107394e-05,
"loss": 0.6221,
"step": 830
},
{
"epoch": 1.468941382327209,
"grad_norm": 0.6328125,
"learning_rate": 3.872678892804563e-05,
"loss": 0.6118,
"step": 840
},
{
"epoch": 1.4864391951006124,
"grad_norm": 0.60546875,
"learning_rate": 3.868879931872218e-05,
"loss": 0.6137,
"step": 850
},
{
"epoch": 1.5039370078740157,
"grad_norm": 0.58203125,
"learning_rate": 3.865027265562789e-05,
"loss": 0.61,
"step": 860
},
{
"epoch": 1.521434820647419,
"grad_norm": 0.60546875,
"learning_rate": 3.8611210178569746e-05,
"loss": 0.6285,
"step": 870
},
{
"epoch": 1.5389326334208224,
"grad_norm": 0.59765625,
"learning_rate": 3.857161314459745e-05,
"loss": 0.6211,
"step": 880
},
{
"epoch": 1.5564304461942258,
"grad_norm": 0.58203125,
"learning_rate": 3.853148282796306e-05,
"loss": 0.6291,
"step": 890
},
{
"epoch": 1.5739282589676291,
"grad_norm": 0.59375,
"learning_rate": 3.849082052007988e-05,
"loss": 0.6074,
"step": 900
},
{
"epoch": 1.5914260717410325,
"grad_norm": 0.6640625,
"learning_rate": 3.844962752948099e-05,
"loss": 0.6314,
"step": 910
},
{
"epoch": 1.6089238845144358,
"grad_norm": 0.5859375,
"learning_rate": 3.8407905181777095e-05,
"loss": 0.6167,
"step": 920
},
{
"epoch": 1.6264216972878391,
"grad_norm": 0.6484375,
"learning_rate": 3.836565481961386e-05,
"loss": 0.6235,
"step": 930
},
{
"epoch": 1.6439195100612425,
"grad_norm": 0.5859375,
"learning_rate": 3.8322877802628704e-05,
"loss": 0.6106,
"step": 940
},
{
"epoch": 1.6614173228346458,
"grad_norm": 0.5546875,
"learning_rate": 3.827957550740711e-05,
"loss": 0.6041,
"step": 950
},
{
"epoch": 1.678915135608049,
"grad_norm": 0.59375,
"learning_rate": 3.82357493274382e-05,
"loss": 0.6176,
"step": 960
},
{
"epoch": 1.6964129483814523,
"grad_norm": 0.5625,
"learning_rate": 3.8191400673070034e-05,
"loss": 0.6105,
"step": 970
},
{
"epoch": 1.7139107611548556,
"grad_norm": 0.55859375,
"learning_rate": 3.81465309714641e-05,
"loss": 0.6278,
"step": 980
},
{
"epoch": 1.731408573928259,
"grad_norm": 0.6484375,
"learning_rate": 3.810114166654948e-05,
"loss": 0.6189,
"step": 990
},
{
"epoch": 1.7489063867016623,
"grad_norm": 0.66796875,
"learning_rate": 3.805523421897633e-05,
"loss": 0.6299,
"step": 1000
},
{
"epoch": 1.7664041994750657,
"grad_norm": 0.62109375,
"learning_rate": 3.800881010606889e-05,
"loss": 0.6059,
"step": 1010
},
{
"epoch": 1.7839020122484688,
"grad_norm": 0.58203125,
"learning_rate": 3.7961870821777946e-05,
"loss": 0.6199,
"step": 1020
},
{
"epoch": 1.8013998250218721,
"grad_norm": 0.58984375,
"learning_rate": 3.791441787663276e-05,
"loss": 0.5948,
"step": 1030
},
{
"epoch": 1.8188976377952755,
"grad_norm": 0.55859375,
"learning_rate": 3.786645279769244e-05,
"loss": 0.608,
"step": 1040
},
{
"epoch": 1.8363954505686788,
"grad_norm": 0.63671875,
"learning_rate": 3.781797712849683e-05,
"loss": 0.6327,
"step": 1050
},
{
"epoch": 1.8538932633420822,
"grad_norm": 0.66015625,
"learning_rate": 3.7768992429016796e-05,
"loss": 0.6392,
"step": 1060
},
{
"epoch": 1.8713910761154855,
"grad_norm": 0.546875,
"learning_rate": 3.7719500275604074e-05,
"loss": 0.6359,
"step": 1070
},
{
"epoch": 1.8888888888888888,
"grad_norm": 0.6171875,
"learning_rate": 3.7669502260940495e-05,
"loss": 0.6156,
"step": 1080
},
{
"epoch": 1.9063867016622922,
"grad_norm": 0.5390625,
"learning_rate": 3.761899999398677e-05,
"loss": 0.6142,
"step": 1090
},
{
"epoch": 1.9238845144356955,
"grad_norm": 0.62890625,
"learning_rate": 3.756799509993069e-05,
"loss": 0.6282,
"step": 1100
},
{
"epoch": 1.9413823272090989,
"grad_norm": 0.63671875,
"learning_rate": 3.751648922013482e-05,
"loss": 0.6388,
"step": 1110
},
{
"epoch": 1.9588801399825022,
"grad_norm": 0.578125,
"learning_rate": 3.746448401208373e-05,
"loss": 0.612,
"step": 1120
},
{
"epoch": 1.9763779527559056,
"grad_norm": 0.54296875,
"learning_rate": 3.7411981149330575e-05,
"loss": 0.622,
"step": 1130
},
{
"epoch": 1.993875765529309,
"grad_norm": 0.625,
"learning_rate": 3.735898232144331e-05,
"loss": 0.6204,
"step": 1140
},
{
"epoch": 2.010498687664042,
"grad_norm": 0.609375,
"learning_rate": 3.7305489233950284e-05,
"loss": 0.5708,
"step": 1150
},
{
"epoch": 2.027996500437445,
"grad_norm": 0.6953125,
"learning_rate": 3.725150360828537e-05,
"loss": 0.5258,
"step": 1160
},
{
"epoch": 2.0454943132108485,
"grad_norm": 0.609375,
"learning_rate": 3.7197027181732556e-05,
"loss": 0.5357,
"step": 1170
},
{
"epoch": 2.062992125984252,
"grad_norm": 0.57421875,
"learning_rate": 3.714206170737003e-05,
"loss": 0.5301,
"step": 1180
},
{
"epoch": 2.080489938757655,
"grad_norm": 0.640625,
"learning_rate": 3.708660895401381e-05,
"loss": 0.5334,
"step": 1190
},
{
"epoch": 2.0979877515310585,
"grad_norm": 0.6015625,
"learning_rate": 3.703067070616076e-05,
"loss": 0.5308,
"step": 1200
},
{
"epoch": 2.115485564304462,
"grad_norm": 0.53515625,
"learning_rate": 3.697424876393122e-05,
"loss": 0.5243,
"step": 1210
},
{
"epoch": 2.1329833770778652,
"grad_norm": 0.6171875,
"learning_rate": 3.6917344943011043e-05,
"loss": 0.5139,
"step": 1220
},
{
"epoch": 2.1504811898512686,
"grad_norm": 0.6171875,
"learning_rate": 3.685996107459317e-05,
"loss": 0.533,
"step": 1230
},
{
"epoch": 2.167979002624672,
"grad_norm": 0.65625,
"learning_rate": 3.680209900531872e-05,
"loss": 0.5098,
"step": 1240
},
{
"epoch": 2.1854768153980753,
"grad_norm": 0.6015625,
"learning_rate": 3.6743760597217536e-05,
"loss": 0.5318,
"step": 1250
},
{
"epoch": 2.2029746281714786,
"grad_norm": 0.59375,
"learning_rate": 3.668494772764828e-05,
"loss": 0.5223,
"step": 1260
},
{
"epoch": 2.220472440944882,
"grad_norm": 0.59765625,
"learning_rate": 3.662566228923805e-05,
"loss": 0.5241,
"step": 1270
},
{
"epoch": 2.2379702537182853,
"grad_norm": 0.58984375,
"learning_rate": 3.656590618982138e-05,
"loss": 0.5249,
"step": 1280
},
{
"epoch": 2.2554680664916886,
"grad_norm": 0.6328125,
"learning_rate": 3.650568135237897e-05,
"loss": 0.5412,
"step": 1290
},
{
"epoch": 2.272965879265092,
"grad_norm": 0.66796875,
"learning_rate": 3.644498971497569e-05,
"loss": 0.53,
"step": 1300
},
{
"epoch": 2.2904636920384953,
"grad_norm": 0.67578125,
"learning_rate": 3.63838332306983e-05,
"loss": 0.5382,
"step": 1310
},
{
"epoch": 2.3079615048118987,
"grad_norm": 0.61328125,
"learning_rate": 3.6322213867592536e-05,
"loss": 0.5278,
"step": 1320
},
{
"epoch": 2.325459317585302,
"grad_norm": 0.98828125,
"learning_rate": 3.626013360859982e-05,
"loss": 0.5409,
"step": 1330
},
{
"epoch": 2.3429571303587053,
"grad_norm": 0.64453125,
"learning_rate": 3.619759445149341e-05,
"loss": 0.5277,
"step": 1340
},
{
"epoch": 2.3604549431321082,
"grad_norm": 0.55078125,
"learning_rate": 3.613459840881415e-05,
"loss": 0.5344,
"step": 1350
},
{
"epoch": 2.377952755905512,
"grad_norm": 0.6328125,
"learning_rate": 3.607114750780568e-05,
"loss": 0.5364,
"step": 1360
},
{
"epoch": 2.395450568678915,
"grad_norm": 0.6171875,
"learning_rate": 3.6007243790349205e-05,
"loss": 0.5238,
"step": 1370
},
{
"epoch": 2.4129483814523183,
"grad_norm": 0.57421875,
"learning_rate": 3.594288931289778e-05,
"loss": 0.5322,
"step": 1380
},
{
"epoch": 2.4304461942257216,
"grad_norm": 0.59765625,
"learning_rate": 3.587808614641015e-05,
"loss": 0.5418,
"step": 1390
},
{
"epoch": 2.447944006999125,
"grad_norm": 0.61328125,
"learning_rate": 3.581283637628409e-05,
"loss": 0.537,
"step": 1400
},
{
"epoch": 2.4654418197725283,
"grad_norm": 0.609375,
"learning_rate": 3.574714210228929e-05,
"loss": 0.5286,
"step": 1410
},
{
"epoch": 2.4829396325459316,
"grad_norm": 0.61328125,
"learning_rate": 3.5681005438499795e-05,
"loss": 0.5196,
"step": 1420
},
{
"epoch": 2.500437445319335,
"grad_norm": 0.62109375,
"learning_rate": 3.561442851322599e-05,
"loss": 0.5559,
"step": 1430
},
{
"epoch": 2.5179352580927383,
"grad_norm": 0.59375,
"learning_rate": 3.554741346894606e-05,
"loss": 0.5325,
"step": 1440
},
{
"epoch": 2.5354330708661417,
"grad_norm": 0.578125,
"learning_rate": 3.5479962462237116e-05,
"loss": 0.5368,
"step": 1450
},
{
"epoch": 2.552930883639545,
"grad_norm": 0.59375,
"learning_rate": 3.541207766370572e-05,
"loss": 0.5368,
"step": 1460
},
{
"epoch": 2.5704286964129484,
"grad_norm": 0.73046875,
"learning_rate": 3.534376125791807e-05,
"loss": 0.5385,
"step": 1470
},
{
"epoch": 2.5879265091863517,
"grad_norm": 0.54296875,
"learning_rate": 3.527501544332972e-05,
"loss": 0.5365,
"step": 1480
},
{
"epoch": 2.605424321959755,
"grad_norm": 0.6484375,
"learning_rate": 3.520584243221479e-05,
"loss": 0.5439,
"step": 1490
},
{
"epoch": 2.6229221347331584,
"grad_norm": 0.5390625,
"learning_rate": 3.5136244450594814e-05,
"loss": 0.5214,
"step": 1500
},
{
"epoch": 2.6404199475065617,
"grad_norm": 0.6484375,
"learning_rate": 3.506622373816706e-05,
"loss": 0.5334,
"step": 1510
},
{
"epoch": 2.657917760279965,
"grad_norm": 0.55859375,
"learning_rate": 3.499578254823249e-05,
"loss": 0.544,
"step": 1520
},
{
"epoch": 2.6754155730533684,
"grad_norm": 0.54296875,
"learning_rate": 3.4924923147623265e-05,
"loss": 0.5254,
"step": 1530
},
{
"epoch": 2.6929133858267718,
"grad_norm": 0.71484375,
"learning_rate": 3.485364781662974e-05,
"loss": 0.536,
"step": 1540
},
{
"epoch": 2.710411198600175,
"grad_norm": 0.66796875,
"learning_rate": 3.478195884892712e-05,
"loss": 0.5471,
"step": 1550
},
{
"epoch": 2.7279090113735784,
"grad_norm": 0.61328125,
"learning_rate": 3.4709858551501664e-05,
"loss": 0.5235,
"step": 1560
},
{
"epoch": 2.745406824146982,
"grad_norm": 0.58203125,
"learning_rate": 3.463734924457638e-05,
"loss": 0.5486,
"step": 1570
},
{
"epoch": 2.7629046369203847,
"grad_norm": 0.5625,
"learning_rate": 3.4564433261536446e-05,
"loss": 0.5218,
"step": 1580
},
{
"epoch": 2.7804024496937885,
"grad_norm": 0.5703125,
"learning_rate": 3.449111294885407e-05,
"loss": 0.5256,
"step": 1590
},
{
"epoch": 2.7979002624671914,
"grad_norm": 0.62109375,
"learning_rate": 3.4417390666012966e-05,
"loss": 0.5296,
"step": 1600
},
{
"epoch": 2.815398075240595,
"grad_norm": 0.57421875,
"learning_rate": 3.434326878543246e-05,
"loss": 0.5365,
"step": 1610
},
{
"epoch": 2.832895888013998,
"grad_norm": 0.58984375,
"learning_rate": 3.426874969239113e-05,
"loss": 0.5226,
"step": 1620
},
{
"epoch": 2.850393700787402,
"grad_norm": 0.61328125,
"learning_rate": 3.419383578495005e-05,
"loss": 0.5351,
"step": 1630
},
{
"epoch": 2.8678915135608047,
"grad_norm": 0.625,
"learning_rate": 3.411852947387559e-05,
"loss": 0.5297,
"step": 1640
},
{
"epoch": 2.885389326334208,
"grad_norm": 0.66015625,
"learning_rate": 3.40428331825619e-05,
"loss": 0.5248,
"step": 1650
},
{
"epoch": 2.9028871391076114,
"grad_norm": 0.60546875,
"learning_rate": 3.3966749346952856e-05,
"loss": 0.5327,
"step": 1660
},
{
"epoch": 2.9203849518810148,
"grad_norm": 0.5546875,
"learning_rate": 3.389028041546372e-05,
"loss": 0.5315,
"step": 1670
},
{
"epoch": 2.937882764654418,
"grad_norm": 0.6015625,
"learning_rate": 3.381342884890231e-05,
"loss": 0.5327,
"step": 1680
},
{
"epoch": 2.9553805774278215,
"grad_norm": 0.60546875,
"learning_rate": 3.373619712038985e-05,
"loss": 0.5399,
"step": 1690
},
{
"epoch": 2.972878390201225,
"grad_norm": 0.60546875,
"learning_rate": 3.365858771528135e-05,
"loss": 0.5307,
"step": 1700
},
{
"epoch": 2.990376202974628,
"grad_norm": 0.5859375,
"learning_rate": 3.358060313108564e-05,
"loss": 0.5427,
"step": 1710
},
{
"epoch": 3.0069991251093615,
"grad_norm": 0.65234375,
"learning_rate": 3.3502245877384986e-05,
"loss": 0.4888,
"step": 1720
},
{
"epoch": 3.024496937882765,
"grad_norm": 0.6640625,
"learning_rate": 3.3423518475754376e-05,
"loss": 0.4435,
"step": 1730
},
{
"epoch": 3.041994750656168,
"grad_norm": 0.71875,
"learning_rate": 3.33444234596803e-05,
"loss": 0.429,
"step": 1740
},
{
"epoch": 3.059492563429571,
"grad_norm": 0.67578125,
"learning_rate": 3.326496337447928e-05,
"loss": 0.437,
"step": 1750
},
{
"epoch": 3.0769903762029744,
"grad_norm": 0.65625,
"learning_rate": 3.3185140777215956e-05,
"loss": 0.4387,
"step": 1760
},
{
"epoch": 3.094488188976378,
"grad_norm": 0.7109375,
"learning_rate": 3.3104958236620755e-05,
"loss": 0.4211,
"step": 1770
},
{
"epoch": 3.111986001749781,
"grad_norm": 0.58203125,
"learning_rate": 3.302441833300728e-05,
"loss": 0.4265,
"step": 1780
},
{
"epoch": 3.1294838145231845,
"grad_norm": 0.640625,
"learning_rate": 3.2943523658189246e-05,
"loss": 0.438,
"step": 1790
},
{
"epoch": 3.146981627296588,
"grad_norm": 0.6171875,
"learning_rate": 3.286227681539708e-05,
"loss": 0.4379,
"step": 1800
},
{
"epoch": 3.164479440069991,
"grad_norm": 0.7109375,
"learning_rate": 3.278068041919415e-05,
"loss": 0.4284,
"step": 1810
},
{
"epoch": 3.1819772528433945,
"grad_norm": 0.65234375,
"learning_rate": 3.2698737095392593e-05,
"loss": 0.4299,
"step": 1820
},
{
"epoch": 3.199475065616798,
"grad_norm": 0.61328125,
"learning_rate": 3.261644948096891e-05,
"loss": 0.4351,
"step": 1830
},
{
"epoch": 3.216972878390201,
"grad_norm": 0.62890625,
"learning_rate": 3.2533820223978996e-05,
"loss": 0.4336,
"step": 1840
},
{
"epoch": 3.2344706911636045,
"grad_norm": 0.63671875,
"learning_rate": 3.245085198347298e-05,
"loss": 0.4376,
"step": 1850
},
{
"epoch": 3.251968503937008,
"grad_norm": 0.56640625,
"learning_rate": 3.236754742940965e-05,
"loss": 0.4461,
"step": 1860
},
{
"epoch": 3.269466316710411,
"grad_norm": 0.6484375,
"learning_rate": 3.228390924257053e-05,
"loss": 0.4472,
"step": 1870
},
{
"epoch": 3.2869641294838146,
"grad_norm": 0.68359375,
"learning_rate": 3.2199940114473615e-05,
"loss": 0.4404,
"step": 1880
},
{
"epoch": 3.304461942257218,
"grad_norm": 0.671875,
"learning_rate": 3.211564274728676e-05,
"loss": 0.4357,
"step": 1890
},
{
"epoch": 3.3219597550306212,
"grad_norm": 0.640625,
"learning_rate": 3.203101985374069e-05,
"loss": 0.4405,
"step": 1900
},
{
"epoch": 3.3394575678040246,
"grad_norm": 0.6640625,
"learning_rate": 3.194607415704174e-05,
"loss": 0.4524,
"step": 1910
},
{
"epoch": 3.356955380577428,
"grad_norm": 0.6328125,
"learning_rate": 3.186080839078423e-05,
"loss": 0.4311,
"step": 1920
},
{
"epoch": 3.3744531933508313,
"grad_norm": 0.6640625,
"learning_rate": 3.177522529886244e-05,
"loss": 0.4257,
"step": 1930
},
{
"epoch": 3.3919510061242346,
"grad_norm": 0.87109375,
"learning_rate": 3.1689327635382374e-05,
"loss": 0.4305,
"step": 1940
},
{
"epoch": 3.409448818897638,
"grad_norm": 0.71875,
"learning_rate": 3.160311816457309e-05,
"loss": 0.4473,
"step": 1950
},
{
"epoch": 3.4269466316710413,
"grad_norm": 0.63671875,
"learning_rate": 3.151659966069777e-05,
"loss": 0.4543,
"step": 1960
},
{
"epoch": 3.4444444444444446,
"grad_norm": 0.59765625,
"learning_rate": 3.142977490796441e-05,
"loss": 0.4444,
"step": 1970
},
{
"epoch": 3.4619422572178475,
"grad_norm": 0.65625,
"learning_rate": 3.134264670043627e-05,
"loss": 0.4543,
"step": 1980
},
{
"epoch": 3.4794400699912513,
"grad_norm": 0.6484375,
"learning_rate": 3.125521784194192e-05,
"loss": 0.4433,
"step": 1990
},
{
"epoch": 3.4969378827646542,
"grad_norm": 0.625,
"learning_rate": 3.1167491145985035e-05,
"loss": 0.4403,
"step": 2000
},
{
"epoch": 3.514435695538058,
"grad_norm": 0.62109375,
"learning_rate": 3.107946943565384e-05,
"loss": 0.4516,
"step": 2010
},
{
"epoch": 3.531933508311461,
"grad_norm": 0.66015625,
"learning_rate": 3.099115554353027e-05,
"loss": 0.4326,
"step": 2020
},
{
"epoch": 3.5494313210848643,
"grad_norm": 0.61328125,
"learning_rate": 3.090255231159881e-05,
"loss": 0.4309,
"step": 2030
},
{
"epoch": 3.5669291338582676,
"grad_norm": 0.6171875,
"learning_rate": 3.081366259115504e-05,
"loss": 0.4328,
"step": 2040
},
{
"epoch": 3.584426946631671,
"grad_norm": 0.6484375,
"learning_rate": 3.072448924271389e-05,
"loss": 0.4333,
"step": 2050
},
{
"epoch": 3.6019247594050743,
"grad_norm": 0.6171875,
"learning_rate": 3.063503513591758e-05,
"loss": 0.4518,
"step": 2060
},
{
"epoch": 3.6194225721784776,
"grad_norm": 0.65234375,
"learning_rate": 3.0545303149443265e-05,
"loss": 0.4383,
"step": 2070
},
{
"epoch": 3.636920384951881,
"grad_norm": 0.6484375,
"learning_rate": 3.0455296170910404e-05,
"loss": 0.436,
"step": 2080
},
{
"epoch": 3.6544181977252843,
"grad_norm": 0.671875,
"learning_rate": 3.0365017096787834e-05,
"loss": 0.4363,
"step": 2090
},
{
"epoch": 3.6719160104986877,
"grad_norm": 0.6484375,
"learning_rate": 3.0274468832300576e-05,
"loss": 0.4463,
"step": 2100
},
{
"epoch": 3.689413823272091,
"grad_norm": 0.609375,
"learning_rate": 3.0183654291336313e-05,
"loss": 0.4533,
"step": 2110
},
{
"epoch": 3.7069116360454943,
"grad_norm": 0.6796875,
"learning_rate": 3.009257639635164e-05,
"loss": 0.4255,
"step": 2120
},
{
"epoch": 3.7244094488188977,
"grad_norm": 0.65234375,
"learning_rate": 3.0001238078278022e-05,
"loss": 0.4452,
"step": 2130
},
{
"epoch": 3.741907261592301,
"grad_norm": 0.625,
"learning_rate": 2.990964227642746e-05,
"loss": 0.4344,
"step": 2140
},
{
"epoch": 3.7594050743657044,
"grad_norm": 0.70703125,
"learning_rate": 2.9817791938397902e-05,
"loss": 0.4327,
"step": 2150
},
{
"epoch": 3.7769028871391077,
"grad_norm": 0.66015625,
"learning_rate": 2.9725690019978422e-05,
"loss": 0.4461,
"step": 2160
},
{
"epoch": 3.794400699912511,
"grad_norm": 0.6640625,
"learning_rate": 2.9633339485054037e-05,
"loss": 0.443,
"step": 2170
},
{
"epoch": 3.8118985126859144,
"grad_norm": 0.6171875,
"learning_rate": 2.9540743305510406e-05,
"loss": 0.4387,
"step": 2180
},
{
"epoch": 3.8293963254593177,
"grad_norm": 0.703125,
"learning_rate": 2.9447904461138114e-05,
"loss": 0.4355,
"step": 2190
},
{
"epoch": 3.846894138232721,
"grad_norm": 0.625,
"learning_rate": 2.9354825939536854e-05,
"loss": 0.4425,
"step": 2200
},
{
"epoch": 3.864391951006124,
"grad_norm": 0.6171875,
"learning_rate": 2.9261510736019222e-05,
"loss": 0.4518,
"step": 2210
},
{
"epoch": 3.8818897637795278,
"grad_norm": 0.6171875,
"learning_rate": 2.9167961853514368e-05,
"loss": 0.4552,
"step": 2220
},
{
"epoch": 3.8993875765529307,
"grad_norm": 0.71875,
"learning_rate": 2.9074182302471347e-05,
"loss": 0.4506,
"step": 2230
},
{
"epoch": 3.9168853893263345,
"grad_norm": 0.703125,
"learning_rate": 2.898017510076224e-05,
"loss": 0.4388,
"step": 2240
},
{
"epoch": 3.9343832020997374,
"grad_norm": 0.671875,
"learning_rate": 2.8885943273585037e-05,
"loss": 0.4373,
"step": 2250
},
{
"epoch": 3.9518810148731407,
"grad_norm": 0.61328125,
"learning_rate": 2.8791489853366284e-05,
"loss": 0.4359,
"step": 2260
},
{
"epoch": 3.969378827646544,
"grad_norm": 0.69140625,
"learning_rate": 2.8696817879663515e-05,
"loss": 0.4538,
"step": 2270
},
{
"epoch": 3.9868766404199474,
"grad_norm": 0.73046875,
"learning_rate": 2.8601930399067393e-05,
"loss": 0.4394,
"step": 2280
},
{
"epoch": 4.003499562554681,
"grad_norm": 0.62109375,
"learning_rate": 2.8506830465103723e-05,
"loss": 0.4332,
"step": 2290
},
{
"epoch": 4.020997375328084,
"grad_norm": 0.76953125,
"learning_rate": 2.8411521138135154e-05,
"loss": 0.3484,
"step": 2300
},
{
"epoch": 4.038495188101487,
"grad_norm": 0.6015625,
"learning_rate": 2.8316005485262717e-05,
"loss": 0.3652,
"step": 2310
},
{
"epoch": 4.05599300087489,
"grad_norm": 0.76171875,
"learning_rate": 2.8220286580227093e-05,
"loss": 0.3603,
"step": 2320
},
{
"epoch": 4.073490813648294,
"grad_norm": 0.69921875,
"learning_rate": 2.8124367503309736e-05,
"loss": 0.3401,
"step": 2330
},
{
"epoch": 4.090988626421697,
"grad_norm": 0.71875,
"learning_rate": 2.802825134123371e-05,
"loss": 0.3446,
"step": 2340
},
{
"epoch": 4.108486439195101,
"grad_norm": 0.7265625,
"learning_rate": 2.7931941187064402e-05,
"loss": 0.3732,
"step": 2350
},
{
"epoch": 4.125984251968504,
"grad_norm": 0.66015625,
"learning_rate": 2.783544014010993e-05,
"loss": 0.3567,
"step": 2360
},
{
"epoch": 4.1434820647419075,
"grad_norm": 0.765625,
"learning_rate": 2.773875130582146e-05,
"loss": 0.3575,
"step": 2370
},
{
"epoch": 4.16097987751531,
"grad_norm": 0.7265625,
"learning_rate": 2.764187779569324e-05,
"loss": 0.36,
"step": 2380
},
{
"epoch": 4.178477690288714,
"grad_norm": 0.6875,
"learning_rate": 2.7544822727162475e-05,
"loss": 0.345,
"step": 2390
},
{
"epoch": 4.195975503062117,
"grad_norm": 0.6953125,
"learning_rate": 2.7447589223509004e-05,
"loss": 0.3614,
"step": 2400
},
{
"epoch": 4.213473315835521,
"grad_norm": 0.6171875,
"learning_rate": 2.7350180413754798e-05,
"loss": 0.3584,
"step": 2410
},
{
"epoch": 4.230971128608924,
"grad_norm": 0.68359375,
"learning_rate": 2.7252599432563278e-05,
"loss": 0.3567,
"step": 2420
},
{
"epoch": 4.248468941382328,
"grad_norm": 0.66796875,
"learning_rate": 2.7154849420138405e-05,
"loss": 0.3513,
"step": 2430
},
{
"epoch": 4.2659667541557305,
"grad_norm": 0.75390625,
"learning_rate": 2.7056933522123677e-05,
"loss": 0.3572,
"step": 2440
},
{
"epoch": 4.283464566929134,
"grad_norm": 0.69921875,
"learning_rate": 2.695885488950083e-05,
"loss": 0.3461,
"step": 2450
},
{
"epoch": 4.300962379702537,
"grad_norm": 0.76171875,
"learning_rate": 2.686061667848852e-05,
"loss": 0.3534,
"step": 2460
},
{
"epoch": 4.318460192475941,
"grad_norm": 0.7578125,
"learning_rate": 2.67622220504407e-05,
"loss": 0.3608,
"step": 2470
},
{
"epoch": 4.335958005249344,
"grad_norm": 0.7578125,
"learning_rate": 2.6663674171744887e-05,
"loss": 0.3549,
"step": 2480
},
{
"epoch": 4.353455818022747,
"grad_norm": 0.69921875,
"learning_rate": 2.6564976213720292e-05,
"loss": 0.3533,
"step": 2490
},
{
"epoch": 4.3709536307961505,
"grad_norm": 0.69140625,
"learning_rate": 2.646613135251576e-05,
"loss": 0.3542,
"step": 2500
},
{
"epoch": 4.388451443569553,
"grad_norm": 0.859375,
"learning_rate": 2.6367142769007533e-05,
"loss": 0.3426,
"step": 2510
},
{
"epoch": 4.405949256342957,
"grad_norm": 0.66796875,
"learning_rate": 2.626801364869693e-05,
"loss": 0.3511,
"step": 2520
},
{
"epoch": 4.42344706911636,
"grad_norm": 0.72265625,
"learning_rate": 2.6168747181607795e-05,
"loss": 0.3528,
"step": 2530
},
{
"epoch": 4.440944881889764,
"grad_norm": 0.73046875,
"learning_rate": 2.606934656218386e-05,
"loss": 0.3491,
"step": 2540
},
{
"epoch": 4.458442694663167,
"grad_norm": 0.7734375,
"learning_rate": 2.5969814989185965e-05,
"loss": 0.3563,
"step": 2550
},
{
"epoch": 4.475940507436571,
"grad_norm": 0.71875,
"learning_rate": 2.5870155665589066e-05,
"loss": 0.3603,
"step": 2560
},
{
"epoch": 4.4934383202099735,
"grad_norm": 0.7265625,
"learning_rate": 2.5770371798479234e-05,
"loss": 0.3525,
"step": 2570
},
{
"epoch": 4.510936132983377,
"grad_norm": 0.734375,
"learning_rate": 2.5670466598950385e-05,
"loss": 0.3487,
"step": 2580
},
{
"epoch": 4.52843394575678,
"grad_norm": 0.71875,
"learning_rate": 2.5570443282000974e-05,
"loss": 0.3606,
"step": 2590
},
{
"epoch": 4.545931758530184,
"grad_norm": 0.73046875,
"learning_rate": 2.547030506643055e-05,
"loss": 0.3569,
"step": 2600
},
{
"epoch": 4.563429571303587,
"grad_norm": 0.74609375,
"learning_rate": 2.5370055174736144e-05,
"loss": 0.3509,
"step": 2610
},
{
"epoch": 4.580927384076991,
"grad_norm": 0.6328125,
"learning_rate": 2.5269696833008577e-05,
"loss": 0.3549,
"step": 2620
},
{
"epoch": 4.5984251968503935,
"grad_norm": 0.7265625,
"learning_rate": 2.516923327082865e-05,
"loss": 0.3521,
"step": 2630
},
{
"epoch": 4.615923009623797,
"grad_norm": 0.671875,
"learning_rate": 2.5068667721163213e-05,
"loss": 0.3549,
"step": 2640
},
{
"epoch": 4.6334208223972,
"grad_norm": 0.7734375,
"learning_rate": 2.496800342026112e-05,
"loss": 0.3448,
"step": 2650
},
{
"epoch": 4.650918635170604,
"grad_norm": 0.6640625,
"learning_rate": 2.4867243607549085e-05,
"loss": 0.3517,
"step": 2660
},
{
"epoch": 4.668416447944007,
"grad_norm": 0.6953125,
"learning_rate": 2.4766391525527436e-05,
"loss": 0.3548,
"step": 2670
},
{
"epoch": 4.685914260717411,
"grad_norm": 0.65234375,
"learning_rate": 2.4665450419665783e-05,
"loss": 0.3605,
"step": 2680
},
{
"epoch": 4.703412073490814,
"grad_norm": 0.80078125,
"learning_rate": 2.4564423538298567e-05,
"loss": 0.3565,
"step": 2690
},
{
"epoch": 4.7209098862642165,
"grad_norm": 0.69921875,
"learning_rate": 2.4463314132520522e-05,
"loss": 0.344,
"step": 2700
},
{
"epoch": 4.73840769903762,
"grad_norm": 0.68359375,
"learning_rate": 2.4362125456082058e-05,
"loss": 0.3496,
"step": 2710
},
{
"epoch": 4.755905511811024,
"grad_norm": 0.8125,
"learning_rate": 2.426086076528455e-05,
"loss": 0.3584,
"step": 2720
},
{
"epoch": 4.773403324584427,
"grad_norm": 0.6875,
"learning_rate": 2.4159523318875562e-05,
"loss": 0.3627,
"step": 2730
},
{
"epoch": 4.79090113735783,
"grad_norm": 0.7734375,
"learning_rate": 2.405811637794396e-05,
"loss": 0.3585,
"step": 2740
},
{
"epoch": 4.808398950131234,
"grad_norm": 0.73046875,
"learning_rate": 2.3956643205814994e-05,
"loss": 0.3651,
"step": 2750
},
{
"epoch": 4.8258967629046365,
"grad_norm": 0.66796875,
"learning_rate": 2.385510706794524e-05,
"loss": 0.3537,
"step": 2760
},
{
"epoch": 4.84339457567804,
"grad_norm": 0.7734375,
"learning_rate": 2.375351123181756e-05,
"loss": 0.3476,
"step": 2770
},
{
"epoch": 4.860892388451443,
"grad_norm": 0.75,
"learning_rate": 2.365185896683595e-05,
"loss": 0.3416,
"step": 2780
},
{
"epoch": 4.878390201224847,
"grad_norm": 0.7578125,
"learning_rate": 2.3550153544220273e-05,
"loss": 0.3569,
"step": 2790
},
{
"epoch": 4.89588801399825,
"grad_norm": 0.69921875,
"learning_rate": 2.3448398236901073e-05,
"loss": 0.3502,
"step": 2800
},
{
"epoch": 4.913385826771654,
"grad_norm": 0.77734375,
"learning_rate": 2.3346596319414173e-05,
"loss": 0.3538,
"step": 2810
},
{
"epoch": 4.930883639545057,
"grad_norm": 0.67578125,
"learning_rate": 2.3244751067795366e-05,
"loss": 0.3635,
"step": 2820
},
{
"epoch": 4.94838145231846,
"grad_norm": 0.71484375,
"learning_rate": 2.314286575947494e-05,
"loss": 0.3567,
"step": 2830
},
{
"epoch": 4.965879265091863,
"grad_norm": 0.8515625,
"learning_rate": 2.304094367317223e-05,
"loss": 0.3293,
"step": 2840
},
{
"epoch": 4.983377077865267,
"grad_norm": 0.7421875,
"learning_rate": 2.293898808879011e-05,
"loss": 0.3656,
"step": 2850
},
{
"epoch": 5.0,
"grad_norm": 1.0546875,
"learning_rate": 2.2837002287309426e-05,
"loss": 0.3726,
"step": 2860
}
],
"logging_steps": 10,
"max_steps": 5710,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 7.691223286459776e+18,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}