{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 2860, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.017497812773403325, "grad_norm": 4.1875, "learning_rate": 2.0930232558139536e-06, "loss": 1.0581, "step": 10 }, { "epoch": 0.03499562554680665, "grad_norm": 3.328125, "learning_rate": 4.418604651162791e-06, "loss": 1.0503, "step": 20 }, { "epoch": 0.05249343832020997, "grad_norm": 2.71875, "learning_rate": 6.744186046511628e-06, "loss": 0.9804, "step": 30 }, { "epoch": 0.0699912510936133, "grad_norm": 1.4296875, "learning_rate": 9.069767441860465e-06, "loss": 0.8786, "step": 40 }, { "epoch": 0.08748906386701662, "grad_norm": 1.0, "learning_rate": 1.1395348837209304e-05, "loss": 0.8081, "step": 50 }, { "epoch": 0.10498687664041995, "grad_norm": 0.8203125, "learning_rate": 1.372093023255814e-05, "loss": 0.8047, "step": 60 }, { "epoch": 0.12248468941382328, "grad_norm": 0.765625, "learning_rate": 1.6046511627906977e-05, "loss": 0.7453, "step": 70 }, { "epoch": 0.1399825021872266, "grad_norm": 0.8125, "learning_rate": 1.8372093023255815e-05, "loss": 0.7581, "step": 80 }, { "epoch": 0.15748031496062992, "grad_norm": 0.734375, "learning_rate": 2.0697674418604654e-05, "loss": 0.7399, "step": 90 }, { "epoch": 0.17497812773403323, "grad_norm": 0.71875, "learning_rate": 2.3023255813953492e-05, "loss": 0.7476, "step": 100 }, { "epoch": 0.19247594050743658, "grad_norm": 0.7109375, "learning_rate": 2.5348837209302327e-05, "loss": 0.7247, "step": 110 }, { "epoch": 0.2099737532808399, "grad_norm": 0.81640625, "learning_rate": 2.7674418604651166e-05, "loss": 0.7179, "step": 120 }, { "epoch": 0.2274715660542432, "grad_norm": 0.77734375, "learning_rate": 3.0000000000000004e-05, "loss": 0.6949, "step": 130 }, { "epoch": 0.24496937882764655, "grad_norm": 0.82421875, "learning_rate": 3.232558139534884e-05, "loss": 0.7043, "step": 140 }, { "epoch": 0.26246719160104987, "grad_norm": 0.69921875, "learning_rate": 3.4651162790697674e-05, "loss": 0.6934, "step": 150 }, { "epoch": 0.2799650043744532, "grad_norm": 0.6796875, "learning_rate": 3.697674418604651e-05, "loss": 0.6993, "step": 160 }, { "epoch": 0.2974628171478565, "grad_norm": 0.66015625, "learning_rate": 3.930232558139535e-05, "loss": 0.7027, "step": 170 }, { "epoch": 0.31496062992125984, "grad_norm": 0.66015625, "learning_rate": 3.9999858083836754e-05, "loss": 0.6935, "step": 180 }, { "epoch": 0.3324584426946632, "grad_norm": 0.6875, "learning_rate": 3.999916298964876e-05, "loss": 0.7137, "step": 190 }, { "epoch": 0.34995625546806647, "grad_norm": 0.7578125, "learning_rate": 3.999788867354247e-05, "loss": 0.6916, "step": 200 }, { "epoch": 0.3674540682414698, "grad_norm": 0.69921875, "learning_rate": 3.9996035176526014e-05, "loss": 0.6947, "step": 210 }, { "epoch": 0.38495188101487315, "grad_norm": 0.77734375, "learning_rate": 3.9993602558245834e-05, "loss": 0.7045, "step": 220 }, { "epoch": 0.40244969378827644, "grad_norm": 0.6484375, "learning_rate": 3.9990590896984766e-05, "loss": 0.7016, "step": 230 }, { "epoch": 0.4199475065616798, "grad_norm": 0.6328125, "learning_rate": 3.998700028965957e-05, "loss": 0.7072, "step": 240 }, { "epoch": 0.4374453193350831, "grad_norm": 0.65234375, "learning_rate": 3.9982830851817736e-05, "loss": 0.6986, "step": 250 }, { "epoch": 0.4549431321084864, "grad_norm": 0.66015625, "learning_rate": 3.997808271763385e-05, "loss": 0.6804, "step": 260 }, { "epoch": 0.47244094488188976, "grad_norm": 0.6640625, "learning_rate": 3.997275603990519e-05, "loss": 0.7016, "step": 270 }, { "epoch": 0.4899387576552931, "grad_norm": 0.703125, "learning_rate": 3.996685099004686e-05, "loss": 0.6921, "step": 280 }, { "epoch": 0.5074365704286964, "grad_norm": 0.6875, "learning_rate": 3.996036775808628e-05, "loss": 0.6857, "step": 290 }, { "epoch": 0.5249343832020997, "grad_norm": 0.6875, "learning_rate": 3.995330655265704e-05, "loss": 0.6967, "step": 300 }, { "epoch": 0.5424321959755031, "grad_norm": 0.72265625, "learning_rate": 3.994566760099221e-05, "loss": 0.6969, "step": 310 }, { "epoch": 0.5599300087489064, "grad_norm": 0.59765625, "learning_rate": 3.9937451148917e-05, "loss": 0.6862, "step": 320 }, { "epoch": 0.5774278215223098, "grad_norm": 0.625, "learning_rate": 3.992865746084089e-05, "loss": 0.695, "step": 330 }, { "epoch": 0.594925634295713, "grad_norm": 0.62109375, "learning_rate": 3.991928681974908e-05, "loss": 0.7036, "step": 340 }, { "epoch": 0.6124234470691163, "grad_norm": 0.734375, "learning_rate": 3.9909339527193416e-05, "loss": 0.7064, "step": 350 }, { "epoch": 0.6299212598425197, "grad_norm": 0.6171875, "learning_rate": 3.989881590328265e-05, "loss": 0.6768, "step": 360 }, { "epoch": 0.647419072615923, "grad_norm": 0.6640625, "learning_rate": 3.98877162866722e-05, "loss": 0.7161, "step": 370 }, { "epoch": 0.6649168853893264, "grad_norm": 0.625, "learning_rate": 3.9876041034553165e-05, "loss": 0.6731, "step": 380 }, { "epoch": 0.6824146981627297, "grad_norm": 0.65625, "learning_rate": 3.9863790522640926e-05, "loss": 0.6905, "step": 390 }, { "epoch": 0.6999125109361329, "grad_norm": 0.66796875, "learning_rate": 3.9850965145162954e-05, "loss": 0.7067, "step": 400 }, { "epoch": 0.7174103237095363, "grad_norm": 0.6015625, "learning_rate": 3.983756531484622e-05, "loss": 0.6982, "step": 410 }, { "epoch": 0.7349081364829396, "grad_norm": 0.59375, "learning_rate": 3.9823591462903856e-05, "loss": 0.6932, "step": 420 }, { "epoch": 0.752405949256343, "grad_norm": 0.6953125, "learning_rate": 3.9809044039021295e-05, "loss": 0.7004, "step": 430 }, { "epoch": 0.7699037620297463, "grad_norm": 0.63671875, "learning_rate": 3.979392351134178e-05, "loss": 0.7002, "step": 440 }, { "epoch": 0.7874015748031497, "grad_norm": 0.63671875, "learning_rate": 3.977823036645132e-05, "loss": 0.7065, "step": 450 }, { "epoch": 0.8048993875765529, "grad_norm": 0.62109375, "learning_rate": 3.976196510936307e-05, "loss": 0.692, "step": 460 }, { "epoch": 0.8223972003499562, "grad_norm": 0.58984375, "learning_rate": 3.9745128263500976e-05, "loss": 0.6847, "step": 470 }, { "epoch": 0.8398950131233596, "grad_norm": 0.75, "learning_rate": 3.972772037068303e-05, "loss": 0.6868, "step": 480 }, { "epoch": 0.8573928258967629, "grad_norm": 0.69140625, "learning_rate": 3.970974199110378e-05, "loss": 0.6727, "step": 490 }, { "epoch": 0.8748906386701663, "grad_norm": 0.54296875, "learning_rate": 3.9691193703316336e-05, "loss": 0.69, "step": 500 }, { "epoch": 0.8923884514435696, "grad_norm": 0.578125, "learning_rate": 3.9672076104213706e-05, "loss": 0.6696, "step": 510 }, { "epoch": 0.9098862642169728, "grad_norm": 0.61328125, "learning_rate": 3.965238980900965e-05, "loss": 0.6852, "step": 520 }, { "epoch": 0.9273840769903762, "grad_norm": 0.6328125, "learning_rate": 3.963213545121881e-05, "loss": 0.6852, "step": 530 }, { "epoch": 0.9448818897637795, "grad_norm": 0.63671875, "learning_rate": 3.9611313682636395e-05, "loss": 0.6907, "step": 540 }, { "epoch": 0.9623797025371829, "grad_norm": 0.59375, "learning_rate": 3.9589925173317165e-05, "loss": 0.6753, "step": 550 }, { "epoch": 0.9798775153105862, "grad_norm": 0.63671875, "learning_rate": 3.956797061155386e-05, "loss": 0.6796, "step": 560 }, { "epoch": 0.9973753280839895, "grad_norm": 0.640625, "learning_rate": 3.954545070385508e-05, "loss": 0.6875, "step": 570 }, { "epoch": 1.0139982502187226, "grad_norm": 0.59765625, "learning_rate": 3.952236617492253e-05, "loss": 0.6204, "step": 580 }, { "epoch": 1.031496062992126, "grad_norm": 0.59765625, "learning_rate": 3.94987177676277e-05, "loss": 0.5968, "step": 590 }, { "epoch": 1.0489938757655293, "grad_norm": 0.671875, "learning_rate": 3.947450624298799e-05, "loss": 0.6147, "step": 600 }, { "epoch": 1.0664916885389326, "grad_norm": 0.65234375, "learning_rate": 3.944973238014215e-05, "loss": 0.6052, "step": 610 }, { "epoch": 1.083989501312336, "grad_norm": 0.6640625, "learning_rate": 3.942439697632528e-05, "loss": 0.6152, "step": 620 }, { "epoch": 1.1014873140857393, "grad_norm": 0.62109375, "learning_rate": 3.9398500846843136e-05, "loss": 0.622, "step": 630 }, { "epoch": 1.1189851268591426, "grad_norm": 0.578125, "learning_rate": 3.9372044825045905e-05, "loss": 0.6361, "step": 640 }, { "epoch": 1.136482939632546, "grad_norm": 0.5703125, "learning_rate": 3.934502976230138e-05, "loss": 0.6015, "step": 650 }, { "epoch": 1.1539807524059493, "grad_norm": 0.6171875, "learning_rate": 3.9317456527967566e-05, "loss": 0.6231, "step": 660 }, { "epoch": 1.1714785651793527, "grad_norm": 0.640625, "learning_rate": 3.92893260093647e-05, "loss": 0.6325, "step": 670 }, { "epoch": 1.188976377952756, "grad_norm": 0.58203125, "learning_rate": 3.926063911174673e-05, "loss": 0.6, "step": 680 }, { "epoch": 1.2064741907261591, "grad_norm": 0.58984375, "learning_rate": 3.92313967582721e-05, "loss": 0.6158, "step": 690 }, { "epoch": 1.2239720034995625, "grad_norm": 0.62109375, "learning_rate": 3.9201599889974155e-05, "loss": 0.6184, "step": 700 }, { "epoch": 1.2414698162729658, "grad_norm": 0.6015625, "learning_rate": 3.917124946573076e-05, "loss": 0.6159, "step": 710 }, { "epoch": 1.2589676290463692, "grad_norm": 0.57421875, "learning_rate": 3.91403464622335e-05, "loss": 0.6221, "step": 720 }, { "epoch": 1.2764654418197725, "grad_norm": 0.640625, "learning_rate": 3.910889187395622e-05, "loss": 0.6393, "step": 730 }, { "epoch": 1.2939632545931758, "grad_norm": 0.6484375, "learning_rate": 3.9076886713123056e-05, "loss": 0.6168, "step": 740 }, { "epoch": 1.3114610673665792, "grad_norm": 0.578125, "learning_rate": 3.9044332009675816e-05, "loss": 0.611, "step": 750 }, { "epoch": 1.3289588801399825, "grad_norm": 0.66796875, "learning_rate": 3.901122881124087e-05, "loss": 0.625, "step": 760 }, { "epoch": 1.3464566929133859, "grad_norm": 0.6640625, "learning_rate": 3.897757818309543e-05, "loss": 0.6196, "step": 770 }, { "epoch": 1.3639545056867892, "grad_norm": 0.62109375, "learning_rate": 3.8943381208133266e-05, "loss": 0.6451, "step": 780 }, { "epoch": 1.3814523184601923, "grad_norm": 0.62109375, "learning_rate": 3.890863898682984e-05, "loss": 0.6255, "step": 790 }, { "epoch": 1.3989501312335957, "grad_norm": 0.6171875, "learning_rate": 3.887335263720693e-05, "loss": 0.6182, "step": 800 }, { "epoch": 1.416447944006999, "grad_norm": 0.57421875, "learning_rate": 3.883752329479662e-05, "loss": 0.6206, "step": 810 }, { "epoch": 1.4339457567804024, "grad_norm": 0.65625, "learning_rate": 3.880115211260478e-05, "loss": 0.6068, "step": 820 }, { "epoch": 1.4514435695538057, "grad_norm": 0.52734375, "learning_rate": 3.876424026107394e-05, "loss": 0.6221, "step": 830 }, { "epoch": 1.468941382327209, "grad_norm": 0.6328125, "learning_rate": 3.872678892804563e-05, "loss": 0.6118, "step": 840 }, { "epoch": 1.4864391951006124, "grad_norm": 0.60546875, "learning_rate": 3.868879931872218e-05, "loss": 0.6137, "step": 850 }, { "epoch": 1.5039370078740157, "grad_norm": 0.58203125, "learning_rate": 3.865027265562789e-05, "loss": 0.61, "step": 860 }, { "epoch": 1.521434820647419, "grad_norm": 0.60546875, "learning_rate": 3.8611210178569746e-05, "loss": 0.6285, "step": 870 }, { "epoch": 1.5389326334208224, "grad_norm": 0.59765625, "learning_rate": 3.857161314459745e-05, "loss": 0.6211, "step": 880 }, { "epoch": 1.5564304461942258, "grad_norm": 0.58203125, "learning_rate": 3.853148282796306e-05, "loss": 0.6291, "step": 890 }, { "epoch": 1.5739282589676291, "grad_norm": 0.59375, "learning_rate": 3.849082052007988e-05, "loss": 0.6074, "step": 900 }, { "epoch": 1.5914260717410325, "grad_norm": 0.6640625, "learning_rate": 3.844962752948099e-05, "loss": 0.6314, "step": 910 }, { "epoch": 1.6089238845144358, "grad_norm": 0.5859375, "learning_rate": 3.8407905181777095e-05, "loss": 0.6167, "step": 920 }, { "epoch": 1.6264216972878391, "grad_norm": 0.6484375, "learning_rate": 3.836565481961386e-05, "loss": 0.6235, "step": 930 }, { "epoch": 1.6439195100612425, "grad_norm": 0.5859375, "learning_rate": 3.8322877802628704e-05, "loss": 0.6106, "step": 940 }, { "epoch": 1.6614173228346458, "grad_norm": 0.5546875, "learning_rate": 3.827957550740711e-05, "loss": 0.6041, "step": 950 }, { "epoch": 1.678915135608049, "grad_norm": 0.59375, "learning_rate": 3.82357493274382e-05, "loss": 0.6176, "step": 960 }, { "epoch": 1.6964129483814523, "grad_norm": 0.5625, "learning_rate": 3.8191400673070034e-05, "loss": 0.6105, "step": 970 }, { "epoch": 1.7139107611548556, "grad_norm": 0.55859375, "learning_rate": 3.81465309714641e-05, "loss": 0.6278, "step": 980 }, { "epoch": 1.731408573928259, "grad_norm": 0.6484375, "learning_rate": 3.810114166654948e-05, "loss": 0.6189, "step": 990 }, { "epoch": 1.7489063867016623, "grad_norm": 0.66796875, "learning_rate": 3.805523421897633e-05, "loss": 0.6299, "step": 1000 }, { "epoch": 1.7664041994750657, "grad_norm": 0.62109375, "learning_rate": 3.800881010606889e-05, "loss": 0.6059, "step": 1010 }, { "epoch": 1.7839020122484688, "grad_norm": 0.58203125, "learning_rate": 3.7961870821777946e-05, "loss": 0.6199, "step": 1020 }, { "epoch": 1.8013998250218721, "grad_norm": 0.58984375, "learning_rate": 3.791441787663276e-05, "loss": 0.5948, "step": 1030 }, { "epoch": 1.8188976377952755, "grad_norm": 0.55859375, "learning_rate": 3.786645279769244e-05, "loss": 0.608, "step": 1040 }, { "epoch": 1.8363954505686788, "grad_norm": 0.63671875, "learning_rate": 3.781797712849683e-05, "loss": 0.6327, "step": 1050 }, { "epoch": 1.8538932633420822, "grad_norm": 0.66015625, "learning_rate": 3.7768992429016796e-05, "loss": 0.6392, "step": 1060 }, { "epoch": 1.8713910761154855, "grad_norm": 0.546875, "learning_rate": 3.7719500275604074e-05, "loss": 0.6359, "step": 1070 }, { "epoch": 1.8888888888888888, "grad_norm": 0.6171875, "learning_rate": 3.7669502260940495e-05, "loss": 0.6156, "step": 1080 }, { "epoch": 1.9063867016622922, "grad_norm": 0.5390625, "learning_rate": 3.761899999398677e-05, "loss": 0.6142, "step": 1090 }, { "epoch": 1.9238845144356955, "grad_norm": 0.62890625, "learning_rate": 3.756799509993069e-05, "loss": 0.6282, "step": 1100 }, { "epoch": 1.9413823272090989, "grad_norm": 0.63671875, "learning_rate": 3.751648922013482e-05, "loss": 0.6388, "step": 1110 }, { "epoch": 1.9588801399825022, "grad_norm": 0.578125, "learning_rate": 3.746448401208373e-05, "loss": 0.612, "step": 1120 }, { "epoch": 1.9763779527559056, "grad_norm": 0.54296875, "learning_rate": 3.7411981149330575e-05, "loss": 0.622, "step": 1130 }, { "epoch": 1.993875765529309, "grad_norm": 0.625, "learning_rate": 3.735898232144331e-05, "loss": 0.6204, "step": 1140 }, { "epoch": 2.010498687664042, "grad_norm": 0.609375, "learning_rate": 3.7305489233950284e-05, "loss": 0.5708, "step": 1150 }, { "epoch": 2.027996500437445, "grad_norm": 0.6953125, "learning_rate": 3.725150360828537e-05, "loss": 0.5258, "step": 1160 }, { "epoch": 2.0454943132108485, "grad_norm": 0.609375, "learning_rate": 3.7197027181732556e-05, "loss": 0.5357, "step": 1170 }, { "epoch": 2.062992125984252, "grad_norm": 0.57421875, "learning_rate": 3.714206170737003e-05, "loss": 0.5301, "step": 1180 }, { "epoch": 2.080489938757655, "grad_norm": 0.640625, "learning_rate": 3.708660895401381e-05, "loss": 0.5334, "step": 1190 }, { "epoch": 2.0979877515310585, "grad_norm": 0.6015625, "learning_rate": 3.703067070616076e-05, "loss": 0.5308, "step": 1200 }, { "epoch": 2.115485564304462, "grad_norm": 0.53515625, "learning_rate": 3.697424876393122e-05, "loss": 0.5243, "step": 1210 }, { "epoch": 2.1329833770778652, "grad_norm": 0.6171875, "learning_rate": 3.6917344943011043e-05, "loss": 0.5139, "step": 1220 }, { "epoch": 2.1504811898512686, "grad_norm": 0.6171875, "learning_rate": 3.685996107459317e-05, "loss": 0.533, "step": 1230 }, { "epoch": 2.167979002624672, "grad_norm": 0.65625, "learning_rate": 3.680209900531872e-05, "loss": 0.5098, "step": 1240 }, { "epoch": 2.1854768153980753, "grad_norm": 0.6015625, "learning_rate": 3.6743760597217536e-05, "loss": 0.5318, "step": 1250 }, { "epoch": 2.2029746281714786, "grad_norm": 0.59375, "learning_rate": 3.668494772764828e-05, "loss": 0.5223, "step": 1260 }, { "epoch": 2.220472440944882, "grad_norm": 0.59765625, "learning_rate": 3.662566228923805e-05, "loss": 0.5241, "step": 1270 }, { "epoch": 2.2379702537182853, "grad_norm": 0.58984375, "learning_rate": 3.656590618982138e-05, "loss": 0.5249, "step": 1280 }, { "epoch": 2.2554680664916886, "grad_norm": 0.6328125, "learning_rate": 3.650568135237897e-05, "loss": 0.5412, "step": 1290 }, { "epoch": 2.272965879265092, "grad_norm": 0.66796875, "learning_rate": 3.644498971497569e-05, "loss": 0.53, "step": 1300 }, { "epoch": 2.2904636920384953, "grad_norm": 0.67578125, "learning_rate": 3.63838332306983e-05, "loss": 0.5382, "step": 1310 }, { "epoch": 2.3079615048118987, "grad_norm": 0.61328125, "learning_rate": 3.6322213867592536e-05, "loss": 0.5278, "step": 1320 }, { "epoch": 2.325459317585302, "grad_norm": 0.98828125, "learning_rate": 3.626013360859982e-05, "loss": 0.5409, "step": 1330 }, { "epoch": 2.3429571303587053, "grad_norm": 0.64453125, "learning_rate": 3.619759445149341e-05, "loss": 0.5277, "step": 1340 }, { "epoch": 2.3604549431321082, "grad_norm": 0.55078125, "learning_rate": 3.613459840881415e-05, "loss": 0.5344, "step": 1350 }, { "epoch": 2.377952755905512, "grad_norm": 0.6328125, "learning_rate": 3.607114750780568e-05, "loss": 0.5364, "step": 1360 }, { "epoch": 2.395450568678915, "grad_norm": 0.6171875, "learning_rate": 3.6007243790349205e-05, "loss": 0.5238, "step": 1370 }, { "epoch": 2.4129483814523183, "grad_norm": 0.57421875, "learning_rate": 3.594288931289778e-05, "loss": 0.5322, "step": 1380 }, { "epoch": 2.4304461942257216, "grad_norm": 0.59765625, "learning_rate": 3.587808614641015e-05, "loss": 0.5418, "step": 1390 }, { "epoch": 2.447944006999125, "grad_norm": 0.61328125, "learning_rate": 3.581283637628409e-05, "loss": 0.537, "step": 1400 }, { "epoch": 2.4654418197725283, "grad_norm": 0.609375, "learning_rate": 3.574714210228929e-05, "loss": 0.5286, "step": 1410 }, { "epoch": 2.4829396325459316, "grad_norm": 0.61328125, "learning_rate": 3.5681005438499795e-05, "loss": 0.5196, "step": 1420 }, { "epoch": 2.500437445319335, "grad_norm": 0.62109375, "learning_rate": 3.561442851322599e-05, "loss": 0.5559, "step": 1430 }, { "epoch": 2.5179352580927383, "grad_norm": 0.59375, "learning_rate": 3.554741346894606e-05, "loss": 0.5325, "step": 1440 }, { "epoch": 2.5354330708661417, "grad_norm": 0.578125, "learning_rate": 3.5479962462237116e-05, "loss": 0.5368, "step": 1450 }, { "epoch": 2.552930883639545, "grad_norm": 0.59375, "learning_rate": 3.541207766370572e-05, "loss": 0.5368, "step": 1460 }, { "epoch": 2.5704286964129484, "grad_norm": 0.73046875, "learning_rate": 3.534376125791807e-05, "loss": 0.5385, "step": 1470 }, { "epoch": 2.5879265091863517, "grad_norm": 0.54296875, "learning_rate": 3.527501544332972e-05, "loss": 0.5365, "step": 1480 }, { "epoch": 2.605424321959755, "grad_norm": 0.6484375, "learning_rate": 3.520584243221479e-05, "loss": 0.5439, "step": 1490 }, { "epoch": 2.6229221347331584, "grad_norm": 0.5390625, "learning_rate": 3.5136244450594814e-05, "loss": 0.5214, "step": 1500 }, { "epoch": 2.6404199475065617, "grad_norm": 0.6484375, "learning_rate": 3.506622373816706e-05, "loss": 0.5334, "step": 1510 }, { "epoch": 2.657917760279965, "grad_norm": 0.55859375, "learning_rate": 3.499578254823249e-05, "loss": 0.544, "step": 1520 }, { "epoch": 2.6754155730533684, "grad_norm": 0.54296875, "learning_rate": 3.4924923147623265e-05, "loss": 0.5254, "step": 1530 }, { "epoch": 2.6929133858267718, "grad_norm": 0.71484375, "learning_rate": 3.485364781662974e-05, "loss": 0.536, "step": 1540 }, { "epoch": 2.710411198600175, "grad_norm": 0.66796875, "learning_rate": 3.478195884892712e-05, "loss": 0.5471, "step": 1550 }, { "epoch": 2.7279090113735784, "grad_norm": 0.61328125, "learning_rate": 3.4709858551501664e-05, "loss": 0.5235, "step": 1560 }, { "epoch": 2.745406824146982, "grad_norm": 0.58203125, "learning_rate": 3.463734924457638e-05, "loss": 0.5486, "step": 1570 }, { "epoch": 2.7629046369203847, "grad_norm": 0.5625, "learning_rate": 3.4564433261536446e-05, "loss": 0.5218, "step": 1580 }, { "epoch": 2.7804024496937885, "grad_norm": 0.5703125, "learning_rate": 3.449111294885407e-05, "loss": 0.5256, "step": 1590 }, { "epoch": 2.7979002624671914, "grad_norm": 0.62109375, "learning_rate": 3.4417390666012966e-05, "loss": 0.5296, "step": 1600 }, { "epoch": 2.815398075240595, "grad_norm": 0.57421875, "learning_rate": 3.434326878543246e-05, "loss": 0.5365, "step": 1610 }, { "epoch": 2.832895888013998, "grad_norm": 0.58984375, "learning_rate": 3.426874969239113e-05, "loss": 0.5226, "step": 1620 }, { "epoch": 2.850393700787402, "grad_norm": 0.61328125, "learning_rate": 3.419383578495005e-05, "loss": 0.5351, "step": 1630 }, { "epoch": 2.8678915135608047, "grad_norm": 0.625, "learning_rate": 3.411852947387559e-05, "loss": 0.5297, "step": 1640 }, { "epoch": 2.885389326334208, "grad_norm": 0.66015625, "learning_rate": 3.40428331825619e-05, "loss": 0.5248, "step": 1650 }, { "epoch": 2.9028871391076114, "grad_norm": 0.60546875, "learning_rate": 3.3966749346952856e-05, "loss": 0.5327, "step": 1660 }, { "epoch": 2.9203849518810148, "grad_norm": 0.5546875, "learning_rate": 3.389028041546372e-05, "loss": 0.5315, "step": 1670 }, { "epoch": 2.937882764654418, "grad_norm": 0.6015625, "learning_rate": 3.381342884890231e-05, "loss": 0.5327, "step": 1680 }, { "epoch": 2.9553805774278215, "grad_norm": 0.60546875, "learning_rate": 3.373619712038985e-05, "loss": 0.5399, "step": 1690 }, { "epoch": 2.972878390201225, "grad_norm": 0.60546875, "learning_rate": 3.365858771528135e-05, "loss": 0.5307, "step": 1700 }, { "epoch": 2.990376202974628, "grad_norm": 0.5859375, "learning_rate": 3.358060313108564e-05, "loss": 0.5427, "step": 1710 }, { "epoch": 3.0069991251093615, "grad_norm": 0.65234375, "learning_rate": 3.3502245877384986e-05, "loss": 0.4888, "step": 1720 }, { "epoch": 3.024496937882765, "grad_norm": 0.6640625, "learning_rate": 3.3423518475754376e-05, "loss": 0.4435, "step": 1730 }, { "epoch": 3.041994750656168, "grad_norm": 0.71875, "learning_rate": 3.33444234596803e-05, "loss": 0.429, "step": 1740 }, { "epoch": 3.059492563429571, "grad_norm": 0.67578125, "learning_rate": 3.326496337447928e-05, "loss": 0.437, "step": 1750 }, { "epoch": 3.0769903762029744, "grad_norm": 0.65625, "learning_rate": 3.3185140777215956e-05, "loss": 0.4387, "step": 1760 }, { "epoch": 3.094488188976378, "grad_norm": 0.7109375, "learning_rate": 3.3104958236620755e-05, "loss": 0.4211, "step": 1770 }, { "epoch": 3.111986001749781, "grad_norm": 0.58203125, "learning_rate": 3.302441833300728e-05, "loss": 0.4265, "step": 1780 }, { "epoch": 3.1294838145231845, "grad_norm": 0.640625, "learning_rate": 3.2943523658189246e-05, "loss": 0.438, "step": 1790 }, { "epoch": 3.146981627296588, "grad_norm": 0.6171875, "learning_rate": 3.286227681539708e-05, "loss": 0.4379, "step": 1800 }, { "epoch": 3.164479440069991, "grad_norm": 0.7109375, "learning_rate": 3.278068041919415e-05, "loss": 0.4284, "step": 1810 }, { "epoch": 3.1819772528433945, "grad_norm": 0.65234375, "learning_rate": 3.2698737095392593e-05, "loss": 0.4299, "step": 1820 }, { "epoch": 3.199475065616798, "grad_norm": 0.61328125, "learning_rate": 3.261644948096891e-05, "loss": 0.4351, "step": 1830 }, { "epoch": 3.216972878390201, "grad_norm": 0.62890625, "learning_rate": 3.2533820223978996e-05, "loss": 0.4336, "step": 1840 }, { "epoch": 3.2344706911636045, "grad_norm": 0.63671875, "learning_rate": 3.245085198347298e-05, "loss": 0.4376, "step": 1850 }, { "epoch": 3.251968503937008, "grad_norm": 0.56640625, "learning_rate": 3.236754742940965e-05, "loss": 0.4461, "step": 1860 }, { "epoch": 3.269466316710411, "grad_norm": 0.6484375, "learning_rate": 3.228390924257053e-05, "loss": 0.4472, "step": 1870 }, { "epoch": 3.2869641294838146, "grad_norm": 0.68359375, "learning_rate": 3.2199940114473615e-05, "loss": 0.4404, "step": 1880 }, { "epoch": 3.304461942257218, "grad_norm": 0.671875, "learning_rate": 3.211564274728676e-05, "loss": 0.4357, "step": 1890 }, { "epoch": 3.3219597550306212, "grad_norm": 0.640625, "learning_rate": 3.203101985374069e-05, "loss": 0.4405, "step": 1900 }, { "epoch": 3.3394575678040246, "grad_norm": 0.6640625, "learning_rate": 3.194607415704174e-05, "loss": 0.4524, "step": 1910 }, { "epoch": 3.356955380577428, "grad_norm": 0.6328125, "learning_rate": 3.186080839078423e-05, "loss": 0.4311, "step": 1920 }, { "epoch": 3.3744531933508313, "grad_norm": 0.6640625, "learning_rate": 3.177522529886244e-05, "loss": 0.4257, "step": 1930 }, { "epoch": 3.3919510061242346, "grad_norm": 0.87109375, "learning_rate": 3.1689327635382374e-05, "loss": 0.4305, "step": 1940 }, { "epoch": 3.409448818897638, "grad_norm": 0.71875, "learning_rate": 3.160311816457309e-05, "loss": 0.4473, "step": 1950 }, { "epoch": 3.4269466316710413, "grad_norm": 0.63671875, "learning_rate": 3.151659966069777e-05, "loss": 0.4543, "step": 1960 }, { "epoch": 3.4444444444444446, "grad_norm": 0.59765625, "learning_rate": 3.142977490796441e-05, "loss": 0.4444, "step": 1970 }, { "epoch": 3.4619422572178475, "grad_norm": 0.65625, "learning_rate": 3.134264670043627e-05, "loss": 0.4543, "step": 1980 }, { "epoch": 3.4794400699912513, "grad_norm": 0.6484375, "learning_rate": 3.125521784194192e-05, "loss": 0.4433, "step": 1990 }, { "epoch": 3.4969378827646542, "grad_norm": 0.625, "learning_rate": 3.1167491145985035e-05, "loss": 0.4403, "step": 2000 }, { "epoch": 3.514435695538058, "grad_norm": 0.62109375, "learning_rate": 3.107946943565384e-05, "loss": 0.4516, "step": 2010 }, { "epoch": 3.531933508311461, "grad_norm": 0.66015625, "learning_rate": 3.099115554353027e-05, "loss": 0.4326, "step": 2020 }, { "epoch": 3.5494313210848643, "grad_norm": 0.61328125, "learning_rate": 3.090255231159881e-05, "loss": 0.4309, "step": 2030 }, { "epoch": 3.5669291338582676, "grad_norm": 0.6171875, "learning_rate": 3.081366259115504e-05, "loss": 0.4328, "step": 2040 }, { "epoch": 3.584426946631671, "grad_norm": 0.6484375, "learning_rate": 3.072448924271389e-05, "loss": 0.4333, "step": 2050 }, { "epoch": 3.6019247594050743, "grad_norm": 0.6171875, "learning_rate": 3.063503513591758e-05, "loss": 0.4518, "step": 2060 }, { "epoch": 3.6194225721784776, "grad_norm": 0.65234375, "learning_rate": 3.0545303149443265e-05, "loss": 0.4383, "step": 2070 }, { "epoch": 3.636920384951881, "grad_norm": 0.6484375, "learning_rate": 3.0455296170910404e-05, "loss": 0.436, "step": 2080 }, { "epoch": 3.6544181977252843, "grad_norm": 0.671875, "learning_rate": 3.0365017096787834e-05, "loss": 0.4363, "step": 2090 }, { "epoch": 3.6719160104986877, "grad_norm": 0.6484375, "learning_rate": 3.0274468832300576e-05, "loss": 0.4463, "step": 2100 }, { "epoch": 3.689413823272091, "grad_norm": 0.609375, "learning_rate": 3.0183654291336313e-05, "loss": 0.4533, "step": 2110 }, { "epoch": 3.7069116360454943, "grad_norm": 0.6796875, "learning_rate": 3.009257639635164e-05, "loss": 0.4255, "step": 2120 }, { "epoch": 3.7244094488188977, "grad_norm": 0.65234375, "learning_rate": 3.0001238078278022e-05, "loss": 0.4452, "step": 2130 }, { "epoch": 3.741907261592301, "grad_norm": 0.625, "learning_rate": 2.990964227642746e-05, "loss": 0.4344, "step": 2140 }, { "epoch": 3.7594050743657044, "grad_norm": 0.70703125, "learning_rate": 2.9817791938397902e-05, "loss": 0.4327, "step": 2150 }, { "epoch": 3.7769028871391077, "grad_norm": 0.66015625, "learning_rate": 2.9725690019978422e-05, "loss": 0.4461, "step": 2160 }, { "epoch": 3.794400699912511, "grad_norm": 0.6640625, "learning_rate": 2.9633339485054037e-05, "loss": 0.443, "step": 2170 }, { "epoch": 3.8118985126859144, "grad_norm": 0.6171875, "learning_rate": 2.9540743305510406e-05, "loss": 0.4387, "step": 2180 }, { "epoch": 3.8293963254593177, "grad_norm": 0.703125, "learning_rate": 2.9447904461138114e-05, "loss": 0.4355, "step": 2190 }, { "epoch": 3.846894138232721, "grad_norm": 0.625, "learning_rate": 2.9354825939536854e-05, "loss": 0.4425, "step": 2200 }, { "epoch": 3.864391951006124, "grad_norm": 0.6171875, "learning_rate": 2.9261510736019222e-05, "loss": 0.4518, "step": 2210 }, { "epoch": 3.8818897637795278, "grad_norm": 0.6171875, "learning_rate": 2.9167961853514368e-05, "loss": 0.4552, "step": 2220 }, { "epoch": 3.8993875765529307, "grad_norm": 0.71875, "learning_rate": 2.9074182302471347e-05, "loss": 0.4506, "step": 2230 }, { "epoch": 3.9168853893263345, "grad_norm": 0.703125, "learning_rate": 2.898017510076224e-05, "loss": 0.4388, "step": 2240 }, { "epoch": 3.9343832020997374, "grad_norm": 0.671875, "learning_rate": 2.8885943273585037e-05, "loss": 0.4373, "step": 2250 }, { "epoch": 3.9518810148731407, "grad_norm": 0.61328125, "learning_rate": 2.8791489853366284e-05, "loss": 0.4359, "step": 2260 }, { "epoch": 3.969378827646544, "grad_norm": 0.69140625, "learning_rate": 2.8696817879663515e-05, "loss": 0.4538, "step": 2270 }, { "epoch": 3.9868766404199474, "grad_norm": 0.73046875, "learning_rate": 2.8601930399067393e-05, "loss": 0.4394, "step": 2280 }, { "epoch": 4.003499562554681, "grad_norm": 0.62109375, "learning_rate": 2.8506830465103723e-05, "loss": 0.4332, "step": 2290 }, { "epoch": 4.020997375328084, "grad_norm": 0.76953125, "learning_rate": 2.8411521138135154e-05, "loss": 0.3484, "step": 2300 }, { "epoch": 4.038495188101487, "grad_norm": 0.6015625, "learning_rate": 2.8316005485262717e-05, "loss": 0.3652, "step": 2310 }, { "epoch": 4.05599300087489, "grad_norm": 0.76171875, "learning_rate": 2.8220286580227093e-05, "loss": 0.3603, "step": 2320 }, { "epoch": 4.073490813648294, "grad_norm": 0.69921875, "learning_rate": 2.8124367503309736e-05, "loss": 0.3401, "step": 2330 }, { "epoch": 4.090988626421697, "grad_norm": 0.71875, "learning_rate": 2.802825134123371e-05, "loss": 0.3446, "step": 2340 }, { "epoch": 4.108486439195101, "grad_norm": 0.7265625, "learning_rate": 2.7931941187064402e-05, "loss": 0.3732, "step": 2350 }, { "epoch": 4.125984251968504, "grad_norm": 0.66015625, "learning_rate": 2.783544014010993e-05, "loss": 0.3567, "step": 2360 }, { "epoch": 4.1434820647419075, "grad_norm": 0.765625, "learning_rate": 2.773875130582146e-05, "loss": 0.3575, "step": 2370 }, { "epoch": 4.16097987751531, "grad_norm": 0.7265625, "learning_rate": 2.764187779569324e-05, "loss": 0.36, "step": 2380 }, { "epoch": 4.178477690288714, "grad_norm": 0.6875, "learning_rate": 2.7544822727162475e-05, "loss": 0.345, "step": 2390 }, { "epoch": 4.195975503062117, "grad_norm": 0.6953125, "learning_rate": 2.7447589223509004e-05, "loss": 0.3614, "step": 2400 }, { "epoch": 4.213473315835521, "grad_norm": 0.6171875, "learning_rate": 2.7350180413754798e-05, "loss": 0.3584, "step": 2410 }, { "epoch": 4.230971128608924, "grad_norm": 0.68359375, "learning_rate": 2.7252599432563278e-05, "loss": 0.3567, "step": 2420 }, { "epoch": 4.248468941382328, "grad_norm": 0.66796875, "learning_rate": 2.7154849420138405e-05, "loss": 0.3513, "step": 2430 }, { "epoch": 4.2659667541557305, "grad_norm": 0.75390625, "learning_rate": 2.7056933522123677e-05, "loss": 0.3572, "step": 2440 }, { "epoch": 4.283464566929134, "grad_norm": 0.69921875, "learning_rate": 2.695885488950083e-05, "loss": 0.3461, "step": 2450 }, { "epoch": 4.300962379702537, "grad_norm": 0.76171875, "learning_rate": 2.686061667848852e-05, "loss": 0.3534, "step": 2460 }, { "epoch": 4.318460192475941, "grad_norm": 0.7578125, "learning_rate": 2.67622220504407e-05, "loss": 0.3608, "step": 2470 }, { "epoch": 4.335958005249344, "grad_norm": 0.7578125, "learning_rate": 2.6663674171744887e-05, "loss": 0.3549, "step": 2480 }, { "epoch": 4.353455818022747, "grad_norm": 0.69921875, "learning_rate": 2.6564976213720292e-05, "loss": 0.3533, "step": 2490 }, { "epoch": 4.3709536307961505, "grad_norm": 0.69140625, "learning_rate": 2.646613135251576e-05, "loss": 0.3542, "step": 2500 }, { "epoch": 4.388451443569553, "grad_norm": 0.859375, "learning_rate": 2.6367142769007533e-05, "loss": 0.3426, "step": 2510 }, { "epoch": 4.405949256342957, "grad_norm": 0.66796875, "learning_rate": 2.626801364869693e-05, "loss": 0.3511, "step": 2520 }, { "epoch": 4.42344706911636, "grad_norm": 0.72265625, "learning_rate": 2.6168747181607795e-05, "loss": 0.3528, "step": 2530 }, { "epoch": 4.440944881889764, "grad_norm": 0.73046875, "learning_rate": 2.606934656218386e-05, "loss": 0.3491, "step": 2540 }, { "epoch": 4.458442694663167, "grad_norm": 0.7734375, "learning_rate": 2.5969814989185965e-05, "loss": 0.3563, "step": 2550 }, { "epoch": 4.475940507436571, "grad_norm": 0.71875, "learning_rate": 2.5870155665589066e-05, "loss": 0.3603, "step": 2560 }, { "epoch": 4.4934383202099735, "grad_norm": 0.7265625, "learning_rate": 2.5770371798479234e-05, "loss": 0.3525, "step": 2570 }, { "epoch": 4.510936132983377, "grad_norm": 0.734375, "learning_rate": 2.5670466598950385e-05, "loss": 0.3487, "step": 2580 }, { "epoch": 4.52843394575678, "grad_norm": 0.71875, "learning_rate": 2.5570443282000974e-05, "loss": 0.3606, "step": 2590 }, { "epoch": 4.545931758530184, "grad_norm": 0.73046875, "learning_rate": 2.547030506643055e-05, "loss": 0.3569, "step": 2600 }, { "epoch": 4.563429571303587, "grad_norm": 0.74609375, "learning_rate": 2.5370055174736144e-05, "loss": 0.3509, "step": 2610 }, { "epoch": 4.580927384076991, "grad_norm": 0.6328125, "learning_rate": 2.5269696833008577e-05, "loss": 0.3549, "step": 2620 }, { "epoch": 4.5984251968503935, "grad_norm": 0.7265625, "learning_rate": 2.516923327082865e-05, "loss": 0.3521, "step": 2630 }, { "epoch": 4.615923009623797, "grad_norm": 0.671875, "learning_rate": 2.5068667721163213e-05, "loss": 0.3549, "step": 2640 }, { "epoch": 4.6334208223972, "grad_norm": 0.7734375, "learning_rate": 2.496800342026112e-05, "loss": 0.3448, "step": 2650 }, { "epoch": 4.650918635170604, "grad_norm": 0.6640625, "learning_rate": 2.4867243607549085e-05, "loss": 0.3517, "step": 2660 }, { "epoch": 4.668416447944007, "grad_norm": 0.6953125, "learning_rate": 2.4766391525527436e-05, "loss": 0.3548, "step": 2670 }, { "epoch": 4.685914260717411, "grad_norm": 0.65234375, "learning_rate": 2.4665450419665783e-05, "loss": 0.3605, "step": 2680 }, { "epoch": 4.703412073490814, "grad_norm": 0.80078125, "learning_rate": 2.4564423538298567e-05, "loss": 0.3565, "step": 2690 }, { "epoch": 4.7209098862642165, "grad_norm": 0.69921875, "learning_rate": 2.4463314132520522e-05, "loss": 0.344, "step": 2700 }, { "epoch": 4.73840769903762, "grad_norm": 0.68359375, "learning_rate": 2.4362125456082058e-05, "loss": 0.3496, "step": 2710 }, { "epoch": 4.755905511811024, "grad_norm": 0.8125, "learning_rate": 2.426086076528455e-05, "loss": 0.3584, "step": 2720 }, { "epoch": 4.773403324584427, "grad_norm": 0.6875, "learning_rate": 2.4159523318875562e-05, "loss": 0.3627, "step": 2730 }, { "epoch": 4.79090113735783, "grad_norm": 0.7734375, "learning_rate": 2.405811637794396e-05, "loss": 0.3585, "step": 2740 }, { "epoch": 4.808398950131234, "grad_norm": 0.73046875, "learning_rate": 2.3956643205814994e-05, "loss": 0.3651, "step": 2750 }, { "epoch": 4.8258967629046365, "grad_norm": 0.66796875, "learning_rate": 2.385510706794524e-05, "loss": 0.3537, "step": 2760 }, { "epoch": 4.84339457567804, "grad_norm": 0.7734375, "learning_rate": 2.375351123181756e-05, "loss": 0.3476, "step": 2770 }, { "epoch": 4.860892388451443, "grad_norm": 0.75, "learning_rate": 2.365185896683595e-05, "loss": 0.3416, "step": 2780 }, { "epoch": 4.878390201224847, "grad_norm": 0.7578125, "learning_rate": 2.3550153544220273e-05, "loss": 0.3569, "step": 2790 }, { "epoch": 4.89588801399825, "grad_norm": 0.69921875, "learning_rate": 2.3448398236901073e-05, "loss": 0.3502, "step": 2800 }, { "epoch": 4.913385826771654, "grad_norm": 0.77734375, "learning_rate": 2.3346596319414173e-05, "loss": 0.3538, "step": 2810 }, { "epoch": 4.930883639545057, "grad_norm": 0.67578125, "learning_rate": 2.3244751067795366e-05, "loss": 0.3635, "step": 2820 }, { "epoch": 4.94838145231846, "grad_norm": 0.71484375, "learning_rate": 2.314286575947494e-05, "loss": 0.3567, "step": 2830 }, { "epoch": 4.965879265091863, "grad_norm": 0.8515625, "learning_rate": 2.304094367317223e-05, "loss": 0.3293, "step": 2840 }, { "epoch": 4.983377077865267, "grad_norm": 0.7421875, "learning_rate": 2.293898808879011e-05, "loss": 0.3656, "step": 2850 }, { "epoch": 5.0, "grad_norm": 1.0546875, "learning_rate": 2.2837002287309426e-05, "loss": 0.3726, "step": 2860 } ], "logging_steps": 10, "max_steps": 5710, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 7.691223286459776e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }