diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,14821 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.6299212598425197, + "eval_steps": 20, + "global_step": 2000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00031496062992125983, + "grad_norm": NaN, + "learning_rate": 1e-05, + "loss": 0.6279, + "step": 1 + }, + { + "epoch": 0.0006299212598425197, + "grad_norm": NaN, + "learning_rate": 1e-05, + "loss": 0.7073, + "step": 2 + }, + { + "epoch": 0.0009448818897637795, + "grad_norm": Infinity, + "learning_rate": 1e-05, + "loss": 0.4724, + "step": 3 + }, + { + "epoch": 0.0012598425196850393, + "grad_norm": Infinity, + "learning_rate": 1e-05, + "loss": 0.8548, + "step": 4 + }, + { + "epoch": 0.0015748031496062992, + "grad_norm": 241.3546600341797, + "learning_rate": 9.99999842519685e-06, + "loss": 0.9812, + "step": 5 + }, + { + "epoch": 0.001889763779527559, + "grad_norm": 68.62373352050781, + "learning_rate": 9.999996850393701e-06, + "loss": 0.6851, + "step": 6 + }, + { + "epoch": 0.002204724409448819, + "grad_norm": 45.99556350708008, + "learning_rate": 9.999995275590552e-06, + "loss": 0.7491, + "step": 7 + }, + { + "epoch": 0.0025196850393700786, + "grad_norm": 27.5998592376709, + "learning_rate": 9.999993700787403e-06, + "loss": 0.6691, + "step": 8 + }, + { + "epoch": 0.0028346456692913387, + "grad_norm": 53.23931121826172, + "learning_rate": 9.999992125984252e-06, + "loss": 0.7518, + "step": 9 + }, + { + "epoch": 0.0031496062992125984, + "grad_norm": 21.858654022216797, + "learning_rate": 9.999990551181103e-06, + "loss": 0.6986, + "step": 10 + }, + { + "epoch": 0.0034645669291338585, + "grad_norm": 38.456905364990234, + "learning_rate": 9.999988976377953e-06, + "loss": 0.6296, + "step": 11 + }, + { + "epoch": 0.003779527559055118, + "grad_norm": 36.96352005004883, + "learning_rate": 9.999987401574804e-06, + "loss": 0.6674, + "step": 12 + }, + { + "epoch": 0.004094488188976378, + "grad_norm": 50.633941650390625, + "learning_rate": 9.999985826771655e-06, + "loss": 0.6214, + "step": 13 + }, + { + "epoch": 0.004409448818897638, + "grad_norm": 90.81790161132812, + "learning_rate": 9.999984251968506e-06, + "loss": 0.7667, + "step": 14 + }, + { + "epoch": 0.004724409448818898, + "grad_norm": 157.15757751464844, + "learning_rate": 9.999982677165355e-06, + "loss": 1.0109, + "step": 15 + }, + { + "epoch": 0.005039370078740157, + "grad_norm": 57.94607925415039, + "learning_rate": 9.999981102362206e-06, + "loss": 0.6623, + "step": 16 + }, + { + "epoch": 0.005354330708661417, + "grad_norm": 96.26383972167969, + "learning_rate": 9.999979527559057e-06, + "loss": 0.7511, + "step": 17 + }, + { + "epoch": 0.005669291338582677, + "grad_norm": 63.17537307739258, + "learning_rate": 9.999977952755906e-06, + "loss": 0.7292, + "step": 18 + }, + { + "epoch": 0.005984251968503937, + "grad_norm": 28.436891555786133, + "learning_rate": 9.999976377952757e-06, + "loss": 0.6612, + "step": 19 + }, + { + "epoch": 0.006299212598425197, + "grad_norm": 46.60204315185547, + "learning_rate": 9.999974803149607e-06, + "loss": 0.471, + "step": 20 + }, + { + "epoch": 0.006299212598425197, + "eval_loss": 0.6495372653007507, + "eval_runtime": 323.1698, + "eval_samples_per_second": 0.362, + "eval_steps_per_second": 0.362, + "step": 20 + }, + { + "epoch": 0.006614173228346456, + "grad_norm": 34.931636810302734, + "learning_rate": 9.999973228346457e-06, + "loss": 0.6546, + "step": 21 + }, + { + "epoch": 0.006929133858267717, + "grad_norm": 37.24106216430664, + "learning_rate": 9.999971653543308e-06, + "loss": 0.5786, + "step": 22 + }, + { + "epoch": 0.007244094488188977, + "grad_norm": 49.713348388671875, + "learning_rate": 9.99997007874016e-06, + "loss": 0.605, + "step": 23 + }, + { + "epoch": 0.007559055118110236, + "grad_norm": 54.44386291503906, + "learning_rate": 9.999968503937009e-06, + "loss": 0.3443, + "step": 24 + }, + { + "epoch": 0.007874015748031496, + "grad_norm": 79.05270385742188, + "learning_rate": 9.999966929133858e-06, + "loss": 0.6105, + "step": 25 + }, + { + "epoch": 0.008188976377952756, + "grad_norm": 77.4219741821289, + "learning_rate": 9.999965354330709e-06, + "loss": 0.864, + "step": 26 + }, + { + "epoch": 0.008503937007874015, + "grad_norm": 137.48190307617188, + "learning_rate": 9.99996377952756e-06, + "loss": 1.1885, + "step": 27 + }, + { + "epoch": 0.008818897637795276, + "grad_norm": 156.0934600830078, + "learning_rate": 9.999962204724411e-06, + "loss": 1.1261, + "step": 28 + }, + { + "epoch": 0.009133858267716535, + "grad_norm": 117.86957550048828, + "learning_rate": 9.99996062992126e-06, + "loss": 0.6776, + "step": 29 + }, + { + "epoch": 0.009448818897637795, + "grad_norm": 107.25189971923828, + "learning_rate": 9.999959055118111e-06, + "loss": 0.7123, + "step": 30 + }, + { + "epoch": 0.009763779527559056, + "grad_norm": 53.983299255371094, + "learning_rate": 9.99995748031496e-06, + "loss": 0.6213, + "step": 31 + }, + { + "epoch": 0.010078740157480314, + "grad_norm": 60.340389251708984, + "learning_rate": 9.999955905511812e-06, + "loss": 0.7145, + "step": 32 + }, + { + "epoch": 0.010393700787401575, + "grad_norm": 76.8556137084961, + "learning_rate": 9.999954330708663e-06, + "loss": 0.6812, + "step": 33 + }, + { + "epoch": 0.010708661417322834, + "grad_norm": 45.225807189941406, + "learning_rate": 9.999952755905514e-06, + "loss": 0.6361, + "step": 34 + }, + { + "epoch": 0.011023622047244094, + "grad_norm": 65.44268798828125, + "learning_rate": 9.999951181102363e-06, + "loss": 0.6421, + "step": 35 + }, + { + "epoch": 0.011338582677165355, + "grad_norm": 42.70692825317383, + "learning_rate": 9.999949606299212e-06, + "loss": 0.6809, + "step": 36 + }, + { + "epoch": 0.011653543307086614, + "grad_norm": 43.51832580566406, + "learning_rate": 9.999948031496063e-06, + "loss": 0.6612, + "step": 37 + }, + { + "epoch": 0.011968503937007874, + "grad_norm": 37.09170913696289, + "learning_rate": 9.999946456692914e-06, + "loss": 0.7119, + "step": 38 + }, + { + "epoch": 0.012283464566929133, + "grad_norm": 29.477069854736328, + "learning_rate": 9.999944881889765e-06, + "loss": 0.6933, + "step": 39 + }, + { + "epoch": 0.012598425196850394, + "grad_norm": 44.96734619140625, + "learning_rate": 9.999943307086614e-06, + "loss": 0.7328, + "step": 40 + }, + { + "epoch": 0.012598425196850394, + "eval_loss": 0.6671837568283081, + "eval_runtime": 309.18, + "eval_samples_per_second": 0.378, + "eval_steps_per_second": 0.378, + "step": 40 + }, + { + "epoch": 0.012913385826771654, + "grad_norm": 19.817779541015625, + "learning_rate": 9.999941732283465e-06, + "loss": 0.6673, + "step": 41 + }, + { + "epoch": 0.013228346456692913, + "grad_norm": 23.819435119628906, + "learning_rate": 9.999940157480316e-06, + "loss": 0.6966, + "step": 42 + }, + { + "epoch": 0.013543307086614173, + "grad_norm": 21.729511260986328, + "learning_rate": 9.999938582677167e-06, + "loss": 0.6791, + "step": 43 + }, + { + "epoch": 0.013858267716535434, + "grad_norm": 18.30646514892578, + "learning_rate": 9.999937007874017e-06, + "loss": 0.6786, + "step": 44 + }, + { + "epoch": 0.014173228346456693, + "grad_norm": 11.647773742675781, + "learning_rate": 9.999935433070866e-06, + "loss": 0.6956, + "step": 45 + }, + { + "epoch": 0.014488188976377953, + "grad_norm": 15.509359359741211, + "learning_rate": 9.999933858267717e-06, + "loss": 0.6616, + "step": 46 + }, + { + "epoch": 0.014803149606299212, + "grad_norm": 20.342838287353516, + "learning_rate": 9.999932283464568e-06, + "loss": 0.6184, + "step": 47 + }, + { + "epoch": 0.015118110236220473, + "grad_norm": 15.10333251953125, + "learning_rate": 9.999930708661419e-06, + "loss": 0.6671, + "step": 48 + }, + { + "epoch": 0.015433070866141733, + "grad_norm": 22.806962966918945, + "learning_rate": 9.999929133858268e-06, + "loss": 0.7039, + "step": 49 + }, + { + "epoch": 0.015748031496062992, + "grad_norm": 22.755117416381836, + "learning_rate": 9.99992755905512e-06, + "loss": 0.6512, + "step": 50 + }, + { + "epoch": 0.016062992125984252, + "grad_norm": 30.809261322021484, + "learning_rate": 9.999925984251969e-06, + "loss": 0.7041, + "step": 51 + }, + { + "epoch": 0.016377952755905513, + "grad_norm": 30.779508590698242, + "learning_rate": 9.99992440944882e-06, + "loss": 0.6321, + "step": 52 + }, + { + "epoch": 0.01669291338582677, + "grad_norm": 41.0311279296875, + "learning_rate": 9.99992283464567e-06, + "loss": 0.7655, + "step": 53 + }, + { + "epoch": 0.01700787401574803, + "grad_norm": 38.755794525146484, + "learning_rate": 9.999921259842522e-06, + "loss": 0.5663, + "step": 54 + }, + { + "epoch": 0.01732283464566929, + "grad_norm": 76.18267059326172, + "learning_rate": 9.99991968503937e-06, + "loss": 0.817, + "step": 55 + }, + { + "epoch": 0.01763779527559055, + "grad_norm": 64.29126739501953, + "learning_rate": 9.99991811023622e-06, + "loss": 0.7239, + "step": 56 + }, + { + "epoch": 0.017952755905511812, + "grad_norm": 64.24166107177734, + "learning_rate": 9.999916535433071e-06, + "loss": 0.5583, + "step": 57 + }, + { + "epoch": 0.01826771653543307, + "grad_norm": 35.76716232299805, + "learning_rate": 9.999914960629922e-06, + "loss": 0.6473, + "step": 58 + }, + { + "epoch": 0.01858267716535433, + "grad_norm": 31.879194259643555, + "learning_rate": 9.999913385826773e-06, + "loss": 0.5875, + "step": 59 + }, + { + "epoch": 0.01889763779527559, + "grad_norm": 43.807613372802734, + "learning_rate": 9.999911811023622e-06, + "loss": 0.6888, + "step": 60 + }, + { + "epoch": 0.01889763779527559, + "eval_loss": 0.6864338517189026, + "eval_runtime": 308.0649, + "eval_samples_per_second": 0.38, + "eval_steps_per_second": 0.38, + "step": 60 + }, + { + "epoch": 0.01921259842519685, + "grad_norm": 68.39881896972656, + "learning_rate": 9.999910236220473e-06, + "loss": 0.7587, + "step": 61 + }, + { + "epoch": 0.01952755905511811, + "grad_norm": 43.71537780761719, + "learning_rate": 9.999908661417323e-06, + "loss": 0.6054, + "step": 62 + }, + { + "epoch": 0.01984251968503937, + "grad_norm": 38.36960220336914, + "learning_rate": 9.999907086614175e-06, + "loss": 0.527, + "step": 63 + }, + { + "epoch": 0.02015748031496063, + "grad_norm": 53.25741958618164, + "learning_rate": 9.999905511811025e-06, + "loss": 0.8724, + "step": 64 + }, + { + "epoch": 0.02047244094488189, + "grad_norm": 38.55160140991211, + "learning_rate": 9.999903937007874e-06, + "loss": 0.6693, + "step": 65 + }, + { + "epoch": 0.02078740157480315, + "grad_norm": 36.05056381225586, + "learning_rate": 9.999902362204725e-06, + "loss": 0.6024, + "step": 66 + }, + { + "epoch": 0.02110236220472441, + "grad_norm": 83.4520492553711, + "learning_rate": 9.999900787401576e-06, + "loss": 0.6391, + "step": 67 + }, + { + "epoch": 0.021417322834645668, + "grad_norm": 23.22808837890625, + "learning_rate": 9.999899212598427e-06, + "loss": 0.6275, + "step": 68 + }, + { + "epoch": 0.021732283464566928, + "grad_norm": 42.370445251464844, + "learning_rate": 9.999897637795276e-06, + "loss": 0.3825, + "step": 69 + }, + { + "epoch": 0.02204724409448819, + "grad_norm": 38.64667892456055, + "learning_rate": 9.999896062992127e-06, + "loss": 0.6354, + "step": 70 + }, + { + "epoch": 0.02236220472440945, + "grad_norm": 44.61943054199219, + "learning_rate": 9.999894488188977e-06, + "loss": 0.6041, + "step": 71 + }, + { + "epoch": 0.02267716535433071, + "grad_norm": 36.52523422241211, + "learning_rate": 9.999892913385828e-06, + "loss": 0.5188, + "step": 72 + }, + { + "epoch": 0.022992125984251967, + "grad_norm": 38.4240608215332, + "learning_rate": 9.999891338582679e-06, + "loss": 0.612, + "step": 73 + }, + { + "epoch": 0.023307086614173227, + "grad_norm": 112.46929168701172, + "learning_rate": 9.99988976377953e-06, + "loss": 0.9316, + "step": 74 + }, + { + "epoch": 0.023622047244094488, + "grad_norm": 91.35350799560547, + "learning_rate": 9.999888188976379e-06, + "loss": 0.4754, + "step": 75 + }, + { + "epoch": 0.02393700787401575, + "grad_norm": 136.8651123046875, + "learning_rate": 9.999886614173228e-06, + "loss": 0.8443, + "step": 76 + }, + { + "epoch": 0.02425196850393701, + "grad_norm": 64.04878997802734, + "learning_rate": 9.999885039370079e-06, + "loss": 0.5659, + "step": 77 + }, + { + "epoch": 0.024566929133858266, + "grad_norm": 127.41741180419922, + "learning_rate": 9.99988346456693e-06, + "loss": 0.5924, + "step": 78 + }, + { + "epoch": 0.024881889763779527, + "grad_norm": 88.72442626953125, + "learning_rate": 9.999881889763781e-06, + "loss": 0.6118, + "step": 79 + }, + { + "epoch": 0.025196850393700787, + "grad_norm": 91.45403289794922, + "learning_rate": 9.99988031496063e-06, + "loss": 0.7566, + "step": 80 + }, + { + "epoch": 0.025196850393700787, + "eval_loss": 0.620968222618103, + "eval_runtime": 308.6891, + "eval_samples_per_second": 0.379, + "eval_steps_per_second": 0.379, + "step": 80 + }, + { + "epoch": 0.025511811023622048, + "grad_norm": 40.077823638916016, + "learning_rate": 9.999878740157481e-06, + "loss": 0.3255, + "step": 81 + }, + { + "epoch": 0.025826771653543308, + "grad_norm": 86.21344757080078, + "learning_rate": 9.99987716535433e-06, + "loss": 0.5621, + "step": 82 + }, + { + "epoch": 0.02614173228346457, + "grad_norm": 102.60726165771484, + "learning_rate": 9.999875590551182e-06, + "loss": 0.8275, + "step": 83 + }, + { + "epoch": 0.026456692913385826, + "grad_norm": 118.94241333007812, + "learning_rate": 9.999874015748033e-06, + "loss": 0.5316, + "step": 84 + }, + { + "epoch": 0.026771653543307086, + "grad_norm": 44.944576263427734, + "learning_rate": 9.999872440944882e-06, + "loss": 0.6057, + "step": 85 + }, + { + "epoch": 0.027086614173228347, + "grad_norm": 43.299503326416016, + "learning_rate": 9.999870866141733e-06, + "loss": 0.3967, + "step": 86 + }, + { + "epoch": 0.027401574803149607, + "grad_norm": 67.91696166992188, + "learning_rate": 9.999869291338584e-06, + "loss": 0.5094, + "step": 87 + }, + { + "epoch": 0.027716535433070868, + "grad_norm": 54.101783752441406, + "learning_rate": 9.999867716535435e-06, + "loss": 0.7315, + "step": 88 + }, + { + "epoch": 0.028031496062992125, + "grad_norm": 72.56822204589844, + "learning_rate": 9.999866141732284e-06, + "loss": 0.7512, + "step": 89 + }, + { + "epoch": 0.028346456692913385, + "grad_norm": 74.34241485595703, + "learning_rate": 9.999864566929135e-06, + "loss": 0.6363, + "step": 90 + }, + { + "epoch": 0.028661417322834646, + "grad_norm": 63.87611770629883, + "learning_rate": 9.999862992125984e-06, + "loss": 0.7365, + "step": 91 + }, + { + "epoch": 0.028976377952755906, + "grad_norm": 90.9892807006836, + "learning_rate": 9.999861417322835e-06, + "loss": 0.8111, + "step": 92 + }, + { + "epoch": 0.029291338582677167, + "grad_norm": 51.74814987182617, + "learning_rate": 9.999859842519686e-06, + "loss": 0.4766, + "step": 93 + }, + { + "epoch": 0.029606299212598424, + "grad_norm": 48.99016571044922, + "learning_rate": 9.999858267716537e-06, + "loss": 0.4001, + "step": 94 + }, + { + "epoch": 0.029921259842519685, + "grad_norm": 35.5272216796875, + "learning_rate": 9.999856692913387e-06, + "loss": 0.4208, + "step": 95 + }, + { + "epoch": 0.030236220472440945, + "grad_norm": 53.10519790649414, + "learning_rate": 9.999855118110236e-06, + "loss": 0.7011, + "step": 96 + }, + { + "epoch": 0.030551181102362206, + "grad_norm": 51.19492721557617, + "learning_rate": 9.999853543307087e-06, + "loss": 0.5737, + "step": 97 + }, + { + "epoch": 0.030866141732283466, + "grad_norm": 40.27799987792969, + "learning_rate": 9.999851968503938e-06, + "loss": 0.4889, + "step": 98 + }, + { + "epoch": 0.031181102362204723, + "grad_norm": 53.10594940185547, + "learning_rate": 9.999850393700789e-06, + "loss": 0.2341, + "step": 99 + }, + { + "epoch": 0.031496062992125984, + "grad_norm": 64.59747314453125, + "learning_rate": 9.999848818897638e-06, + "loss": 0.5309, + "step": 100 + }, + { + "epoch": 0.031496062992125984, + "eval_loss": 0.6238653659820557, + "eval_runtime": 308.8997, + "eval_samples_per_second": 0.379, + "eval_steps_per_second": 0.379, + "step": 100 + }, + { + "epoch": 0.03181102362204724, + "grad_norm": 56.20338821411133, + "learning_rate": 9.99984724409449e-06, + "loss": 0.5554, + "step": 101 + }, + { + "epoch": 0.032125984251968505, + "grad_norm": 78.87137603759766, + "learning_rate": 9.999845669291339e-06, + "loss": 0.6822, + "step": 102 + }, + { + "epoch": 0.03244094488188976, + "grad_norm": 76.1572036743164, + "learning_rate": 9.99984409448819e-06, + "loss": 0.4211, + "step": 103 + }, + { + "epoch": 0.032755905511811026, + "grad_norm": 53.714942932128906, + "learning_rate": 9.99984251968504e-06, + "loss": 0.2694, + "step": 104 + }, + { + "epoch": 0.03307086614173228, + "grad_norm": 100.88641357421875, + "learning_rate": 9.99984094488189e-06, + "loss": 0.6587, + "step": 105 + }, + { + "epoch": 0.03338582677165354, + "grad_norm": 59.46120071411133, + "learning_rate": 9.999839370078741e-06, + "loss": 0.3308, + "step": 106 + }, + { + "epoch": 0.033700787401574804, + "grad_norm": 142.22496032714844, + "learning_rate": 9.99983779527559e-06, + "loss": 0.9014, + "step": 107 + }, + { + "epoch": 0.03401574803149606, + "grad_norm": 116.70782470703125, + "learning_rate": 9.999836220472441e-06, + "loss": 0.7636, + "step": 108 + }, + { + "epoch": 0.034330708661417325, + "grad_norm": 122.21369171142578, + "learning_rate": 9.999834645669292e-06, + "loss": 0.6502, + "step": 109 + }, + { + "epoch": 0.03464566929133858, + "grad_norm": 146.69210815429688, + "learning_rate": 9.999833070866143e-06, + "loss": 1.088, + "step": 110 + }, + { + "epoch": 0.03496062992125984, + "grad_norm": 90.0801010131836, + "learning_rate": 9.999831496062992e-06, + "loss": 0.5323, + "step": 111 + }, + { + "epoch": 0.0352755905511811, + "grad_norm": 64.70466613769531, + "learning_rate": 9.999829921259843e-06, + "loss": 0.4768, + "step": 112 + }, + { + "epoch": 0.03559055118110236, + "grad_norm": 43.21613311767578, + "learning_rate": 9.999828346456694e-06, + "loss": 0.5867, + "step": 113 + }, + { + "epoch": 0.035905511811023624, + "grad_norm": 98.97393798828125, + "learning_rate": 9.999826771653545e-06, + "loss": 0.9022, + "step": 114 + }, + { + "epoch": 0.03622047244094488, + "grad_norm": 49.0715446472168, + "learning_rate": 9.999825196850395e-06, + "loss": 0.4157, + "step": 115 + }, + { + "epoch": 0.03653543307086614, + "grad_norm": 49.2851676940918, + "learning_rate": 9.999823622047244e-06, + "loss": 0.4487, + "step": 116 + }, + { + "epoch": 0.0368503937007874, + "grad_norm": 37.42869567871094, + "learning_rate": 9.999822047244095e-06, + "loss": 0.576, + "step": 117 + }, + { + "epoch": 0.03716535433070866, + "grad_norm": 43.0858154296875, + "learning_rate": 9.999820472440946e-06, + "loss": 0.6744, + "step": 118 + }, + { + "epoch": 0.037480314960629924, + "grad_norm": 51.691558837890625, + "learning_rate": 9.999818897637797e-06, + "loss": 0.4399, + "step": 119 + }, + { + "epoch": 0.03779527559055118, + "grad_norm": 48.11525344848633, + "learning_rate": 9.999817322834646e-06, + "loss": 0.7208, + "step": 120 + }, + { + "epoch": 0.03779527559055118, + "eval_loss": 0.6145237684249878, + "eval_runtime": 297.9679, + "eval_samples_per_second": 0.393, + "eval_steps_per_second": 0.393, + "step": 120 + }, + { + "epoch": 0.03811023622047244, + "grad_norm": 59.46645736694336, + "learning_rate": 9.999815748031497e-06, + "loss": 0.6338, + "step": 121 + }, + { + "epoch": 0.0384251968503937, + "grad_norm": 41.5179443359375, + "learning_rate": 9.999814173228347e-06, + "loss": 0.6118, + "step": 122 + }, + { + "epoch": 0.03874015748031496, + "grad_norm": 30.39054298400879, + "learning_rate": 9.999812598425198e-06, + "loss": 0.589, + "step": 123 + }, + { + "epoch": 0.03905511811023622, + "grad_norm": 47.73324966430664, + "learning_rate": 9.999811023622049e-06, + "loss": 0.4502, + "step": 124 + }, + { + "epoch": 0.03937007874015748, + "grad_norm": 63.116180419921875, + "learning_rate": 9.999809448818898e-06, + "loss": 0.5543, + "step": 125 + }, + { + "epoch": 0.03968503937007874, + "grad_norm": 48.51982879638672, + "learning_rate": 9.999807874015749e-06, + "loss": 0.3345, + "step": 126 + }, + { + "epoch": 0.04, + "grad_norm": 70.61180114746094, + "learning_rate": 9.999806299212598e-06, + "loss": 0.408, + "step": 127 + }, + { + "epoch": 0.04031496062992126, + "grad_norm": 100.2572021484375, + "learning_rate": 9.999804724409449e-06, + "loss": 0.5952, + "step": 128 + }, + { + "epoch": 0.04062992125984252, + "grad_norm": 120.8499984741211, + "learning_rate": 9.9998031496063e-06, + "loss": 0.5751, + "step": 129 + }, + { + "epoch": 0.04094488188976378, + "grad_norm": 116.19609069824219, + "learning_rate": 9.999801574803151e-06, + "loss": 0.9086, + "step": 130 + }, + { + "epoch": 0.041259842519685036, + "grad_norm": 203.10231018066406, + "learning_rate": 9.9998e-06, + "loss": 0.907, + "step": 131 + }, + { + "epoch": 0.0415748031496063, + "grad_norm": 90.6951904296875, + "learning_rate": 9.999798425196851e-06, + "loss": 0.4064, + "step": 132 + }, + { + "epoch": 0.04188976377952756, + "grad_norm": 67.41472625732422, + "learning_rate": 9.999796850393702e-06, + "loss": 0.4828, + "step": 133 + }, + { + "epoch": 0.04220472440944882, + "grad_norm": 137.20547485351562, + "learning_rate": 9.999795275590553e-06, + "loss": 0.7428, + "step": 134 + }, + { + "epoch": 0.04251968503937008, + "grad_norm": 290.8530578613281, + "learning_rate": 9.999793700787403e-06, + "loss": 0.8161, + "step": 135 + }, + { + "epoch": 0.042834645669291335, + "grad_norm": 149.33602905273438, + "learning_rate": 9.999792125984252e-06, + "loss": 0.5009, + "step": 136 + }, + { + "epoch": 0.0431496062992126, + "grad_norm": 114.50115203857422, + "learning_rate": 9.999790551181103e-06, + "loss": 1.0042, + "step": 137 + }, + { + "epoch": 0.043464566929133856, + "grad_norm": 101.2292251586914, + "learning_rate": 9.999788976377954e-06, + "loss": 0.4708, + "step": 138 + }, + { + "epoch": 0.04377952755905512, + "grad_norm": 91.65269470214844, + "learning_rate": 9.999787401574805e-06, + "loss": 0.6723, + "step": 139 + }, + { + "epoch": 0.04409448818897638, + "grad_norm": 54.82379913330078, + "learning_rate": 9.999785826771654e-06, + "loss": 0.4811, + "step": 140 + }, + { + "epoch": 0.04409448818897638, + "eval_loss": 0.5761768221855164, + "eval_runtime": 296.7123, + "eval_samples_per_second": 0.394, + "eval_steps_per_second": 0.394, + "step": 140 + }, + { + "epoch": 0.044409448818897634, + "grad_norm": 30.785768508911133, + "learning_rate": 9.999784251968505e-06, + "loss": 0.2963, + "step": 141 + }, + { + "epoch": 0.0447244094488189, + "grad_norm": 44.203250885009766, + "learning_rate": 9.999782677165354e-06, + "loss": 0.6864, + "step": 142 + }, + { + "epoch": 0.045039370078740155, + "grad_norm": 67.31315612792969, + "learning_rate": 9.999781102362205e-06, + "loss": 0.6552, + "step": 143 + }, + { + "epoch": 0.04535433070866142, + "grad_norm": 36.43077087402344, + "learning_rate": 9.999779527559056e-06, + "loss": 0.6163, + "step": 144 + }, + { + "epoch": 0.04566929133858268, + "grad_norm": 81.87039184570312, + "learning_rate": 9.999777952755906e-06, + "loss": 1.0169, + "step": 145 + }, + { + "epoch": 0.045984251968503934, + "grad_norm": 32.918399810791016, + "learning_rate": 9.999776377952757e-06, + "loss": 0.4143, + "step": 146 + }, + { + "epoch": 0.0462992125984252, + "grad_norm": 40.705284118652344, + "learning_rate": 9.999774803149606e-06, + "loss": 0.5362, + "step": 147 + }, + { + "epoch": 0.046614173228346455, + "grad_norm": 25.83769416809082, + "learning_rate": 9.999773228346457e-06, + "loss": 0.4726, + "step": 148 + }, + { + "epoch": 0.04692913385826772, + "grad_norm": 51.383758544921875, + "learning_rate": 9.999771653543308e-06, + "loss": 0.6378, + "step": 149 + }, + { + "epoch": 0.047244094488188976, + "grad_norm": 59.23312759399414, + "learning_rate": 9.999770078740159e-06, + "loss": 0.4571, + "step": 150 + }, + { + "epoch": 0.04755905511811023, + "grad_norm": 32.09741973876953, + "learning_rate": 9.999768503937008e-06, + "loss": 0.3173, + "step": 151 + }, + { + "epoch": 0.0478740157480315, + "grad_norm": 40.37042999267578, + "learning_rate": 9.99976692913386e-06, + "loss": 0.6127, + "step": 152 + }, + { + "epoch": 0.048188976377952754, + "grad_norm": 50.806793212890625, + "learning_rate": 9.999765354330709e-06, + "loss": 0.6617, + "step": 153 + }, + { + "epoch": 0.04850393700787402, + "grad_norm": 42.13128662109375, + "learning_rate": 9.99976377952756e-06, + "loss": 0.4311, + "step": 154 + }, + { + "epoch": 0.048818897637795275, + "grad_norm": 51.54093933105469, + "learning_rate": 9.99976220472441e-06, + "loss": 0.7501, + "step": 155 + }, + { + "epoch": 0.04913385826771653, + "grad_norm": 71.40542602539062, + "learning_rate": 9.99976062992126e-06, + "loss": 1.0623, + "step": 156 + }, + { + "epoch": 0.049448818897637796, + "grad_norm": 23.955883026123047, + "learning_rate": 9.999759055118111e-06, + "loss": 0.118, + "step": 157 + }, + { + "epoch": 0.04976377952755905, + "grad_norm": 46.93206024169922, + "learning_rate": 9.999757480314962e-06, + "loss": 0.6759, + "step": 158 + }, + { + "epoch": 0.05007874015748032, + "grad_norm": 40.86898422241211, + "learning_rate": 9.999755905511813e-06, + "loss": 0.5849, + "step": 159 + }, + { + "epoch": 0.050393700787401574, + "grad_norm": 54.818450927734375, + "learning_rate": 9.999754330708662e-06, + "loss": 0.4963, + "step": 160 + }, + { + "epoch": 0.050393700787401574, + "eval_loss": 0.646256148815155, + "eval_runtime": 299.1934, + "eval_samples_per_second": 0.391, + "eval_steps_per_second": 0.391, + "step": 160 + }, + { + "epoch": 0.05070866141732284, + "grad_norm": 39.67280578613281, + "learning_rate": 9.999752755905513e-06, + "loss": 0.6, + "step": 161 + }, + { + "epoch": 0.051023622047244095, + "grad_norm": 44.99142074584961, + "learning_rate": 9.999751181102362e-06, + "loss": 0.4204, + "step": 162 + }, + { + "epoch": 0.05133858267716535, + "grad_norm": 47.1932373046875, + "learning_rate": 9.999749606299213e-06, + "loss": 0.5264, + "step": 163 + }, + { + "epoch": 0.051653543307086616, + "grad_norm": 59.98406219482422, + "learning_rate": 9.999748031496064e-06, + "loss": 0.5378, + "step": 164 + }, + { + "epoch": 0.05196850393700787, + "grad_norm": 54.76002883911133, + "learning_rate": 9.999746456692914e-06, + "loss": 0.4653, + "step": 165 + }, + { + "epoch": 0.05228346456692914, + "grad_norm": 65.97516632080078, + "learning_rate": 9.999744881889765e-06, + "loss": 0.493, + "step": 166 + }, + { + "epoch": 0.052598425196850394, + "grad_norm": 74.50453186035156, + "learning_rate": 9.999743307086614e-06, + "loss": 0.5024, + "step": 167 + }, + { + "epoch": 0.05291338582677165, + "grad_norm": 79.50423431396484, + "learning_rate": 9.999741732283465e-06, + "loss": 0.683, + "step": 168 + }, + { + "epoch": 0.053228346456692915, + "grad_norm": 30.747211456298828, + "learning_rate": 9.999740157480316e-06, + "loss": 0.1244, + "step": 169 + }, + { + "epoch": 0.05354330708661417, + "grad_norm": 127.78273010253906, + "learning_rate": 9.999738582677167e-06, + "loss": 1.0797, + "step": 170 + }, + { + "epoch": 0.053858267716535436, + "grad_norm": 33.981021881103516, + "learning_rate": 9.999737007874016e-06, + "loss": 0.1601, + "step": 171 + }, + { + "epoch": 0.054173228346456694, + "grad_norm": 89.67857360839844, + "learning_rate": 9.999735433070867e-06, + "loss": 0.5669, + "step": 172 + }, + { + "epoch": 0.05448818897637795, + "grad_norm": 44.71755599975586, + "learning_rate": 9.999733858267717e-06, + "loss": 0.2556, + "step": 173 + }, + { + "epoch": 0.054803149606299215, + "grad_norm": 63.42751693725586, + "learning_rate": 9.999732283464568e-06, + "loss": 0.353, + "step": 174 + }, + { + "epoch": 0.05511811023622047, + "grad_norm": 145.7510986328125, + "learning_rate": 9.999730708661419e-06, + "loss": 0.6861, + "step": 175 + }, + { + "epoch": 0.055433070866141736, + "grad_norm": 86.73828125, + "learning_rate": 9.999729133858268e-06, + "loss": 0.7471, + "step": 176 + }, + { + "epoch": 0.05574803149606299, + "grad_norm": 78.96038055419922, + "learning_rate": 9.999727559055119e-06, + "loss": 0.5997, + "step": 177 + }, + { + "epoch": 0.05606299212598425, + "grad_norm": 77.76657104492188, + "learning_rate": 9.999725984251968e-06, + "loss": 0.4139, + "step": 178 + }, + { + "epoch": 0.056377952755905514, + "grad_norm": 56.66273880004883, + "learning_rate": 9.999724409448819e-06, + "loss": 0.3804, + "step": 179 + }, + { + "epoch": 0.05669291338582677, + "grad_norm": 45.65275192260742, + "learning_rate": 9.99972283464567e-06, + "loss": 0.2486, + "step": 180 + }, + { + "epoch": 0.05669291338582677, + "eval_loss": 0.6303219795227051, + "eval_runtime": 297.0224, + "eval_samples_per_second": 0.394, + "eval_steps_per_second": 0.394, + "step": 180 + }, + { + "epoch": 0.057007874015748035, + "grad_norm": 110.07376098632812, + "learning_rate": 9.999721259842521e-06, + "loss": 0.9565, + "step": 181 + }, + { + "epoch": 0.05732283464566929, + "grad_norm": 83.31719970703125, + "learning_rate": 9.99971968503937e-06, + "loss": 0.7403, + "step": 182 + }, + { + "epoch": 0.05763779527559055, + "grad_norm": 74.98153686523438, + "learning_rate": 9.999718110236221e-06, + "loss": 0.9278, + "step": 183 + }, + { + "epoch": 0.05795275590551181, + "grad_norm": 86.86173248291016, + "learning_rate": 9.999716535433072e-06, + "loss": 0.5827, + "step": 184 + }, + { + "epoch": 0.05826771653543307, + "grad_norm": 33.1976318359375, + "learning_rate": 9.999714960629922e-06, + "loss": 0.3653, + "step": 185 + }, + { + "epoch": 0.058582677165354334, + "grad_norm": 80.26583862304688, + "learning_rate": 9.999713385826773e-06, + "loss": 0.4716, + "step": 186 + }, + { + "epoch": 0.05889763779527559, + "grad_norm": 58.74989700317383, + "learning_rate": 9.999711811023622e-06, + "loss": 0.3977, + "step": 187 + }, + { + "epoch": 0.05921259842519685, + "grad_norm": 75.13705444335938, + "learning_rate": 9.999710236220473e-06, + "loss": 0.595, + "step": 188 + }, + { + "epoch": 0.05952755905511811, + "grad_norm": 44.060882568359375, + "learning_rate": 9.999708661417324e-06, + "loss": 0.3714, + "step": 189 + }, + { + "epoch": 0.05984251968503937, + "grad_norm": 35.92017364501953, + "learning_rate": 9.999707086614175e-06, + "loss": 0.4465, + "step": 190 + }, + { + "epoch": 0.06015748031496063, + "grad_norm": 35.966800689697266, + "learning_rate": 9.999705511811024e-06, + "loss": 0.4875, + "step": 191 + }, + { + "epoch": 0.06047244094488189, + "grad_norm": 48.5458869934082, + "learning_rate": 9.999703937007875e-06, + "loss": 0.6907, + "step": 192 + }, + { + "epoch": 0.06078740157480315, + "grad_norm": 38.40484619140625, + "learning_rate": 9.999702362204725e-06, + "loss": 0.3482, + "step": 193 + }, + { + "epoch": 0.06110236220472441, + "grad_norm": 87.23228454589844, + "learning_rate": 9.999700787401576e-06, + "loss": 0.4824, + "step": 194 + }, + { + "epoch": 0.06141732283464567, + "grad_norm": 62.46897888183594, + "learning_rate": 9.999699212598427e-06, + "loss": 0.3294, + "step": 195 + }, + { + "epoch": 0.06173228346456693, + "grad_norm": 33.665218353271484, + "learning_rate": 9.999697637795276e-06, + "loss": 0.222, + "step": 196 + }, + { + "epoch": 0.06204724409448819, + "grad_norm": 91.13434600830078, + "learning_rate": 9.999696062992127e-06, + "loss": 0.9194, + "step": 197 + }, + { + "epoch": 0.06236220472440945, + "grad_norm": 31.107872009277344, + "learning_rate": 9.999694488188976e-06, + "loss": 0.2354, + "step": 198 + }, + { + "epoch": 0.06267716535433071, + "grad_norm": 99.7812728881836, + "learning_rate": 9.999692913385827e-06, + "loss": 0.6351, + "step": 199 + }, + { + "epoch": 0.06299212598425197, + "grad_norm": 41.42717361450195, + "learning_rate": 9.999691338582678e-06, + "loss": 0.1414, + "step": 200 + }, + { + "epoch": 0.06299212598425197, + "eval_loss": 0.6573231220245361, + "eval_runtime": 302.1543, + "eval_samples_per_second": 0.387, + "eval_steps_per_second": 0.387, + "step": 200 + }, + { + "epoch": 0.06330708661417322, + "grad_norm": 61.020408630371094, + "learning_rate": 9.999689763779529e-06, + "loss": 0.6242, + "step": 201 + }, + { + "epoch": 0.06362204724409448, + "grad_norm": 70.84980010986328, + "learning_rate": 9.999688188976378e-06, + "loss": 0.1907, + "step": 202 + }, + { + "epoch": 0.06393700787401575, + "grad_norm": 174.39080810546875, + "learning_rate": 9.99968661417323e-06, + "loss": 0.9131, + "step": 203 + }, + { + "epoch": 0.06425196850393701, + "grad_norm": 47.28941345214844, + "learning_rate": 9.99968503937008e-06, + "loss": 0.2425, + "step": 204 + }, + { + "epoch": 0.06456692913385827, + "grad_norm": 51.628211975097656, + "learning_rate": 9.99968346456693e-06, + "loss": 0.2873, + "step": 205 + }, + { + "epoch": 0.06488188976377952, + "grad_norm": 63.0713996887207, + "learning_rate": 9.99968188976378e-06, + "loss": 0.6194, + "step": 206 + }, + { + "epoch": 0.06519685039370078, + "grad_norm": 164.83543395996094, + "learning_rate": 9.99968031496063e-06, + "loss": 0.5043, + "step": 207 + }, + { + "epoch": 0.06551181102362205, + "grad_norm": 83.96135711669922, + "learning_rate": 9.999678740157481e-06, + "loss": 0.6309, + "step": 208 + }, + { + "epoch": 0.06582677165354331, + "grad_norm": 81.93275451660156, + "learning_rate": 9.999677165354332e-06, + "loss": 0.4175, + "step": 209 + }, + { + "epoch": 0.06614173228346457, + "grad_norm": 129.6193389892578, + "learning_rate": 9.999675590551183e-06, + "loss": 0.5378, + "step": 210 + }, + { + "epoch": 0.06645669291338582, + "grad_norm": 137.78428649902344, + "learning_rate": 9.999674015748032e-06, + "loss": 1.2212, + "step": 211 + }, + { + "epoch": 0.06677165354330708, + "grad_norm": 42.24091339111328, + "learning_rate": 9.999672440944883e-06, + "loss": 0.173, + "step": 212 + }, + { + "epoch": 0.06708661417322835, + "grad_norm": 68.79737091064453, + "learning_rate": 9.999670866141732e-06, + "loss": 0.1975, + "step": 213 + }, + { + "epoch": 0.06740157480314961, + "grad_norm": 125.35755920410156, + "learning_rate": 9.999669291338583e-06, + "loss": 0.7453, + "step": 214 + }, + { + "epoch": 0.06771653543307087, + "grad_norm": 56.881229400634766, + "learning_rate": 9.999667716535434e-06, + "loss": 0.3073, + "step": 215 + }, + { + "epoch": 0.06803149606299212, + "grad_norm": 121.92823028564453, + "learning_rate": 9.999666141732284e-06, + "loss": 1.0795, + "step": 216 + }, + { + "epoch": 0.06834645669291338, + "grad_norm": 44.86691665649414, + "learning_rate": 9.999664566929135e-06, + "loss": 0.4312, + "step": 217 + }, + { + "epoch": 0.06866141732283465, + "grad_norm": 142.288330078125, + "learning_rate": 9.999662992125984e-06, + "loss": 0.6631, + "step": 218 + }, + { + "epoch": 0.06897637795275591, + "grad_norm": 200.1629180908203, + "learning_rate": 9.999661417322835e-06, + "loss": 0.5105, + "step": 219 + }, + { + "epoch": 0.06929133858267716, + "grad_norm": 83.08853149414062, + "learning_rate": 9.999659842519686e-06, + "loss": 0.6537, + "step": 220 + }, + { + "epoch": 0.06929133858267716, + "eval_loss": 0.7070333957672119, + "eval_runtime": 307.2171, + "eval_samples_per_second": 0.381, + "eval_steps_per_second": 0.381, + "step": 220 + }, + { + "epoch": 0.06960629921259842, + "grad_norm": 131.77316284179688, + "learning_rate": 9.999658267716537e-06, + "loss": 0.7263, + "step": 221 + }, + { + "epoch": 0.06992125984251968, + "grad_norm": 103.4114761352539, + "learning_rate": 9.999656692913386e-06, + "loss": 1.0884, + "step": 222 + }, + { + "epoch": 0.07023622047244095, + "grad_norm": 68.44525909423828, + "learning_rate": 9.999655118110237e-06, + "loss": 0.6249, + "step": 223 + }, + { + "epoch": 0.0705511811023622, + "grad_norm": 61.6135139465332, + "learning_rate": 9.999653543307087e-06, + "loss": 0.5604, + "step": 224 + }, + { + "epoch": 0.07086614173228346, + "grad_norm": 86.59762573242188, + "learning_rate": 9.999651968503938e-06, + "loss": 0.5263, + "step": 225 + }, + { + "epoch": 0.07118110236220472, + "grad_norm": 42.36429214477539, + "learning_rate": 9.999650393700789e-06, + "loss": 0.4644, + "step": 226 + }, + { + "epoch": 0.07149606299212598, + "grad_norm": 38.47148132324219, + "learning_rate": 9.999648818897638e-06, + "loss": 0.4286, + "step": 227 + }, + { + "epoch": 0.07181102362204725, + "grad_norm": 23.091997146606445, + "learning_rate": 9.999647244094489e-06, + "loss": 0.3635, + "step": 228 + }, + { + "epoch": 0.0721259842519685, + "grad_norm": 48.05474090576172, + "learning_rate": 9.99964566929134e-06, + "loss": 0.3296, + "step": 229 + }, + { + "epoch": 0.07244094488188976, + "grad_norm": 56.6866569519043, + "learning_rate": 9.99964409448819e-06, + "loss": 0.5982, + "step": 230 + }, + { + "epoch": 0.07275590551181102, + "grad_norm": 34.4522705078125, + "learning_rate": 9.99964251968504e-06, + "loss": 0.5496, + "step": 231 + }, + { + "epoch": 0.07307086614173228, + "grad_norm": 36.2459831237793, + "learning_rate": 9.999640944881891e-06, + "loss": 0.4069, + "step": 232 + }, + { + "epoch": 0.07338582677165355, + "grad_norm": 57.894195556640625, + "learning_rate": 9.99963937007874e-06, + "loss": 0.4789, + "step": 233 + }, + { + "epoch": 0.0737007874015748, + "grad_norm": 100.86152648925781, + "learning_rate": 9.999637795275591e-06, + "loss": 0.3101, + "step": 234 + }, + { + "epoch": 0.07401574803149606, + "grad_norm": 49.66980743408203, + "learning_rate": 9.999636220472442e-06, + "loss": 0.516, + "step": 235 + }, + { + "epoch": 0.07433070866141732, + "grad_norm": 52.82820510864258, + "learning_rate": 9.999634645669292e-06, + "loss": 0.3732, + "step": 236 + }, + { + "epoch": 0.07464566929133858, + "grad_norm": 56.593467712402344, + "learning_rate": 9.999633070866143e-06, + "loss": 0.3873, + "step": 237 + }, + { + "epoch": 0.07496062992125985, + "grad_norm": 20.434045791625977, + "learning_rate": 9.999631496062992e-06, + "loss": 0.2495, + "step": 238 + }, + { + "epoch": 0.0752755905511811, + "grad_norm": 65.34156799316406, + "learning_rate": 9.999629921259843e-06, + "loss": 0.6973, + "step": 239 + }, + { + "epoch": 0.07559055118110236, + "grad_norm": 32.1629638671875, + "learning_rate": 9.999628346456694e-06, + "loss": 0.3504, + "step": 240 + }, + { + "epoch": 0.07559055118110236, + "eval_loss": 0.7387034296989441, + "eval_runtime": 307.2747, + "eval_samples_per_second": 0.381, + "eval_steps_per_second": 0.381, + "step": 240 + }, + { + "epoch": 0.07590551181102362, + "grad_norm": 78.75101470947266, + "learning_rate": 9.999626771653545e-06, + "loss": 0.5959, + "step": 241 + }, + { + "epoch": 0.07622047244094488, + "grad_norm": 52.39651870727539, + "learning_rate": 9.999625196850394e-06, + "loss": 0.3459, + "step": 242 + }, + { + "epoch": 0.07653543307086615, + "grad_norm": 100.4014663696289, + "learning_rate": 9.999623622047245e-06, + "loss": 1.4185, + "step": 243 + }, + { + "epoch": 0.0768503937007874, + "grad_norm": 63.76593780517578, + "learning_rate": 9.999622047244095e-06, + "loss": 0.2557, + "step": 244 + }, + { + "epoch": 0.07716535433070866, + "grad_norm": 141.5568084716797, + "learning_rate": 9.999620472440946e-06, + "loss": 0.4333, + "step": 245 + }, + { + "epoch": 0.07748031496062992, + "grad_norm": 104.44050598144531, + "learning_rate": 9.999618897637797e-06, + "loss": 0.5989, + "step": 246 + }, + { + "epoch": 0.07779527559055117, + "grad_norm": 43.31072998046875, + "learning_rate": 9.999617322834646e-06, + "loss": 0.579, + "step": 247 + }, + { + "epoch": 0.07811023622047245, + "grad_norm": 47.9522819519043, + "learning_rate": 9.999615748031497e-06, + "loss": 0.2994, + "step": 248 + }, + { + "epoch": 0.0784251968503937, + "grad_norm": 40.461368560791016, + "learning_rate": 9.999614173228346e-06, + "loss": 0.4243, + "step": 249 + }, + { + "epoch": 0.07874015748031496, + "grad_norm": 26.12702751159668, + "learning_rate": 9.999612598425197e-06, + "loss": 0.3175, + "step": 250 + }, + { + "epoch": 0.07905511811023622, + "grad_norm": 76.74534606933594, + "learning_rate": 9.999611023622048e-06, + "loss": 0.6386, + "step": 251 + }, + { + "epoch": 0.07937007874015747, + "grad_norm": 61.15847396850586, + "learning_rate": 9.999609448818899e-06, + "loss": 0.2725, + "step": 252 + }, + { + "epoch": 0.07968503937007874, + "grad_norm": 119.91480255126953, + "learning_rate": 9.999607874015748e-06, + "loss": 0.2381, + "step": 253 + }, + { + "epoch": 0.08, + "grad_norm": 138.52313232421875, + "learning_rate": 9.9996062992126e-06, + "loss": 1.0807, + "step": 254 + }, + { + "epoch": 0.08031496062992126, + "grad_norm": 99.42451477050781, + "learning_rate": 9.99960472440945e-06, + "loss": 0.4007, + "step": 255 + }, + { + "epoch": 0.08062992125984252, + "grad_norm": 51.6858024597168, + "learning_rate": 9.9996031496063e-06, + "loss": 0.2963, + "step": 256 + }, + { + "epoch": 0.08094488188976377, + "grad_norm": 52.566734313964844, + "learning_rate": 9.99960157480315e-06, + "loss": 0.2889, + "step": 257 + }, + { + "epoch": 0.08125984251968504, + "grad_norm": 94.96017456054688, + "learning_rate": 9.9996e-06, + "loss": 1.354, + "step": 258 + }, + { + "epoch": 0.0815748031496063, + "grad_norm": 54.514915466308594, + "learning_rate": 9.999598425196851e-06, + "loss": 0.1296, + "step": 259 + }, + { + "epoch": 0.08188976377952756, + "grad_norm": 165.0517578125, + "learning_rate": 9.999596850393702e-06, + "loss": 0.7929, + "step": 260 + }, + { + "epoch": 0.08188976377952756, + "eval_loss": 0.767807126045227, + "eval_runtime": 307.1459, + "eval_samples_per_second": 0.381, + "eval_steps_per_second": 0.381, + "step": 260 + }, + { + "epoch": 0.08220472440944881, + "grad_norm": 62.38746643066406, + "learning_rate": 9.999595275590553e-06, + "loss": 0.2907, + "step": 261 + }, + { + "epoch": 0.08251968503937007, + "grad_norm": 107.24059295654297, + "learning_rate": 9.999593700787402e-06, + "loss": 0.9582, + "step": 262 + }, + { + "epoch": 0.08283464566929134, + "grad_norm": 182.7991943359375, + "learning_rate": 9.999592125984253e-06, + "loss": 1.1002, + "step": 263 + }, + { + "epoch": 0.0831496062992126, + "grad_norm": 39.42921829223633, + "learning_rate": 9.999590551181102e-06, + "loss": 0.3591, + "step": 264 + }, + { + "epoch": 0.08346456692913386, + "grad_norm": 78.57293701171875, + "learning_rate": 9.999588976377953e-06, + "loss": 0.5395, + "step": 265 + }, + { + "epoch": 0.08377952755905511, + "grad_norm": 46.55572509765625, + "learning_rate": 9.999587401574804e-06, + "loss": 0.4244, + "step": 266 + }, + { + "epoch": 0.08409448818897637, + "grad_norm": 72.65052795410156, + "learning_rate": 9.999585826771654e-06, + "loss": 0.5564, + "step": 267 + }, + { + "epoch": 0.08440944881889764, + "grad_norm": 40.577850341796875, + "learning_rate": 9.999584251968505e-06, + "loss": 0.319, + "step": 268 + }, + { + "epoch": 0.0847244094488189, + "grad_norm": 60.15317916870117, + "learning_rate": 9.999582677165354e-06, + "loss": 0.2762, + "step": 269 + }, + { + "epoch": 0.08503937007874016, + "grad_norm": 64.16828918457031, + "learning_rate": 9.999581102362205e-06, + "loss": 0.5016, + "step": 270 + }, + { + "epoch": 0.08535433070866141, + "grad_norm": 62.69063949584961, + "learning_rate": 9.999579527559056e-06, + "loss": 0.4175, + "step": 271 + }, + { + "epoch": 0.08566929133858267, + "grad_norm": 49.130157470703125, + "learning_rate": 9.999577952755907e-06, + "loss": 0.3619, + "step": 272 + }, + { + "epoch": 0.08598425196850394, + "grad_norm": 71.22623443603516, + "learning_rate": 9.999576377952756e-06, + "loss": 0.3924, + "step": 273 + }, + { + "epoch": 0.0862992125984252, + "grad_norm": 77.7140884399414, + "learning_rate": 9.999574803149607e-06, + "loss": 0.4717, + "step": 274 + }, + { + "epoch": 0.08661417322834646, + "grad_norm": 78.33636474609375, + "learning_rate": 9.999573228346458e-06, + "loss": 0.5174, + "step": 275 + }, + { + "epoch": 0.08692913385826771, + "grad_norm": 48.37542724609375, + "learning_rate": 9.999571653543308e-06, + "loss": 0.2733, + "step": 276 + }, + { + "epoch": 0.08724409448818897, + "grad_norm": 57.93960189819336, + "learning_rate": 9.999570078740159e-06, + "loss": 0.6626, + "step": 277 + }, + { + "epoch": 0.08755905511811024, + "grad_norm": 58.80123519897461, + "learning_rate": 9.999568503937008e-06, + "loss": 0.4598, + "step": 278 + }, + { + "epoch": 0.0878740157480315, + "grad_norm": 49.037818908691406, + "learning_rate": 9.999566929133859e-06, + "loss": 0.4452, + "step": 279 + }, + { + "epoch": 0.08818897637795275, + "grad_norm": 62.81136703491211, + "learning_rate": 9.99956535433071e-06, + "loss": 0.348, + "step": 280 + }, + { + "epoch": 0.08818897637795275, + "eval_loss": 0.618428111076355, + "eval_runtime": 306.6914, + "eval_samples_per_second": 0.381, + "eval_steps_per_second": 0.381, + "step": 280 + }, + { + "epoch": 0.08850393700787401, + "grad_norm": 61.92344284057617, + "learning_rate": 9.999563779527561e-06, + "loss": 0.2919, + "step": 281 + }, + { + "epoch": 0.08881889763779527, + "grad_norm": 75.2997817993164, + "learning_rate": 9.99956220472441e-06, + "loss": 0.4883, + "step": 282 + }, + { + "epoch": 0.08913385826771654, + "grad_norm": 34.474639892578125, + "learning_rate": 9.999560629921261e-06, + "loss": 0.1744, + "step": 283 + }, + { + "epoch": 0.0894488188976378, + "grad_norm": 79.71351623535156, + "learning_rate": 9.99955905511811e-06, + "loss": 0.4965, + "step": 284 + }, + { + "epoch": 0.08976377952755905, + "grad_norm": 127.44145202636719, + "learning_rate": 9.999557480314961e-06, + "loss": 0.7171, + "step": 285 + }, + { + "epoch": 0.09007874015748031, + "grad_norm": 57.06454086303711, + "learning_rate": 9.999555905511812e-06, + "loss": 0.5155, + "step": 286 + }, + { + "epoch": 0.09039370078740157, + "grad_norm": 86.23944854736328, + "learning_rate": 9.999554330708662e-06, + "loss": 0.613, + "step": 287 + }, + { + "epoch": 0.09070866141732284, + "grad_norm": 188.0777587890625, + "learning_rate": 9.999552755905513e-06, + "loss": 0.3885, + "step": 288 + }, + { + "epoch": 0.0910236220472441, + "grad_norm": 49.08794403076172, + "learning_rate": 9.999551181102362e-06, + "loss": 0.3903, + "step": 289 + }, + { + "epoch": 0.09133858267716535, + "grad_norm": 77.4734115600586, + "learning_rate": 9.999549606299213e-06, + "loss": 0.629, + "step": 290 + }, + { + "epoch": 0.09165354330708661, + "grad_norm": 31.26721954345703, + "learning_rate": 9.999548031496064e-06, + "loss": 0.2508, + "step": 291 + }, + { + "epoch": 0.09196850393700787, + "grad_norm": 59.19281768798828, + "learning_rate": 9.999546456692915e-06, + "loss": 0.1862, + "step": 292 + }, + { + "epoch": 0.09228346456692914, + "grad_norm": 97.84364318847656, + "learning_rate": 9.999544881889764e-06, + "loss": 0.8973, + "step": 293 + }, + { + "epoch": 0.0925984251968504, + "grad_norm": 94.82438659667969, + "learning_rate": 9.999543307086615e-06, + "loss": 0.8992, + "step": 294 + }, + { + "epoch": 0.09291338582677165, + "grad_norm": 57.59076690673828, + "learning_rate": 9.999541732283465e-06, + "loss": 0.6437, + "step": 295 + }, + { + "epoch": 0.09322834645669291, + "grad_norm": 37.8861198425293, + "learning_rate": 9.999540157480316e-06, + "loss": 0.4185, + "step": 296 + }, + { + "epoch": 0.09354330708661417, + "grad_norm": 28.94227409362793, + "learning_rate": 9.999538582677167e-06, + "loss": 0.285, + "step": 297 + }, + { + "epoch": 0.09385826771653544, + "grad_norm": 50.66032409667969, + "learning_rate": 9.999537007874016e-06, + "loss": 0.6087, + "step": 298 + }, + { + "epoch": 0.0941732283464567, + "grad_norm": 24.23774528503418, + "learning_rate": 9.999535433070867e-06, + "loss": 0.2489, + "step": 299 + }, + { + "epoch": 0.09448818897637795, + "grad_norm": 50.08018493652344, + "learning_rate": 9.999533858267718e-06, + "loss": 0.6251, + "step": 300 + }, + { + "epoch": 0.09448818897637795, + "eval_loss": 0.5396940112113953, + "eval_runtime": 294.9781, + "eval_samples_per_second": 0.397, + "eval_steps_per_second": 0.397, + "step": 300 + }, + { + "epoch": 0.09480314960629921, + "grad_norm": 87.98992919921875, + "learning_rate": 9.999532283464569e-06, + "loss": 0.6764, + "step": 301 + }, + { + "epoch": 0.09511811023622047, + "grad_norm": 47.72505187988281, + "learning_rate": 9.999530708661418e-06, + "loss": 0.754, + "step": 302 + }, + { + "epoch": 0.09543307086614174, + "grad_norm": 29.56645393371582, + "learning_rate": 9.999529133858269e-06, + "loss": 0.2754, + "step": 303 + }, + { + "epoch": 0.095748031496063, + "grad_norm": 66.290283203125, + "learning_rate": 9.999527559055118e-06, + "loss": 0.5663, + "step": 304 + }, + { + "epoch": 0.09606299212598425, + "grad_norm": 38.929725646972656, + "learning_rate": 9.99952598425197e-06, + "loss": 0.3187, + "step": 305 + }, + { + "epoch": 0.09637795275590551, + "grad_norm": 55.97653579711914, + "learning_rate": 9.99952440944882e-06, + "loss": 0.4682, + "step": 306 + }, + { + "epoch": 0.09669291338582676, + "grad_norm": 41.88676071166992, + "learning_rate": 9.99952283464567e-06, + "loss": 0.4863, + "step": 307 + }, + { + "epoch": 0.09700787401574804, + "grad_norm": 39.72370529174805, + "learning_rate": 9.99952125984252e-06, + "loss": 0.3445, + "step": 308 + }, + { + "epoch": 0.09732283464566929, + "grad_norm": 48.20722579956055, + "learning_rate": 9.99951968503937e-06, + "loss": 0.4213, + "step": 309 + }, + { + "epoch": 0.09763779527559055, + "grad_norm": 54.454715728759766, + "learning_rate": 9.999518110236221e-06, + "loss": 0.4397, + "step": 310 + }, + { + "epoch": 0.0979527559055118, + "grad_norm": 64.91082000732422, + "learning_rate": 9.999516535433072e-06, + "loss": 0.2707, + "step": 311 + }, + { + "epoch": 0.09826771653543306, + "grad_norm": 94.55459594726562, + "learning_rate": 9.999514960629923e-06, + "loss": 0.3693, + "step": 312 + }, + { + "epoch": 0.09858267716535433, + "grad_norm": 96.01959991455078, + "learning_rate": 9.999513385826772e-06, + "loss": 0.5796, + "step": 313 + }, + { + "epoch": 0.09889763779527559, + "grad_norm": 44.687355041503906, + "learning_rate": 9.999511811023623e-06, + "loss": 0.4414, + "step": 314 + }, + { + "epoch": 0.09921259842519685, + "grad_norm": 108.15480041503906, + "learning_rate": 9.999510236220473e-06, + "loss": 0.4173, + "step": 315 + }, + { + "epoch": 0.0995275590551181, + "grad_norm": 42.95850372314453, + "learning_rate": 9.999508661417324e-06, + "loss": 0.1891, + "step": 316 + }, + { + "epoch": 0.09984251968503936, + "grad_norm": 66.48217010498047, + "learning_rate": 9.999507086614174e-06, + "loss": 0.2298, + "step": 317 + }, + { + "epoch": 0.10015748031496063, + "grad_norm": 152.71719360351562, + "learning_rate": 9.999505511811024e-06, + "loss": 0.4982, + "step": 318 + }, + { + "epoch": 0.10047244094488189, + "grad_norm": 68.98497772216797, + "learning_rate": 9.999503937007875e-06, + "loss": 0.6307, + "step": 319 + }, + { + "epoch": 0.10078740157480315, + "grad_norm": 68.25215911865234, + "learning_rate": 9.999502362204724e-06, + "loss": 0.2943, + "step": 320 + }, + { + "epoch": 0.10078740157480315, + "eval_loss": 0.6371558904647827, + "eval_runtime": 297.2013, + "eval_samples_per_second": 0.394, + "eval_steps_per_second": 0.394, + "step": 320 + }, + { + "epoch": 0.1011023622047244, + "grad_norm": 64.41675567626953, + "learning_rate": 9.999500787401577e-06, + "loss": 0.3071, + "step": 321 + }, + { + "epoch": 0.10141732283464568, + "grad_norm": 146.19937133789062, + "learning_rate": 9.999499212598426e-06, + "loss": 0.1765, + "step": 322 + }, + { + "epoch": 0.10173228346456693, + "grad_norm": 81.2210464477539, + "learning_rate": 9.999497637795277e-06, + "loss": 0.854, + "step": 323 + }, + { + "epoch": 0.10204724409448819, + "grad_norm": 142.67649841308594, + "learning_rate": 9.999496062992126e-06, + "loss": 0.431, + "step": 324 + }, + { + "epoch": 0.10236220472440945, + "grad_norm": 23.75156593322754, + "learning_rate": 9.999494488188977e-06, + "loss": 0.0743, + "step": 325 + }, + { + "epoch": 0.1026771653543307, + "grad_norm": 154.65882873535156, + "learning_rate": 9.999492913385828e-06, + "loss": 0.7567, + "step": 326 + }, + { + "epoch": 0.10299212598425198, + "grad_norm": 284.99822998046875, + "learning_rate": 9.999491338582678e-06, + "loss": 1.2282, + "step": 327 + }, + { + "epoch": 0.10330708661417323, + "grad_norm": 87.0584945678711, + "learning_rate": 9.999489763779529e-06, + "loss": 1.0155, + "step": 328 + }, + { + "epoch": 0.10362204724409449, + "grad_norm": 54.77091598510742, + "learning_rate": 9.999488188976378e-06, + "loss": 0.2331, + "step": 329 + }, + { + "epoch": 0.10393700787401575, + "grad_norm": 93.18637084960938, + "learning_rate": 9.999486614173229e-06, + "loss": 0.5502, + "step": 330 + }, + { + "epoch": 0.104251968503937, + "grad_norm": 70.45845031738281, + "learning_rate": 9.99948503937008e-06, + "loss": 0.5354, + "step": 331 + }, + { + "epoch": 0.10456692913385827, + "grad_norm": 55.51031494140625, + "learning_rate": 9.999483464566931e-06, + "loss": 0.5554, + "step": 332 + }, + { + "epoch": 0.10488188976377953, + "grad_norm": 93.07254028320312, + "learning_rate": 9.99948188976378e-06, + "loss": 0.521, + "step": 333 + }, + { + "epoch": 0.10519685039370079, + "grad_norm": 31.241605758666992, + "learning_rate": 9.999480314960631e-06, + "loss": 0.2664, + "step": 334 + }, + { + "epoch": 0.10551181102362205, + "grad_norm": 54.83103942871094, + "learning_rate": 9.99947874015748e-06, + "loss": 0.4913, + "step": 335 + }, + { + "epoch": 0.1058267716535433, + "grad_norm": 55.371360778808594, + "learning_rate": 9.999477165354331e-06, + "loss": 0.5836, + "step": 336 + }, + { + "epoch": 0.10614173228346457, + "grad_norm": 77.42748260498047, + "learning_rate": 9.999475590551182e-06, + "loss": 0.4551, + "step": 337 + }, + { + "epoch": 0.10645669291338583, + "grad_norm": 52.34659194946289, + "learning_rate": 9.999474015748032e-06, + "loss": 0.3793, + "step": 338 + }, + { + "epoch": 0.10677165354330709, + "grad_norm": 69.15506744384766, + "learning_rate": 9.999472440944883e-06, + "loss": 0.4266, + "step": 339 + }, + { + "epoch": 0.10708661417322834, + "grad_norm": 59.99565887451172, + "learning_rate": 9.999470866141732e-06, + "loss": 0.3804, + "step": 340 + }, + { + "epoch": 0.10708661417322834, + "eval_loss": 0.5206155776977539, + "eval_runtime": 293.7622, + "eval_samples_per_second": 0.398, + "eval_steps_per_second": 0.398, + "step": 340 + }, + { + "epoch": 0.1074015748031496, + "grad_norm": 62.315391540527344, + "learning_rate": 9.999469291338583e-06, + "loss": 0.5939, + "step": 341 + }, + { + "epoch": 0.10771653543307087, + "grad_norm": 45.72246551513672, + "learning_rate": 9.999467716535434e-06, + "loss": 0.2962, + "step": 342 + }, + { + "epoch": 0.10803149606299213, + "grad_norm": 72.52598571777344, + "learning_rate": 9.999466141732285e-06, + "loss": 0.773, + "step": 343 + }, + { + "epoch": 0.10834645669291339, + "grad_norm": 50.78411102294922, + "learning_rate": 9.999464566929134e-06, + "loss": 0.4392, + "step": 344 + }, + { + "epoch": 0.10866141732283464, + "grad_norm": 70.54865264892578, + "learning_rate": 9.999462992125985e-06, + "loss": 0.5495, + "step": 345 + }, + { + "epoch": 0.1089763779527559, + "grad_norm": 42.886756896972656, + "learning_rate": 9.999461417322836e-06, + "loss": 0.3558, + "step": 346 + }, + { + "epoch": 0.10929133858267717, + "grad_norm": 61.35227584838867, + "learning_rate": 9.999459842519686e-06, + "loss": 0.583, + "step": 347 + }, + { + "epoch": 0.10960629921259843, + "grad_norm": 75.92544555664062, + "learning_rate": 9.999458267716537e-06, + "loss": 0.6637, + "step": 348 + }, + { + "epoch": 0.10992125984251969, + "grad_norm": 47.078548431396484, + "learning_rate": 9.999456692913386e-06, + "loss": 0.5437, + "step": 349 + }, + { + "epoch": 0.11023622047244094, + "grad_norm": 46.12405014038086, + "learning_rate": 9.999455118110237e-06, + "loss": 0.463, + "step": 350 + }, + { + "epoch": 0.1105511811023622, + "grad_norm": 49.5578727722168, + "learning_rate": 9.999453543307088e-06, + "loss": 0.2595, + "step": 351 + }, + { + "epoch": 0.11086614173228347, + "grad_norm": 42.03670883178711, + "learning_rate": 9.999451968503939e-06, + "loss": 0.3669, + "step": 352 + }, + { + "epoch": 0.11118110236220473, + "grad_norm": 55.1522102355957, + "learning_rate": 9.999450393700788e-06, + "loss": 0.5018, + "step": 353 + }, + { + "epoch": 0.11149606299212599, + "grad_norm": 125.36481475830078, + "learning_rate": 9.999448818897639e-06, + "loss": 0.4888, + "step": 354 + }, + { + "epoch": 0.11181102362204724, + "grad_norm": 81.62045288085938, + "learning_rate": 9.999447244094488e-06, + "loss": 0.4195, + "step": 355 + }, + { + "epoch": 0.1121259842519685, + "grad_norm": 85.13298797607422, + "learning_rate": 9.99944566929134e-06, + "loss": 0.4851, + "step": 356 + }, + { + "epoch": 0.11244094488188977, + "grad_norm": 245.27197265625, + "learning_rate": 9.99944409448819e-06, + "loss": 0.5387, + "step": 357 + }, + { + "epoch": 0.11275590551181103, + "grad_norm": 71.68444061279297, + "learning_rate": 9.99944251968504e-06, + "loss": 0.2865, + "step": 358 + }, + { + "epoch": 0.11307086614173228, + "grad_norm": 44.38494110107422, + "learning_rate": 9.99944094488189e-06, + "loss": 0.4339, + "step": 359 + }, + { + "epoch": 0.11338582677165354, + "grad_norm": 62.24411392211914, + "learning_rate": 9.99943937007874e-06, + "loss": 0.4798, + "step": 360 + }, + { + "epoch": 0.11338582677165354, + "eval_loss": 0.5544171929359436, + "eval_runtime": 293.2522, + "eval_samples_per_second": 0.399, + "eval_steps_per_second": 0.399, + "step": 360 + }, + { + "epoch": 0.1137007874015748, + "grad_norm": 97.97586059570312, + "learning_rate": 9.999437795275591e-06, + "loss": 0.8155, + "step": 361 + }, + { + "epoch": 0.11401574803149607, + "grad_norm": 101.96649169921875, + "learning_rate": 9.999436220472442e-06, + "loss": 0.3751, + "step": 362 + }, + { + "epoch": 0.11433070866141733, + "grad_norm": 109.22547912597656, + "learning_rate": 9.999434645669293e-06, + "loss": 0.9483, + "step": 363 + }, + { + "epoch": 0.11464566929133858, + "grad_norm": 101.05289459228516, + "learning_rate": 9.999433070866142e-06, + "loss": 0.4469, + "step": 364 + }, + { + "epoch": 0.11496062992125984, + "grad_norm": 97.29914855957031, + "learning_rate": 9.999431496062993e-06, + "loss": 0.4986, + "step": 365 + }, + { + "epoch": 0.1152755905511811, + "grad_norm": 52.88810729980469, + "learning_rate": 9.999429921259843e-06, + "loss": 0.239, + "step": 366 + }, + { + "epoch": 0.11559055118110237, + "grad_norm": 220.28375244140625, + "learning_rate": 9.999428346456694e-06, + "loss": 0.9343, + "step": 367 + }, + { + "epoch": 0.11590551181102363, + "grad_norm": 128.3485565185547, + "learning_rate": 9.999426771653545e-06, + "loss": 0.6493, + "step": 368 + }, + { + "epoch": 0.11622047244094488, + "grad_norm": 68.30548095703125, + "learning_rate": 9.999425196850394e-06, + "loss": 0.6118, + "step": 369 + }, + { + "epoch": 0.11653543307086614, + "grad_norm": 85.01860809326172, + "learning_rate": 9.999423622047245e-06, + "loss": 0.9292, + "step": 370 + }, + { + "epoch": 0.1168503937007874, + "grad_norm": 42.967952728271484, + "learning_rate": 9.999422047244096e-06, + "loss": 0.3607, + "step": 371 + }, + { + "epoch": 0.11716535433070867, + "grad_norm": 31.546159744262695, + "learning_rate": 9.999420472440947e-06, + "loss": 0.1924, + "step": 372 + }, + { + "epoch": 0.11748031496062993, + "grad_norm": 49.90913391113281, + "learning_rate": 9.999418897637796e-06, + "loss": 0.3681, + "step": 373 + }, + { + "epoch": 0.11779527559055118, + "grad_norm": 43.57588195800781, + "learning_rate": 9.999417322834647e-06, + "loss": 0.2731, + "step": 374 + }, + { + "epoch": 0.11811023622047244, + "grad_norm": 53.5254020690918, + "learning_rate": 9.999415748031496e-06, + "loss": 0.5743, + "step": 375 + }, + { + "epoch": 0.1184251968503937, + "grad_norm": 57.816184997558594, + "learning_rate": 9.999414173228347e-06, + "loss": 0.6925, + "step": 376 + }, + { + "epoch": 0.11874015748031497, + "grad_norm": 50.099021911621094, + "learning_rate": 9.999412598425198e-06, + "loss": 0.4436, + "step": 377 + }, + { + "epoch": 0.11905511811023622, + "grad_norm": 38.81980514526367, + "learning_rate": 9.999411023622048e-06, + "loss": 0.3597, + "step": 378 + }, + { + "epoch": 0.11937007874015748, + "grad_norm": 60.50627517700195, + "learning_rate": 9.999409448818899e-06, + "loss": 0.3588, + "step": 379 + }, + { + "epoch": 0.11968503937007874, + "grad_norm": 88.20054626464844, + "learning_rate": 9.999407874015748e-06, + "loss": 0.3192, + "step": 380 + }, + { + "epoch": 0.11968503937007874, + "eval_loss": 0.5467623472213745, + "eval_runtime": 292.898, + "eval_samples_per_second": 0.399, + "eval_steps_per_second": 0.399, + "step": 380 + }, + { + "epoch": 0.12, + "grad_norm": 39.78269577026367, + "learning_rate": 9.999406299212599e-06, + "loss": 0.4447, + "step": 381 + }, + { + "epoch": 0.12031496062992127, + "grad_norm": 61.480377197265625, + "learning_rate": 9.99940472440945e-06, + "loss": 0.3143, + "step": 382 + }, + { + "epoch": 0.12062992125984252, + "grad_norm": 200.373779296875, + "learning_rate": 9.999403149606301e-06, + "loss": 1.3858, + "step": 383 + }, + { + "epoch": 0.12094488188976378, + "grad_norm": 162.78187561035156, + "learning_rate": 9.99940157480315e-06, + "loss": 0.6117, + "step": 384 + }, + { + "epoch": 0.12125984251968504, + "grad_norm": 42.38726043701172, + "learning_rate": 9.999400000000001e-06, + "loss": 0.3125, + "step": 385 + }, + { + "epoch": 0.1215748031496063, + "grad_norm": 47.82020568847656, + "learning_rate": 9.99939842519685e-06, + "loss": 0.3962, + "step": 386 + }, + { + "epoch": 0.12188976377952757, + "grad_norm": 106.18301391601562, + "learning_rate": 9.999396850393701e-06, + "loss": 0.6107, + "step": 387 + }, + { + "epoch": 0.12220472440944882, + "grad_norm": 57.39361572265625, + "learning_rate": 9.999395275590552e-06, + "loss": 0.2136, + "step": 388 + }, + { + "epoch": 0.12251968503937008, + "grad_norm": 32.14236068725586, + "learning_rate": 9.999393700787402e-06, + "loss": 0.0976, + "step": 389 + }, + { + "epoch": 0.12283464566929134, + "grad_norm": 13.914974212646484, + "learning_rate": 9.999392125984253e-06, + "loss": 0.0404, + "step": 390 + }, + { + "epoch": 0.1231496062992126, + "grad_norm": 67.5541000366211, + "learning_rate": 9.999390551181104e-06, + "loss": 0.5028, + "step": 391 + }, + { + "epoch": 0.12346456692913386, + "grad_norm": 108.35496520996094, + "learning_rate": 9.999388976377955e-06, + "loss": 0.686, + "step": 392 + }, + { + "epoch": 0.12377952755905512, + "grad_norm": 69.92194366455078, + "learning_rate": 9.999387401574804e-06, + "loss": 0.2082, + "step": 393 + }, + { + "epoch": 0.12409448818897638, + "grad_norm": 23.27518081665039, + "learning_rate": 9.999385826771655e-06, + "loss": 0.0539, + "step": 394 + }, + { + "epoch": 0.12440944881889764, + "grad_norm": 48.74710464477539, + "learning_rate": 9.999384251968504e-06, + "loss": 0.3422, + "step": 395 + }, + { + "epoch": 0.1247244094488189, + "grad_norm": 138.21983337402344, + "learning_rate": 9.999382677165355e-06, + "loss": 0.8565, + "step": 396 + }, + { + "epoch": 0.12503937007874016, + "grad_norm": 112.90072631835938, + "learning_rate": 9.999381102362206e-06, + "loss": 0.2024, + "step": 397 + }, + { + "epoch": 0.12535433070866142, + "grad_norm": 256.4333190917969, + "learning_rate": 9.999379527559056e-06, + "loss": 1.8303, + "step": 398 + }, + { + "epoch": 0.12566929133858268, + "grad_norm": 112.24787902832031, + "learning_rate": 9.999377952755907e-06, + "loss": 0.3992, + "step": 399 + }, + { + "epoch": 0.12598425196850394, + "grad_norm": 131.27490234375, + "learning_rate": 9.999376377952756e-06, + "loss": 0.5066, + "step": 400 + }, + { + "epoch": 0.12598425196850394, + "eval_loss": 0.6591930985450745, + "eval_runtime": 304.46, + "eval_samples_per_second": 0.384, + "eval_steps_per_second": 0.384, + "step": 400 + }, + { + "epoch": 0.1262992125984252, + "grad_norm": 194.88157653808594, + "learning_rate": 9.999374803149607e-06, + "loss": 1.5517, + "step": 401 + }, + { + "epoch": 0.12661417322834645, + "grad_norm": 57.20365524291992, + "learning_rate": 9.999373228346458e-06, + "loss": 0.4244, + "step": 402 + }, + { + "epoch": 0.1269291338582677, + "grad_norm": 36.50507736206055, + "learning_rate": 9.999371653543309e-06, + "loss": 0.1584, + "step": 403 + }, + { + "epoch": 0.12724409448818896, + "grad_norm": 133.4579620361328, + "learning_rate": 9.999370078740158e-06, + "loss": 1.333, + "step": 404 + }, + { + "epoch": 0.12755905511811025, + "grad_norm": 128.0843505859375, + "learning_rate": 9.999368503937009e-06, + "loss": 0.4923, + "step": 405 + }, + { + "epoch": 0.1278740157480315, + "grad_norm": 100.4643325805664, + "learning_rate": 9.999366929133858e-06, + "loss": 0.4229, + "step": 406 + }, + { + "epoch": 0.12818897637795276, + "grad_norm": 51.65610885620117, + "learning_rate": 9.99936535433071e-06, + "loss": 0.3144, + "step": 407 + }, + { + "epoch": 0.12850393700787402, + "grad_norm": 86.17060852050781, + "learning_rate": 9.99936377952756e-06, + "loss": 0.5492, + "step": 408 + }, + { + "epoch": 0.12881889763779528, + "grad_norm": 11.016338348388672, + "learning_rate": 9.99936220472441e-06, + "loss": 0.0502, + "step": 409 + }, + { + "epoch": 0.12913385826771653, + "grad_norm": 75.9980697631836, + "learning_rate": 9.99936062992126e-06, + "loss": 0.47, + "step": 410 + }, + { + "epoch": 0.1294488188976378, + "grad_norm": 64.87591552734375, + "learning_rate": 9.99935905511811e-06, + "loss": 0.2592, + "step": 411 + }, + { + "epoch": 0.12976377952755905, + "grad_norm": 109.087646484375, + "learning_rate": 9.999357480314961e-06, + "loss": 0.7454, + "step": 412 + }, + { + "epoch": 0.1300787401574803, + "grad_norm": 46.490875244140625, + "learning_rate": 9.999355905511812e-06, + "loss": 0.2048, + "step": 413 + }, + { + "epoch": 0.13039370078740156, + "grad_norm": 156.15184020996094, + "learning_rate": 9.999354330708663e-06, + "loss": 0.4924, + "step": 414 + }, + { + "epoch": 0.13070866141732285, + "grad_norm": 120.5556640625, + "learning_rate": 9.999352755905512e-06, + "loss": 0.8374, + "step": 415 + }, + { + "epoch": 0.1310236220472441, + "grad_norm": 151.4144744873047, + "learning_rate": 9.999351181102363e-06, + "loss": 0.6144, + "step": 416 + }, + { + "epoch": 0.13133858267716536, + "grad_norm": 90.7903060913086, + "learning_rate": 9.999349606299214e-06, + "loss": 0.2709, + "step": 417 + }, + { + "epoch": 0.13165354330708662, + "grad_norm": 27.181320190429688, + "learning_rate": 9.999348031496064e-06, + "loss": 0.3197, + "step": 418 + }, + { + "epoch": 0.13196850393700787, + "grad_norm": 98.53707885742188, + "learning_rate": 9.999346456692915e-06, + "loss": 0.8255, + "step": 419 + }, + { + "epoch": 0.13228346456692913, + "grad_norm": 34.432132720947266, + "learning_rate": 9.999344881889764e-06, + "loss": 0.1117, + "step": 420 + }, + { + "epoch": 0.13228346456692913, + "eval_loss": 0.535347580909729, + "eval_runtime": 306.3308, + "eval_samples_per_second": 0.382, + "eval_steps_per_second": 0.382, + "step": 420 + }, + { + "epoch": 0.1325984251968504, + "grad_norm": 42.67380142211914, + "learning_rate": 9.999343307086615e-06, + "loss": 0.3283, + "step": 421 + }, + { + "epoch": 0.13291338582677165, + "grad_norm": 71.41397857666016, + "learning_rate": 9.999341732283466e-06, + "loss": 0.295, + "step": 422 + }, + { + "epoch": 0.1332283464566929, + "grad_norm": 28.39377784729004, + "learning_rate": 9.999340157480317e-06, + "loss": 0.1143, + "step": 423 + }, + { + "epoch": 0.13354330708661416, + "grad_norm": 61.82292556762695, + "learning_rate": 9.999338582677166e-06, + "loss": 0.5727, + "step": 424 + }, + { + "epoch": 0.13385826771653545, + "grad_norm": 69.15302276611328, + "learning_rate": 9.999337007874017e-06, + "loss": 0.3849, + "step": 425 + }, + { + "epoch": 0.1341732283464567, + "grad_norm": 38.20407485961914, + "learning_rate": 9.999335433070866e-06, + "loss": 0.334, + "step": 426 + }, + { + "epoch": 0.13448818897637796, + "grad_norm": 25.310161590576172, + "learning_rate": 9.999333858267717e-06, + "loss": 0.1519, + "step": 427 + }, + { + "epoch": 0.13480314960629922, + "grad_norm": 28.452911376953125, + "learning_rate": 9.999332283464568e-06, + "loss": 0.2238, + "step": 428 + }, + { + "epoch": 0.13511811023622047, + "grad_norm": 127.5716552734375, + "learning_rate": 9.999330708661418e-06, + "loss": 1.3774, + "step": 429 + }, + { + "epoch": 0.13543307086614173, + "grad_norm": 60.44778060913086, + "learning_rate": 9.999329133858269e-06, + "loss": 0.7446, + "step": 430 + }, + { + "epoch": 0.135748031496063, + "grad_norm": 54.13089370727539, + "learning_rate": 9.999327559055118e-06, + "loss": 0.7198, + "step": 431 + }, + { + "epoch": 0.13606299212598424, + "grad_norm": 44.978485107421875, + "learning_rate": 9.999325984251969e-06, + "loss": 0.457, + "step": 432 + }, + { + "epoch": 0.1363779527559055, + "grad_norm": 78.32373809814453, + "learning_rate": 9.99932440944882e-06, + "loss": 0.5859, + "step": 433 + }, + { + "epoch": 0.13669291338582676, + "grad_norm": 132.4387969970703, + "learning_rate": 9.999322834645671e-06, + "loss": 0.4658, + "step": 434 + }, + { + "epoch": 0.13700787401574804, + "grad_norm": 14.64592170715332, + "learning_rate": 9.99932125984252e-06, + "loss": 0.0912, + "step": 435 + }, + { + "epoch": 0.1373228346456693, + "grad_norm": 57.1812858581543, + "learning_rate": 9.999319685039371e-06, + "loss": 0.4572, + "step": 436 + }, + { + "epoch": 0.13763779527559056, + "grad_norm": 46.73292541503906, + "learning_rate": 9.99931811023622e-06, + "loss": 0.4464, + "step": 437 + }, + { + "epoch": 0.13795275590551181, + "grad_norm": 21.240659713745117, + "learning_rate": 9.999316535433072e-06, + "loss": 0.1034, + "step": 438 + }, + { + "epoch": 0.13826771653543307, + "grad_norm": 46.250614166259766, + "learning_rate": 9.999314960629922e-06, + "loss": 0.2897, + "step": 439 + }, + { + "epoch": 0.13858267716535433, + "grad_norm": 73.32185363769531, + "learning_rate": 9.999313385826772e-06, + "loss": 0.5788, + "step": 440 + }, + { + "epoch": 0.13858267716535433, + "eval_loss": 0.5959511399269104, + "eval_runtime": 304.5113, + "eval_samples_per_second": 0.384, + "eval_steps_per_second": 0.384, + "step": 440 + }, + { + "epoch": 0.13889763779527559, + "grad_norm": 7.073561191558838, + "learning_rate": 9.999311811023623e-06, + "loss": 0.0235, + "step": 441 + }, + { + "epoch": 0.13921259842519684, + "grad_norm": 130.6144561767578, + "learning_rate": 9.999310236220474e-06, + "loss": 1.088, + "step": 442 + }, + { + "epoch": 0.1395275590551181, + "grad_norm": 105.4767837524414, + "learning_rate": 9.999308661417325e-06, + "loss": 1.3411, + "step": 443 + }, + { + "epoch": 0.13984251968503936, + "grad_norm": 45.18183898925781, + "learning_rate": 9.999307086614174e-06, + "loss": 0.2972, + "step": 444 + }, + { + "epoch": 0.14015748031496064, + "grad_norm": 84.5200424194336, + "learning_rate": 9.999305511811025e-06, + "loss": 0.3241, + "step": 445 + }, + { + "epoch": 0.1404724409448819, + "grad_norm": 33.606468200683594, + "learning_rate": 9.999303937007874e-06, + "loss": 0.3225, + "step": 446 + }, + { + "epoch": 0.14078740157480316, + "grad_norm": 156.7371063232422, + "learning_rate": 9.999302362204725e-06, + "loss": 0.8385, + "step": 447 + }, + { + "epoch": 0.1411023622047244, + "grad_norm": 51.86471176147461, + "learning_rate": 9.999300787401576e-06, + "loss": 0.3886, + "step": 448 + }, + { + "epoch": 0.14141732283464567, + "grad_norm": 98.36966705322266, + "learning_rate": 9.999299212598426e-06, + "loss": 0.3667, + "step": 449 + }, + { + "epoch": 0.14173228346456693, + "grad_norm": 163.55325317382812, + "learning_rate": 9.999297637795277e-06, + "loss": 0.9837, + "step": 450 + }, + { + "epoch": 0.14204724409448818, + "grad_norm": 52.89970779418945, + "learning_rate": 9.999296062992126e-06, + "loss": 0.3358, + "step": 451 + }, + { + "epoch": 0.14236220472440944, + "grad_norm": 71.64704132080078, + "learning_rate": 9.999294488188977e-06, + "loss": 0.7998, + "step": 452 + }, + { + "epoch": 0.1426771653543307, + "grad_norm": 19.4017276763916, + "learning_rate": 9.999292913385828e-06, + "loss": 0.0742, + "step": 453 + }, + { + "epoch": 0.14299212598425196, + "grad_norm": 89.31649017333984, + "learning_rate": 9.999291338582679e-06, + "loss": 0.557, + "step": 454 + }, + { + "epoch": 0.14330708661417324, + "grad_norm": 47.06895446777344, + "learning_rate": 9.999289763779528e-06, + "loss": 0.5652, + "step": 455 + }, + { + "epoch": 0.1436220472440945, + "grad_norm": 60.04147720336914, + "learning_rate": 9.99928818897638e-06, + "loss": 0.4523, + "step": 456 + }, + { + "epoch": 0.14393700787401575, + "grad_norm": 33.64701843261719, + "learning_rate": 9.999286614173228e-06, + "loss": 0.3136, + "step": 457 + }, + { + "epoch": 0.144251968503937, + "grad_norm": 24.181163787841797, + "learning_rate": 9.99928503937008e-06, + "loss": 0.1441, + "step": 458 + }, + { + "epoch": 0.14456692913385827, + "grad_norm": 121.68936157226562, + "learning_rate": 9.99928346456693e-06, + "loss": 0.9653, + "step": 459 + }, + { + "epoch": 0.14488188976377953, + "grad_norm": 48.23858642578125, + "learning_rate": 9.99928188976378e-06, + "loss": 0.4576, + "step": 460 + }, + { + "epoch": 0.14488188976377953, + "eval_loss": 0.6207642555236816, + "eval_runtime": 295.7863, + "eval_samples_per_second": 0.396, + "eval_steps_per_second": 0.396, + "step": 460 + }, + { + "epoch": 0.14519685039370078, + "grad_norm": 59.632354736328125, + "learning_rate": 9.99928031496063e-06, + "loss": 0.4868, + "step": 461 + }, + { + "epoch": 0.14551181102362204, + "grad_norm": 103.69953918457031, + "learning_rate": 9.999278740157482e-06, + "loss": 0.5991, + "step": 462 + }, + { + "epoch": 0.1458267716535433, + "grad_norm": 36.89337158203125, + "learning_rate": 9.999277165354333e-06, + "loss": 0.2331, + "step": 463 + }, + { + "epoch": 0.14614173228346455, + "grad_norm": 42.319114685058594, + "learning_rate": 9.999275590551182e-06, + "loss": 0.3162, + "step": 464 + }, + { + "epoch": 0.14645669291338584, + "grad_norm": 36.906063079833984, + "learning_rate": 9.999274015748033e-06, + "loss": 0.3396, + "step": 465 + }, + { + "epoch": 0.1467716535433071, + "grad_norm": 25.45098304748535, + "learning_rate": 9.999272440944882e-06, + "loss": 0.1806, + "step": 466 + }, + { + "epoch": 0.14708661417322835, + "grad_norm": 56.87234878540039, + "learning_rate": 9.999270866141733e-06, + "loss": 0.2911, + "step": 467 + }, + { + "epoch": 0.1474015748031496, + "grad_norm": 20.72125244140625, + "learning_rate": 9.999269291338584e-06, + "loss": 0.0734, + "step": 468 + }, + { + "epoch": 0.14771653543307087, + "grad_norm": 69.94297790527344, + "learning_rate": 9.999267716535434e-06, + "loss": 0.3137, + "step": 469 + }, + { + "epoch": 0.14803149606299212, + "grad_norm": 55.783897399902344, + "learning_rate": 9.999266141732285e-06, + "loss": 0.3248, + "step": 470 + }, + { + "epoch": 0.14834645669291338, + "grad_norm": 119.89505004882812, + "learning_rate": 9.999264566929134e-06, + "loss": 0.7729, + "step": 471 + }, + { + "epoch": 0.14866141732283464, + "grad_norm": 45.10006332397461, + "learning_rate": 9.999262992125985e-06, + "loss": 0.287, + "step": 472 + }, + { + "epoch": 0.1489763779527559, + "grad_norm": 146.70803833007812, + "learning_rate": 9.999261417322836e-06, + "loss": 1.0979, + "step": 473 + }, + { + "epoch": 0.14929133858267715, + "grad_norm": 156.35951232910156, + "learning_rate": 9.999259842519687e-06, + "loss": 0.8397, + "step": 474 + }, + { + "epoch": 0.14960629921259844, + "grad_norm": 99.72000122070312, + "learning_rate": 9.999258267716536e-06, + "loss": 0.3907, + "step": 475 + }, + { + "epoch": 0.1499212598425197, + "grad_norm": 119.9110336303711, + "learning_rate": 9.999256692913387e-06, + "loss": 0.8118, + "step": 476 + }, + { + "epoch": 0.15023622047244095, + "grad_norm": 160.22637939453125, + "learning_rate": 9.999255118110236e-06, + "loss": 0.6305, + "step": 477 + }, + { + "epoch": 0.1505511811023622, + "grad_norm": 53.594276428222656, + "learning_rate": 9.999253543307087e-06, + "loss": 0.2044, + "step": 478 + }, + { + "epoch": 0.15086614173228347, + "grad_norm": 76.94689178466797, + "learning_rate": 9.999251968503938e-06, + "loss": 1.0085, + "step": 479 + }, + { + "epoch": 0.15118110236220472, + "grad_norm": 133.71861267089844, + "learning_rate": 9.999250393700788e-06, + "loss": 0.715, + "step": 480 + }, + { + "epoch": 0.15118110236220472, + "eval_loss": 0.5550094246864319, + "eval_runtime": 306.9821, + "eval_samples_per_second": 0.381, + "eval_steps_per_second": 0.381, + "step": 480 + }, + { + "epoch": 0.15149606299212598, + "grad_norm": 65.10472106933594, + "learning_rate": 9.999248818897639e-06, + "loss": 0.3845, + "step": 481 + }, + { + "epoch": 0.15181102362204724, + "grad_norm": 64.05907440185547, + "learning_rate": 9.999247244094488e-06, + "loss": 0.4831, + "step": 482 + }, + { + "epoch": 0.1521259842519685, + "grad_norm": 29.167266845703125, + "learning_rate": 9.999245669291339e-06, + "loss": 0.1163, + "step": 483 + }, + { + "epoch": 0.15244094488188975, + "grad_norm": 81.4149398803711, + "learning_rate": 9.99924409448819e-06, + "loss": 0.5063, + "step": 484 + }, + { + "epoch": 0.15275590551181104, + "grad_norm": 29.339014053344727, + "learning_rate": 9.999242519685041e-06, + "loss": 0.3148, + "step": 485 + }, + { + "epoch": 0.1530708661417323, + "grad_norm": 56.6673698425293, + "learning_rate": 9.99924094488189e-06, + "loss": 0.5286, + "step": 486 + }, + { + "epoch": 0.15338582677165355, + "grad_norm": 52.65031051635742, + "learning_rate": 9.999239370078741e-06, + "loss": 0.2759, + "step": 487 + }, + { + "epoch": 0.1537007874015748, + "grad_norm": 68.89445495605469, + "learning_rate": 9.999237795275592e-06, + "loss": 0.4206, + "step": 488 + }, + { + "epoch": 0.15401574803149606, + "grad_norm": 57.06834030151367, + "learning_rate": 9.999236220472442e-06, + "loss": 0.3788, + "step": 489 + }, + { + "epoch": 0.15433070866141732, + "grad_norm": 30.39971923828125, + "learning_rate": 9.999234645669293e-06, + "loss": 0.4785, + "step": 490 + }, + { + "epoch": 0.15464566929133858, + "grad_norm": 58.35342025756836, + "learning_rate": 9.999233070866142e-06, + "loss": 0.5897, + "step": 491 + }, + { + "epoch": 0.15496062992125983, + "grad_norm": 58.403533935546875, + "learning_rate": 9.999231496062993e-06, + "loss": 0.8812, + "step": 492 + }, + { + "epoch": 0.1552755905511811, + "grad_norm": 71.63230895996094, + "learning_rate": 9.999229921259844e-06, + "loss": 0.1902, + "step": 493 + }, + { + "epoch": 0.15559055118110235, + "grad_norm": 31.834192276000977, + "learning_rate": 9.999228346456695e-06, + "loss": 0.7004, + "step": 494 + }, + { + "epoch": 0.15590551181102363, + "grad_norm": 35.32748794555664, + "learning_rate": 9.999226771653544e-06, + "loss": 0.5292, + "step": 495 + }, + { + "epoch": 0.1562204724409449, + "grad_norm": 63.80234909057617, + "learning_rate": 9.999225196850395e-06, + "loss": 0.69, + "step": 496 + }, + { + "epoch": 0.15653543307086615, + "grad_norm": 47.65721893310547, + "learning_rate": 9.999223622047244e-06, + "loss": 0.4307, + "step": 497 + }, + { + "epoch": 0.1568503937007874, + "grad_norm": 23.857507705688477, + "learning_rate": 9.999222047244095e-06, + "loss": 0.4278, + "step": 498 + }, + { + "epoch": 0.15716535433070866, + "grad_norm": 36.79646301269531, + "learning_rate": 9.999220472440946e-06, + "loss": 0.3957, + "step": 499 + }, + { + "epoch": 0.15748031496062992, + "grad_norm": 37.079471588134766, + "learning_rate": 9.999218897637796e-06, + "loss": 0.5078, + "step": 500 + }, + { + "epoch": 0.15748031496062992, + "eval_loss": 0.5599373579025269, + "eval_runtime": 312.5913, + "eval_samples_per_second": 0.374, + "eval_steps_per_second": 0.374, + "step": 500 + }, + { + "epoch": 0.15779527559055118, + "grad_norm": 31.857145309448242, + "learning_rate": 9.999217322834647e-06, + "loss": 0.3308, + "step": 501 + }, + { + "epoch": 0.15811023622047243, + "grad_norm": 31.417692184448242, + "learning_rate": 9.999215748031496e-06, + "loss": 0.4725, + "step": 502 + }, + { + "epoch": 0.1584251968503937, + "grad_norm": 100.06877899169922, + "learning_rate": 9.999214173228347e-06, + "loss": 0.8381, + "step": 503 + }, + { + "epoch": 0.15874015748031495, + "grad_norm": 29.754446029663086, + "learning_rate": 9.999212598425198e-06, + "loss": 0.422, + "step": 504 + }, + { + "epoch": 0.15905511811023623, + "grad_norm": 36.77560043334961, + "learning_rate": 9.999211023622049e-06, + "loss": 0.2553, + "step": 505 + }, + { + "epoch": 0.1593700787401575, + "grad_norm": 45.064300537109375, + "learning_rate": 9.999209448818898e-06, + "loss": 0.4116, + "step": 506 + }, + { + "epoch": 0.15968503937007875, + "grad_norm": 82.82674407958984, + "learning_rate": 9.999207874015747e-06, + "loss": 1.244, + "step": 507 + }, + { + "epoch": 0.16, + "grad_norm": 30.511226654052734, + "learning_rate": 9.999206299212598e-06, + "loss": 0.2208, + "step": 508 + }, + { + "epoch": 0.16031496062992126, + "grad_norm": 26.348651885986328, + "learning_rate": 9.99920472440945e-06, + "loss": 0.2132, + "step": 509 + }, + { + "epoch": 0.16062992125984252, + "grad_norm": 88.1176986694336, + "learning_rate": 9.9992031496063e-06, + "loss": 0.5883, + "step": 510 + }, + { + "epoch": 0.16094488188976377, + "grad_norm": 36.90149688720703, + "learning_rate": 9.99920157480315e-06, + "loss": 0.3298, + "step": 511 + }, + { + "epoch": 0.16125984251968503, + "grad_norm": 30.67510414123535, + "learning_rate": 9.9992e-06, + "loss": 0.2843, + "step": 512 + }, + { + "epoch": 0.1615748031496063, + "grad_norm": 89.83963012695312, + "learning_rate": 9.999198425196852e-06, + "loss": 0.4022, + "step": 513 + }, + { + "epoch": 0.16188976377952755, + "grad_norm": 46.62565612792969, + "learning_rate": 9.999196850393703e-06, + "loss": 0.491, + "step": 514 + }, + { + "epoch": 0.16220472440944883, + "grad_norm": 27.551525115966797, + "learning_rate": 9.999195275590552e-06, + "loss": 0.1321, + "step": 515 + }, + { + "epoch": 0.1625196850393701, + "grad_norm": 15.754891395568848, + "learning_rate": 9.999193700787403e-06, + "loss": 0.0748, + "step": 516 + }, + { + "epoch": 0.16283464566929134, + "grad_norm": 55.909244537353516, + "learning_rate": 9.999192125984252e-06, + "loss": 0.2661, + "step": 517 + }, + { + "epoch": 0.1631496062992126, + "grad_norm": 37.772239685058594, + "learning_rate": 9.999190551181103e-06, + "loss": 0.1733, + "step": 518 + }, + { + "epoch": 0.16346456692913386, + "grad_norm": 66.4226303100586, + "learning_rate": 9.999188976377954e-06, + "loss": 0.4846, + "step": 519 + }, + { + "epoch": 0.16377952755905512, + "grad_norm": 37.98040771484375, + "learning_rate": 9.999187401574804e-06, + "loss": 0.1224, + "step": 520 + }, + { + "epoch": 0.16377952755905512, + "eval_loss": 0.6723836064338684, + "eval_runtime": 307.1583, + "eval_samples_per_second": 0.381, + "eval_steps_per_second": 0.381, + "step": 520 + }, + { + "epoch": 0.16409448818897637, + "grad_norm": 53.96063995361328, + "learning_rate": 9.999185826771655e-06, + "loss": 0.4602, + "step": 521 + }, + { + "epoch": 0.16440944881889763, + "grad_norm": 32.147621154785156, + "learning_rate": 9.999184251968504e-06, + "loss": 0.0799, + "step": 522 + }, + { + "epoch": 0.1647244094488189, + "grad_norm": 104.01305389404297, + "learning_rate": 9.999182677165355e-06, + "loss": 0.2182, + "step": 523 + }, + { + "epoch": 0.16503937007874014, + "grad_norm": 106.27403259277344, + "learning_rate": 9.999181102362206e-06, + "loss": 1.2292, + "step": 524 + }, + { + "epoch": 0.16535433070866143, + "grad_norm": 33.72996139526367, + "learning_rate": 9.999179527559057e-06, + "loss": 0.2182, + "step": 525 + }, + { + "epoch": 0.16566929133858269, + "grad_norm": 83.75643920898438, + "learning_rate": 9.999177952755906e-06, + "loss": 1.4189, + "step": 526 + }, + { + "epoch": 0.16598425196850394, + "grad_norm": 81.67135620117188, + "learning_rate": 9.999176377952755e-06, + "loss": 0.3019, + "step": 527 + }, + { + "epoch": 0.1662992125984252, + "grad_norm": 160.24029541015625, + "learning_rate": 9.999174803149606e-06, + "loss": 0.2741, + "step": 528 + }, + { + "epoch": 0.16661417322834646, + "grad_norm": 73.85230255126953, + "learning_rate": 9.999173228346457e-06, + "loss": 1.037, + "step": 529 + }, + { + "epoch": 0.16692913385826771, + "grad_norm": 87.96981811523438, + "learning_rate": 9.999171653543308e-06, + "loss": 0.3799, + "step": 530 + }, + { + "epoch": 0.16724409448818897, + "grad_norm": 62.46072769165039, + "learning_rate": 9.999170078740158e-06, + "loss": 0.5531, + "step": 531 + }, + { + "epoch": 0.16755905511811023, + "grad_norm": 151.0035858154297, + "learning_rate": 9.999168503937009e-06, + "loss": 1.4022, + "step": 532 + }, + { + "epoch": 0.16787401574803149, + "grad_norm": 60.152671813964844, + "learning_rate": 9.99916692913386e-06, + "loss": 0.6236, + "step": 533 + }, + { + "epoch": 0.16818897637795274, + "grad_norm": 102.74165344238281, + "learning_rate": 9.99916535433071e-06, + "loss": 1.2304, + "step": 534 + }, + { + "epoch": 0.16850393700787403, + "grad_norm": 20.071189880371094, + "learning_rate": 9.99916377952756e-06, + "loss": 0.1347, + "step": 535 + }, + { + "epoch": 0.16881889763779528, + "grad_norm": 43.08108139038086, + "learning_rate": 9.999162204724411e-06, + "loss": 0.4371, + "step": 536 + }, + { + "epoch": 0.16913385826771654, + "grad_norm": 64.68496704101562, + "learning_rate": 9.99916062992126e-06, + "loss": 0.6083, + "step": 537 + }, + { + "epoch": 0.1694488188976378, + "grad_norm": 39.30328369140625, + "learning_rate": 9.999159055118111e-06, + "loss": 0.5559, + "step": 538 + }, + { + "epoch": 0.16976377952755906, + "grad_norm": 56.48727035522461, + "learning_rate": 9.999157480314962e-06, + "loss": 0.7363, + "step": 539 + }, + { + "epoch": 0.1700787401574803, + "grad_norm": 26.32372283935547, + "learning_rate": 9.999155905511812e-06, + "loss": 0.3275, + "step": 540 + }, + { + "epoch": 0.1700787401574803, + "eval_loss": 0.6582661867141724, + "eval_runtime": 302.7589, + "eval_samples_per_second": 0.386, + "eval_steps_per_second": 0.386, + "step": 540 + }, + { + "epoch": 0.17039370078740157, + "grad_norm": 69.46321868896484, + "learning_rate": 9.999154330708663e-06, + "loss": 0.3794, + "step": 541 + }, + { + "epoch": 0.17070866141732283, + "grad_norm": 36.254520416259766, + "learning_rate": 9.999152755905512e-06, + "loss": 0.6954, + "step": 542 + }, + { + "epoch": 0.17102362204724408, + "grad_norm": 55.22049331665039, + "learning_rate": 9.999151181102363e-06, + "loss": 0.4438, + "step": 543 + }, + { + "epoch": 0.17133858267716534, + "grad_norm": 24.40268325805664, + "learning_rate": 9.999149606299214e-06, + "loss": 0.3439, + "step": 544 + }, + { + "epoch": 0.17165354330708663, + "grad_norm": 29.554643630981445, + "learning_rate": 9.999148031496065e-06, + "loss": 0.4619, + "step": 545 + }, + { + "epoch": 0.17196850393700788, + "grad_norm": 34.69717788696289, + "learning_rate": 9.999146456692914e-06, + "loss": 0.3959, + "step": 546 + }, + { + "epoch": 0.17228346456692914, + "grad_norm": 49.8066520690918, + "learning_rate": 9.999144881889763e-06, + "loss": 0.6021, + "step": 547 + }, + { + "epoch": 0.1725984251968504, + "grad_norm": 29.715484619140625, + "learning_rate": 9.999143307086614e-06, + "loss": 0.5588, + "step": 548 + }, + { + "epoch": 0.17291338582677165, + "grad_norm": 24.99655532836914, + "learning_rate": 9.999141732283465e-06, + "loss": 0.5414, + "step": 549 + }, + { + "epoch": 0.1732283464566929, + "grad_norm": 21.271596908569336, + "learning_rate": 9.999140157480316e-06, + "loss": 0.4187, + "step": 550 + }, + { + "epoch": 0.17354330708661417, + "grad_norm": 51.398712158203125, + "learning_rate": 9.999138582677166e-06, + "loss": 0.6258, + "step": 551 + }, + { + "epoch": 0.17385826771653543, + "grad_norm": 51.94355010986328, + "learning_rate": 9.999137007874017e-06, + "loss": 0.3318, + "step": 552 + }, + { + "epoch": 0.17417322834645668, + "grad_norm": 30.253637313842773, + "learning_rate": 9.999135433070866e-06, + "loss": 0.2938, + "step": 553 + }, + { + "epoch": 0.17448818897637794, + "grad_norm": 44.27308654785156, + "learning_rate": 9.999133858267717e-06, + "loss": 0.5734, + "step": 554 + }, + { + "epoch": 0.17480314960629922, + "grad_norm": 33.18519973754883, + "learning_rate": 9.999132283464568e-06, + "loss": 0.566, + "step": 555 + }, + { + "epoch": 0.17511811023622048, + "grad_norm": 48.19463348388672, + "learning_rate": 9.999130708661419e-06, + "loss": 0.6241, + "step": 556 + }, + { + "epoch": 0.17543307086614174, + "grad_norm": 34.98441696166992, + "learning_rate": 9.999129133858268e-06, + "loss": 0.4501, + "step": 557 + }, + { + "epoch": 0.175748031496063, + "grad_norm": 30.51637077331543, + "learning_rate": 9.99912755905512e-06, + "loss": 0.2408, + "step": 558 + }, + { + "epoch": 0.17606299212598425, + "grad_norm": 65.26117706298828, + "learning_rate": 9.99912598425197e-06, + "loss": 0.6767, + "step": 559 + }, + { + "epoch": 0.1763779527559055, + "grad_norm": 37.751888275146484, + "learning_rate": 9.99912440944882e-06, + "loss": 0.3929, + "step": 560 + }, + { + "epoch": 0.1763779527559055, + "eval_loss": 0.5896673798561096, + "eval_runtime": 296.0217, + "eval_samples_per_second": 0.395, + "eval_steps_per_second": 0.395, + "step": 560 + }, + { + "epoch": 0.17669291338582677, + "grad_norm": 47.340492248535156, + "learning_rate": 9.99912283464567e-06, + "loss": 0.3066, + "step": 561 + }, + { + "epoch": 0.17700787401574802, + "grad_norm": 86.48099517822266, + "learning_rate": 9.99912125984252e-06, + "loss": 0.4933, + "step": 562 + }, + { + "epoch": 0.17732283464566928, + "grad_norm": 33.50886917114258, + "learning_rate": 9.99911968503937e-06, + "loss": 0.1638, + "step": 563 + }, + { + "epoch": 0.17763779527559054, + "grad_norm": 48.44023895263672, + "learning_rate": 9.999118110236222e-06, + "loss": 0.4082, + "step": 564 + }, + { + "epoch": 0.17795275590551182, + "grad_norm": 54.50333786010742, + "learning_rate": 9.999116535433073e-06, + "loss": 0.3993, + "step": 565 + }, + { + "epoch": 0.17826771653543308, + "grad_norm": 82.37716674804688, + "learning_rate": 9.999114960629922e-06, + "loss": 0.9308, + "step": 566 + }, + { + "epoch": 0.17858267716535434, + "grad_norm": 53.5648307800293, + "learning_rate": 9.999113385826771e-06, + "loss": 0.757, + "step": 567 + }, + { + "epoch": 0.1788976377952756, + "grad_norm": 51.68220901489258, + "learning_rate": 9.999111811023622e-06, + "loss": 0.408, + "step": 568 + }, + { + "epoch": 0.17921259842519685, + "grad_norm": 32.468666076660156, + "learning_rate": 9.999110236220473e-06, + "loss": 0.1988, + "step": 569 + }, + { + "epoch": 0.1795275590551181, + "grad_norm": 92.7658462524414, + "learning_rate": 9.999108661417324e-06, + "loss": 0.737, + "step": 570 + }, + { + "epoch": 0.17984251968503936, + "grad_norm": 94.98796844482422, + "learning_rate": 9.999107086614174e-06, + "loss": 0.2773, + "step": 571 + }, + { + "epoch": 0.18015748031496062, + "grad_norm": 45.95973587036133, + "learning_rate": 9.999105511811025e-06, + "loss": 0.417, + "step": 572 + }, + { + "epoch": 0.18047244094488188, + "grad_norm": 42.59912872314453, + "learning_rate": 9.999103937007874e-06, + "loss": 0.437, + "step": 573 + }, + { + "epoch": 0.18078740157480314, + "grad_norm": 73.75167083740234, + "learning_rate": 9.999102362204725e-06, + "loss": 0.2676, + "step": 574 + }, + { + "epoch": 0.18110236220472442, + "grad_norm": 46.682533264160156, + "learning_rate": 9.999100787401576e-06, + "loss": 0.1902, + "step": 575 + }, + { + "epoch": 0.18141732283464568, + "grad_norm": 35.30620574951172, + "learning_rate": 9.999099212598427e-06, + "loss": 0.2953, + "step": 576 + }, + { + "epoch": 0.18173228346456693, + "grad_norm": 113.94246673583984, + "learning_rate": 9.999097637795276e-06, + "loss": 0.7046, + "step": 577 + }, + { + "epoch": 0.1820472440944882, + "grad_norm": 207.06141662597656, + "learning_rate": 9.999096062992125e-06, + "loss": 0.5952, + "step": 578 + }, + { + "epoch": 0.18236220472440945, + "grad_norm": 34.89611053466797, + "learning_rate": 9.999094488188978e-06, + "loss": 0.2947, + "step": 579 + }, + { + "epoch": 0.1826771653543307, + "grad_norm": 108.61929321289062, + "learning_rate": 9.999092913385827e-06, + "loss": 0.2417, + "step": 580 + }, + { + "epoch": 0.1826771653543307, + "eval_loss": 0.5409280061721802, + "eval_runtime": 307.2823, + "eval_samples_per_second": 0.381, + "eval_steps_per_second": 0.381, + "step": 580 + }, + { + "epoch": 0.18299212598425196, + "grad_norm": 15.958565711975098, + "learning_rate": 9.999091338582678e-06, + "loss": 0.1013, + "step": 581 + }, + { + "epoch": 0.18330708661417322, + "grad_norm": 68.7283706665039, + "learning_rate": 9.999089763779528e-06, + "loss": 0.8978, + "step": 582 + }, + { + "epoch": 0.18362204724409448, + "grad_norm": 80.29981994628906, + "learning_rate": 9.999088188976379e-06, + "loss": 0.5132, + "step": 583 + }, + { + "epoch": 0.18393700787401573, + "grad_norm": 25.062374114990234, + "learning_rate": 9.99908661417323e-06, + "loss": 0.1291, + "step": 584 + }, + { + "epoch": 0.18425196850393702, + "grad_norm": 58.20054244995117, + "learning_rate": 9.99908503937008e-06, + "loss": 0.7141, + "step": 585 + }, + { + "epoch": 0.18456692913385828, + "grad_norm": 73.24835968017578, + "learning_rate": 9.99908346456693e-06, + "loss": 0.5306, + "step": 586 + }, + { + "epoch": 0.18488188976377953, + "grad_norm": 145.0093994140625, + "learning_rate": 9.99908188976378e-06, + "loss": 0.6498, + "step": 587 + }, + { + "epoch": 0.1851968503937008, + "grad_norm": 62.22865295410156, + "learning_rate": 9.99908031496063e-06, + "loss": 0.6765, + "step": 588 + }, + { + "epoch": 0.18551181102362205, + "grad_norm": 11.53496265411377, + "learning_rate": 9.999078740157481e-06, + "loss": 0.0789, + "step": 589 + }, + { + "epoch": 0.1858267716535433, + "grad_norm": 104.29961395263672, + "learning_rate": 9.999077165354332e-06, + "loss": 0.444, + "step": 590 + }, + { + "epoch": 0.18614173228346456, + "grad_norm": 26.030893325805664, + "learning_rate": 9.999075590551182e-06, + "loss": 0.0719, + "step": 591 + }, + { + "epoch": 0.18645669291338582, + "grad_norm": 59.57289123535156, + "learning_rate": 9.999074015748033e-06, + "loss": 0.4676, + "step": 592 + }, + { + "epoch": 0.18677165354330708, + "grad_norm": 47.10686111450195, + "learning_rate": 9.999072440944882e-06, + "loss": 0.6408, + "step": 593 + }, + { + "epoch": 0.18708661417322833, + "grad_norm": 97.91781616210938, + "learning_rate": 9.999070866141733e-06, + "loss": 0.8239, + "step": 594 + }, + { + "epoch": 0.18740157480314962, + "grad_norm": 82.05168151855469, + "learning_rate": 9.999069291338584e-06, + "loss": 0.2311, + "step": 595 + }, + { + "epoch": 0.18771653543307087, + "grad_norm": 73.30006408691406, + "learning_rate": 9.999067716535435e-06, + "loss": 0.6648, + "step": 596 + }, + { + "epoch": 0.18803149606299213, + "grad_norm": 36.88441467285156, + "learning_rate": 9.999066141732284e-06, + "loss": 0.411, + "step": 597 + }, + { + "epoch": 0.1883464566929134, + "grad_norm": 21.77279281616211, + "learning_rate": 9.999064566929133e-06, + "loss": 0.0762, + "step": 598 + }, + { + "epoch": 0.18866141732283465, + "grad_norm": 41.469337463378906, + "learning_rate": 9.999062992125984e-06, + "loss": 0.2768, + "step": 599 + }, + { + "epoch": 0.1889763779527559, + "grad_norm": 23.42574119567871, + "learning_rate": 9.999061417322835e-06, + "loss": 0.3597, + "step": 600 + }, + { + "epoch": 0.1889763779527559, + "eval_loss": 0.5386444330215454, + "eval_runtime": 309.0194, + "eval_samples_per_second": 0.379, + "eval_steps_per_second": 0.379, + "step": 600 + }, + { + "epoch": 0.18929133858267716, + "grad_norm": 49.9908332824707, + "learning_rate": 9.999059842519686e-06, + "loss": 0.3259, + "step": 601 + }, + { + "epoch": 0.18960629921259842, + "grad_norm": 111.2554702758789, + "learning_rate": 9.999058267716536e-06, + "loss": 0.451, + "step": 602 + }, + { + "epoch": 0.18992125984251967, + "grad_norm": 87.17556762695312, + "learning_rate": 9.999056692913387e-06, + "loss": 0.2611, + "step": 603 + }, + { + "epoch": 0.19023622047244093, + "grad_norm": 24.059478759765625, + "learning_rate": 9.999055118110238e-06, + "loss": 0.3407, + "step": 604 + }, + { + "epoch": 0.19055118110236222, + "grad_norm": 121.00428009033203, + "learning_rate": 9.999053543307089e-06, + "loss": 1.3334, + "step": 605 + }, + { + "epoch": 0.19086614173228347, + "grad_norm": 26.074909210205078, + "learning_rate": 9.999051968503938e-06, + "loss": 0.2243, + "step": 606 + }, + { + "epoch": 0.19118110236220473, + "grad_norm": 60.16206359863281, + "learning_rate": 9.999050393700787e-06, + "loss": 0.1694, + "step": 607 + }, + { + "epoch": 0.191496062992126, + "grad_norm": 94.56045532226562, + "learning_rate": 9.999048818897638e-06, + "loss": 0.453, + "step": 608 + }, + { + "epoch": 0.19181102362204724, + "grad_norm": 106.34618377685547, + "learning_rate": 9.99904724409449e-06, + "loss": 0.5918, + "step": 609 + }, + { + "epoch": 0.1921259842519685, + "grad_norm": 33.97660827636719, + "learning_rate": 9.99904566929134e-06, + "loss": 0.6824, + "step": 610 + }, + { + "epoch": 0.19244094488188976, + "grad_norm": 25.390705108642578, + "learning_rate": 9.99904409448819e-06, + "loss": 0.4475, + "step": 611 + }, + { + "epoch": 0.19275590551181102, + "grad_norm": 77.29833984375, + "learning_rate": 9.99904251968504e-06, + "loss": 0.5393, + "step": 612 + }, + { + "epoch": 0.19307086614173227, + "grad_norm": 56.867801666259766, + "learning_rate": 9.99904094488189e-06, + "loss": 0.4419, + "step": 613 + }, + { + "epoch": 0.19338582677165353, + "grad_norm": 62.77841567993164, + "learning_rate": 9.99903937007874e-06, + "loss": 0.7734, + "step": 614 + }, + { + "epoch": 0.19370078740157481, + "grad_norm": 76.52714538574219, + "learning_rate": 9.999037795275592e-06, + "loss": 0.7938, + "step": 615 + }, + { + "epoch": 0.19401574803149607, + "grad_norm": 26.540964126586914, + "learning_rate": 9.999036220472443e-06, + "loss": 0.2187, + "step": 616 + }, + { + "epoch": 0.19433070866141733, + "grad_norm": 20.736865997314453, + "learning_rate": 9.999034645669292e-06, + "loss": 0.1165, + "step": 617 + }, + { + "epoch": 0.19464566929133859, + "grad_norm": 86.58599853515625, + "learning_rate": 9.999033070866141e-06, + "loss": 0.4113, + "step": 618 + }, + { + "epoch": 0.19496062992125984, + "grad_norm": 17.32464599609375, + "learning_rate": 9.999031496062992e-06, + "loss": 0.2206, + "step": 619 + }, + { + "epoch": 0.1952755905511811, + "grad_norm": 54.347904205322266, + "learning_rate": 9.999029921259843e-06, + "loss": 0.6016, + "step": 620 + }, + { + "epoch": 0.1952755905511811, + "eval_loss": 0.5922896265983582, + "eval_runtime": 309.8531, + "eval_samples_per_second": 0.378, + "eval_steps_per_second": 0.378, + "step": 620 + }, + { + "epoch": 0.19559055118110236, + "grad_norm": 62.3914794921875, + "learning_rate": 9.999028346456694e-06, + "loss": 0.5295, + "step": 621 + }, + { + "epoch": 0.1959055118110236, + "grad_norm": 29.178783416748047, + "learning_rate": 9.999026771653544e-06, + "loss": 0.2871, + "step": 622 + }, + { + "epoch": 0.19622047244094487, + "grad_norm": 28.286312103271484, + "learning_rate": 9.999025196850395e-06, + "loss": 0.3118, + "step": 623 + }, + { + "epoch": 0.19653543307086613, + "grad_norm": 25.735105514526367, + "learning_rate": 9.999023622047244e-06, + "loss": 0.2462, + "step": 624 + }, + { + "epoch": 0.1968503937007874, + "grad_norm": 173.3253936767578, + "learning_rate": 9.999022047244095e-06, + "loss": 0.9632, + "step": 625 + }, + { + "epoch": 0.19716535433070867, + "grad_norm": 85.50897979736328, + "learning_rate": 9.999020472440946e-06, + "loss": 0.643, + "step": 626 + }, + { + "epoch": 0.19748031496062993, + "grad_norm": 29.653247833251953, + "learning_rate": 9.999018897637795e-06, + "loss": 0.4286, + "step": 627 + }, + { + "epoch": 0.19779527559055118, + "grad_norm": 37.991817474365234, + "learning_rate": 9.999017322834646e-06, + "loss": 0.3509, + "step": 628 + }, + { + "epoch": 0.19811023622047244, + "grad_norm": 97.16632843017578, + "learning_rate": 9.999015748031497e-06, + "loss": 0.3752, + "step": 629 + }, + { + "epoch": 0.1984251968503937, + "grad_norm": 50.35764694213867, + "learning_rate": 9.999014173228348e-06, + "loss": 0.5741, + "step": 630 + }, + { + "epoch": 0.19874015748031496, + "grad_norm": 28.755027770996094, + "learning_rate": 9.999012598425197e-06, + "loss": 0.278, + "step": 631 + }, + { + "epoch": 0.1990551181102362, + "grad_norm": 74.70952606201172, + "learning_rate": 9.999011023622048e-06, + "loss": 0.6146, + "step": 632 + }, + { + "epoch": 0.19937007874015747, + "grad_norm": 78.7210693359375, + "learning_rate": 9.999009448818898e-06, + "loss": 0.2827, + "step": 633 + }, + { + "epoch": 0.19968503937007873, + "grad_norm": 17.010251998901367, + "learning_rate": 9.999007874015749e-06, + "loss": 0.239, + "step": 634 + }, + { + "epoch": 0.2, + "grad_norm": 81.048095703125, + "learning_rate": 9.9990062992126e-06, + "loss": 1.0794, + "step": 635 + }, + { + "epoch": 0.20031496062992127, + "grad_norm": 30.376026153564453, + "learning_rate": 9.99900472440945e-06, + "loss": 0.2238, + "step": 636 + }, + { + "epoch": 0.20062992125984253, + "grad_norm": 45.60966873168945, + "learning_rate": 9.9990031496063e-06, + "loss": 0.7123, + "step": 637 + }, + { + "epoch": 0.20094488188976378, + "grad_norm": 47.69063186645508, + "learning_rate": 9.99900157480315e-06, + "loss": 0.5227, + "step": 638 + }, + { + "epoch": 0.20125984251968504, + "grad_norm": 25.39598274230957, + "learning_rate": 9.999e-06, + "loss": 0.0982, + "step": 639 + }, + { + "epoch": 0.2015748031496063, + "grad_norm": 36.95790481567383, + "learning_rate": 9.998998425196851e-06, + "loss": 0.4809, + "step": 640 + }, + { + "epoch": 0.2015748031496063, + "eval_loss": 0.5429102182388306, + "eval_runtime": 295.5911, + "eval_samples_per_second": 0.396, + "eval_steps_per_second": 0.396, + "step": 640 + }, + { + "epoch": 0.20188976377952755, + "grad_norm": 47.62435531616211, + "learning_rate": 9.998996850393702e-06, + "loss": 0.6058, + "step": 641 + }, + { + "epoch": 0.2022047244094488, + "grad_norm": 31.27633285522461, + "learning_rate": 9.998995275590552e-06, + "loss": 0.2615, + "step": 642 + }, + { + "epoch": 0.20251968503937007, + "grad_norm": 47.4228630065918, + "learning_rate": 9.998993700787403e-06, + "loss": 0.6306, + "step": 643 + }, + { + "epoch": 0.20283464566929135, + "grad_norm": 67.43081665039062, + "learning_rate": 9.998992125984252e-06, + "loss": 0.654, + "step": 644 + }, + { + "epoch": 0.2031496062992126, + "grad_norm": 57.79238510131836, + "learning_rate": 9.998990551181103e-06, + "loss": 0.3423, + "step": 645 + }, + { + "epoch": 0.20346456692913387, + "grad_norm": 38.009735107421875, + "learning_rate": 9.998988976377954e-06, + "loss": 0.5406, + "step": 646 + }, + { + "epoch": 0.20377952755905512, + "grad_norm": 36.380531311035156, + "learning_rate": 9.998987401574805e-06, + "loss": 0.3381, + "step": 647 + }, + { + "epoch": 0.20409448818897638, + "grad_norm": 33.59734344482422, + "learning_rate": 9.998985826771654e-06, + "loss": 0.2425, + "step": 648 + }, + { + "epoch": 0.20440944881889764, + "grad_norm": 40.288902282714844, + "learning_rate": 9.998984251968505e-06, + "loss": 0.4959, + "step": 649 + }, + { + "epoch": 0.2047244094488189, + "grad_norm": 51.55220413208008, + "learning_rate": 9.998982677165356e-06, + "loss": 0.4607, + "step": 650 + }, + { + "epoch": 0.20503937007874015, + "grad_norm": 51.30996322631836, + "learning_rate": 9.998981102362205e-06, + "loss": 0.4445, + "step": 651 + }, + { + "epoch": 0.2053543307086614, + "grad_norm": 139.98907470703125, + "learning_rate": 9.998979527559056e-06, + "loss": 1.0222, + "step": 652 + }, + { + "epoch": 0.20566929133858267, + "grad_norm": 42.455169677734375, + "learning_rate": 9.998977952755906e-06, + "loss": 0.3876, + "step": 653 + }, + { + "epoch": 0.20598425196850395, + "grad_norm": 22.130889892578125, + "learning_rate": 9.998976377952757e-06, + "loss": 0.3178, + "step": 654 + }, + { + "epoch": 0.2062992125984252, + "grad_norm": 27.28899574279785, + "learning_rate": 9.998974803149608e-06, + "loss": 0.236, + "step": 655 + }, + { + "epoch": 0.20661417322834646, + "grad_norm": 19.113969802856445, + "learning_rate": 9.998973228346459e-06, + "loss": 0.2199, + "step": 656 + }, + { + "epoch": 0.20692913385826772, + "grad_norm": 70.42593383789062, + "learning_rate": 9.998971653543308e-06, + "loss": 0.5523, + "step": 657 + }, + { + "epoch": 0.20724409448818898, + "grad_norm": 51.796207427978516, + "learning_rate": 9.998970078740157e-06, + "loss": 0.5754, + "step": 658 + }, + { + "epoch": 0.20755905511811024, + "grad_norm": 95.93733215332031, + "learning_rate": 9.998968503937008e-06, + "loss": 0.7566, + "step": 659 + }, + { + "epoch": 0.2078740157480315, + "grad_norm": 27.715049743652344, + "learning_rate": 9.99896692913386e-06, + "loss": 0.2745, + "step": 660 + }, + { + "epoch": 0.2078740157480315, + "eval_loss": 0.4993188977241516, + "eval_runtime": 309.1118, + "eval_samples_per_second": 0.379, + "eval_steps_per_second": 0.379, + "step": 660 + }, + { + "epoch": 0.20818897637795275, + "grad_norm": 43.9770393371582, + "learning_rate": 9.99896535433071e-06, + "loss": 0.5356, + "step": 661 + }, + { + "epoch": 0.208503937007874, + "grad_norm": 28.040441513061523, + "learning_rate": 9.99896377952756e-06, + "loss": 0.315, + "step": 662 + }, + { + "epoch": 0.20881889763779526, + "grad_norm": 45.72781753540039, + "learning_rate": 9.99896220472441e-06, + "loss": 0.4161, + "step": 663 + }, + { + "epoch": 0.20913385826771655, + "grad_norm": 27.964317321777344, + "learning_rate": 9.99896062992126e-06, + "loss": 0.2185, + "step": 664 + }, + { + "epoch": 0.2094488188976378, + "grad_norm": 32.59132766723633, + "learning_rate": 9.99895905511811e-06, + "loss": 0.3825, + "step": 665 + }, + { + "epoch": 0.20976377952755906, + "grad_norm": 38.677207946777344, + "learning_rate": 9.998957480314962e-06, + "loss": 0.2456, + "step": 666 + }, + { + "epoch": 0.21007874015748032, + "grad_norm": 14.392455101013184, + "learning_rate": 9.998955905511813e-06, + "loss": 0.0907, + "step": 667 + }, + { + "epoch": 0.21039370078740158, + "grad_norm": 50.628910064697266, + "learning_rate": 9.998954330708662e-06, + "loss": 0.7256, + "step": 668 + }, + { + "epoch": 0.21070866141732283, + "grad_norm": 60.51618194580078, + "learning_rate": 9.998952755905511e-06, + "loss": 0.4601, + "step": 669 + }, + { + "epoch": 0.2110236220472441, + "grad_norm": 40.12845230102539, + "learning_rate": 9.998951181102362e-06, + "loss": 0.5639, + "step": 670 + }, + { + "epoch": 0.21133858267716535, + "grad_norm": 45.10561752319336, + "learning_rate": 9.998949606299213e-06, + "loss": 0.4125, + "step": 671 + }, + { + "epoch": 0.2116535433070866, + "grad_norm": 79.58858489990234, + "learning_rate": 9.998948031496064e-06, + "loss": 0.8415, + "step": 672 + }, + { + "epoch": 0.21196850393700786, + "grad_norm": 35.8792724609375, + "learning_rate": 9.998946456692914e-06, + "loss": 0.3386, + "step": 673 + }, + { + "epoch": 0.21228346456692915, + "grad_norm": 13.449446678161621, + "learning_rate": 9.998944881889765e-06, + "loss": 0.1095, + "step": 674 + }, + { + "epoch": 0.2125984251968504, + "grad_norm": 82.7651596069336, + "learning_rate": 9.998943307086616e-06, + "loss": 0.331, + "step": 675 + }, + { + "epoch": 0.21291338582677166, + "grad_norm": 80.85045623779297, + "learning_rate": 9.998941732283467e-06, + "loss": 0.9251, + "step": 676 + }, + { + "epoch": 0.21322834645669292, + "grad_norm": 34.32810974121094, + "learning_rate": 9.998940157480316e-06, + "loss": 0.2547, + "step": 677 + }, + { + "epoch": 0.21354330708661418, + "grad_norm": 39.6412239074707, + "learning_rate": 9.998938582677165e-06, + "loss": 0.3042, + "step": 678 + }, + { + "epoch": 0.21385826771653543, + "grad_norm": 29.662391662597656, + "learning_rate": 9.998937007874016e-06, + "loss": 0.353, + "step": 679 + }, + { + "epoch": 0.2141732283464567, + "grad_norm": 15.640954971313477, + "learning_rate": 9.998935433070867e-06, + "loss": 0.0869, + "step": 680 + }, + { + "epoch": 0.2141732283464567, + "eval_loss": 0.5842440128326416, + "eval_runtime": 303.7569, + "eval_samples_per_second": 0.385, + "eval_steps_per_second": 0.385, + "step": 680 + }, + { + "epoch": 0.21448818897637795, + "grad_norm": 47.02978515625, + "learning_rate": 9.998933858267718e-06, + "loss": 0.6137, + "step": 681 + }, + { + "epoch": 0.2148031496062992, + "grad_norm": 23.671756744384766, + "learning_rate": 9.998932283464568e-06, + "loss": 0.2595, + "step": 682 + }, + { + "epoch": 0.21511811023622046, + "grad_norm": 35.9589729309082, + "learning_rate": 9.998930708661418e-06, + "loss": 0.3242, + "step": 683 + }, + { + "epoch": 0.21543307086614175, + "grad_norm": 42.95161819458008, + "learning_rate": 9.998929133858268e-06, + "loss": 0.2721, + "step": 684 + }, + { + "epoch": 0.215748031496063, + "grad_norm": 58.640968322753906, + "learning_rate": 9.998927559055119e-06, + "loss": 0.3166, + "step": 685 + }, + { + "epoch": 0.21606299212598426, + "grad_norm": 35.683467864990234, + "learning_rate": 9.99892598425197e-06, + "loss": 0.3474, + "step": 686 + }, + { + "epoch": 0.21637795275590552, + "grad_norm": 9.069628715515137, + "learning_rate": 9.99892440944882e-06, + "loss": 0.0522, + "step": 687 + }, + { + "epoch": 0.21669291338582677, + "grad_norm": 29.538806915283203, + "learning_rate": 9.99892283464567e-06, + "loss": 0.3609, + "step": 688 + }, + { + "epoch": 0.21700787401574803, + "grad_norm": 40.99892044067383, + "learning_rate": 9.99892125984252e-06, + "loss": 0.5832, + "step": 689 + }, + { + "epoch": 0.2173228346456693, + "grad_norm": 48.680538177490234, + "learning_rate": 9.99891968503937e-06, + "loss": 0.5598, + "step": 690 + }, + { + "epoch": 0.21763779527559055, + "grad_norm": 37.50501251220703, + "learning_rate": 9.998918110236221e-06, + "loss": 0.211, + "step": 691 + }, + { + "epoch": 0.2179527559055118, + "grad_norm": 145.6537322998047, + "learning_rate": 9.998916535433072e-06, + "loss": 1.0281, + "step": 692 + }, + { + "epoch": 0.21826771653543306, + "grad_norm": 82.14835357666016, + "learning_rate": 9.998914960629922e-06, + "loss": 0.4626, + "step": 693 + }, + { + "epoch": 0.21858267716535434, + "grad_norm": 114.24347686767578, + "learning_rate": 9.998913385826773e-06, + "loss": 0.4935, + "step": 694 + }, + { + "epoch": 0.2188976377952756, + "grad_norm": 44.75710678100586, + "learning_rate": 9.998911811023622e-06, + "loss": 0.2071, + "step": 695 + }, + { + "epoch": 0.21921259842519686, + "grad_norm": 35.995880126953125, + "learning_rate": 9.998910236220473e-06, + "loss": 0.1709, + "step": 696 + }, + { + "epoch": 0.21952755905511812, + "grad_norm": 74.69808197021484, + "learning_rate": 9.998908661417324e-06, + "loss": 0.7127, + "step": 697 + }, + { + "epoch": 0.21984251968503937, + "grad_norm": 46.23347854614258, + "learning_rate": 9.998907086614173e-06, + "loss": 0.2524, + "step": 698 + }, + { + "epoch": 0.22015748031496063, + "grad_norm": 23.196392059326172, + "learning_rate": 9.998905511811024e-06, + "loss": 0.0743, + "step": 699 + }, + { + "epoch": 0.2204724409448819, + "grad_norm": 68.11265563964844, + "learning_rate": 9.998903937007875e-06, + "loss": 0.5743, + "step": 700 + }, + { + "epoch": 0.2204724409448819, + "eval_loss": 0.6227251291275024, + "eval_runtime": 306.9107, + "eval_samples_per_second": 0.381, + "eval_steps_per_second": 0.381, + "step": 700 + }, + { + "epoch": 0.22078740157480314, + "grad_norm": 49.117034912109375, + "learning_rate": 9.998902362204726e-06, + "loss": 0.5395, + "step": 701 + }, + { + "epoch": 0.2211023622047244, + "grad_norm": 72.21612548828125, + "learning_rate": 9.998900787401575e-06, + "loss": 0.1053, + "step": 702 + }, + { + "epoch": 0.22141732283464566, + "grad_norm": 106.6301498413086, + "learning_rate": 9.998899212598426e-06, + "loss": 0.4932, + "step": 703 + }, + { + "epoch": 0.22173228346456694, + "grad_norm": 171.39988708496094, + "learning_rate": 9.998897637795276e-06, + "loss": 0.9959, + "step": 704 + }, + { + "epoch": 0.2220472440944882, + "grad_norm": 33.84798812866211, + "learning_rate": 9.998896062992127e-06, + "loss": 0.3472, + "step": 705 + }, + { + "epoch": 0.22236220472440946, + "grad_norm": 35.71989059448242, + "learning_rate": 9.998894488188978e-06, + "loss": 0.1522, + "step": 706 + }, + { + "epoch": 0.22267716535433071, + "grad_norm": 65.45905303955078, + "learning_rate": 9.998892913385829e-06, + "loss": 0.9923, + "step": 707 + }, + { + "epoch": 0.22299212598425197, + "grad_norm": 88.24420166015625, + "learning_rate": 9.998891338582678e-06, + "loss": 0.7105, + "step": 708 + }, + { + "epoch": 0.22330708661417323, + "grad_norm": 71.33065032958984, + "learning_rate": 9.998889763779527e-06, + "loss": 1.2078, + "step": 709 + }, + { + "epoch": 0.22362204724409449, + "grad_norm": 25.207542419433594, + "learning_rate": 9.998888188976378e-06, + "loss": 0.1815, + "step": 710 + }, + { + "epoch": 0.22393700787401574, + "grad_norm": 23.143169403076172, + "learning_rate": 9.99888661417323e-06, + "loss": 0.1984, + "step": 711 + }, + { + "epoch": 0.224251968503937, + "grad_norm": 32.797752380371094, + "learning_rate": 9.99888503937008e-06, + "loss": 0.3693, + "step": 712 + }, + { + "epoch": 0.22456692913385826, + "grad_norm": 35.14696502685547, + "learning_rate": 9.99888346456693e-06, + "loss": 0.4894, + "step": 713 + }, + { + "epoch": 0.22488188976377954, + "grad_norm": 24.498680114746094, + "learning_rate": 9.99888188976378e-06, + "loss": 0.3493, + "step": 714 + }, + { + "epoch": 0.2251968503937008, + "grad_norm": 57.850730895996094, + "learning_rate": 9.99888031496063e-06, + "loss": 0.483, + "step": 715 + }, + { + "epoch": 0.22551181102362206, + "grad_norm": 47.62955093383789, + "learning_rate": 9.998878740157481e-06, + "loss": 0.3677, + "step": 716 + }, + { + "epoch": 0.2258267716535433, + "grad_norm": 25.785797119140625, + "learning_rate": 9.998877165354332e-06, + "loss": 0.324, + "step": 717 + }, + { + "epoch": 0.22614173228346457, + "grad_norm": 44.619041442871094, + "learning_rate": 9.998875590551181e-06, + "loss": 0.3471, + "step": 718 + }, + { + "epoch": 0.22645669291338583, + "grad_norm": 10.855001449584961, + "learning_rate": 9.998874015748032e-06, + "loss": 0.205, + "step": 719 + }, + { + "epoch": 0.22677165354330708, + "grad_norm": 53.096466064453125, + "learning_rate": 9.998872440944883e-06, + "loss": 0.4501, + "step": 720 + }, + { + "epoch": 0.22677165354330708, + "eval_loss": 0.5146042704582214, + "eval_runtime": 316.9221, + "eval_samples_per_second": 0.369, + "eval_steps_per_second": 0.369, + "step": 720 + }, + { + "epoch": 0.22708661417322834, + "grad_norm": 44.9933967590332, + "learning_rate": 9.998870866141734e-06, + "loss": 0.4339, + "step": 721 + }, + { + "epoch": 0.2274015748031496, + "grad_norm": 46.176700592041016, + "learning_rate": 9.998869291338583e-06, + "loss": 0.4345, + "step": 722 + }, + { + "epoch": 0.22771653543307085, + "grad_norm": 21.4019832611084, + "learning_rate": 9.998867716535434e-06, + "loss": 0.5765, + "step": 723 + }, + { + "epoch": 0.22803149606299214, + "grad_norm": 17.84157943725586, + "learning_rate": 9.998866141732284e-06, + "loss": 0.1562, + "step": 724 + }, + { + "epoch": 0.2283464566929134, + "grad_norm": 18.121217727661133, + "learning_rate": 9.998864566929135e-06, + "loss": 0.1687, + "step": 725 + }, + { + "epoch": 0.22866141732283465, + "grad_norm": 32.00659942626953, + "learning_rate": 9.998862992125986e-06, + "loss": 0.1804, + "step": 726 + }, + { + "epoch": 0.2289763779527559, + "grad_norm": 32.684757232666016, + "learning_rate": 9.998861417322837e-06, + "loss": 0.2921, + "step": 727 + }, + { + "epoch": 0.22929133858267717, + "grad_norm": 65.45207214355469, + "learning_rate": 9.998859842519686e-06, + "loss": 0.237, + "step": 728 + }, + { + "epoch": 0.22960629921259842, + "grad_norm": 24.639812469482422, + "learning_rate": 9.998858267716535e-06, + "loss": 0.1602, + "step": 729 + }, + { + "epoch": 0.22992125984251968, + "grad_norm": 105.74681854248047, + "learning_rate": 9.998856692913386e-06, + "loss": 0.6624, + "step": 730 + }, + { + "epoch": 0.23023622047244094, + "grad_norm": 123.29035949707031, + "learning_rate": 9.998855118110237e-06, + "loss": 0.1592, + "step": 731 + }, + { + "epoch": 0.2305511811023622, + "grad_norm": 75.88599395751953, + "learning_rate": 9.998853543307088e-06, + "loss": 0.8749, + "step": 732 + }, + { + "epoch": 0.23086614173228345, + "grad_norm": 93.44769287109375, + "learning_rate": 9.998851968503938e-06, + "loss": 0.2961, + "step": 733 + }, + { + "epoch": 0.23118110236220474, + "grad_norm": 149.9324188232422, + "learning_rate": 9.998850393700789e-06, + "loss": 1.0564, + "step": 734 + }, + { + "epoch": 0.231496062992126, + "grad_norm": 158.12449645996094, + "learning_rate": 9.998848818897638e-06, + "loss": 1.6293, + "step": 735 + }, + { + "epoch": 0.23181102362204725, + "grad_norm": 149.76597595214844, + "learning_rate": 9.998847244094489e-06, + "loss": 1.6253, + "step": 736 + }, + { + "epoch": 0.2321259842519685, + "grad_norm": 40.33038330078125, + "learning_rate": 9.99884566929134e-06, + "loss": 0.5073, + "step": 737 + }, + { + "epoch": 0.23244094488188977, + "grad_norm": 67.39472198486328, + "learning_rate": 9.998844094488189e-06, + "loss": 0.303, + "step": 738 + }, + { + "epoch": 0.23275590551181102, + "grad_norm": 48.84297561645508, + "learning_rate": 9.99884251968504e-06, + "loss": 0.3308, + "step": 739 + }, + { + "epoch": 0.23307086614173228, + "grad_norm": 50.47072219848633, + "learning_rate": 9.99884094488189e-06, + "loss": 0.5499, + "step": 740 + }, + { + "epoch": 0.23307086614173228, + "eval_loss": 0.563422441482544, + "eval_runtime": 300.8217, + "eval_samples_per_second": 0.389, + "eval_steps_per_second": 0.389, + "step": 740 + }, + { + "epoch": 0.23338582677165354, + "grad_norm": 40.9046745300293, + "learning_rate": 9.99883937007874e-06, + "loss": 0.275, + "step": 741 + }, + { + "epoch": 0.2337007874015748, + "grad_norm": 68.57015991210938, + "learning_rate": 9.998837795275591e-06, + "loss": 0.5954, + "step": 742 + }, + { + "epoch": 0.23401574803149605, + "grad_norm": 32.3016357421875, + "learning_rate": 9.998836220472442e-06, + "loss": 0.1714, + "step": 743 + }, + { + "epoch": 0.23433070866141734, + "grad_norm": 35.811279296875, + "learning_rate": 9.998834645669292e-06, + "loss": 0.3879, + "step": 744 + }, + { + "epoch": 0.2346456692913386, + "grad_norm": 25.600717544555664, + "learning_rate": 9.998833070866143e-06, + "loss": 0.0832, + "step": 745 + }, + { + "epoch": 0.23496062992125985, + "grad_norm": 38.07957458496094, + "learning_rate": 9.998831496062994e-06, + "loss": 0.6504, + "step": 746 + }, + { + "epoch": 0.2352755905511811, + "grad_norm": 62.875614166259766, + "learning_rate": 9.998829921259845e-06, + "loss": 0.6823, + "step": 747 + }, + { + "epoch": 0.23559055118110236, + "grad_norm": 95.92493438720703, + "learning_rate": 9.998828346456694e-06, + "loss": 1.5914, + "step": 748 + }, + { + "epoch": 0.23590551181102362, + "grad_norm": 40.44503402709961, + "learning_rate": 9.998826771653543e-06, + "loss": 0.6454, + "step": 749 + }, + { + "epoch": 0.23622047244094488, + "grad_norm": 16.39349365234375, + "learning_rate": 9.998825196850394e-06, + "loss": 0.2515, + "step": 750 + }, + { + "epoch": 0.23653543307086614, + "grad_norm": 54.119468688964844, + "learning_rate": 9.998823622047245e-06, + "loss": 0.497, + "step": 751 + }, + { + "epoch": 0.2368503937007874, + "grad_norm": 55.42496871948242, + "learning_rate": 9.998822047244096e-06, + "loss": 0.2491, + "step": 752 + }, + { + "epoch": 0.23716535433070865, + "grad_norm": 55.21287155151367, + "learning_rate": 9.998820472440945e-06, + "loss": 0.9017, + "step": 753 + }, + { + "epoch": 0.23748031496062993, + "grad_norm": 26.37238311767578, + "learning_rate": 9.998818897637796e-06, + "loss": 0.4861, + "step": 754 + }, + { + "epoch": 0.2377952755905512, + "grad_norm": 23.099788665771484, + "learning_rate": 9.998817322834646e-06, + "loss": 0.4609, + "step": 755 + }, + { + "epoch": 0.23811023622047245, + "grad_norm": 35.12017822265625, + "learning_rate": 9.998815748031497e-06, + "loss": 0.4897, + "step": 756 + }, + { + "epoch": 0.2384251968503937, + "grad_norm": 13.813502311706543, + "learning_rate": 9.998814173228348e-06, + "loss": 0.1271, + "step": 757 + }, + { + "epoch": 0.23874015748031496, + "grad_norm": 29.46474266052246, + "learning_rate": 9.998812598425197e-06, + "loss": 0.505, + "step": 758 + }, + { + "epoch": 0.23905511811023622, + "grad_norm": 26.503684997558594, + "learning_rate": 9.998811023622048e-06, + "loss": 0.424, + "step": 759 + }, + { + "epoch": 0.23937007874015748, + "grad_norm": 36.68299102783203, + "learning_rate": 9.998809448818897e-06, + "loss": 0.5153, + "step": 760 + }, + { + "epoch": 0.23937007874015748, + "eval_loss": 0.467477947473526, + "eval_runtime": 299.9771, + "eval_samples_per_second": 0.39, + "eval_steps_per_second": 0.39, + "step": 760 + }, + { + "epoch": 0.23968503937007873, + "grad_norm": 27.404645919799805, + "learning_rate": 9.998807874015748e-06, + "loss": 0.4891, + "step": 761 + }, + { + "epoch": 0.24, + "grad_norm": 35.5542106628418, + "learning_rate": 9.9988062992126e-06, + "loss": 0.6146, + "step": 762 + }, + { + "epoch": 0.24031496062992125, + "grad_norm": 64.38270568847656, + "learning_rate": 9.99880472440945e-06, + "loss": 0.7001, + "step": 763 + }, + { + "epoch": 0.24062992125984253, + "grad_norm": 11.979158401489258, + "learning_rate": 9.9988031496063e-06, + "loss": 0.1519, + "step": 764 + }, + { + "epoch": 0.2409448818897638, + "grad_norm": 45.081756591796875, + "learning_rate": 9.99880157480315e-06, + "loss": 0.5738, + "step": 765 + }, + { + "epoch": 0.24125984251968505, + "grad_norm": 16.82332992553711, + "learning_rate": 9.998800000000002e-06, + "loss": 0.2367, + "step": 766 + }, + { + "epoch": 0.2415748031496063, + "grad_norm": 58.26005554199219, + "learning_rate": 9.998798425196853e-06, + "loss": 0.6255, + "step": 767 + }, + { + "epoch": 0.24188976377952756, + "grad_norm": 33.454200744628906, + "learning_rate": 9.998796850393702e-06, + "loss": 0.282, + "step": 768 + }, + { + "epoch": 0.24220472440944882, + "grad_norm": 26.192066192626953, + "learning_rate": 9.998795275590551e-06, + "loss": 0.2512, + "step": 769 + }, + { + "epoch": 0.24251968503937008, + "grad_norm": 24.451866149902344, + "learning_rate": 9.998793700787402e-06, + "loss": 0.1288, + "step": 770 + }, + { + "epoch": 0.24283464566929133, + "grad_norm": 17.908388137817383, + "learning_rate": 9.998792125984253e-06, + "loss": 0.1228, + "step": 771 + }, + { + "epoch": 0.2431496062992126, + "grad_norm": 30.55709457397461, + "learning_rate": 9.998790551181104e-06, + "loss": 0.4937, + "step": 772 + }, + { + "epoch": 0.24346456692913385, + "grad_norm": 47.04617691040039, + "learning_rate": 9.998788976377953e-06, + "loss": 0.7825, + "step": 773 + }, + { + "epoch": 0.24377952755905513, + "grad_norm": 66.30603790283203, + "learning_rate": 9.998787401574804e-06, + "loss": 0.3853, + "step": 774 + }, + { + "epoch": 0.2440944881889764, + "grad_norm": 28.97539520263672, + "learning_rate": 9.998785826771654e-06, + "loss": 0.2909, + "step": 775 + }, + { + "epoch": 0.24440944881889765, + "grad_norm": 71.36101531982422, + "learning_rate": 9.998784251968505e-06, + "loss": 0.6988, + "step": 776 + }, + { + "epoch": 0.2447244094488189, + "grad_norm": 35.699249267578125, + "learning_rate": 9.998782677165356e-06, + "loss": 0.4579, + "step": 777 + }, + { + "epoch": 0.24503937007874016, + "grad_norm": 4.760364532470703, + "learning_rate": 9.998781102362205e-06, + "loss": 0.0287, + "step": 778 + }, + { + "epoch": 0.24535433070866142, + "grad_norm": 30.33928108215332, + "learning_rate": 9.998779527559056e-06, + "loss": 0.5643, + "step": 779 + }, + { + "epoch": 0.24566929133858267, + "grad_norm": 43.770694732666016, + "learning_rate": 9.998777952755905e-06, + "loss": 0.4347, + "step": 780 + }, + { + "epoch": 0.24566929133858267, + "eval_loss": 0.5863191485404968, + "eval_runtime": 301.5247, + "eval_samples_per_second": 0.388, + "eval_steps_per_second": 0.388, + "step": 780 + }, + { + "epoch": 0.24598425196850393, + "grad_norm": 38.59556198120117, + "learning_rate": 9.998776377952756e-06, + "loss": 0.2601, + "step": 781 + }, + { + "epoch": 0.2462992125984252, + "grad_norm": 42.7963981628418, + "learning_rate": 9.998774803149607e-06, + "loss": 0.329, + "step": 782 + }, + { + "epoch": 0.24661417322834644, + "grad_norm": 48.59577941894531, + "learning_rate": 9.998773228346458e-06, + "loss": 0.5232, + "step": 783 + }, + { + "epoch": 0.24692913385826773, + "grad_norm": 51.12166976928711, + "learning_rate": 9.998771653543308e-06, + "loss": 0.955, + "step": 784 + }, + { + "epoch": 0.247244094488189, + "grad_norm": 29.71060562133789, + "learning_rate": 9.998770078740159e-06, + "loss": 0.1399, + "step": 785 + }, + { + "epoch": 0.24755905511811024, + "grad_norm": 73.006103515625, + "learning_rate": 9.998768503937008e-06, + "loss": 0.8551, + "step": 786 + }, + { + "epoch": 0.2478740157480315, + "grad_norm": 24.96092414855957, + "learning_rate": 9.998766929133859e-06, + "loss": 0.379, + "step": 787 + }, + { + "epoch": 0.24818897637795276, + "grad_norm": 23.48893165588379, + "learning_rate": 9.99876535433071e-06, + "loss": 0.1811, + "step": 788 + }, + { + "epoch": 0.24850393700787402, + "grad_norm": 38.28635025024414, + "learning_rate": 9.998763779527559e-06, + "loss": 0.4841, + "step": 789 + }, + { + "epoch": 0.24881889763779527, + "grad_norm": 53.374549865722656, + "learning_rate": 9.99876220472441e-06, + "loss": 0.5892, + "step": 790 + }, + { + "epoch": 0.24913385826771653, + "grad_norm": 32.456485748291016, + "learning_rate": 9.998760629921261e-06, + "loss": 0.4001, + "step": 791 + }, + { + "epoch": 0.2494488188976378, + "grad_norm": 75.20841217041016, + "learning_rate": 9.998759055118112e-06, + "loss": 0.5947, + "step": 792 + }, + { + "epoch": 0.24976377952755904, + "grad_norm": 43.24211502075195, + "learning_rate": 9.998757480314961e-06, + "loss": 0.4858, + "step": 793 + }, + { + "epoch": 0.25007874015748033, + "grad_norm": 50.684852600097656, + "learning_rate": 9.998755905511812e-06, + "loss": 0.3602, + "step": 794 + }, + { + "epoch": 0.2503937007874016, + "grad_norm": 24.13330078125, + "learning_rate": 9.998754330708662e-06, + "loss": 0.3898, + "step": 795 + }, + { + "epoch": 0.25070866141732284, + "grad_norm": 32.628719329833984, + "learning_rate": 9.998752755905513e-06, + "loss": 0.547, + "step": 796 + }, + { + "epoch": 0.2510236220472441, + "grad_norm": 24.97728157043457, + "learning_rate": 9.998751181102364e-06, + "loss": 0.536, + "step": 797 + }, + { + "epoch": 0.25133858267716536, + "grad_norm": 21.705215454101562, + "learning_rate": 9.998749606299213e-06, + "loss": 0.315, + "step": 798 + }, + { + "epoch": 0.2516535433070866, + "grad_norm": 27.06643295288086, + "learning_rate": 9.998748031496064e-06, + "loss": 0.4722, + "step": 799 + }, + { + "epoch": 0.25196850393700787, + "grad_norm": 23.71688461303711, + "learning_rate": 9.998746456692913e-06, + "loss": 0.5445, + "step": 800 + }, + { + "epoch": 0.25196850393700787, + "eval_loss": 0.51589435338974, + "eval_runtime": 316.4111, + "eval_samples_per_second": 0.37, + "eval_steps_per_second": 0.37, + "step": 800 + }, + { + "epoch": 0.2522834645669291, + "grad_norm": 35.45967483520508, + "learning_rate": 9.998744881889764e-06, + "loss": 0.5146, + "step": 801 + }, + { + "epoch": 0.2525984251968504, + "grad_norm": 24.311609268188477, + "learning_rate": 9.998743307086615e-06, + "loss": 0.596, + "step": 802 + }, + { + "epoch": 0.25291338582677164, + "grad_norm": 37.100257873535156, + "learning_rate": 9.998741732283466e-06, + "loss": 0.4307, + "step": 803 + }, + { + "epoch": 0.2532283464566929, + "grad_norm": 21.675411224365234, + "learning_rate": 9.998740157480315e-06, + "loss": 0.3512, + "step": 804 + }, + { + "epoch": 0.25354330708661416, + "grad_norm": 21.883447647094727, + "learning_rate": 9.998738582677166e-06, + "loss": 0.4789, + "step": 805 + }, + { + "epoch": 0.2538582677165354, + "grad_norm": 26.242074966430664, + "learning_rate": 9.998737007874016e-06, + "loss": 0.6633, + "step": 806 + }, + { + "epoch": 0.25417322834645667, + "grad_norm": 36.36134338378906, + "learning_rate": 9.998735433070867e-06, + "loss": 0.2947, + "step": 807 + }, + { + "epoch": 0.2544881889763779, + "grad_norm": 58.036354064941406, + "learning_rate": 9.998733858267718e-06, + "loss": 0.5375, + "step": 808 + }, + { + "epoch": 0.25480314960629924, + "grad_norm": 77.08882141113281, + "learning_rate": 9.998732283464567e-06, + "loss": 0.2146, + "step": 809 + }, + { + "epoch": 0.2551181102362205, + "grad_norm": 25.003931045532227, + "learning_rate": 9.998730708661418e-06, + "loss": 0.3593, + "step": 810 + }, + { + "epoch": 0.25543307086614175, + "grad_norm": 35.63140869140625, + "learning_rate": 9.998729133858267e-06, + "loss": 0.3855, + "step": 811 + }, + { + "epoch": 0.255748031496063, + "grad_norm": 13.581232070922852, + "learning_rate": 9.998727559055118e-06, + "loss": 0.1482, + "step": 812 + }, + { + "epoch": 0.25606299212598427, + "grad_norm": 46.440670013427734, + "learning_rate": 9.99872598425197e-06, + "loss": 0.5954, + "step": 813 + }, + { + "epoch": 0.2563779527559055, + "grad_norm": 24.534271240234375, + "learning_rate": 9.99872440944882e-06, + "loss": 0.2221, + "step": 814 + }, + { + "epoch": 0.2566929133858268, + "grad_norm": 28.46855926513672, + "learning_rate": 9.99872283464567e-06, + "loss": 0.1972, + "step": 815 + }, + { + "epoch": 0.25700787401574804, + "grad_norm": 27.41106605529785, + "learning_rate": 9.99872125984252e-06, + "loss": 0.2933, + "step": 816 + }, + { + "epoch": 0.2573228346456693, + "grad_norm": 77.73954772949219, + "learning_rate": 9.998719685039372e-06, + "loss": 0.5044, + "step": 817 + }, + { + "epoch": 0.25763779527559055, + "grad_norm": 48.21875, + "learning_rate": 9.998718110236221e-06, + "loss": 0.5613, + "step": 818 + }, + { + "epoch": 0.2579527559055118, + "grad_norm": 33.06459045410156, + "learning_rate": 9.998716535433072e-06, + "loss": 0.4511, + "step": 819 + }, + { + "epoch": 0.25826771653543307, + "grad_norm": 91.82710266113281, + "learning_rate": 9.998714960629921e-06, + "loss": 1.3236, + "step": 820 + }, + { + "epoch": 0.25826771653543307, + "eval_loss": 0.4647013545036316, + "eval_runtime": 297.9126, + "eval_samples_per_second": 0.393, + "eval_steps_per_second": 0.393, + "step": 820 + }, + { + "epoch": 0.2585826771653543, + "grad_norm": 80.5277328491211, + "learning_rate": 9.998713385826772e-06, + "loss": 0.7502, + "step": 821 + }, + { + "epoch": 0.2588976377952756, + "grad_norm": 200.29324340820312, + "learning_rate": 9.998711811023623e-06, + "loss": 1.2537, + "step": 822 + }, + { + "epoch": 0.25921259842519684, + "grad_norm": 22.48586654663086, + "learning_rate": 9.998710236220474e-06, + "loss": 0.1774, + "step": 823 + }, + { + "epoch": 0.2595275590551181, + "grad_norm": 43.362831115722656, + "learning_rate": 9.998708661417323e-06, + "loss": 0.336, + "step": 824 + }, + { + "epoch": 0.25984251968503935, + "grad_norm": 27.631332397460938, + "learning_rate": 9.998707086614174e-06, + "loss": 0.2185, + "step": 825 + }, + { + "epoch": 0.2601574803149606, + "grad_norm": 44.31364440917969, + "learning_rate": 9.998705511811024e-06, + "loss": 0.3823, + "step": 826 + }, + { + "epoch": 0.26047244094488187, + "grad_norm": 78.68717193603516, + "learning_rate": 9.998703937007875e-06, + "loss": 0.5907, + "step": 827 + }, + { + "epoch": 0.2607874015748031, + "grad_norm": 86.36324310302734, + "learning_rate": 9.998702362204726e-06, + "loss": 0.7028, + "step": 828 + }, + { + "epoch": 0.26110236220472444, + "grad_norm": 108.75439453125, + "learning_rate": 9.998700787401575e-06, + "loss": 0.9105, + "step": 829 + }, + { + "epoch": 0.2614173228346457, + "grad_norm": 33.248592376708984, + "learning_rate": 9.998699212598426e-06, + "loss": 0.2809, + "step": 830 + }, + { + "epoch": 0.26173228346456695, + "grad_norm": 61.382259368896484, + "learning_rate": 9.998697637795275e-06, + "loss": 0.5749, + "step": 831 + }, + { + "epoch": 0.2620472440944882, + "grad_norm": 34.68425750732422, + "learning_rate": 9.998696062992126e-06, + "loss": 0.5513, + "step": 832 + }, + { + "epoch": 0.26236220472440946, + "grad_norm": 25.614322662353516, + "learning_rate": 9.998694488188977e-06, + "loss": 0.0668, + "step": 833 + }, + { + "epoch": 0.2626771653543307, + "grad_norm": 42.22956848144531, + "learning_rate": 9.998692913385828e-06, + "loss": 0.3998, + "step": 834 + }, + { + "epoch": 0.262992125984252, + "grad_norm": 34.24924087524414, + "learning_rate": 9.998691338582678e-06, + "loss": 0.2085, + "step": 835 + }, + { + "epoch": 0.26330708661417324, + "grad_norm": 72.84844970703125, + "learning_rate": 9.998689763779529e-06, + "loss": 0.7099, + "step": 836 + }, + { + "epoch": 0.2636220472440945, + "grad_norm": 16.073625564575195, + "learning_rate": 9.99868818897638e-06, + "loss": 0.1643, + "step": 837 + }, + { + "epoch": 0.26393700787401575, + "grad_norm": 12.518115997314453, + "learning_rate": 9.998686614173229e-06, + "loss": 0.1489, + "step": 838 + }, + { + "epoch": 0.264251968503937, + "grad_norm": 19.30543327331543, + "learning_rate": 9.99868503937008e-06, + "loss": 0.2271, + "step": 839 + }, + { + "epoch": 0.26456692913385826, + "grad_norm": 80.19608306884766, + "learning_rate": 9.998683464566929e-06, + "loss": 0.3996, + "step": 840 + }, + { + "epoch": 0.26456692913385826, + "eval_loss": 0.46735909581184387, + "eval_runtime": 308.0881, + "eval_samples_per_second": 0.38, + "eval_steps_per_second": 0.38, + "step": 840 + }, + { + "epoch": 0.2648818897637795, + "grad_norm": 47.69651412963867, + "learning_rate": 9.99868188976378e-06, + "loss": 0.4229, + "step": 841 + }, + { + "epoch": 0.2651968503937008, + "grad_norm": 40.88669204711914, + "learning_rate": 9.998680314960631e-06, + "loss": 0.594, + "step": 842 + }, + { + "epoch": 0.26551181102362204, + "grad_norm": 44.67588806152344, + "learning_rate": 9.998678740157482e-06, + "loss": 0.1496, + "step": 843 + }, + { + "epoch": 0.2658267716535433, + "grad_norm": 24.644241333007812, + "learning_rate": 9.998677165354331e-06, + "loss": 0.3401, + "step": 844 + }, + { + "epoch": 0.26614173228346455, + "grad_norm": 18.624984741210938, + "learning_rate": 9.998675590551182e-06, + "loss": 0.2221, + "step": 845 + }, + { + "epoch": 0.2664566929133858, + "grad_norm": 17.709203720092773, + "learning_rate": 9.998674015748032e-06, + "loss": 0.1178, + "step": 846 + }, + { + "epoch": 0.26677165354330706, + "grad_norm": 24.67478370666504, + "learning_rate": 9.998672440944883e-06, + "loss": 0.1389, + "step": 847 + }, + { + "epoch": 0.2670866141732283, + "grad_norm": 31.281604766845703, + "learning_rate": 9.998670866141734e-06, + "loss": 0.2201, + "step": 848 + }, + { + "epoch": 0.26740157480314963, + "grad_norm": 37.749542236328125, + "learning_rate": 9.998669291338583e-06, + "loss": 0.2498, + "step": 849 + }, + { + "epoch": 0.2677165354330709, + "grad_norm": 77.09476470947266, + "learning_rate": 9.998667716535434e-06, + "loss": 1.0283, + "step": 850 + }, + { + "epoch": 0.26803149606299215, + "grad_norm": 63.68134307861328, + "learning_rate": 9.998666141732283e-06, + "loss": 0.4823, + "step": 851 + }, + { + "epoch": 0.2683464566929134, + "grad_norm": 78.47075653076172, + "learning_rate": 9.998664566929134e-06, + "loss": 0.42, + "step": 852 + }, + { + "epoch": 0.26866141732283466, + "grad_norm": 116.93570709228516, + "learning_rate": 9.998662992125985e-06, + "loss": 0.2312, + "step": 853 + }, + { + "epoch": 0.2689763779527559, + "grad_norm": 20.844566345214844, + "learning_rate": 9.998661417322836e-06, + "loss": 0.1281, + "step": 854 + }, + { + "epoch": 0.2692913385826772, + "grad_norm": 26.171772003173828, + "learning_rate": 9.998659842519686e-06, + "loss": 0.2235, + "step": 855 + }, + { + "epoch": 0.26960629921259843, + "grad_norm": 63.36984634399414, + "learning_rate": 9.998658267716537e-06, + "loss": 0.6844, + "step": 856 + }, + { + "epoch": 0.2699212598425197, + "grad_norm": 76.0230941772461, + "learning_rate": 9.998656692913386e-06, + "loss": 0.5884, + "step": 857 + }, + { + "epoch": 0.27023622047244095, + "grad_norm": 57.21022033691406, + "learning_rate": 9.998655118110237e-06, + "loss": 0.6529, + "step": 858 + }, + { + "epoch": 0.2705511811023622, + "grad_norm": 17.419769287109375, + "learning_rate": 9.998653543307088e-06, + "loss": 0.0747, + "step": 859 + }, + { + "epoch": 0.27086614173228346, + "grad_norm": 87.31539154052734, + "learning_rate": 9.998651968503937e-06, + "loss": 0.6404, + "step": 860 + }, + { + "epoch": 0.27086614173228346, + "eval_loss": 0.5078207850456238, + "eval_runtime": 304.5599, + "eval_samples_per_second": 0.384, + "eval_steps_per_second": 0.384, + "step": 860 + }, + { + "epoch": 0.2711811023622047, + "grad_norm": 39.05491638183594, + "learning_rate": 9.998650393700788e-06, + "loss": 0.3182, + "step": 861 + }, + { + "epoch": 0.271496062992126, + "grad_norm": 43.85835647583008, + "learning_rate": 9.998648818897639e-06, + "loss": 0.2367, + "step": 862 + }, + { + "epoch": 0.27181102362204723, + "grad_norm": 61.60994338989258, + "learning_rate": 9.99864724409449e-06, + "loss": 0.7357, + "step": 863 + }, + { + "epoch": 0.2721259842519685, + "grad_norm": 61.263484954833984, + "learning_rate": 9.99864566929134e-06, + "loss": 0.1954, + "step": 864 + }, + { + "epoch": 0.27244094488188975, + "grad_norm": 41.59515380859375, + "learning_rate": 9.99864409448819e-06, + "loss": 0.4512, + "step": 865 + }, + { + "epoch": 0.272755905511811, + "grad_norm": 24.524341583251953, + "learning_rate": 9.99864251968504e-06, + "loss": 0.1307, + "step": 866 + }, + { + "epoch": 0.27307086614173226, + "grad_norm": 58.72618865966797, + "learning_rate": 9.99864094488189e-06, + "loss": 0.707, + "step": 867 + }, + { + "epoch": 0.2733858267716535, + "grad_norm": 46.32933044433594, + "learning_rate": 9.998639370078742e-06, + "loss": 0.218, + "step": 868 + }, + { + "epoch": 0.27370078740157483, + "grad_norm": 36.803565979003906, + "learning_rate": 9.998637795275591e-06, + "loss": 0.1957, + "step": 869 + }, + { + "epoch": 0.2740157480314961, + "grad_norm": 29.207927703857422, + "learning_rate": 9.998636220472442e-06, + "loss": 0.3488, + "step": 870 + }, + { + "epoch": 0.27433070866141734, + "grad_norm": 26.461669921875, + "learning_rate": 9.998634645669291e-06, + "loss": 0.1342, + "step": 871 + }, + { + "epoch": 0.2746456692913386, + "grad_norm": 98.34436798095703, + "learning_rate": 9.998633070866142e-06, + "loss": 0.2521, + "step": 872 + }, + { + "epoch": 0.27496062992125986, + "grad_norm": 40.949153900146484, + "learning_rate": 9.998631496062993e-06, + "loss": 0.3699, + "step": 873 + }, + { + "epoch": 0.2752755905511811, + "grad_norm": 26.012012481689453, + "learning_rate": 9.998629921259844e-06, + "loss": 0.2719, + "step": 874 + }, + { + "epoch": 0.2755905511811024, + "grad_norm": 28.45779800415039, + "learning_rate": 9.998628346456693e-06, + "loss": 0.1321, + "step": 875 + }, + { + "epoch": 0.27590551181102363, + "grad_norm": 63.54460144042969, + "learning_rate": 9.998626771653544e-06, + "loss": 0.5929, + "step": 876 + }, + { + "epoch": 0.2762204724409449, + "grad_norm": 101.21806335449219, + "learning_rate": 9.998625196850394e-06, + "loss": 0.6633, + "step": 877 + }, + { + "epoch": 0.27653543307086614, + "grad_norm": 8.460577964782715, + "learning_rate": 9.998623622047245e-06, + "loss": 0.0187, + "step": 878 + }, + { + "epoch": 0.2768503937007874, + "grad_norm": 101.49215698242188, + "learning_rate": 9.998622047244096e-06, + "loss": 0.7255, + "step": 879 + }, + { + "epoch": 0.27716535433070866, + "grad_norm": 110.3086929321289, + "learning_rate": 9.998620472440945e-06, + "loss": 0.9578, + "step": 880 + }, + { + "epoch": 0.27716535433070866, + "eval_loss": 0.49943187832832336, + "eval_runtime": 306.9283, + "eval_samples_per_second": 0.381, + "eval_steps_per_second": 0.381, + "step": 880 + }, + { + "epoch": 0.2774803149606299, + "grad_norm": 6.745794773101807, + "learning_rate": 9.998618897637796e-06, + "loss": 0.0379, + "step": 881 + }, + { + "epoch": 0.27779527559055117, + "grad_norm": 51.96881866455078, + "learning_rate": 9.998617322834645e-06, + "loss": 0.7563, + "step": 882 + }, + { + "epoch": 0.27811023622047243, + "grad_norm": 52.106834411621094, + "learning_rate": 9.998615748031496e-06, + "loss": 0.4385, + "step": 883 + }, + { + "epoch": 0.2784251968503937, + "grad_norm": 69.72996520996094, + "learning_rate": 9.998614173228347e-06, + "loss": 0.8885, + "step": 884 + }, + { + "epoch": 0.27874015748031494, + "grad_norm": 67.1977310180664, + "learning_rate": 9.998612598425198e-06, + "loss": 0.814, + "step": 885 + }, + { + "epoch": 0.2790551181102362, + "grad_norm": 111.88276672363281, + "learning_rate": 9.998611023622048e-06, + "loss": 0.6336, + "step": 886 + }, + { + "epoch": 0.27937007874015746, + "grad_norm": 40.74708557128906, + "learning_rate": 9.998609448818899e-06, + "loss": 0.2043, + "step": 887 + }, + { + "epoch": 0.2796850393700787, + "grad_norm": 42.77908706665039, + "learning_rate": 9.99860787401575e-06, + "loss": 0.2083, + "step": 888 + }, + { + "epoch": 0.28, + "grad_norm": 57.296024322509766, + "learning_rate": 9.998606299212599e-06, + "loss": 0.1104, + "step": 889 + }, + { + "epoch": 0.2803149606299213, + "grad_norm": 32.91524124145508, + "learning_rate": 9.99860472440945e-06, + "loss": 0.498, + "step": 890 + }, + { + "epoch": 0.28062992125984254, + "grad_norm": 71.90412139892578, + "learning_rate": 9.9986031496063e-06, + "loss": 0.8651, + "step": 891 + }, + { + "epoch": 0.2809448818897638, + "grad_norm": 11.306217193603516, + "learning_rate": 9.99860157480315e-06, + "loss": 0.0642, + "step": 892 + }, + { + "epoch": 0.28125984251968505, + "grad_norm": 15.339298248291016, + "learning_rate": 9.998600000000001e-06, + "loss": 0.0829, + "step": 893 + }, + { + "epoch": 0.2815748031496063, + "grad_norm": 59.414466857910156, + "learning_rate": 9.998598425196852e-06, + "loss": 0.5543, + "step": 894 + }, + { + "epoch": 0.28188976377952757, + "grad_norm": 66.70774841308594, + "learning_rate": 9.998596850393701e-06, + "loss": 0.3054, + "step": 895 + }, + { + "epoch": 0.2822047244094488, + "grad_norm": 20.609098434448242, + "learning_rate": 9.998595275590552e-06, + "loss": 0.1729, + "step": 896 + }, + { + "epoch": 0.2825196850393701, + "grad_norm": 7.707085609436035, + "learning_rate": 9.998593700787402e-06, + "loss": 0.0473, + "step": 897 + }, + { + "epoch": 0.28283464566929134, + "grad_norm": 41.48631286621094, + "learning_rate": 9.998592125984253e-06, + "loss": 0.4578, + "step": 898 + }, + { + "epoch": 0.2831496062992126, + "grad_norm": 23.614355087280273, + "learning_rate": 9.998590551181104e-06, + "loss": 0.193, + "step": 899 + }, + { + "epoch": 0.28346456692913385, + "grad_norm": 15.626468658447266, + "learning_rate": 9.998588976377953e-06, + "loss": 0.0845, + "step": 900 + }, + { + "epoch": 0.28346456692913385, + "eval_loss": 0.5890966653823853, + "eval_runtime": 302.1316, + "eval_samples_per_second": 0.387, + "eval_steps_per_second": 0.387, + "step": 900 + }, + { + "epoch": 0.2837795275590551, + "grad_norm": 60.57646942138672, + "learning_rate": 9.998587401574804e-06, + "loss": 0.3963, + "step": 901 + }, + { + "epoch": 0.28409448818897637, + "grad_norm": 47.83597946166992, + "learning_rate": 9.998585826771653e-06, + "loss": 0.6909, + "step": 902 + }, + { + "epoch": 0.2844094488188976, + "grad_norm": 26.267818450927734, + "learning_rate": 9.998584251968504e-06, + "loss": 0.0956, + "step": 903 + }, + { + "epoch": 0.2847244094488189, + "grad_norm": 63.959110260009766, + "learning_rate": 9.998582677165355e-06, + "loss": 0.4272, + "step": 904 + }, + { + "epoch": 0.28503937007874014, + "grad_norm": 59.13768005371094, + "learning_rate": 9.998581102362206e-06, + "loss": 0.3178, + "step": 905 + }, + { + "epoch": 0.2853543307086614, + "grad_norm": 39.657814025878906, + "learning_rate": 9.998579527559056e-06, + "loss": 0.6433, + "step": 906 + }, + { + "epoch": 0.28566929133858265, + "grad_norm": 90.1864013671875, + "learning_rate": 9.998577952755907e-06, + "loss": 0.6963, + "step": 907 + }, + { + "epoch": 0.2859842519685039, + "grad_norm": 121.48163604736328, + "learning_rate": 9.998576377952758e-06, + "loss": 1.8743, + "step": 908 + }, + { + "epoch": 0.2862992125984252, + "grad_norm": 37.84361267089844, + "learning_rate": 9.998574803149607e-06, + "loss": 0.3433, + "step": 909 + }, + { + "epoch": 0.2866141732283465, + "grad_norm": 42.82717514038086, + "learning_rate": 9.998573228346458e-06, + "loss": 0.3181, + "step": 910 + }, + { + "epoch": 0.28692913385826774, + "grad_norm": 55.58892822265625, + "learning_rate": 9.998571653543307e-06, + "loss": 1.2413, + "step": 911 + }, + { + "epoch": 0.287244094488189, + "grad_norm": 56.79817581176758, + "learning_rate": 9.998570078740158e-06, + "loss": 0.3707, + "step": 912 + }, + { + "epoch": 0.28755905511811025, + "grad_norm": 54.66569900512695, + "learning_rate": 9.998568503937009e-06, + "loss": 0.7417, + "step": 913 + }, + { + "epoch": 0.2878740157480315, + "grad_norm": 22.454544067382812, + "learning_rate": 9.99856692913386e-06, + "loss": 0.1303, + "step": 914 + }, + { + "epoch": 0.28818897637795277, + "grad_norm": 72.16681671142578, + "learning_rate": 9.99856535433071e-06, + "loss": 0.4007, + "step": 915 + }, + { + "epoch": 0.288503937007874, + "grad_norm": 52.15703201293945, + "learning_rate": 9.99856377952756e-06, + "loss": 0.7637, + "step": 916 + }, + { + "epoch": 0.2888188976377953, + "grad_norm": 8.107488632202148, + "learning_rate": 9.99856220472441e-06, + "loss": 0.0643, + "step": 917 + }, + { + "epoch": 0.28913385826771654, + "grad_norm": 47.568267822265625, + "learning_rate": 9.99856062992126e-06, + "loss": 0.4449, + "step": 918 + }, + { + "epoch": 0.2894488188976378, + "grad_norm": 21.596525192260742, + "learning_rate": 9.998559055118112e-06, + "loss": 0.1228, + "step": 919 + }, + { + "epoch": 0.28976377952755905, + "grad_norm": 40.52389144897461, + "learning_rate": 9.998557480314961e-06, + "loss": 0.446, + "step": 920 + }, + { + "epoch": 0.28976377952755905, + "eval_loss": 0.5098508596420288, + "eval_runtime": 307.5399, + "eval_samples_per_second": 0.38, + "eval_steps_per_second": 0.38, + "step": 920 + }, + { + "epoch": 0.2900787401574803, + "grad_norm": 43.565303802490234, + "learning_rate": 9.998555905511812e-06, + "loss": 0.5184, + "step": 921 + }, + { + "epoch": 0.29039370078740157, + "grad_norm": 54.92490768432617, + "learning_rate": 9.998554330708661e-06, + "loss": 0.7516, + "step": 922 + }, + { + "epoch": 0.2907086614173228, + "grad_norm": 47.38011169433594, + "learning_rate": 9.998552755905512e-06, + "loss": 0.4989, + "step": 923 + }, + { + "epoch": 0.2910236220472441, + "grad_norm": 63.52509689331055, + "learning_rate": 9.998551181102363e-06, + "loss": 0.4689, + "step": 924 + }, + { + "epoch": 0.29133858267716534, + "grad_norm": 38.14700698852539, + "learning_rate": 9.998549606299214e-06, + "loss": 0.5137, + "step": 925 + }, + { + "epoch": 0.2916535433070866, + "grad_norm": 63.95713806152344, + "learning_rate": 9.998548031496063e-06, + "loss": 0.3511, + "step": 926 + }, + { + "epoch": 0.29196850393700785, + "grad_norm": 38.79820251464844, + "learning_rate": 9.998546456692914e-06, + "loss": 0.3497, + "step": 927 + }, + { + "epoch": 0.2922834645669291, + "grad_norm": 76.02424621582031, + "learning_rate": 9.998544881889764e-06, + "loss": 0.416, + "step": 928 + }, + { + "epoch": 0.2925984251968504, + "grad_norm": 45.44684982299805, + "learning_rate": 9.998543307086615e-06, + "loss": 0.4091, + "step": 929 + }, + { + "epoch": 0.2929133858267717, + "grad_norm": 32.81157684326172, + "learning_rate": 9.998541732283466e-06, + "loss": 0.4614, + "step": 930 + }, + { + "epoch": 0.29322834645669293, + "grad_norm": 45.81043243408203, + "learning_rate": 9.998540157480315e-06, + "loss": 0.4832, + "step": 931 + }, + { + "epoch": 0.2935433070866142, + "grad_norm": 42.5070915222168, + "learning_rate": 9.998538582677166e-06, + "loss": 0.352, + "step": 932 + }, + { + "epoch": 0.29385826771653545, + "grad_norm": 40.78940200805664, + "learning_rate": 9.998537007874017e-06, + "loss": 0.9412, + "step": 933 + }, + { + "epoch": 0.2941732283464567, + "grad_norm": 44.96437454223633, + "learning_rate": 9.998535433070868e-06, + "loss": 0.9758, + "step": 934 + }, + { + "epoch": 0.29448818897637796, + "grad_norm": 8.247536659240723, + "learning_rate": 9.998533858267717e-06, + "loss": 0.0863, + "step": 935 + }, + { + "epoch": 0.2948031496062992, + "grad_norm": 72.65038299560547, + "learning_rate": 9.998532283464568e-06, + "loss": 0.6317, + "step": 936 + }, + { + "epoch": 0.2951181102362205, + "grad_norm": 28.256349563598633, + "learning_rate": 9.998530708661418e-06, + "loss": 0.3451, + "step": 937 + }, + { + "epoch": 0.29543307086614173, + "grad_norm": 17.122854232788086, + "learning_rate": 9.998529133858269e-06, + "loss": 0.2172, + "step": 938 + }, + { + "epoch": 0.295748031496063, + "grad_norm": 25.158649444580078, + "learning_rate": 9.99852755905512e-06, + "loss": 0.3431, + "step": 939 + }, + { + "epoch": 0.29606299212598425, + "grad_norm": 48.18063735961914, + "learning_rate": 9.998525984251969e-06, + "loss": 0.6472, + "step": 940 + }, + { + "epoch": 0.29606299212598425, + "eval_loss": 0.4980691373348236, + "eval_runtime": 301.7364, + "eval_samples_per_second": 0.388, + "eval_steps_per_second": 0.388, + "step": 940 + }, + { + "epoch": 0.2963779527559055, + "grad_norm": 93.8964614868164, + "learning_rate": 9.99852440944882e-06, + "loss": 0.8136, + "step": 941 + }, + { + "epoch": 0.29669291338582676, + "grad_norm": 19.963516235351562, + "learning_rate": 9.99852283464567e-06, + "loss": 0.194, + "step": 942 + }, + { + "epoch": 0.297007874015748, + "grad_norm": 20.99871826171875, + "learning_rate": 9.99852125984252e-06, + "loss": 0.294, + "step": 943 + }, + { + "epoch": 0.2973228346456693, + "grad_norm": 43.991703033447266, + "learning_rate": 9.998519685039371e-06, + "loss": 0.4679, + "step": 944 + }, + { + "epoch": 0.29763779527559053, + "grad_norm": 48.08595275878906, + "learning_rate": 9.998518110236222e-06, + "loss": 0.6872, + "step": 945 + }, + { + "epoch": 0.2979527559055118, + "grad_norm": 14.398991584777832, + "learning_rate": 9.998516535433071e-06, + "loss": 0.1213, + "step": 946 + }, + { + "epoch": 0.29826771653543305, + "grad_norm": 60.385154724121094, + "learning_rate": 9.998514960629922e-06, + "loss": 0.9587, + "step": 947 + }, + { + "epoch": 0.2985826771653543, + "grad_norm": 23.951004028320312, + "learning_rate": 9.998513385826772e-06, + "loss": 0.1824, + "step": 948 + }, + { + "epoch": 0.2988976377952756, + "grad_norm": 24.64979362487793, + "learning_rate": 9.998511811023623e-06, + "loss": 0.2494, + "step": 949 + }, + { + "epoch": 0.2992125984251969, + "grad_norm": 33.05177307128906, + "learning_rate": 9.998510236220474e-06, + "loss": 0.8105, + "step": 950 + }, + { + "epoch": 0.29952755905511813, + "grad_norm": 59.83413314819336, + "learning_rate": 9.998508661417323e-06, + "loss": 0.6087, + "step": 951 + }, + { + "epoch": 0.2998425196850394, + "grad_norm": 79.51978302001953, + "learning_rate": 9.998507086614174e-06, + "loss": 0.8573, + "step": 952 + }, + { + "epoch": 0.30015748031496065, + "grad_norm": 59.22591018676758, + "learning_rate": 9.998505511811023e-06, + "loss": 0.7464, + "step": 953 + }, + { + "epoch": 0.3004724409448819, + "grad_norm": 53.090614318847656, + "learning_rate": 9.998503937007876e-06, + "loss": 0.6472, + "step": 954 + }, + { + "epoch": 0.30078740157480316, + "grad_norm": 35.13800048828125, + "learning_rate": 9.998502362204725e-06, + "loss": 0.2764, + "step": 955 + }, + { + "epoch": 0.3011023622047244, + "grad_norm": 18.853023529052734, + "learning_rate": 9.998500787401576e-06, + "loss": 0.1602, + "step": 956 + }, + { + "epoch": 0.3014173228346457, + "grad_norm": 13.263671875, + "learning_rate": 9.998499212598426e-06, + "loss": 0.1145, + "step": 957 + }, + { + "epoch": 0.30173228346456693, + "grad_norm": 38.00738525390625, + "learning_rate": 9.998497637795277e-06, + "loss": 0.1779, + "step": 958 + }, + { + "epoch": 0.3020472440944882, + "grad_norm": 29.51807403564453, + "learning_rate": 9.998496062992128e-06, + "loss": 0.3998, + "step": 959 + }, + { + "epoch": 0.30236220472440944, + "grad_norm": 49.63967514038086, + "learning_rate": 9.998494488188977e-06, + "loss": 1.2491, + "step": 960 + }, + { + "epoch": 0.30236220472440944, + "eval_loss": 0.505746066570282, + "eval_runtime": 303.3875, + "eval_samples_per_second": 0.386, + "eval_steps_per_second": 0.386, + "step": 960 + }, + { + "epoch": 0.3026771653543307, + "grad_norm": 22.166236877441406, + "learning_rate": 9.998492913385828e-06, + "loss": 0.2884, + "step": 961 + }, + { + "epoch": 0.30299212598425196, + "grad_norm": 9.977280616760254, + "learning_rate": 9.998491338582677e-06, + "loss": 0.0985, + "step": 962 + }, + { + "epoch": 0.3033070866141732, + "grad_norm": 51.85643005371094, + "learning_rate": 9.998489763779528e-06, + "loss": 0.4792, + "step": 963 + }, + { + "epoch": 0.3036220472440945, + "grad_norm": 11.42037582397461, + "learning_rate": 9.998488188976379e-06, + "loss": 0.0609, + "step": 964 + }, + { + "epoch": 0.30393700787401573, + "grad_norm": 29.179784774780273, + "learning_rate": 9.99848661417323e-06, + "loss": 0.2065, + "step": 965 + }, + { + "epoch": 0.304251968503937, + "grad_norm": 99.96176147460938, + "learning_rate": 9.99848503937008e-06, + "loss": 0.1855, + "step": 966 + }, + { + "epoch": 0.30456692913385824, + "grad_norm": 67.8636245727539, + "learning_rate": 9.99848346456693e-06, + "loss": 0.4098, + "step": 967 + }, + { + "epoch": 0.3048818897637795, + "grad_norm": 67.14287567138672, + "learning_rate": 9.99848188976378e-06, + "loss": 0.8962, + "step": 968 + }, + { + "epoch": 0.3051968503937008, + "grad_norm": 52.694175720214844, + "learning_rate": 9.99848031496063e-06, + "loss": 0.3033, + "step": 969 + }, + { + "epoch": 0.30551181102362207, + "grad_norm": 124.31986236572266, + "learning_rate": 9.998478740157482e-06, + "loss": 0.6216, + "step": 970 + }, + { + "epoch": 0.30582677165354333, + "grad_norm": 59.26445770263672, + "learning_rate": 9.998477165354331e-06, + "loss": 0.4527, + "step": 971 + }, + { + "epoch": 0.3061417322834646, + "grad_norm": 28.591516494750977, + "learning_rate": 9.998475590551182e-06, + "loss": 0.1072, + "step": 972 + }, + { + "epoch": 0.30645669291338584, + "grad_norm": 117.83760833740234, + "learning_rate": 9.998474015748031e-06, + "loss": 0.622, + "step": 973 + }, + { + "epoch": 0.3067716535433071, + "grad_norm": 34.50752639770508, + "learning_rate": 9.998472440944882e-06, + "loss": 0.4129, + "step": 974 + }, + { + "epoch": 0.30708661417322836, + "grad_norm": 8.890520095825195, + "learning_rate": 9.998470866141733e-06, + "loss": 0.0409, + "step": 975 + }, + { + "epoch": 0.3074015748031496, + "grad_norm": 52.015785217285156, + "learning_rate": 9.998469291338584e-06, + "loss": 0.4855, + "step": 976 + }, + { + "epoch": 0.30771653543307087, + "grad_norm": 23.1546688079834, + "learning_rate": 9.998467716535434e-06, + "loss": 0.1683, + "step": 977 + }, + { + "epoch": 0.3080314960629921, + "grad_norm": 26.71653938293457, + "learning_rate": 9.998466141732285e-06, + "loss": 0.1988, + "step": 978 + }, + { + "epoch": 0.3083464566929134, + "grad_norm": 55.756309509277344, + "learning_rate": 9.998464566929135e-06, + "loss": 0.6107, + "step": 979 + }, + { + "epoch": 0.30866141732283464, + "grad_norm": 38.31888198852539, + "learning_rate": 9.998462992125985e-06, + "loss": 0.3716, + "step": 980 + }, + { + "epoch": 0.30866141732283464, + "eval_loss": 0.4945707321166992, + "eval_runtime": 303.1038, + "eval_samples_per_second": 0.386, + "eval_steps_per_second": 0.386, + "step": 980 + }, + { + "epoch": 0.3089763779527559, + "grad_norm": 136.65208435058594, + "learning_rate": 9.998461417322836e-06, + "loss": 1.04, + "step": 981 + }, + { + "epoch": 0.30929133858267716, + "grad_norm": 70.81889343261719, + "learning_rate": 9.998459842519685e-06, + "loss": 0.7749, + "step": 982 + }, + { + "epoch": 0.3096062992125984, + "grad_norm": 81.47750854492188, + "learning_rate": 9.998458267716536e-06, + "loss": 0.4565, + "step": 983 + }, + { + "epoch": 0.30992125984251967, + "grad_norm": 84.3187255859375, + "learning_rate": 9.998456692913387e-06, + "loss": 1.1115, + "step": 984 + }, + { + "epoch": 0.3102362204724409, + "grad_norm": 121.83627319335938, + "learning_rate": 9.998455118110238e-06, + "loss": 0.4089, + "step": 985 + }, + { + "epoch": 0.3105511811023622, + "grad_norm": 46.44225311279297, + "learning_rate": 9.998453543307087e-06, + "loss": 0.6732, + "step": 986 + }, + { + "epoch": 0.31086614173228344, + "grad_norm": 60.36497497558594, + "learning_rate": 9.998451968503938e-06, + "loss": 0.3421, + "step": 987 + }, + { + "epoch": 0.3111811023622047, + "grad_norm": 34.300819396972656, + "learning_rate": 9.998450393700788e-06, + "loss": 0.1633, + "step": 988 + }, + { + "epoch": 0.311496062992126, + "grad_norm": 29.893108367919922, + "learning_rate": 9.998448818897639e-06, + "loss": 0.2435, + "step": 989 + }, + { + "epoch": 0.31181102362204727, + "grad_norm": 56.01438522338867, + "learning_rate": 9.99844724409449e-06, + "loss": 0.2676, + "step": 990 + }, + { + "epoch": 0.3121259842519685, + "grad_norm": 58.90812301635742, + "learning_rate": 9.998445669291339e-06, + "loss": 0.362, + "step": 991 + }, + { + "epoch": 0.3124409448818898, + "grad_norm": 54.015743255615234, + "learning_rate": 9.99844409448819e-06, + "loss": 0.3922, + "step": 992 + }, + { + "epoch": 0.31275590551181104, + "grad_norm": 43.66928482055664, + "learning_rate": 9.99844251968504e-06, + "loss": 0.3873, + "step": 993 + }, + { + "epoch": 0.3130708661417323, + "grad_norm": 37.3400764465332, + "learning_rate": 9.99844094488189e-06, + "loss": 0.1752, + "step": 994 + }, + { + "epoch": 0.31338582677165355, + "grad_norm": 22.93543243408203, + "learning_rate": 9.998439370078741e-06, + "loss": 0.2166, + "step": 995 + }, + { + "epoch": 0.3137007874015748, + "grad_norm": 54.434444427490234, + "learning_rate": 9.998437795275592e-06, + "loss": 0.6486, + "step": 996 + }, + { + "epoch": 0.31401574803149607, + "grad_norm": 29.223251342773438, + "learning_rate": 9.998436220472441e-06, + "loss": 0.2945, + "step": 997 + }, + { + "epoch": 0.3143307086614173, + "grad_norm": 40.362152099609375, + "learning_rate": 9.998434645669292e-06, + "loss": 0.6047, + "step": 998 + }, + { + "epoch": 0.3146456692913386, + "grad_norm": 17.493072509765625, + "learning_rate": 9.998433070866142e-06, + "loss": 0.3557, + "step": 999 + }, + { + "epoch": 0.31496062992125984, + "grad_norm": 32.33248519897461, + "learning_rate": 9.998431496062993e-06, + "loss": 0.1763, + "step": 1000 + }, + { + "epoch": 0.31496062992125984, + "eval_loss": 0.4581963121891022, + "eval_runtime": 306.4597, + "eval_samples_per_second": 0.382, + "eval_steps_per_second": 0.382, + "step": 1000 + }, + { + "epoch": 0.3152755905511811, + "grad_norm": 41.288719177246094, + "learning_rate": 9.998429921259844e-06, + "loss": 0.3175, + "step": 1001 + }, + { + "epoch": 0.31559055118110235, + "grad_norm": 21.213180541992188, + "learning_rate": 9.998428346456693e-06, + "loss": 0.1175, + "step": 1002 + }, + { + "epoch": 0.3159055118110236, + "grad_norm": 35.53864669799805, + "learning_rate": 9.998426771653544e-06, + "loss": 0.2783, + "step": 1003 + }, + { + "epoch": 0.31622047244094487, + "grad_norm": 46.64375305175781, + "learning_rate": 9.998425196850395e-06, + "loss": 0.4337, + "step": 1004 + }, + { + "epoch": 0.3165354330708661, + "grad_norm": 25.14255142211914, + "learning_rate": 9.998423622047246e-06, + "loss": 0.3534, + "step": 1005 + }, + { + "epoch": 0.3168503937007874, + "grad_norm": 64.11309051513672, + "learning_rate": 9.998422047244095e-06, + "loss": 0.1756, + "step": 1006 + }, + { + "epoch": 0.31716535433070864, + "grad_norm": 54.7104377746582, + "learning_rate": 9.998420472440946e-06, + "loss": 0.9319, + "step": 1007 + }, + { + "epoch": 0.3174803149606299, + "grad_norm": 84.07182312011719, + "learning_rate": 9.998418897637796e-06, + "loss": 0.9422, + "step": 1008 + }, + { + "epoch": 0.3177952755905512, + "grad_norm": 46.56437683105469, + "learning_rate": 9.998417322834647e-06, + "loss": 0.2973, + "step": 1009 + }, + { + "epoch": 0.31811023622047246, + "grad_norm": 39.493682861328125, + "learning_rate": 9.998415748031498e-06, + "loss": 0.4972, + "step": 1010 + }, + { + "epoch": 0.3184251968503937, + "grad_norm": 51.46127700805664, + "learning_rate": 9.998414173228347e-06, + "loss": 0.6015, + "step": 1011 + }, + { + "epoch": 0.318740157480315, + "grad_norm": 27.754695892333984, + "learning_rate": 9.998412598425198e-06, + "loss": 0.2744, + "step": 1012 + }, + { + "epoch": 0.31905511811023624, + "grad_norm": 26.403295516967773, + "learning_rate": 9.998411023622047e-06, + "loss": 0.2464, + "step": 1013 + }, + { + "epoch": 0.3193700787401575, + "grad_norm": 66.5213623046875, + "learning_rate": 9.998409448818898e-06, + "loss": 0.3692, + "step": 1014 + }, + { + "epoch": 0.31968503937007875, + "grad_norm": 35.4731559753418, + "learning_rate": 9.998407874015749e-06, + "loss": 0.3412, + "step": 1015 + }, + { + "epoch": 0.32, + "grad_norm": 23.23394203186035, + "learning_rate": 9.9984062992126e-06, + "loss": 0.4104, + "step": 1016 + }, + { + "epoch": 0.32031496062992126, + "grad_norm": 44.30449676513672, + "learning_rate": 9.99840472440945e-06, + "loss": 0.4564, + "step": 1017 + }, + { + "epoch": 0.3206299212598425, + "grad_norm": 18.734786987304688, + "learning_rate": 9.9984031496063e-06, + "loss": 0.2802, + "step": 1018 + }, + { + "epoch": 0.3209448818897638, + "grad_norm": 40.950653076171875, + "learning_rate": 9.99840157480315e-06, + "loss": 0.3802, + "step": 1019 + }, + { + "epoch": 0.32125984251968503, + "grad_norm": 97.77069091796875, + "learning_rate": 9.9984e-06, + "loss": 0.467, + "step": 1020 + }, + { + "epoch": 0.32125984251968503, + "eval_loss": 0.4887102544307709, + "eval_runtime": 304.7183, + "eval_samples_per_second": 0.384, + "eval_steps_per_second": 0.384, + "step": 1020 + }, + { + "epoch": 0.3215748031496063, + "grad_norm": 131.0202178955078, + "learning_rate": 9.998398425196852e-06, + "loss": 0.9191, + "step": 1021 + }, + { + "epoch": 0.32188976377952755, + "grad_norm": 72.55668640136719, + "learning_rate": 9.998396850393701e-06, + "loss": 0.9462, + "step": 1022 + }, + { + "epoch": 0.3222047244094488, + "grad_norm": 47.58323669433594, + "learning_rate": 9.998395275590552e-06, + "loss": 0.7648, + "step": 1023 + }, + { + "epoch": 0.32251968503937006, + "grad_norm": 15.815386772155762, + "learning_rate": 9.998393700787403e-06, + "loss": 0.1747, + "step": 1024 + }, + { + "epoch": 0.3228346456692913, + "grad_norm": 23.02329444885254, + "learning_rate": 9.998392125984254e-06, + "loss": 0.2737, + "step": 1025 + }, + { + "epoch": 0.3231496062992126, + "grad_norm": 69.93936157226562, + "learning_rate": 9.998390551181103e-06, + "loss": 0.6328, + "step": 1026 + }, + { + "epoch": 0.32346456692913383, + "grad_norm": 44.680335998535156, + "learning_rate": 9.998388976377954e-06, + "loss": 0.2808, + "step": 1027 + }, + { + "epoch": 0.3237795275590551, + "grad_norm": 61.480506896972656, + "learning_rate": 9.998387401574804e-06, + "loss": 0.3987, + "step": 1028 + }, + { + "epoch": 0.3240944881889764, + "grad_norm": 49.2515754699707, + "learning_rate": 9.998385826771655e-06, + "loss": 0.2713, + "step": 1029 + }, + { + "epoch": 0.32440944881889766, + "grad_norm": 39.322608947753906, + "learning_rate": 9.998384251968506e-06, + "loss": 0.5236, + "step": 1030 + }, + { + "epoch": 0.3247244094488189, + "grad_norm": 50.91118621826172, + "learning_rate": 9.998382677165355e-06, + "loss": 0.3365, + "step": 1031 + }, + { + "epoch": 0.3250393700787402, + "grad_norm": 57.56412887573242, + "learning_rate": 9.998381102362206e-06, + "loss": 0.5609, + "step": 1032 + }, + { + "epoch": 0.32535433070866143, + "grad_norm": 106.12908935546875, + "learning_rate": 9.998379527559055e-06, + "loss": 0.6946, + "step": 1033 + }, + { + "epoch": 0.3256692913385827, + "grad_norm": 28.783655166625977, + "learning_rate": 9.998377952755906e-06, + "loss": 0.3571, + "step": 1034 + }, + { + "epoch": 0.32598425196850395, + "grad_norm": 35.49094772338867, + "learning_rate": 9.998376377952757e-06, + "loss": 0.2201, + "step": 1035 + }, + { + "epoch": 0.3262992125984252, + "grad_norm": 41.67929458618164, + "learning_rate": 9.998374803149608e-06, + "loss": 0.2486, + "step": 1036 + }, + { + "epoch": 0.32661417322834646, + "grad_norm": 55.53398132324219, + "learning_rate": 9.998373228346457e-06, + "loss": 0.6653, + "step": 1037 + }, + { + "epoch": 0.3269291338582677, + "grad_norm": 63.74702835083008, + "learning_rate": 9.998371653543308e-06, + "loss": 0.3952, + "step": 1038 + }, + { + "epoch": 0.327244094488189, + "grad_norm": 43.730770111083984, + "learning_rate": 9.998370078740158e-06, + "loss": 0.2713, + "step": 1039 + }, + { + "epoch": 0.32755905511811023, + "grad_norm": 20.90167808532715, + "learning_rate": 9.998368503937009e-06, + "loss": 0.3298, + "step": 1040 + }, + { + "epoch": 0.32755905511811023, + "eval_loss": 0.553530216217041, + "eval_runtime": 338.2916, + "eval_samples_per_second": 0.346, + "eval_steps_per_second": 0.346, + "step": 1040 + }, + { + "epoch": 0.3278740157480315, + "grad_norm": 53.132686614990234, + "learning_rate": 9.99836692913386e-06, + "loss": 0.2454, + "step": 1041 + }, + { + "epoch": 0.32818897637795275, + "grad_norm": 62.30380630493164, + "learning_rate": 9.998365354330709e-06, + "loss": 0.6391, + "step": 1042 + }, + { + "epoch": 0.328503937007874, + "grad_norm": 34.83534622192383, + "learning_rate": 9.99836377952756e-06, + "loss": 0.2441, + "step": 1043 + }, + { + "epoch": 0.32881889763779526, + "grad_norm": 37.7114372253418, + "learning_rate": 9.99836220472441e-06, + "loss": 0.1986, + "step": 1044 + }, + { + "epoch": 0.3291338582677165, + "grad_norm": 26.971843719482422, + "learning_rate": 9.99836062992126e-06, + "loss": 0.1753, + "step": 1045 + }, + { + "epoch": 0.3294488188976378, + "grad_norm": 64.88057708740234, + "learning_rate": 9.998359055118111e-06, + "loss": 0.3672, + "step": 1046 + }, + { + "epoch": 0.32976377952755903, + "grad_norm": 16.45389747619629, + "learning_rate": 9.998357480314962e-06, + "loss": 0.0662, + "step": 1047 + }, + { + "epoch": 0.3300787401574803, + "grad_norm": 53.416934967041016, + "learning_rate": 9.998355905511811e-06, + "loss": 0.6081, + "step": 1048 + }, + { + "epoch": 0.3303937007874016, + "grad_norm": 44.44949722290039, + "learning_rate": 9.998354330708662e-06, + "loss": 0.2946, + "step": 1049 + }, + { + "epoch": 0.33070866141732286, + "grad_norm": 85.5646743774414, + "learning_rate": 9.998352755905513e-06, + "loss": 0.7022, + "step": 1050 + }, + { + "epoch": 0.3310236220472441, + "grad_norm": 158.97926330566406, + "learning_rate": 9.998351181102363e-06, + "loss": 0.3364, + "step": 1051 + }, + { + "epoch": 0.33133858267716537, + "grad_norm": 53.93890380859375, + "learning_rate": 9.998349606299214e-06, + "loss": 0.0856, + "step": 1052 + }, + { + "epoch": 0.33165354330708663, + "grad_norm": 58.113338470458984, + "learning_rate": 9.998348031496063e-06, + "loss": 0.4051, + "step": 1053 + }, + { + "epoch": 0.3319685039370079, + "grad_norm": 113.30328369140625, + "learning_rate": 9.998346456692914e-06, + "loss": 1.3079, + "step": 1054 + }, + { + "epoch": 0.33228346456692914, + "grad_norm": 52.5933952331543, + "learning_rate": 9.998344881889765e-06, + "loss": 0.2001, + "step": 1055 + }, + { + "epoch": 0.3325984251968504, + "grad_norm": 111.05465698242188, + "learning_rate": 9.998343307086616e-06, + "loss": 0.8016, + "step": 1056 + }, + { + "epoch": 0.33291338582677166, + "grad_norm": 80.2440185546875, + "learning_rate": 9.998341732283465e-06, + "loss": 0.6905, + "step": 1057 + }, + { + "epoch": 0.3332283464566929, + "grad_norm": 46.98655319213867, + "learning_rate": 9.998340157480316e-06, + "loss": 0.645, + "step": 1058 + }, + { + "epoch": 0.33354330708661417, + "grad_norm": 88.55839538574219, + "learning_rate": 9.998338582677166e-06, + "loss": 0.632, + "step": 1059 + }, + { + "epoch": 0.33385826771653543, + "grad_norm": 45.15827178955078, + "learning_rate": 9.998337007874017e-06, + "loss": 0.3994, + "step": 1060 + }, + { + "epoch": 0.33385826771653543, + "eval_loss": 0.49371591210365295, + "eval_runtime": 338.6729, + "eval_samples_per_second": 0.345, + "eval_steps_per_second": 0.345, + "step": 1060 + }, + { + "epoch": 0.3341732283464567, + "grad_norm": 89.61175537109375, + "learning_rate": 9.998335433070868e-06, + "loss": 0.7211, + "step": 1061 + }, + { + "epoch": 0.33448818897637794, + "grad_norm": 43.96299362182617, + "learning_rate": 9.998333858267717e-06, + "loss": 0.2502, + "step": 1062 + }, + { + "epoch": 0.3348031496062992, + "grad_norm": 48.51971435546875, + "learning_rate": 9.998332283464568e-06, + "loss": 0.2617, + "step": 1063 + }, + { + "epoch": 0.33511811023622046, + "grad_norm": 56.80341720581055, + "learning_rate": 9.998330708661417e-06, + "loss": 0.315, + "step": 1064 + }, + { + "epoch": 0.3354330708661417, + "grad_norm": 24.417354583740234, + "learning_rate": 9.998329133858268e-06, + "loss": 0.2498, + "step": 1065 + }, + { + "epoch": 0.33574803149606297, + "grad_norm": 21.887855529785156, + "learning_rate": 9.99832755905512e-06, + "loss": 0.1483, + "step": 1066 + }, + { + "epoch": 0.33606299212598423, + "grad_norm": 52.2086181640625, + "learning_rate": 9.99832598425197e-06, + "loss": 0.3647, + "step": 1067 + }, + { + "epoch": 0.3363779527559055, + "grad_norm": 50.21038818359375, + "learning_rate": 9.99832440944882e-06, + "loss": 0.5818, + "step": 1068 + }, + { + "epoch": 0.3366929133858268, + "grad_norm": 25.27126693725586, + "learning_rate": 9.99832283464567e-06, + "loss": 0.1273, + "step": 1069 + }, + { + "epoch": 0.33700787401574805, + "grad_norm": 36.34380340576172, + "learning_rate": 9.99832125984252e-06, + "loss": 0.3509, + "step": 1070 + }, + { + "epoch": 0.3373228346456693, + "grad_norm": 35.567527770996094, + "learning_rate": 9.99831968503937e-06, + "loss": 0.2472, + "step": 1071 + }, + { + "epoch": 0.33763779527559057, + "grad_norm": 39.77647399902344, + "learning_rate": 9.998318110236222e-06, + "loss": 0.5746, + "step": 1072 + }, + { + "epoch": 0.3379527559055118, + "grad_norm": 10.02043628692627, + "learning_rate": 9.998316535433071e-06, + "loss": 0.0749, + "step": 1073 + }, + { + "epoch": 0.3382677165354331, + "grad_norm": 31.038677215576172, + "learning_rate": 9.998314960629922e-06, + "loss": 0.2673, + "step": 1074 + }, + { + "epoch": 0.33858267716535434, + "grad_norm": 34.27322769165039, + "learning_rate": 9.998313385826773e-06, + "loss": 0.1064, + "step": 1075 + }, + { + "epoch": 0.3388976377952756, + "grad_norm": 91.75303649902344, + "learning_rate": 9.998311811023624e-06, + "loss": 0.7232, + "step": 1076 + }, + { + "epoch": 0.33921259842519685, + "grad_norm": 68.7430419921875, + "learning_rate": 9.998310236220473e-06, + "loss": 0.5586, + "step": 1077 + }, + { + "epoch": 0.3395275590551181, + "grad_norm": 94.78008270263672, + "learning_rate": 9.998308661417324e-06, + "loss": 0.5504, + "step": 1078 + }, + { + "epoch": 0.33984251968503937, + "grad_norm": 54.0759162902832, + "learning_rate": 9.998307086614174e-06, + "loss": 0.1873, + "step": 1079 + }, + { + "epoch": 0.3401574803149606, + "grad_norm": 65.1077651977539, + "learning_rate": 9.998305511811025e-06, + "loss": 1.0425, + "step": 1080 + }, + { + "epoch": 0.3401574803149606, + "eval_loss": 0.6948055028915405, + "eval_runtime": 327.3153, + "eval_samples_per_second": 0.357, + "eval_steps_per_second": 0.357, + "step": 1080 + }, + { + "epoch": 0.3404724409448819, + "grad_norm": 79.78382873535156, + "learning_rate": 9.998303937007876e-06, + "loss": 0.6999, + "step": 1081 + }, + { + "epoch": 0.34078740157480314, + "grad_norm": 97.7957534790039, + "learning_rate": 9.998302362204725e-06, + "loss": 0.4083, + "step": 1082 + }, + { + "epoch": 0.3411023622047244, + "grad_norm": 90.36141967773438, + "learning_rate": 9.998300787401576e-06, + "loss": 0.7225, + "step": 1083 + }, + { + "epoch": 0.34141732283464565, + "grad_norm": 14.357733726501465, + "learning_rate": 9.998299212598425e-06, + "loss": 0.0595, + "step": 1084 + }, + { + "epoch": 0.3417322834645669, + "grad_norm": 47.068233489990234, + "learning_rate": 9.998297637795276e-06, + "loss": 0.7842, + "step": 1085 + }, + { + "epoch": 0.34204724409448817, + "grad_norm": 22.276060104370117, + "learning_rate": 9.998296062992127e-06, + "loss": 0.1631, + "step": 1086 + }, + { + "epoch": 0.3423622047244094, + "grad_norm": 38.76866149902344, + "learning_rate": 9.998294488188978e-06, + "loss": 0.2738, + "step": 1087 + }, + { + "epoch": 0.3426771653543307, + "grad_norm": 45.607505798339844, + "learning_rate": 9.998292913385827e-06, + "loss": 0.4995, + "step": 1088 + }, + { + "epoch": 0.342992125984252, + "grad_norm": 34.9421272277832, + "learning_rate": 9.998291338582678e-06, + "loss": 0.614, + "step": 1089 + }, + { + "epoch": 0.34330708661417325, + "grad_norm": 36.95371627807617, + "learning_rate": 9.998289763779528e-06, + "loss": 0.6949, + "step": 1090 + }, + { + "epoch": 0.3436220472440945, + "grad_norm": 32.992279052734375, + "learning_rate": 9.998288188976379e-06, + "loss": 0.2076, + "step": 1091 + }, + { + "epoch": 0.34393700787401577, + "grad_norm": 15.995903015136719, + "learning_rate": 9.99828661417323e-06, + "loss": 0.1538, + "step": 1092 + }, + { + "epoch": 0.344251968503937, + "grad_norm": 38.209495544433594, + "learning_rate": 9.998285039370079e-06, + "loss": 0.5808, + "step": 1093 + }, + { + "epoch": 0.3445669291338583, + "grad_norm": 52.266441345214844, + "learning_rate": 9.99828346456693e-06, + "loss": 0.3555, + "step": 1094 + }, + { + "epoch": 0.34488188976377954, + "grad_norm": 11.709747314453125, + "learning_rate": 9.998281889763781e-06, + "loss": 0.0999, + "step": 1095 + }, + { + "epoch": 0.3451968503937008, + "grad_norm": 37.43159484863281, + "learning_rate": 9.998280314960632e-06, + "loss": 0.5477, + "step": 1096 + }, + { + "epoch": 0.34551181102362205, + "grad_norm": 31.790834426879883, + "learning_rate": 9.998278740157481e-06, + "loss": 0.4586, + "step": 1097 + }, + { + "epoch": 0.3458267716535433, + "grad_norm": 15.95163631439209, + "learning_rate": 9.998277165354332e-06, + "loss": 0.1477, + "step": 1098 + }, + { + "epoch": 0.34614173228346456, + "grad_norm": 57.9958610534668, + "learning_rate": 9.998275590551182e-06, + "loss": 0.2226, + "step": 1099 + }, + { + "epoch": 0.3464566929133858, + "grad_norm": 31.550888061523438, + "learning_rate": 9.998274015748033e-06, + "loss": 0.3224, + "step": 1100 + }, + { + "epoch": 0.3464566929133858, + "eval_loss": 0.5266521573066711, + "eval_runtime": 337.1873, + "eval_samples_per_second": 0.347, + "eval_steps_per_second": 0.347, + "step": 1100 + }, + { + "epoch": 0.3467716535433071, + "grad_norm": 27.716707229614258, + "learning_rate": 9.998272440944883e-06, + "loss": 0.2458, + "step": 1101 + }, + { + "epoch": 0.34708661417322834, + "grad_norm": 36.678863525390625, + "learning_rate": 9.998270866141733e-06, + "loss": 0.315, + "step": 1102 + }, + { + "epoch": 0.3474015748031496, + "grad_norm": 21.499208450317383, + "learning_rate": 9.998269291338584e-06, + "loss": 0.2061, + "step": 1103 + }, + { + "epoch": 0.34771653543307085, + "grad_norm": 63.6679801940918, + "learning_rate": 9.998267716535433e-06, + "loss": 1.0103, + "step": 1104 + }, + { + "epoch": 0.3480314960629921, + "grad_norm": 34.54896545410156, + "learning_rate": 9.998266141732284e-06, + "loss": 0.5376, + "step": 1105 + }, + { + "epoch": 0.34834645669291336, + "grad_norm": 16.613237380981445, + "learning_rate": 9.998264566929135e-06, + "loss": 0.0988, + "step": 1106 + }, + { + "epoch": 0.3486614173228346, + "grad_norm": 73.5743179321289, + "learning_rate": 9.998262992125986e-06, + "loss": 0.8173, + "step": 1107 + }, + { + "epoch": 0.3489763779527559, + "grad_norm": 6.472119331359863, + "learning_rate": 9.998261417322835e-06, + "loss": 0.0331, + "step": 1108 + }, + { + "epoch": 0.3492913385826772, + "grad_norm": 24.923892974853516, + "learning_rate": 9.998259842519686e-06, + "loss": 0.1636, + "step": 1109 + }, + { + "epoch": 0.34960629921259845, + "grad_norm": 76.55589294433594, + "learning_rate": 9.998258267716536e-06, + "loss": 0.5014, + "step": 1110 + }, + { + "epoch": 0.3499212598425197, + "grad_norm": 15.523338317871094, + "learning_rate": 9.998256692913387e-06, + "loss": 0.0701, + "step": 1111 + }, + { + "epoch": 0.35023622047244096, + "grad_norm": 72.51299285888672, + "learning_rate": 9.998255118110238e-06, + "loss": 0.5609, + "step": 1112 + }, + { + "epoch": 0.3505511811023622, + "grad_norm": 41.60987854003906, + "learning_rate": 9.998253543307087e-06, + "loss": 0.2, + "step": 1113 + }, + { + "epoch": 0.3508661417322835, + "grad_norm": 4.41101598739624, + "learning_rate": 9.998251968503938e-06, + "loss": 0.0192, + "step": 1114 + }, + { + "epoch": 0.35118110236220473, + "grad_norm": 69.58025360107422, + "learning_rate": 9.998250393700787e-06, + "loss": 0.8551, + "step": 1115 + }, + { + "epoch": 0.351496062992126, + "grad_norm": 34.081336975097656, + "learning_rate": 9.998248818897638e-06, + "loss": 0.6141, + "step": 1116 + }, + { + "epoch": 0.35181102362204725, + "grad_norm": 46.96076965332031, + "learning_rate": 9.99824724409449e-06, + "loss": 0.8069, + "step": 1117 + }, + { + "epoch": 0.3521259842519685, + "grad_norm": 88.97361755371094, + "learning_rate": 9.99824566929134e-06, + "loss": 0.3247, + "step": 1118 + }, + { + "epoch": 0.35244094488188976, + "grad_norm": 16.12812042236328, + "learning_rate": 9.99824409448819e-06, + "loss": 0.0924, + "step": 1119 + }, + { + "epoch": 0.352755905511811, + "grad_norm": 3.950244188308716, + "learning_rate": 9.99824251968504e-06, + "loss": 0.0298, + "step": 1120 + }, + { + "epoch": 0.352755905511811, + "eval_loss": 0.5020915865898132, + "eval_runtime": 340.8492, + "eval_samples_per_second": 0.343, + "eval_steps_per_second": 0.343, + "step": 1120 + }, + { + "epoch": 0.3530708661417323, + "grad_norm": 86.11933898925781, + "learning_rate": 9.998240944881891e-06, + "loss": 0.9179, + "step": 1121 + }, + { + "epoch": 0.35338582677165353, + "grad_norm": 41.664955139160156, + "learning_rate": 9.99823937007874e-06, + "loss": 0.1796, + "step": 1122 + }, + { + "epoch": 0.3537007874015748, + "grad_norm": 22.196773529052734, + "learning_rate": 9.998237795275592e-06, + "loss": 0.143, + "step": 1123 + }, + { + "epoch": 0.35401574803149605, + "grad_norm": 57.799415588378906, + "learning_rate": 9.998236220472441e-06, + "loss": 0.441, + "step": 1124 + }, + { + "epoch": 0.3543307086614173, + "grad_norm": 23.602643966674805, + "learning_rate": 9.998234645669292e-06, + "loss": 0.325, + "step": 1125 + }, + { + "epoch": 0.35464566929133856, + "grad_norm": 49.98581314086914, + "learning_rate": 9.998233070866143e-06, + "loss": 0.5312, + "step": 1126 + }, + { + "epoch": 0.3549606299212598, + "grad_norm": 32.001861572265625, + "learning_rate": 9.998231496062994e-06, + "loss": 0.2678, + "step": 1127 + }, + { + "epoch": 0.3552755905511811, + "grad_norm": 22.768354415893555, + "learning_rate": 9.998229921259843e-06, + "loss": 0.2655, + "step": 1128 + }, + { + "epoch": 0.3555905511811024, + "grad_norm": 10.575422286987305, + "learning_rate": 9.998228346456694e-06, + "loss": 0.067, + "step": 1129 + }, + { + "epoch": 0.35590551181102364, + "grad_norm": 23.114152908325195, + "learning_rate": 9.998226771653544e-06, + "loss": 0.1502, + "step": 1130 + }, + { + "epoch": 0.3562204724409449, + "grad_norm": 55.854827880859375, + "learning_rate": 9.998225196850395e-06, + "loss": 0.7007, + "step": 1131 + }, + { + "epoch": 0.35653543307086616, + "grad_norm": 43.69165802001953, + "learning_rate": 9.998223622047246e-06, + "loss": 0.2763, + "step": 1132 + }, + { + "epoch": 0.3568503937007874, + "grad_norm": 10.879396438598633, + "learning_rate": 9.998222047244095e-06, + "loss": 0.0361, + "step": 1133 + }, + { + "epoch": 0.3571653543307087, + "grad_norm": 78.35888671875, + "learning_rate": 9.998220472440946e-06, + "loss": 0.089, + "step": 1134 + }, + { + "epoch": 0.35748031496062993, + "grad_norm": 24.78093147277832, + "learning_rate": 9.998218897637795e-06, + "loss": 0.1069, + "step": 1135 + }, + { + "epoch": 0.3577952755905512, + "grad_norm": 46.83030700683594, + "learning_rate": 9.998217322834646e-06, + "loss": 0.472, + "step": 1136 + }, + { + "epoch": 0.35811023622047244, + "grad_norm": 15.252365112304688, + "learning_rate": 9.998215748031497e-06, + "loss": 0.0778, + "step": 1137 + }, + { + "epoch": 0.3584251968503937, + "grad_norm": 15.580936431884766, + "learning_rate": 9.998214173228348e-06, + "loss": 0.1253, + "step": 1138 + }, + { + "epoch": 0.35874015748031496, + "grad_norm": 2.145813226699829, + "learning_rate": 9.998212598425197e-06, + "loss": 0.0093, + "step": 1139 + }, + { + "epoch": 0.3590551181102362, + "grad_norm": 204.69932556152344, + "learning_rate": 9.998211023622047e-06, + "loss": 1.3589, + "step": 1140 + }, + { + "epoch": 0.3590551181102362, + "eval_loss": 0.5789304971694946, + "eval_runtime": 339.4483, + "eval_samples_per_second": 0.345, + "eval_steps_per_second": 0.345, + "step": 1140 + }, + { + "epoch": 0.3593700787401575, + "grad_norm": 134.10121154785156, + "learning_rate": 9.998209448818898e-06, + "loss": 0.3836, + "step": 1141 + }, + { + "epoch": 0.35968503937007873, + "grad_norm": 49.37288284301758, + "learning_rate": 9.998207874015749e-06, + "loss": 0.5611, + "step": 1142 + }, + { + "epoch": 0.36, + "grad_norm": 42.33864212036133, + "learning_rate": 9.9982062992126e-06, + "loss": 0.1906, + "step": 1143 + }, + { + "epoch": 0.36031496062992124, + "grad_norm": 75.55062866210938, + "learning_rate": 9.998204724409449e-06, + "loss": 0.1209, + "step": 1144 + }, + { + "epoch": 0.3606299212598425, + "grad_norm": 30.669452667236328, + "learning_rate": 9.9982031496063e-06, + "loss": 0.097, + "step": 1145 + }, + { + "epoch": 0.36094488188976376, + "grad_norm": 96.1650390625, + "learning_rate": 9.998201574803151e-06, + "loss": 0.9096, + "step": 1146 + }, + { + "epoch": 0.361259842519685, + "grad_norm": 108.84801483154297, + "learning_rate": 9.998200000000002e-06, + "loss": 0.4399, + "step": 1147 + }, + { + "epoch": 0.36157480314960627, + "grad_norm": 101.94995880126953, + "learning_rate": 9.998198425196851e-06, + "loss": 0.5331, + "step": 1148 + }, + { + "epoch": 0.3618897637795276, + "grad_norm": 42.10065841674805, + "learning_rate": 9.998196850393702e-06, + "loss": 0.2717, + "step": 1149 + }, + { + "epoch": 0.36220472440944884, + "grad_norm": 34.27854919433594, + "learning_rate": 9.998195275590552e-06, + "loss": 0.0578, + "step": 1150 + }, + { + "epoch": 0.3625196850393701, + "grad_norm": 80.90559387207031, + "learning_rate": 9.998193700787403e-06, + "loss": 0.2224, + "step": 1151 + }, + { + "epoch": 0.36283464566929136, + "grad_norm": 89.94649505615234, + "learning_rate": 9.998192125984254e-06, + "loss": 0.2526, + "step": 1152 + }, + { + "epoch": 0.3631496062992126, + "grad_norm": 96.36505126953125, + "learning_rate": 9.998190551181103e-06, + "loss": 0.521, + "step": 1153 + }, + { + "epoch": 0.36346456692913387, + "grad_norm": 177.08819580078125, + "learning_rate": 9.998188976377954e-06, + "loss": 0.6533, + "step": 1154 + }, + { + "epoch": 0.3637795275590551, + "grad_norm": 60.227481842041016, + "learning_rate": 9.998187401574803e-06, + "loss": 0.8652, + "step": 1155 + }, + { + "epoch": 0.3640944881889764, + "grad_norm": 69.29286193847656, + "learning_rate": 9.998185826771654e-06, + "loss": 1.2497, + "step": 1156 + }, + { + "epoch": 0.36440944881889764, + "grad_norm": 59.95965576171875, + "learning_rate": 9.998184251968505e-06, + "loss": 0.4519, + "step": 1157 + }, + { + "epoch": 0.3647244094488189, + "grad_norm": 61.35934066772461, + "learning_rate": 9.998182677165356e-06, + "loss": 0.2029, + "step": 1158 + }, + { + "epoch": 0.36503937007874016, + "grad_norm": 32.42390060424805, + "learning_rate": 9.998181102362205e-06, + "loss": 0.5222, + "step": 1159 + }, + { + "epoch": 0.3653543307086614, + "grad_norm": 58.62075424194336, + "learning_rate": 9.998179527559055e-06, + "loss": 0.6079, + "step": 1160 + }, + { + "epoch": 0.3653543307086614, + "eval_loss": 0.5354205965995789, + "eval_runtime": 339.1978, + "eval_samples_per_second": 0.345, + "eval_steps_per_second": 0.345, + "step": 1160 + }, + { + "epoch": 0.36566929133858267, + "grad_norm": 61.64841079711914, + "learning_rate": 9.998177952755906e-06, + "loss": 0.603, + "step": 1161 + }, + { + "epoch": 0.3659842519685039, + "grad_norm": 34.8085823059082, + "learning_rate": 9.998176377952757e-06, + "loss": 0.3863, + "step": 1162 + }, + { + "epoch": 0.3662992125984252, + "grad_norm": 25.715442657470703, + "learning_rate": 9.998174803149608e-06, + "loss": 0.2857, + "step": 1163 + }, + { + "epoch": 0.36661417322834644, + "grad_norm": 33.884483337402344, + "learning_rate": 9.998173228346457e-06, + "loss": 0.549, + "step": 1164 + }, + { + "epoch": 0.3669291338582677, + "grad_norm": 24.125484466552734, + "learning_rate": 9.998171653543308e-06, + "loss": 0.2794, + "step": 1165 + }, + { + "epoch": 0.36724409448818895, + "grad_norm": 67.11617279052734, + "learning_rate": 9.998170078740159e-06, + "loss": 0.6092, + "step": 1166 + }, + { + "epoch": 0.3675590551181102, + "grad_norm": 23.704805374145508, + "learning_rate": 9.99816850393701e-06, + "loss": 0.4884, + "step": 1167 + }, + { + "epoch": 0.36787401574803147, + "grad_norm": 25.822975158691406, + "learning_rate": 9.99816692913386e-06, + "loss": 0.5046, + "step": 1168 + }, + { + "epoch": 0.3681889763779528, + "grad_norm": 31.311058044433594, + "learning_rate": 9.99816535433071e-06, + "loss": 0.3914, + "step": 1169 + }, + { + "epoch": 0.36850393700787404, + "grad_norm": 18.38756561279297, + "learning_rate": 9.99816377952756e-06, + "loss": 0.4273, + "step": 1170 + }, + { + "epoch": 0.3688188976377953, + "grad_norm": 20.738372802734375, + "learning_rate": 9.99816220472441e-06, + "loss": 0.2612, + "step": 1171 + }, + { + "epoch": 0.36913385826771655, + "grad_norm": 23.450641632080078, + "learning_rate": 9.998160629921261e-06, + "loss": 0.3474, + "step": 1172 + }, + { + "epoch": 0.3694488188976378, + "grad_norm": 38.92578125, + "learning_rate": 9.99815905511811e-06, + "loss": 0.5643, + "step": 1173 + }, + { + "epoch": 0.36976377952755907, + "grad_norm": 59.437496185302734, + "learning_rate": 9.998157480314962e-06, + "loss": 0.7971, + "step": 1174 + }, + { + "epoch": 0.3700787401574803, + "grad_norm": 22.928701400756836, + "learning_rate": 9.998155905511811e-06, + "loss": 0.2843, + "step": 1175 + }, + { + "epoch": 0.3703937007874016, + "grad_norm": 38.7431526184082, + "learning_rate": 9.998154330708662e-06, + "loss": 0.3979, + "step": 1176 + }, + { + "epoch": 0.37070866141732284, + "grad_norm": 5.290953636169434, + "learning_rate": 9.998152755905513e-06, + "loss": 0.0298, + "step": 1177 + }, + { + "epoch": 0.3710236220472441, + "grad_norm": 23.754051208496094, + "learning_rate": 9.998151181102364e-06, + "loss": 0.4325, + "step": 1178 + }, + { + "epoch": 0.37133858267716535, + "grad_norm": 27.173952102661133, + "learning_rate": 9.998149606299213e-06, + "loss": 0.5674, + "step": 1179 + }, + { + "epoch": 0.3716535433070866, + "grad_norm": 42.20391845703125, + "learning_rate": 9.998148031496063e-06, + "loss": 0.5973, + "step": 1180 + }, + { + "epoch": 0.3716535433070866, + "eval_loss": 0.5375993251800537, + "eval_runtime": 339.4527, + "eval_samples_per_second": 0.345, + "eval_steps_per_second": 0.345, + "step": 1180 + }, + { + "epoch": 0.37196850393700787, + "grad_norm": 11.690238952636719, + "learning_rate": 9.998146456692914e-06, + "loss": 0.0917, + "step": 1181 + }, + { + "epoch": 0.3722834645669291, + "grad_norm": 10.012290000915527, + "learning_rate": 9.998144881889765e-06, + "loss": 0.0842, + "step": 1182 + }, + { + "epoch": 0.3725984251968504, + "grad_norm": 21.645814895629883, + "learning_rate": 9.998143307086616e-06, + "loss": 0.152, + "step": 1183 + }, + { + "epoch": 0.37291338582677164, + "grad_norm": 31.5441837310791, + "learning_rate": 9.998141732283465e-06, + "loss": 0.5621, + "step": 1184 + }, + { + "epoch": 0.3732283464566929, + "grad_norm": 41.74612808227539, + "learning_rate": 9.998140157480316e-06, + "loss": 0.7019, + "step": 1185 + }, + { + "epoch": 0.37354330708661415, + "grad_norm": 35.047794342041016, + "learning_rate": 9.998138582677165e-06, + "loss": 0.1264, + "step": 1186 + }, + { + "epoch": 0.3738582677165354, + "grad_norm": 77.93087005615234, + "learning_rate": 9.998137007874016e-06, + "loss": 0.8623, + "step": 1187 + }, + { + "epoch": 0.37417322834645667, + "grad_norm": 32.53571701049805, + "learning_rate": 9.998135433070867e-06, + "loss": 0.2572, + "step": 1188 + }, + { + "epoch": 0.374488188976378, + "grad_norm": 57.9747428894043, + "learning_rate": 9.998133858267718e-06, + "loss": 0.5365, + "step": 1189 + }, + { + "epoch": 0.37480314960629924, + "grad_norm": 46.87409210205078, + "learning_rate": 9.998132283464567e-06, + "loss": 1.2091, + "step": 1190 + }, + { + "epoch": 0.3751181102362205, + "grad_norm": 29.17926788330078, + "learning_rate": 9.998130708661418e-06, + "loss": 0.3407, + "step": 1191 + }, + { + "epoch": 0.37543307086614175, + "grad_norm": 43.51498794555664, + "learning_rate": 9.99812913385827e-06, + "loss": 0.747, + "step": 1192 + }, + { + "epoch": 0.375748031496063, + "grad_norm": 63.52394104003906, + "learning_rate": 9.998127559055119e-06, + "loss": 0.5045, + "step": 1193 + }, + { + "epoch": 0.37606299212598426, + "grad_norm": 82.31373596191406, + "learning_rate": 9.99812598425197e-06, + "loss": 0.9014, + "step": 1194 + }, + { + "epoch": 0.3763779527559055, + "grad_norm": 70.40677642822266, + "learning_rate": 9.998124409448819e-06, + "loss": 0.5663, + "step": 1195 + }, + { + "epoch": 0.3766929133858268, + "grad_norm": 16.95841407775879, + "learning_rate": 9.99812283464567e-06, + "loss": 0.3155, + "step": 1196 + }, + { + "epoch": 0.37700787401574803, + "grad_norm": 46.58156967163086, + "learning_rate": 9.998121259842521e-06, + "loss": 0.6919, + "step": 1197 + }, + { + "epoch": 0.3773228346456693, + "grad_norm": 37.353492736816406, + "learning_rate": 9.998119685039372e-06, + "loss": 0.669, + "step": 1198 + }, + { + "epoch": 0.37763779527559055, + "grad_norm": 23.72784996032715, + "learning_rate": 9.998118110236221e-06, + "loss": 0.3445, + "step": 1199 + }, + { + "epoch": 0.3779527559055118, + "grad_norm": 39.98880386352539, + "learning_rate": 9.99811653543307e-06, + "loss": 0.4265, + "step": 1200 + }, + { + "epoch": 0.3779527559055118, + "eval_loss": 0.5579959154129028, + "eval_runtime": 337.5471, + "eval_samples_per_second": 0.347, + "eval_steps_per_second": 0.347, + "step": 1200 + }, + { + "epoch": 0.37826771653543306, + "grad_norm": 13.343225479125977, + "learning_rate": 9.998114960629922e-06, + "loss": 0.346, + "step": 1201 + }, + { + "epoch": 0.3785826771653543, + "grad_norm": 31.93812370300293, + "learning_rate": 9.998113385826773e-06, + "loss": 0.7538, + "step": 1202 + }, + { + "epoch": 0.3788976377952756, + "grad_norm": 22.874921798706055, + "learning_rate": 9.998111811023624e-06, + "loss": 0.2865, + "step": 1203 + }, + { + "epoch": 0.37921259842519683, + "grad_norm": 50.595577239990234, + "learning_rate": 9.998110236220473e-06, + "loss": 0.7486, + "step": 1204 + }, + { + "epoch": 0.3795275590551181, + "grad_norm": 20.3195858001709, + "learning_rate": 9.998108661417324e-06, + "loss": 0.2986, + "step": 1205 + }, + { + "epoch": 0.37984251968503935, + "grad_norm": 18.59178352355957, + "learning_rate": 9.998107086614173e-06, + "loss": 0.3259, + "step": 1206 + }, + { + "epoch": 0.3801574803149606, + "grad_norm": 23.734966278076172, + "learning_rate": 9.998105511811024e-06, + "loss": 0.4835, + "step": 1207 + }, + { + "epoch": 0.38047244094488186, + "grad_norm": 47.375789642333984, + "learning_rate": 9.998103937007875e-06, + "loss": 0.4708, + "step": 1208 + }, + { + "epoch": 0.3807874015748032, + "grad_norm": 12.326082229614258, + "learning_rate": 9.998102362204726e-06, + "loss": 0.3698, + "step": 1209 + }, + { + "epoch": 0.38110236220472443, + "grad_norm": 30.175519943237305, + "learning_rate": 9.998100787401575e-06, + "loss": 0.4141, + "step": 1210 + }, + { + "epoch": 0.3814173228346457, + "grad_norm": 27.1715087890625, + "learning_rate": 9.998099212598425e-06, + "loss": 0.4674, + "step": 1211 + }, + { + "epoch": 0.38173228346456695, + "grad_norm": 31.19744300842285, + "learning_rate": 9.998097637795277e-06, + "loss": 0.3239, + "step": 1212 + }, + { + "epoch": 0.3820472440944882, + "grad_norm": 44.07891845703125, + "learning_rate": 9.998096062992127e-06, + "loss": 0.7672, + "step": 1213 + }, + { + "epoch": 0.38236220472440946, + "grad_norm": 15.165576934814453, + "learning_rate": 9.998094488188978e-06, + "loss": 0.2029, + "step": 1214 + }, + { + "epoch": 0.3826771653543307, + "grad_norm": 41.70854187011719, + "learning_rate": 9.998092913385827e-06, + "loss": 0.3442, + "step": 1215 + }, + { + "epoch": 0.382992125984252, + "grad_norm": 25.64063835144043, + "learning_rate": 9.998091338582678e-06, + "loss": 0.3051, + "step": 1216 + }, + { + "epoch": 0.38330708661417323, + "grad_norm": 19.23823356628418, + "learning_rate": 9.998089763779529e-06, + "loss": 0.1983, + "step": 1217 + }, + { + "epoch": 0.3836220472440945, + "grad_norm": 65.09869384765625, + "learning_rate": 9.99808818897638e-06, + "loss": 0.4903, + "step": 1218 + }, + { + "epoch": 0.38393700787401575, + "grad_norm": 14.515801429748535, + "learning_rate": 9.99808661417323e-06, + "loss": 0.1511, + "step": 1219 + }, + { + "epoch": 0.384251968503937, + "grad_norm": 34.14856719970703, + "learning_rate": 9.998085039370079e-06, + "loss": 0.4351, + "step": 1220 + }, + { + "epoch": 0.384251968503937, + "eval_loss": 0.49825048446655273, + "eval_runtime": 340.3122, + "eval_samples_per_second": 0.344, + "eval_steps_per_second": 0.344, + "step": 1220 + }, + { + "epoch": 0.38456692913385826, + "grad_norm": 42.81884765625, + "learning_rate": 9.99808346456693e-06, + "loss": 0.4317, + "step": 1221 + }, + { + "epoch": 0.3848818897637795, + "grad_norm": 40.3559455871582, + "learning_rate": 9.99808188976378e-06, + "loss": 0.5042, + "step": 1222 + }, + { + "epoch": 0.3851968503937008, + "grad_norm": 28.505815505981445, + "learning_rate": 9.998080314960631e-06, + "loss": 0.316, + "step": 1223 + }, + { + "epoch": 0.38551181102362203, + "grad_norm": 20.508024215698242, + "learning_rate": 9.99807874015748e-06, + "loss": 0.15, + "step": 1224 + }, + { + "epoch": 0.3858267716535433, + "grad_norm": 73.8309555053711, + "learning_rate": 9.998077165354332e-06, + "loss": 0.4734, + "step": 1225 + }, + { + "epoch": 0.38614173228346454, + "grad_norm": 24.401338577270508, + "learning_rate": 9.998075590551181e-06, + "loss": 0.2553, + "step": 1226 + }, + { + "epoch": 0.3864566929133858, + "grad_norm": 28.52256202697754, + "learning_rate": 9.998074015748032e-06, + "loss": 0.0865, + "step": 1227 + }, + { + "epoch": 0.38677165354330706, + "grad_norm": 45.155696868896484, + "learning_rate": 9.998072440944883e-06, + "loss": 0.413, + "step": 1228 + }, + { + "epoch": 0.38708661417322837, + "grad_norm": 23.677753448486328, + "learning_rate": 9.998070866141734e-06, + "loss": 0.1919, + "step": 1229 + }, + { + "epoch": 0.38740157480314963, + "grad_norm": 27.503589630126953, + "learning_rate": 9.998069291338583e-06, + "loss": 0.6631, + "step": 1230 + }, + { + "epoch": 0.3877165354330709, + "grad_norm": 33.99694061279297, + "learning_rate": 9.998067716535433e-06, + "loss": 0.5537, + "step": 1231 + }, + { + "epoch": 0.38803149606299214, + "grad_norm": 73.67473602294922, + "learning_rate": 9.998066141732284e-06, + "loss": 0.2761, + "step": 1232 + }, + { + "epoch": 0.3883464566929134, + "grad_norm": 47.667327880859375, + "learning_rate": 9.998064566929135e-06, + "loss": 0.1336, + "step": 1233 + }, + { + "epoch": 0.38866141732283466, + "grad_norm": 32.623802185058594, + "learning_rate": 9.998062992125986e-06, + "loss": 0.1881, + "step": 1234 + }, + { + "epoch": 0.3889763779527559, + "grad_norm": 92.79457092285156, + "learning_rate": 9.998061417322835e-06, + "loss": 0.9026, + "step": 1235 + }, + { + "epoch": 0.38929133858267717, + "grad_norm": 47.63346862792969, + "learning_rate": 9.998059842519686e-06, + "loss": 0.2402, + "step": 1236 + }, + { + "epoch": 0.38960629921259843, + "grad_norm": 113.52471923828125, + "learning_rate": 9.998058267716537e-06, + "loss": 0.7758, + "step": 1237 + }, + { + "epoch": 0.3899212598425197, + "grad_norm": 44.212303161621094, + "learning_rate": 9.998056692913388e-06, + "loss": 0.1154, + "step": 1238 + }, + { + "epoch": 0.39023622047244094, + "grad_norm": 113.48955535888672, + "learning_rate": 9.998055118110237e-06, + "loss": 0.7735, + "step": 1239 + }, + { + "epoch": 0.3905511811023622, + "grad_norm": 95.3028564453125, + "learning_rate": 9.998053543307086e-06, + "loss": 0.5174, + "step": 1240 + }, + { + "epoch": 0.3905511811023622, + "eval_loss": 0.680210292339325, + "eval_runtime": 337.0826, + "eval_samples_per_second": 0.347, + "eval_steps_per_second": 0.347, + "step": 1240 + }, + { + "epoch": 0.39086614173228346, + "grad_norm": 75.17251586914062, + "learning_rate": 9.998051968503937e-06, + "loss": 0.6919, + "step": 1241 + }, + { + "epoch": 0.3911811023622047, + "grad_norm": 94.41243743896484, + "learning_rate": 9.998050393700788e-06, + "loss": 1.1939, + "step": 1242 + }, + { + "epoch": 0.39149606299212597, + "grad_norm": 17.574474334716797, + "learning_rate": 9.99804881889764e-06, + "loss": 0.0841, + "step": 1243 + }, + { + "epoch": 0.3918110236220472, + "grad_norm": 22.532442092895508, + "learning_rate": 9.998047244094489e-06, + "loss": 0.0756, + "step": 1244 + }, + { + "epoch": 0.3921259842519685, + "grad_norm": 48.4405632019043, + "learning_rate": 9.99804566929134e-06, + "loss": 0.225, + "step": 1245 + }, + { + "epoch": 0.39244094488188974, + "grad_norm": 76.005859375, + "learning_rate": 9.998044094488189e-06, + "loss": 0.1798, + "step": 1246 + }, + { + "epoch": 0.392755905511811, + "grad_norm": 90.26568603515625, + "learning_rate": 9.99804251968504e-06, + "loss": 1.1972, + "step": 1247 + }, + { + "epoch": 0.39307086614173226, + "grad_norm": 69.64557647705078, + "learning_rate": 9.998040944881891e-06, + "loss": 0.1896, + "step": 1248 + }, + { + "epoch": 0.39338582677165357, + "grad_norm": 68.00067901611328, + "learning_rate": 9.998039370078742e-06, + "loss": 0.7109, + "step": 1249 + }, + { + "epoch": 0.3937007874015748, + "grad_norm": 57.22831344604492, + "learning_rate": 9.998037795275591e-06, + "loss": 0.8236, + "step": 1250 + }, + { + "epoch": 0.3940157480314961, + "grad_norm": 47.59756851196289, + "learning_rate": 9.99803622047244e-06, + "loss": 0.2775, + "step": 1251 + }, + { + "epoch": 0.39433070866141734, + "grad_norm": 23.361492156982422, + "learning_rate": 9.998034645669292e-06, + "loss": 0.0931, + "step": 1252 + }, + { + "epoch": 0.3946456692913386, + "grad_norm": 40.62185287475586, + "learning_rate": 9.998033070866143e-06, + "loss": 0.3612, + "step": 1253 + }, + { + "epoch": 0.39496062992125985, + "grad_norm": 31.514081954956055, + "learning_rate": 9.998031496062994e-06, + "loss": 0.0952, + "step": 1254 + }, + { + "epoch": 0.3952755905511811, + "grad_norm": 39.17756652832031, + "learning_rate": 9.998029921259843e-06, + "loss": 0.2162, + "step": 1255 + }, + { + "epoch": 0.39559055118110237, + "grad_norm": 19.34990882873535, + "learning_rate": 9.998028346456694e-06, + "loss": 0.1018, + "step": 1256 + }, + { + "epoch": 0.3959055118110236, + "grad_norm": 40.027671813964844, + "learning_rate": 9.998026771653543e-06, + "loss": 0.217, + "step": 1257 + }, + { + "epoch": 0.3962204724409449, + "grad_norm": 97.05489349365234, + "learning_rate": 9.998025196850394e-06, + "loss": 0.3828, + "step": 1258 + }, + { + "epoch": 0.39653543307086614, + "grad_norm": 20.452852249145508, + "learning_rate": 9.998023622047245e-06, + "loss": 0.1665, + "step": 1259 + }, + { + "epoch": 0.3968503937007874, + "grad_norm": 76.07334899902344, + "learning_rate": 9.998022047244094e-06, + "loss": 0.9711, + "step": 1260 + }, + { + "epoch": 0.3968503937007874, + "eval_loss": 0.508359432220459, + "eval_runtime": 337.8333, + "eval_samples_per_second": 0.346, + "eval_steps_per_second": 0.346, + "step": 1260 + }, + { + "epoch": 0.39716535433070865, + "grad_norm": 44.56214904785156, + "learning_rate": 9.998020472440945e-06, + "loss": 0.1862, + "step": 1261 + }, + { + "epoch": 0.3974803149606299, + "grad_norm": 105.54949188232422, + "learning_rate": 9.998018897637796e-06, + "loss": 1.0023, + "step": 1262 + }, + { + "epoch": 0.39779527559055117, + "grad_norm": 64.92313385009766, + "learning_rate": 9.998017322834647e-06, + "loss": 0.9656, + "step": 1263 + }, + { + "epoch": 0.3981102362204724, + "grad_norm": 60.504730224609375, + "learning_rate": 9.998015748031497e-06, + "loss": 0.3303, + "step": 1264 + }, + { + "epoch": 0.3984251968503937, + "grad_norm": 57.36290740966797, + "learning_rate": 9.998014173228348e-06, + "loss": 0.2705, + "step": 1265 + }, + { + "epoch": 0.39874015748031494, + "grad_norm": 49.48656463623047, + "learning_rate": 9.998012598425197e-06, + "loss": 0.5267, + "step": 1266 + }, + { + "epoch": 0.3990551181102362, + "grad_norm": 73.73528289794922, + "learning_rate": 9.998011023622048e-06, + "loss": 0.2752, + "step": 1267 + }, + { + "epoch": 0.39937007874015745, + "grad_norm": 77.58939361572266, + "learning_rate": 9.998009448818899e-06, + "loss": 0.782, + "step": 1268 + }, + { + "epoch": 0.39968503937007877, + "grad_norm": 13.655567169189453, + "learning_rate": 9.99800787401575e-06, + "loss": 0.0997, + "step": 1269 + }, + { + "epoch": 0.4, + "grad_norm": 31.09360122680664, + "learning_rate": 9.9980062992126e-06, + "loss": 0.275, + "step": 1270 + }, + { + "epoch": 0.4003149606299213, + "grad_norm": 57.30973815917969, + "learning_rate": 9.998004724409449e-06, + "loss": 0.5763, + "step": 1271 + }, + { + "epoch": 0.40062992125984254, + "grad_norm": 55.74612808227539, + "learning_rate": 9.9980031496063e-06, + "loss": 0.8855, + "step": 1272 + }, + { + "epoch": 0.4009448818897638, + "grad_norm": 45.396583557128906, + "learning_rate": 9.99800157480315e-06, + "loss": 0.4024, + "step": 1273 + }, + { + "epoch": 0.40125984251968505, + "grad_norm": 17.555898666381836, + "learning_rate": 9.998000000000002e-06, + "loss": 0.128, + "step": 1274 + }, + { + "epoch": 0.4015748031496063, + "grad_norm": 54.747310638427734, + "learning_rate": 9.99799842519685e-06, + "loss": 0.8548, + "step": 1275 + }, + { + "epoch": 0.40188976377952756, + "grad_norm": 36.224857330322266, + "learning_rate": 9.997996850393702e-06, + "loss": 0.2832, + "step": 1276 + }, + { + "epoch": 0.4022047244094488, + "grad_norm": 20.606124877929688, + "learning_rate": 9.997995275590551e-06, + "loss": 0.2242, + "step": 1277 + }, + { + "epoch": 0.4025196850393701, + "grad_norm": 37.51258087158203, + "learning_rate": 9.997993700787402e-06, + "loss": 0.3867, + "step": 1278 + }, + { + "epoch": 0.40283464566929134, + "grad_norm": 14.421310424804688, + "learning_rate": 9.997992125984253e-06, + "loss": 0.1174, + "step": 1279 + }, + { + "epoch": 0.4031496062992126, + "grad_norm": 41.61854553222656, + "learning_rate": 9.997990551181104e-06, + "loss": 0.3572, + "step": 1280 + }, + { + "epoch": 0.4031496062992126, + "eval_loss": 0.48692360520362854, + "eval_runtime": 338.2236, + "eval_samples_per_second": 0.346, + "eval_steps_per_second": 0.346, + "step": 1280 + }, + { + "epoch": 0.40346456692913385, + "grad_norm": 23.846805572509766, + "learning_rate": 9.997988976377953e-06, + "loss": 0.419, + "step": 1281 + }, + { + "epoch": 0.4037795275590551, + "grad_norm": 48.06615447998047, + "learning_rate": 9.997987401574804e-06, + "loss": 0.8735, + "step": 1282 + }, + { + "epoch": 0.40409448818897636, + "grad_norm": 15.782271385192871, + "learning_rate": 9.997985826771655e-06, + "loss": 0.0723, + "step": 1283 + }, + { + "epoch": 0.4044094488188976, + "grad_norm": 19.87238883972168, + "learning_rate": 9.997984251968505e-06, + "loss": 0.1359, + "step": 1284 + }, + { + "epoch": 0.4047244094488189, + "grad_norm": 38.17216110229492, + "learning_rate": 9.997982677165356e-06, + "loss": 0.3926, + "step": 1285 + }, + { + "epoch": 0.40503937007874014, + "grad_norm": 45.54020309448242, + "learning_rate": 9.997981102362205e-06, + "loss": 0.3831, + "step": 1286 + }, + { + "epoch": 0.4053543307086614, + "grad_norm": 26.464305877685547, + "learning_rate": 9.997979527559056e-06, + "loss": 0.4216, + "step": 1287 + }, + { + "epoch": 0.4056692913385827, + "grad_norm": 64.78831481933594, + "learning_rate": 9.997977952755907e-06, + "loss": 0.861, + "step": 1288 + }, + { + "epoch": 0.40598425196850396, + "grad_norm": 38.6380615234375, + "learning_rate": 9.997976377952758e-06, + "loss": 0.3388, + "step": 1289 + }, + { + "epoch": 0.4062992125984252, + "grad_norm": 44.91901779174805, + "learning_rate": 9.997974803149607e-06, + "loss": 0.2208, + "step": 1290 + }, + { + "epoch": 0.4066141732283465, + "grad_norm": 59.079185485839844, + "learning_rate": 9.997973228346456e-06, + "loss": 0.6019, + "step": 1291 + }, + { + "epoch": 0.40692913385826773, + "grad_norm": 65.11566925048828, + "learning_rate": 9.997971653543307e-06, + "loss": 0.64, + "step": 1292 + }, + { + "epoch": 0.407244094488189, + "grad_norm": 45.68169021606445, + "learning_rate": 9.997970078740158e-06, + "loss": 0.6369, + "step": 1293 + }, + { + "epoch": 0.40755905511811025, + "grad_norm": 6.096194744110107, + "learning_rate": 9.99796850393701e-06, + "loss": 0.0484, + "step": 1294 + }, + { + "epoch": 0.4078740157480315, + "grad_norm": 31.807212829589844, + "learning_rate": 9.997966929133859e-06, + "loss": 0.3238, + "step": 1295 + }, + { + "epoch": 0.40818897637795276, + "grad_norm": 66.85626983642578, + "learning_rate": 9.99796535433071e-06, + "loss": 0.6099, + "step": 1296 + }, + { + "epoch": 0.408503937007874, + "grad_norm": 15.900289535522461, + "learning_rate": 9.997963779527559e-06, + "loss": 0.3017, + "step": 1297 + }, + { + "epoch": 0.4088188976377953, + "grad_norm": 58.991703033447266, + "learning_rate": 9.99796220472441e-06, + "loss": 0.441, + "step": 1298 + }, + { + "epoch": 0.40913385826771653, + "grad_norm": 64.84235382080078, + "learning_rate": 9.997960629921261e-06, + "loss": 0.5636, + "step": 1299 + }, + { + "epoch": 0.4094488188976378, + "grad_norm": 48.843505859375, + "learning_rate": 9.997959055118112e-06, + "loss": 0.5989, + "step": 1300 + }, + { + "epoch": 0.4094488188976378, + "eval_loss": 0.5216355323791504, + "eval_runtime": 339.9169, + "eval_samples_per_second": 0.344, + "eval_steps_per_second": 0.344, + "step": 1300 + }, + { + "epoch": 0.40976377952755905, + "grad_norm": 22.199951171875, + "learning_rate": 9.997957480314961e-06, + "loss": 0.331, + "step": 1301 + }, + { + "epoch": 0.4100787401574803, + "grad_norm": 56.680816650390625, + "learning_rate": 9.99795590551181e-06, + "loss": 0.4215, + "step": 1302 + }, + { + "epoch": 0.41039370078740156, + "grad_norm": 33.23557662963867, + "learning_rate": 9.997954330708662e-06, + "loss": 0.444, + "step": 1303 + }, + { + "epoch": 0.4107086614173228, + "grad_norm": 31.659833908081055, + "learning_rate": 9.997952755905513e-06, + "loss": 0.4401, + "step": 1304 + }, + { + "epoch": 0.4110236220472441, + "grad_norm": 14.66598129272461, + "learning_rate": 9.997951181102364e-06, + "loss": 0.1447, + "step": 1305 + }, + { + "epoch": 0.41133858267716533, + "grad_norm": 20.910083770751953, + "learning_rate": 9.997949606299213e-06, + "loss": 0.2887, + "step": 1306 + }, + { + "epoch": 0.4116535433070866, + "grad_norm": 39.282596588134766, + "learning_rate": 9.997948031496064e-06, + "loss": 0.5081, + "step": 1307 + }, + { + "epoch": 0.4119685039370079, + "grad_norm": 37.22988510131836, + "learning_rate": 9.997946456692915e-06, + "loss": 0.6354, + "step": 1308 + }, + { + "epoch": 0.41228346456692916, + "grad_norm": 51.220314025878906, + "learning_rate": 9.997944881889766e-06, + "loss": 0.4205, + "step": 1309 + }, + { + "epoch": 0.4125984251968504, + "grad_norm": 31.497386932373047, + "learning_rate": 9.997943307086615e-06, + "loss": 0.4645, + "step": 1310 + }, + { + "epoch": 0.4129133858267717, + "grad_norm": 48.64023971557617, + "learning_rate": 9.997941732283464e-06, + "loss": 0.5369, + "step": 1311 + }, + { + "epoch": 0.41322834645669293, + "grad_norm": 44.41730499267578, + "learning_rate": 9.997940157480315e-06, + "loss": 0.4616, + "step": 1312 + }, + { + "epoch": 0.4135433070866142, + "grad_norm": 32.24418640136719, + "learning_rate": 9.997938582677166e-06, + "loss": 0.4085, + "step": 1313 + }, + { + "epoch": 0.41385826771653544, + "grad_norm": 25.68863296508789, + "learning_rate": 9.997937007874017e-06, + "loss": 0.1576, + "step": 1314 + }, + { + "epoch": 0.4141732283464567, + "grad_norm": 57.67222595214844, + "learning_rate": 9.997935433070867e-06, + "loss": 0.6927, + "step": 1315 + }, + { + "epoch": 0.41448818897637796, + "grad_norm": 46.1195182800293, + "learning_rate": 9.997933858267718e-06, + "loss": 0.3948, + "step": 1316 + }, + { + "epoch": 0.4148031496062992, + "grad_norm": 16.534828186035156, + "learning_rate": 9.997932283464567e-06, + "loss": 0.1382, + "step": 1317 + }, + { + "epoch": 0.41511811023622047, + "grad_norm": 40.07267379760742, + "learning_rate": 9.997930708661418e-06, + "loss": 0.3056, + "step": 1318 + }, + { + "epoch": 0.41543307086614173, + "grad_norm": 29.960514068603516, + "learning_rate": 9.997929133858269e-06, + "loss": 0.1982, + "step": 1319 + }, + { + "epoch": 0.415748031496063, + "grad_norm": 26.702951431274414, + "learning_rate": 9.99792755905512e-06, + "loss": 0.4233, + "step": 1320 + }, + { + "epoch": 0.415748031496063, + "eval_loss": 0.4609772861003876, + "eval_runtime": 338.1687, + "eval_samples_per_second": 0.346, + "eval_steps_per_second": 0.346, + "step": 1320 + }, + { + "epoch": 0.41606299212598424, + "grad_norm": 31.864824295043945, + "learning_rate": 9.99792598425197e-06, + "loss": 0.5666, + "step": 1321 + }, + { + "epoch": 0.4163779527559055, + "grad_norm": 32.256771087646484, + "learning_rate": 9.997924409448819e-06, + "loss": 0.2308, + "step": 1322 + }, + { + "epoch": 0.41669291338582676, + "grad_norm": 20.752614974975586, + "learning_rate": 9.99792283464567e-06, + "loss": 0.3462, + "step": 1323 + }, + { + "epoch": 0.417007874015748, + "grad_norm": 86.646484375, + "learning_rate": 9.99792125984252e-06, + "loss": 0.7229, + "step": 1324 + }, + { + "epoch": 0.41732283464566927, + "grad_norm": 36.791202545166016, + "learning_rate": 9.997919685039372e-06, + "loss": 0.6036, + "step": 1325 + }, + { + "epoch": 0.41763779527559053, + "grad_norm": 60.902095794677734, + "learning_rate": 9.99791811023622e-06, + "loss": 0.7423, + "step": 1326 + }, + { + "epoch": 0.4179527559055118, + "grad_norm": 15.44100284576416, + "learning_rate": 9.997916535433072e-06, + "loss": 0.0577, + "step": 1327 + }, + { + "epoch": 0.4182677165354331, + "grad_norm": 39.82502365112305, + "learning_rate": 9.997914960629921e-06, + "loss": 0.351, + "step": 1328 + }, + { + "epoch": 0.41858267716535436, + "grad_norm": 19.225820541381836, + "learning_rate": 9.997913385826772e-06, + "loss": 0.1259, + "step": 1329 + }, + { + "epoch": 0.4188976377952756, + "grad_norm": 20.358154296875, + "learning_rate": 9.997911811023623e-06, + "loss": 0.1947, + "step": 1330 + }, + { + "epoch": 0.41921259842519687, + "grad_norm": 15.202303886413574, + "learning_rate": 9.997910236220472e-06, + "loss": 0.0907, + "step": 1331 + }, + { + "epoch": 0.4195275590551181, + "grad_norm": 16.259374618530273, + "learning_rate": 9.997908661417323e-06, + "loss": 0.0561, + "step": 1332 + }, + { + "epoch": 0.4198425196850394, + "grad_norm": 40.858028411865234, + "learning_rate": 9.997907086614174e-06, + "loss": 0.3551, + "step": 1333 + }, + { + "epoch": 0.42015748031496064, + "grad_norm": 21.958782196044922, + "learning_rate": 9.997905511811025e-06, + "loss": 0.1549, + "step": 1334 + }, + { + "epoch": 0.4204724409448819, + "grad_norm": 60.448734283447266, + "learning_rate": 9.997903937007875e-06, + "loss": 0.4945, + "step": 1335 + }, + { + "epoch": 0.42078740157480315, + "grad_norm": 9.495431900024414, + "learning_rate": 9.997902362204726e-06, + "loss": 0.0391, + "step": 1336 + }, + { + "epoch": 0.4211023622047244, + "grad_norm": 96.30876922607422, + "learning_rate": 9.997900787401575e-06, + "loss": 0.5236, + "step": 1337 + }, + { + "epoch": 0.42141732283464567, + "grad_norm": 100.25637817382812, + "learning_rate": 9.997899212598426e-06, + "loss": 0.5044, + "step": 1338 + }, + { + "epoch": 0.4217322834645669, + "grad_norm": 10.620447158813477, + "learning_rate": 9.997897637795277e-06, + "loss": 0.0333, + "step": 1339 + }, + { + "epoch": 0.4220472440944882, + "grad_norm": 99.30320739746094, + "learning_rate": 9.997896062992128e-06, + "loss": 0.4552, + "step": 1340 + }, + { + "epoch": 0.4220472440944882, + "eval_loss": 0.6429303884506226, + "eval_runtime": 339.7084, + "eval_samples_per_second": 0.344, + "eval_steps_per_second": 0.344, + "step": 1340 + }, + { + "epoch": 0.42236220472440944, + "grad_norm": 90.87741088867188, + "learning_rate": 9.997894488188977e-06, + "loss": 0.8964, + "step": 1341 + }, + { + "epoch": 0.4226771653543307, + "grad_norm": 53.61817169189453, + "learning_rate": 9.997892913385827e-06, + "loss": 0.2336, + "step": 1342 + }, + { + "epoch": 0.42299212598425195, + "grad_norm": 69.91616821289062, + "learning_rate": 9.997891338582678e-06, + "loss": 0.3807, + "step": 1343 + }, + { + "epoch": 0.4233070866141732, + "grad_norm": 64.52349853515625, + "learning_rate": 9.997889763779528e-06, + "loss": 0.3968, + "step": 1344 + }, + { + "epoch": 0.42362204724409447, + "grad_norm": 101.8323745727539, + "learning_rate": 9.99788818897638e-06, + "loss": 0.6942, + "step": 1345 + }, + { + "epoch": 0.4239370078740157, + "grad_norm": 137.70166015625, + "learning_rate": 9.997886614173229e-06, + "loss": 1.0158, + "step": 1346 + }, + { + "epoch": 0.424251968503937, + "grad_norm": 95.1364974975586, + "learning_rate": 9.99788503937008e-06, + "loss": 0.4745, + "step": 1347 + }, + { + "epoch": 0.4245669291338583, + "grad_norm": 42.36204147338867, + "learning_rate": 9.997883464566929e-06, + "loss": 0.265, + "step": 1348 + }, + { + "epoch": 0.42488188976377955, + "grad_norm": 37.66621017456055, + "learning_rate": 9.99788188976378e-06, + "loss": 0.5518, + "step": 1349 + }, + { + "epoch": 0.4251968503937008, + "grad_norm": 54.68177032470703, + "learning_rate": 9.997880314960631e-06, + "loss": 0.2369, + "step": 1350 + }, + { + "epoch": 0.42551181102362207, + "grad_norm": 88.8768310546875, + "learning_rate": 9.99787874015748e-06, + "loss": 1.004, + "step": 1351 + }, + { + "epoch": 0.4258267716535433, + "grad_norm": 9.298436164855957, + "learning_rate": 9.997877165354331e-06, + "loss": 0.0451, + "step": 1352 + }, + { + "epoch": 0.4261417322834646, + "grad_norm": 68.624755859375, + "learning_rate": 9.997875590551182e-06, + "loss": 0.9096, + "step": 1353 + }, + { + "epoch": 0.42645669291338584, + "grad_norm": 42.33887481689453, + "learning_rate": 9.997874015748033e-06, + "loss": 0.5008, + "step": 1354 + }, + { + "epoch": 0.4267716535433071, + "grad_norm": 84.9489517211914, + "learning_rate": 9.997872440944883e-06, + "loss": 0.4279, + "step": 1355 + }, + { + "epoch": 0.42708661417322835, + "grad_norm": 26.046403884887695, + "learning_rate": 9.997870866141734e-06, + "loss": 0.2294, + "step": 1356 + }, + { + "epoch": 0.4274015748031496, + "grad_norm": 24.110301971435547, + "learning_rate": 9.997869291338583e-06, + "loss": 0.1951, + "step": 1357 + }, + { + "epoch": 0.42771653543307087, + "grad_norm": 31.500816345214844, + "learning_rate": 9.997867716535434e-06, + "loss": 0.1093, + "step": 1358 + }, + { + "epoch": 0.4280314960629921, + "grad_norm": 37.86301040649414, + "learning_rate": 9.997866141732285e-06, + "loss": 0.5599, + "step": 1359 + }, + { + "epoch": 0.4283464566929134, + "grad_norm": 68.08171081542969, + "learning_rate": 9.997864566929136e-06, + "loss": 1.32, + "step": 1360 + }, + { + "epoch": 0.4283464566929134, + "eval_loss": 0.47423291206359863, + "eval_runtime": 339.2488, + "eval_samples_per_second": 0.345, + "eval_steps_per_second": 0.345, + "step": 1360 + }, + { + "epoch": 0.42866141732283464, + "grad_norm": 32.4645881652832, + "learning_rate": 9.997862992125985e-06, + "loss": 0.2244, + "step": 1361 + }, + { + "epoch": 0.4289763779527559, + "grad_norm": 21.44221305847168, + "learning_rate": 9.997861417322834e-06, + "loss": 0.1674, + "step": 1362 + }, + { + "epoch": 0.42929133858267715, + "grad_norm": 67.8936538696289, + "learning_rate": 9.997859842519685e-06, + "loss": 1.0609, + "step": 1363 + }, + { + "epoch": 0.4296062992125984, + "grad_norm": 44.46934127807617, + "learning_rate": 9.997858267716536e-06, + "loss": 0.4285, + "step": 1364 + }, + { + "epoch": 0.42992125984251967, + "grad_norm": 25.02652359008789, + "learning_rate": 9.997856692913387e-06, + "loss": 0.5787, + "step": 1365 + }, + { + "epoch": 0.4302362204724409, + "grad_norm": 53.81482696533203, + "learning_rate": 9.997855118110237e-06, + "loss": 0.3341, + "step": 1366 + }, + { + "epoch": 0.4305511811023622, + "grad_norm": 47.348201751708984, + "learning_rate": 9.997853543307088e-06, + "loss": 0.6322, + "step": 1367 + }, + { + "epoch": 0.4308661417322835, + "grad_norm": 22.753206253051758, + "learning_rate": 9.997851968503937e-06, + "loss": 0.4667, + "step": 1368 + }, + { + "epoch": 0.43118110236220475, + "grad_norm": 13.805994033813477, + "learning_rate": 9.997850393700788e-06, + "loss": 0.2587, + "step": 1369 + }, + { + "epoch": 0.431496062992126, + "grad_norm": 39.555076599121094, + "learning_rate": 9.997848818897639e-06, + "loss": 0.4924, + "step": 1370 + }, + { + "epoch": 0.43181102362204726, + "grad_norm": 34.64474105834961, + "learning_rate": 9.997847244094488e-06, + "loss": 0.5824, + "step": 1371 + }, + { + "epoch": 0.4321259842519685, + "grad_norm": 19.8635311126709, + "learning_rate": 9.99784566929134e-06, + "loss": 0.3249, + "step": 1372 + }, + { + "epoch": 0.4324409448818898, + "grad_norm": 45.133663177490234, + "learning_rate": 9.997844094488189e-06, + "loss": 0.4831, + "step": 1373 + }, + { + "epoch": 0.43275590551181103, + "grad_norm": 39.303409576416016, + "learning_rate": 9.99784251968504e-06, + "loss": 0.3898, + "step": 1374 + }, + { + "epoch": 0.4330708661417323, + "grad_norm": 34.55439758300781, + "learning_rate": 9.99784094488189e-06, + "loss": 0.4709, + "step": 1375 + }, + { + "epoch": 0.43338582677165355, + "grad_norm": 23.586498260498047, + "learning_rate": 9.997839370078742e-06, + "loss": 0.2142, + "step": 1376 + }, + { + "epoch": 0.4337007874015748, + "grad_norm": 25.976821899414062, + "learning_rate": 9.997837795275591e-06, + "loss": 0.4925, + "step": 1377 + }, + { + "epoch": 0.43401574803149606, + "grad_norm": 42.66455841064453, + "learning_rate": 9.997836220472442e-06, + "loss": 0.7777, + "step": 1378 + }, + { + "epoch": 0.4343307086614173, + "grad_norm": 24.16051483154297, + "learning_rate": 9.997834645669293e-06, + "loss": 0.3197, + "step": 1379 + }, + { + "epoch": 0.4346456692913386, + "grad_norm": 62.70100021362305, + "learning_rate": 9.997833070866144e-06, + "loss": 0.2733, + "step": 1380 + }, + { + "epoch": 0.4346456692913386, + "eval_loss": 0.4844158887863159, + "eval_runtime": 338.7196, + "eval_samples_per_second": 0.345, + "eval_steps_per_second": 0.345, + "step": 1380 + }, + { + "epoch": 0.43496062992125983, + "grad_norm": 43.02194595336914, + "learning_rate": 9.997831496062993e-06, + "loss": 0.3504, + "step": 1381 + }, + { + "epoch": 0.4352755905511811, + "grad_norm": 38.13003158569336, + "learning_rate": 9.997829921259842e-06, + "loss": 0.39, + "step": 1382 + }, + { + "epoch": 0.43559055118110235, + "grad_norm": 65.61643981933594, + "learning_rate": 9.997828346456693e-06, + "loss": 0.7555, + "step": 1383 + }, + { + "epoch": 0.4359055118110236, + "grad_norm": 31.994688034057617, + "learning_rate": 9.997826771653544e-06, + "loss": 0.4433, + "step": 1384 + }, + { + "epoch": 0.43622047244094486, + "grad_norm": 39.22994613647461, + "learning_rate": 9.997825196850395e-06, + "loss": 0.235, + "step": 1385 + }, + { + "epoch": 0.4365354330708661, + "grad_norm": 18.356678009033203, + "learning_rate": 9.997823622047245e-06, + "loss": 0.1587, + "step": 1386 + }, + { + "epoch": 0.4368503937007874, + "grad_norm": 16.878463745117188, + "learning_rate": 9.997822047244096e-06, + "loss": 0.3613, + "step": 1387 + }, + { + "epoch": 0.4371653543307087, + "grad_norm": 27.259780883789062, + "learning_rate": 9.997820472440945e-06, + "loss": 0.3243, + "step": 1388 + }, + { + "epoch": 0.43748031496062995, + "grad_norm": 25.487789154052734, + "learning_rate": 9.997818897637796e-06, + "loss": 0.4375, + "step": 1389 + }, + { + "epoch": 0.4377952755905512, + "grad_norm": 15.02742862701416, + "learning_rate": 9.997817322834647e-06, + "loss": 0.2624, + "step": 1390 + }, + { + "epoch": 0.43811023622047246, + "grad_norm": 26.566652297973633, + "learning_rate": 9.997815748031496e-06, + "loss": 0.4211, + "step": 1391 + }, + { + "epoch": 0.4384251968503937, + "grad_norm": 67.95926666259766, + "learning_rate": 9.997814173228347e-06, + "loss": 0.9252, + "step": 1392 + }, + { + "epoch": 0.438740157480315, + "grad_norm": 29.216794967651367, + "learning_rate": 9.997812598425197e-06, + "loss": 0.4028, + "step": 1393 + }, + { + "epoch": 0.43905511811023623, + "grad_norm": 50.51660919189453, + "learning_rate": 9.997811023622048e-06, + "loss": 0.4433, + "step": 1394 + }, + { + "epoch": 0.4393700787401575, + "grad_norm": 13.636687278747559, + "learning_rate": 9.997809448818899e-06, + "loss": 0.2541, + "step": 1395 + }, + { + "epoch": 0.43968503937007875, + "grad_norm": 26.405738830566406, + "learning_rate": 9.99780787401575e-06, + "loss": 0.3662, + "step": 1396 + }, + { + "epoch": 0.44, + "grad_norm": 31.268871307373047, + "learning_rate": 9.997806299212599e-06, + "loss": 0.2522, + "step": 1397 + }, + { + "epoch": 0.44031496062992126, + "grad_norm": 52.73464584350586, + "learning_rate": 9.99780472440945e-06, + "loss": 0.5821, + "step": 1398 + }, + { + "epoch": 0.4406299212598425, + "grad_norm": 7.454155921936035, + "learning_rate": 9.997803149606299e-06, + "loss": 0.0584, + "step": 1399 + }, + { + "epoch": 0.4409448818897638, + "grad_norm": 51.72380828857422, + "learning_rate": 9.997801574803152e-06, + "loss": 0.4622, + "step": 1400 + }, + { + "epoch": 0.4409448818897638, + "eval_loss": 0.47116619348526, + "eval_runtime": 337.7576, + "eval_samples_per_second": 0.346, + "eval_steps_per_second": 0.346, + "step": 1400 + }, + { + "epoch": 0.44125984251968503, + "grad_norm": 10.400198936462402, + "learning_rate": 9.997800000000001e-06, + "loss": 0.0597, + "step": 1401 + }, + { + "epoch": 0.4415748031496063, + "grad_norm": 48.547332763671875, + "learning_rate": 9.99779842519685e-06, + "loss": 0.3959, + "step": 1402 + }, + { + "epoch": 0.44188976377952754, + "grad_norm": 23.85326385498047, + "learning_rate": 9.997796850393701e-06, + "loss": 0.2755, + "step": 1403 + }, + { + "epoch": 0.4422047244094488, + "grad_norm": 25.128524780273438, + "learning_rate": 9.997795275590552e-06, + "loss": 0.0733, + "step": 1404 + }, + { + "epoch": 0.44251968503937006, + "grad_norm": 51.35587692260742, + "learning_rate": 9.997793700787403e-06, + "loss": 0.6701, + "step": 1405 + }, + { + "epoch": 0.4428346456692913, + "grad_norm": 14.554354667663574, + "learning_rate": 9.997792125984253e-06, + "loss": 0.064, + "step": 1406 + }, + { + "epoch": 0.4431496062992126, + "grad_norm": 67.32514953613281, + "learning_rate": 9.997790551181104e-06, + "loss": 0.4264, + "step": 1407 + }, + { + "epoch": 0.4434645669291339, + "grad_norm": 13.576302528381348, + "learning_rate": 9.997788976377953e-06, + "loss": 0.0385, + "step": 1408 + }, + { + "epoch": 0.44377952755905514, + "grad_norm": 126.57298278808594, + "learning_rate": 9.997787401574804e-06, + "loss": 0.7388, + "step": 1409 + }, + { + "epoch": 0.4440944881889764, + "grad_norm": 9.836527824401855, + "learning_rate": 9.997785826771655e-06, + "loss": 0.0383, + "step": 1410 + }, + { + "epoch": 0.44440944881889766, + "grad_norm": 76.40158081054688, + "learning_rate": 9.997784251968504e-06, + "loss": 0.3938, + "step": 1411 + }, + { + "epoch": 0.4447244094488189, + "grad_norm": 60.20785140991211, + "learning_rate": 9.997782677165355e-06, + "loss": 0.2053, + "step": 1412 + }, + { + "epoch": 0.44503937007874017, + "grad_norm": 14.930899620056152, + "learning_rate": 9.997781102362204e-06, + "loss": 0.0554, + "step": 1413 + }, + { + "epoch": 0.44535433070866143, + "grad_norm": 101.52273559570312, + "learning_rate": 9.997779527559055e-06, + "loss": 0.6495, + "step": 1414 + }, + { + "epoch": 0.4456692913385827, + "grad_norm": 6.1531267166137695, + "learning_rate": 9.997777952755906e-06, + "loss": 0.0153, + "step": 1415 + }, + { + "epoch": 0.44598425196850394, + "grad_norm": 55.18434143066406, + "learning_rate": 9.997776377952757e-06, + "loss": 0.5381, + "step": 1416 + }, + { + "epoch": 0.4462992125984252, + "grad_norm": 5.298098564147949, + "learning_rate": 9.997774803149607e-06, + "loss": 0.0163, + "step": 1417 + }, + { + "epoch": 0.44661417322834646, + "grad_norm": 33.11286163330078, + "learning_rate": 9.997773228346458e-06, + "loss": 0.051, + "step": 1418 + }, + { + "epoch": 0.4469291338582677, + "grad_norm": 86.3932876586914, + "learning_rate": 9.997771653543307e-06, + "loss": 0.7798, + "step": 1419 + }, + { + "epoch": 0.44724409448818897, + "grad_norm": 86.36060333251953, + "learning_rate": 9.997770078740158e-06, + "loss": 0.812, + "step": 1420 + }, + { + "epoch": 0.44724409448818897, + "eval_loss": 0.7094002366065979, + "eval_runtime": 338.8984, + "eval_samples_per_second": 0.345, + "eval_steps_per_second": 0.345, + "step": 1420 + }, + { + "epoch": 0.4475590551181102, + "grad_norm": 85.274169921875, + "learning_rate": 9.997768503937009e-06, + "loss": 1.2007, + "step": 1421 + }, + { + "epoch": 0.4478740157480315, + "grad_norm": 37.372406005859375, + "learning_rate": 9.997766929133858e-06, + "loss": 0.6024, + "step": 1422 + }, + { + "epoch": 0.44818897637795274, + "grad_norm": 72.87973022460938, + "learning_rate": 9.99776535433071e-06, + "loss": 0.9482, + "step": 1423 + }, + { + "epoch": 0.448503937007874, + "grad_norm": 19.026866912841797, + "learning_rate": 9.99776377952756e-06, + "loss": 0.0342, + "step": 1424 + }, + { + "epoch": 0.44881889763779526, + "grad_norm": 124.86366271972656, + "learning_rate": 9.997762204724411e-06, + "loss": 1.1195, + "step": 1425 + }, + { + "epoch": 0.4491338582677165, + "grad_norm": 91.7364273071289, + "learning_rate": 9.99776062992126e-06, + "loss": 0.6728, + "step": 1426 + }, + { + "epoch": 0.44944881889763777, + "grad_norm": 199.09945678710938, + "learning_rate": 9.997759055118112e-06, + "loss": 0.1931, + "step": 1427 + }, + { + "epoch": 0.4497637795275591, + "grad_norm": 88.02843475341797, + "learning_rate": 9.997757480314961e-06, + "loss": 0.8221, + "step": 1428 + }, + { + "epoch": 0.45007874015748034, + "grad_norm": 50.00212860107422, + "learning_rate": 9.997755905511812e-06, + "loss": 0.4988, + "step": 1429 + }, + { + "epoch": 0.4503937007874016, + "grad_norm": 46.531864166259766, + "learning_rate": 9.997754330708663e-06, + "loss": 0.166, + "step": 1430 + }, + { + "epoch": 0.45070866141732285, + "grad_norm": 62.54853439331055, + "learning_rate": 9.997752755905512e-06, + "loss": 0.8291, + "step": 1431 + }, + { + "epoch": 0.4510236220472441, + "grad_norm": 50.34440994262695, + "learning_rate": 9.997751181102363e-06, + "loss": 0.9383, + "step": 1432 + }, + { + "epoch": 0.45133858267716537, + "grad_norm": 93.70585632324219, + "learning_rate": 9.997749606299212e-06, + "loss": 0.7288, + "step": 1433 + }, + { + "epoch": 0.4516535433070866, + "grad_norm": 16.99148941040039, + "learning_rate": 9.997748031496063e-06, + "loss": 0.193, + "step": 1434 + }, + { + "epoch": 0.4519685039370079, + "grad_norm": 51.41777420043945, + "learning_rate": 9.997746456692914e-06, + "loss": 0.2534, + "step": 1435 + }, + { + "epoch": 0.45228346456692914, + "grad_norm": 60.7148323059082, + "learning_rate": 9.997744881889765e-06, + "loss": 0.824, + "step": 1436 + }, + { + "epoch": 0.4525984251968504, + "grad_norm": 42.44105529785156, + "learning_rate": 9.997743307086615e-06, + "loss": 0.5518, + "step": 1437 + }, + { + "epoch": 0.45291338582677165, + "grad_norm": 68.7177963256836, + "learning_rate": 9.997741732283466e-06, + "loss": 0.1653, + "step": 1438 + }, + { + "epoch": 0.4532283464566929, + "grad_norm": 34.2302131652832, + "learning_rate": 9.997740157480315e-06, + "loss": 0.3412, + "step": 1439 + }, + { + "epoch": 0.45354330708661417, + "grad_norm": 18.738046646118164, + "learning_rate": 9.997738582677166e-06, + "loss": 0.3157, + "step": 1440 + }, + { + "epoch": 0.45354330708661417, + "eval_loss": 0.5591472387313843, + "eval_runtime": 340.3548, + "eval_samples_per_second": 0.344, + "eval_steps_per_second": 0.344, + "step": 1440 + }, + { + "epoch": 0.4538582677165354, + "grad_norm": 113.84444427490234, + "learning_rate": 9.997737007874017e-06, + "loss": 1.0174, + "step": 1441 + }, + { + "epoch": 0.4541732283464567, + "grad_norm": 38.3621940612793, + "learning_rate": 9.997735433070866e-06, + "loss": 0.6811, + "step": 1442 + }, + { + "epoch": 0.45448818897637794, + "grad_norm": 34.84929656982422, + "learning_rate": 9.997733858267717e-06, + "loss": 0.2798, + "step": 1443 + }, + { + "epoch": 0.4548031496062992, + "grad_norm": 10.604235649108887, + "learning_rate": 9.997732283464567e-06, + "loss": 0.0995, + "step": 1444 + }, + { + "epoch": 0.45511811023622045, + "grad_norm": 48.81785202026367, + "learning_rate": 9.997730708661418e-06, + "loss": 0.713, + "step": 1445 + }, + { + "epoch": 0.4554330708661417, + "grad_norm": 22.666885375976562, + "learning_rate": 9.997729133858269e-06, + "loss": 0.5139, + "step": 1446 + }, + { + "epoch": 0.45574803149606297, + "grad_norm": 12.92003059387207, + "learning_rate": 9.99772755905512e-06, + "loss": 0.1314, + "step": 1447 + }, + { + "epoch": 0.4560629921259843, + "grad_norm": 29.922138214111328, + "learning_rate": 9.997725984251969e-06, + "loss": 0.4134, + "step": 1448 + }, + { + "epoch": 0.45637795275590554, + "grad_norm": 47.216609954833984, + "learning_rate": 9.99772440944882e-06, + "loss": 1.0959, + "step": 1449 + }, + { + "epoch": 0.4566929133858268, + "grad_norm": 20.693220138549805, + "learning_rate": 9.99772283464567e-06, + "loss": 0.1627, + "step": 1450 + }, + { + "epoch": 0.45700787401574805, + "grad_norm": 27.217304229736328, + "learning_rate": 9.99772125984252e-06, + "loss": 0.2322, + "step": 1451 + }, + { + "epoch": 0.4573228346456693, + "grad_norm": 49.21244430541992, + "learning_rate": 9.997719685039371e-06, + "loss": 0.2963, + "step": 1452 + }, + { + "epoch": 0.45763779527559056, + "grad_norm": 29.46310043334961, + "learning_rate": 9.99771811023622e-06, + "loss": 0.3798, + "step": 1453 + }, + { + "epoch": 0.4579527559055118, + "grad_norm": 30.430831909179688, + "learning_rate": 9.997716535433071e-06, + "loss": 0.3264, + "step": 1454 + }, + { + "epoch": 0.4582677165354331, + "grad_norm": 61.625083923339844, + "learning_rate": 9.997714960629922e-06, + "loss": 1.0914, + "step": 1455 + }, + { + "epoch": 0.45858267716535434, + "grad_norm": 28.40131187438965, + "learning_rate": 9.997713385826773e-06, + "loss": 0.1856, + "step": 1456 + }, + { + "epoch": 0.4588976377952756, + "grad_norm": 30.0435733795166, + "learning_rate": 9.997711811023623e-06, + "loss": 0.5765, + "step": 1457 + }, + { + "epoch": 0.45921259842519685, + "grad_norm": 15.2453031539917, + "learning_rate": 9.997710236220474e-06, + "loss": 0.287, + "step": 1458 + }, + { + "epoch": 0.4595275590551181, + "grad_norm": 36.65151596069336, + "learning_rate": 9.997708661417323e-06, + "loss": 0.2213, + "step": 1459 + }, + { + "epoch": 0.45984251968503936, + "grad_norm": 37.6453971862793, + "learning_rate": 9.997707086614174e-06, + "loss": 0.5465, + "step": 1460 + }, + { + "epoch": 0.45984251968503936, + "eval_loss": 0.49336767196655273, + "eval_runtime": 340.9663, + "eval_samples_per_second": 0.343, + "eval_steps_per_second": 0.343, + "step": 1460 + }, + { + "epoch": 0.4601574803149606, + "grad_norm": 21.586259841918945, + "learning_rate": 9.997705511811025e-06, + "loss": 0.1797, + "step": 1461 + }, + { + "epoch": 0.4604724409448819, + "grad_norm": 15.331527709960938, + "learning_rate": 9.997703937007874e-06, + "loss": 0.2192, + "step": 1462 + }, + { + "epoch": 0.46078740157480313, + "grad_norm": 27.347505569458008, + "learning_rate": 9.997702362204725e-06, + "loss": 0.6414, + "step": 1463 + }, + { + "epoch": 0.4611023622047244, + "grad_norm": 49.80681610107422, + "learning_rate": 9.997700787401575e-06, + "loss": 0.2621, + "step": 1464 + }, + { + "epoch": 0.46141732283464565, + "grad_norm": 20.706789016723633, + "learning_rate": 9.997699212598426e-06, + "loss": 0.1548, + "step": 1465 + }, + { + "epoch": 0.4617322834645669, + "grad_norm": 59.39773941040039, + "learning_rate": 9.997697637795276e-06, + "loss": 0.514, + "step": 1466 + }, + { + "epoch": 0.46204724409448816, + "grad_norm": 29.0017147064209, + "learning_rate": 9.997696062992127e-06, + "loss": 0.4804, + "step": 1467 + }, + { + "epoch": 0.4623622047244095, + "grad_norm": 59.08629608154297, + "learning_rate": 9.997694488188977e-06, + "loss": 0.6536, + "step": 1468 + }, + { + "epoch": 0.46267716535433073, + "grad_norm": 39.09469223022461, + "learning_rate": 9.997692913385828e-06, + "loss": 0.3326, + "step": 1469 + }, + { + "epoch": 0.462992125984252, + "grad_norm": 14.819294929504395, + "learning_rate": 9.997691338582679e-06, + "loss": 0.164, + "step": 1470 + }, + { + "epoch": 0.46330708661417325, + "grad_norm": 23.74578094482422, + "learning_rate": 9.997689763779528e-06, + "loss": 0.3426, + "step": 1471 + }, + { + "epoch": 0.4636220472440945, + "grad_norm": 31.888647079467773, + "learning_rate": 9.997688188976379e-06, + "loss": 0.4113, + "step": 1472 + }, + { + "epoch": 0.46393700787401576, + "grad_norm": 42.581398010253906, + "learning_rate": 9.997686614173228e-06, + "loss": 0.3324, + "step": 1473 + }, + { + "epoch": 0.464251968503937, + "grad_norm": 45.32304382324219, + "learning_rate": 9.99768503937008e-06, + "loss": 0.5497, + "step": 1474 + }, + { + "epoch": 0.4645669291338583, + "grad_norm": 31.52800750732422, + "learning_rate": 9.99768346456693e-06, + "loss": 0.6702, + "step": 1475 + }, + { + "epoch": 0.46488188976377953, + "grad_norm": 41.72283172607422, + "learning_rate": 9.997681889763781e-06, + "loss": 0.2619, + "step": 1476 + }, + { + "epoch": 0.4651968503937008, + "grad_norm": 97.3839340209961, + "learning_rate": 9.99768031496063e-06, + "loss": 0.4892, + "step": 1477 + }, + { + "epoch": 0.46551181102362205, + "grad_norm": 39.59928512573242, + "learning_rate": 9.997678740157482e-06, + "loss": 0.5622, + "step": 1478 + }, + { + "epoch": 0.4658267716535433, + "grad_norm": 23.758737564086914, + "learning_rate": 9.997677165354331e-06, + "loss": 0.3074, + "step": 1479 + }, + { + "epoch": 0.46614173228346456, + "grad_norm": 39.63565444946289, + "learning_rate": 9.997675590551182e-06, + "loss": 0.3154, + "step": 1480 + }, + { + "epoch": 0.46614173228346456, + "eval_loss": 0.4589572548866272, + "eval_runtime": 341.5074, + "eval_samples_per_second": 0.343, + "eval_steps_per_second": 0.343, + "step": 1480 + }, + { + "epoch": 0.4664566929133858, + "grad_norm": 19.873315811157227, + "learning_rate": 9.997674015748033e-06, + "loss": 0.2133, + "step": 1481 + }, + { + "epoch": 0.4667716535433071, + "grad_norm": 32.47504806518555, + "learning_rate": 9.997672440944882e-06, + "loss": 0.3143, + "step": 1482 + }, + { + "epoch": 0.46708661417322833, + "grad_norm": 27.075977325439453, + "learning_rate": 9.997670866141733e-06, + "loss": 0.315, + "step": 1483 + }, + { + "epoch": 0.4674015748031496, + "grad_norm": 18.573650360107422, + "learning_rate": 9.997669291338582e-06, + "loss": 0.1998, + "step": 1484 + }, + { + "epoch": 0.46771653543307085, + "grad_norm": 19.674373626708984, + "learning_rate": 9.997667716535433e-06, + "loss": 0.2505, + "step": 1485 + }, + { + "epoch": 0.4680314960629921, + "grad_norm": 40.249202728271484, + "learning_rate": 9.997666141732284e-06, + "loss": 0.4321, + "step": 1486 + }, + { + "epoch": 0.46834645669291336, + "grad_norm": 12.53464412689209, + "learning_rate": 9.997664566929135e-06, + "loss": 0.069, + "step": 1487 + }, + { + "epoch": 0.4686614173228347, + "grad_norm": 38.792728424072266, + "learning_rate": 9.997662992125985e-06, + "loss": 0.417, + "step": 1488 + }, + { + "epoch": 0.46897637795275593, + "grad_norm": 56.2455940246582, + "learning_rate": 9.997661417322836e-06, + "loss": 0.7037, + "step": 1489 + }, + { + "epoch": 0.4692913385826772, + "grad_norm": 47.736263275146484, + "learning_rate": 9.997659842519685e-06, + "loss": 0.2714, + "step": 1490 + }, + { + "epoch": 0.46960629921259844, + "grad_norm": 75.22129821777344, + "learning_rate": 9.997658267716536e-06, + "loss": 0.5145, + "step": 1491 + }, + { + "epoch": 0.4699212598425197, + "grad_norm": 21.2304744720459, + "learning_rate": 9.997656692913387e-06, + "loss": 0.2883, + "step": 1492 + }, + { + "epoch": 0.47023622047244096, + "grad_norm": 62.4765625, + "learning_rate": 9.997655118110236e-06, + "loss": 0.4325, + "step": 1493 + }, + { + "epoch": 0.4705511811023622, + "grad_norm": 49.23244094848633, + "learning_rate": 9.997653543307087e-06, + "loss": 0.6759, + "step": 1494 + }, + { + "epoch": 0.47086614173228347, + "grad_norm": 21.292394638061523, + "learning_rate": 9.997651968503938e-06, + "loss": 0.1809, + "step": 1495 + }, + { + "epoch": 0.47118110236220473, + "grad_norm": 57.66270446777344, + "learning_rate": 9.99765039370079e-06, + "loss": 0.7726, + "step": 1496 + }, + { + "epoch": 0.471496062992126, + "grad_norm": 61.524253845214844, + "learning_rate": 9.997648818897639e-06, + "loss": 0.3385, + "step": 1497 + }, + { + "epoch": 0.47181102362204724, + "grad_norm": 51.74467468261719, + "learning_rate": 9.99764724409449e-06, + "loss": 0.1262, + "step": 1498 + }, + { + "epoch": 0.4721259842519685, + "grad_norm": 64.45372009277344, + "learning_rate": 9.997645669291339e-06, + "loss": 0.5856, + "step": 1499 + }, + { + "epoch": 0.47244094488188976, + "grad_norm": 33.162811279296875, + "learning_rate": 9.99764409448819e-06, + "loss": 0.8524, + "step": 1500 + }, + { + "epoch": 0.47244094488188976, + "eval_loss": 0.4843520522117615, + "eval_runtime": 337.6077, + "eval_samples_per_second": 0.347, + "eval_steps_per_second": 0.347, + "step": 1500 + }, + { + "epoch": 0.472755905511811, + "grad_norm": 58.66725158691406, + "learning_rate": 9.99764251968504e-06, + "loss": 0.7346, + "step": 1501 + }, + { + "epoch": 0.47307086614173227, + "grad_norm": 113.04570770263672, + "learning_rate": 9.99764094488189e-06, + "loss": 0.4166, + "step": 1502 + }, + { + "epoch": 0.47338582677165353, + "grad_norm": 49.267024993896484, + "learning_rate": 9.997639370078741e-06, + "loss": 0.8303, + "step": 1503 + }, + { + "epoch": 0.4737007874015748, + "grad_norm": 18.804046630859375, + "learning_rate": 9.99763779527559e-06, + "loss": 0.2172, + "step": 1504 + }, + { + "epoch": 0.47401574803149604, + "grad_norm": 23.38251495361328, + "learning_rate": 9.997636220472441e-06, + "loss": 0.1723, + "step": 1505 + }, + { + "epoch": 0.4743307086614173, + "grad_norm": 32.252750396728516, + "learning_rate": 9.997634645669292e-06, + "loss": 0.1733, + "step": 1506 + }, + { + "epoch": 0.47464566929133856, + "grad_norm": 58.86079025268555, + "learning_rate": 9.997633070866143e-06, + "loss": 0.8325, + "step": 1507 + }, + { + "epoch": 0.47496062992125987, + "grad_norm": 18.70965003967285, + "learning_rate": 9.997631496062993e-06, + "loss": 0.1669, + "step": 1508 + }, + { + "epoch": 0.4752755905511811, + "grad_norm": 31.96597671508789, + "learning_rate": 9.997629921259844e-06, + "loss": 0.3987, + "step": 1509 + }, + { + "epoch": 0.4755905511811024, + "grad_norm": 21.83759880065918, + "learning_rate": 9.997628346456693e-06, + "loss": 0.1909, + "step": 1510 + }, + { + "epoch": 0.47590551181102364, + "grad_norm": 27.163360595703125, + "learning_rate": 9.997626771653544e-06, + "loss": 0.3432, + "step": 1511 + }, + { + "epoch": 0.4762204724409449, + "grad_norm": 81.91073608398438, + "learning_rate": 9.997625196850395e-06, + "loss": 0.4286, + "step": 1512 + }, + { + "epoch": 0.47653543307086615, + "grad_norm": 23.85365867614746, + "learning_rate": 9.997623622047244e-06, + "loss": 0.2956, + "step": 1513 + }, + { + "epoch": 0.4768503937007874, + "grad_norm": 20.397815704345703, + "learning_rate": 9.997622047244095e-06, + "loss": 0.4461, + "step": 1514 + }, + { + "epoch": 0.47716535433070867, + "grad_norm": 30.885562896728516, + "learning_rate": 9.997620472440945e-06, + "loss": 0.477, + "step": 1515 + }, + { + "epoch": 0.4774803149606299, + "grad_norm": 33.92880630493164, + "learning_rate": 9.997618897637796e-06, + "loss": 0.3881, + "step": 1516 + }, + { + "epoch": 0.4777952755905512, + "grad_norm": 24.906478881835938, + "learning_rate": 9.997617322834647e-06, + "loss": 0.3944, + "step": 1517 + }, + { + "epoch": 0.47811023622047244, + "grad_norm": 17.711050033569336, + "learning_rate": 9.997615748031498e-06, + "loss": 0.3327, + "step": 1518 + }, + { + "epoch": 0.4784251968503937, + "grad_norm": 28.30522918701172, + "learning_rate": 9.997614173228347e-06, + "loss": 0.4556, + "step": 1519 + }, + { + "epoch": 0.47874015748031495, + "grad_norm": 22.290746688842773, + "learning_rate": 9.997612598425198e-06, + "loss": 0.177, + "step": 1520 + }, + { + "epoch": 0.47874015748031495, + "eval_loss": 0.45914146304130554, + "eval_runtime": 337.9882, + "eval_samples_per_second": 0.346, + "eval_steps_per_second": 0.346, + "step": 1520 + }, + { + "epoch": 0.4790551181102362, + "grad_norm": 51.145206451416016, + "learning_rate": 9.997611023622049e-06, + "loss": 0.2974, + "step": 1521 + }, + { + "epoch": 0.47937007874015747, + "grad_norm": 53.45519256591797, + "learning_rate": 9.997609448818898e-06, + "loss": 0.2501, + "step": 1522 + }, + { + "epoch": 0.4796850393700787, + "grad_norm": 41.410377502441406, + "learning_rate": 9.997607874015749e-06, + "loss": 0.5987, + "step": 1523 + }, + { + "epoch": 0.48, + "grad_norm": 29.04597282409668, + "learning_rate": 9.997606299212598e-06, + "loss": 0.3196, + "step": 1524 + }, + { + "epoch": 0.48031496062992124, + "grad_norm": 37.950008392333984, + "learning_rate": 9.99760472440945e-06, + "loss": 0.444, + "step": 1525 + }, + { + "epoch": 0.4806299212598425, + "grad_norm": 47.16240692138672, + "learning_rate": 9.9976031496063e-06, + "loss": 0.6532, + "step": 1526 + }, + { + "epoch": 0.48094488188976375, + "grad_norm": 33.97493362426758, + "learning_rate": 9.997601574803151e-06, + "loss": 0.4194, + "step": 1527 + }, + { + "epoch": 0.48125984251968507, + "grad_norm": 36.6172981262207, + "learning_rate": 9.9976e-06, + "loss": 0.4505, + "step": 1528 + }, + { + "epoch": 0.4815748031496063, + "grad_norm": 26.39623260498047, + "learning_rate": 9.997598425196852e-06, + "loss": 0.2039, + "step": 1529 + }, + { + "epoch": 0.4818897637795276, + "grad_norm": 24.622024536132812, + "learning_rate": 9.997596850393701e-06, + "loss": 0.2808, + "step": 1530 + }, + { + "epoch": 0.48220472440944884, + "grad_norm": 58.132904052734375, + "learning_rate": 9.997595275590552e-06, + "loss": 0.4782, + "step": 1531 + }, + { + "epoch": 0.4825196850393701, + "grad_norm": 53.385154724121094, + "learning_rate": 9.997593700787403e-06, + "loss": 0.4064, + "step": 1532 + }, + { + "epoch": 0.48283464566929135, + "grad_norm": 18.42597007751465, + "learning_rate": 9.997592125984252e-06, + "loss": 0.0884, + "step": 1533 + }, + { + "epoch": 0.4831496062992126, + "grad_norm": 68.61125183105469, + "learning_rate": 9.997590551181103e-06, + "loss": 0.52, + "step": 1534 + }, + { + "epoch": 0.48346456692913387, + "grad_norm": 19.761259078979492, + "learning_rate": 9.997588976377952e-06, + "loss": 0.1054, + "step": 1535 + }, + { + "epoch": 0.4837795275590551, + "grad_norm": 20.919004440307617, + "learning_rate": 9.997587401574803e-06, + "loss": 0.2819, + "step": 1536 + }, + { + "epoch": 0.4840944881889764, + "grad_norm": 24.11286735534668, + "learning_rate": 9.997585826771654e-06, + "loss": 0.2413, + "step": 1537 + }, + { + "epoch": 0.48440944881889764, + "grad_norm": 34.09194564819336, + "learning_rate": 9.997584251968505e-06, + "loss": 0.2417, + "step": 1538 + }, + { + "epoch": 0.4847244094488189, + "grad_norm": 53.1319580078125, + "learning_rate": 9.997582677165355e-06, + "loss": 0.5983, + "step": 1539 + }, + { + "epoch": 0.48503937007874015, + "grad_norm": 23.620807647705078, + "learning_rate": 9.997581102362206e-06, + "loss": 0.1747, + "step": 1540 + }, + { + "epoch": 0.48503937007874015, + "eval_loss": 0.453780859708786, + "eval_runtime": 336.017, + "eval_samples_per_second": 0.348, + "eval_steps_per_second": 0.348, + "step": 1540 + }, + { + "epoch": 0.4853543307086614, + "grad_norm": 45.581031799316406, + "learning_rate": 9.997579527559057e-06, + "loss": 0.4102, + "step": 1541 + }, + { + "epoch": 0.48566929133858266, + "grad_norm": 49.25587463378906, + "learning_rate": 9.997577952755906e-06, + "loss": 0.71, + "step": 1542 + }, + { + "epoch": 0.4859842519685039, + "grad_norm": 4.6258649826049805, + "learning_rate": 9.997576377952757e-06, + "loss": 0.0368, + "step": 1543 + }, + { + "epoch": 0.4862992125984252, + "grad_norm": 17.629261016845703, + "learning_rate": 9.997574803149606e-06, + "loss": 0.2871, + "step": 1544 + }, + { + "epoch": 0.48661417322834644, + "grad_norm": 53.029666900634766, + "learning_rate": 9.997573228346457e-06, + "loss": 0.4007, + "step": 1545 + }, + { + "epoch": 0.4869291338582677, + "grad_norm": 25.097759246826172, + "learning_rate": 9.997571653543308e-06, + "loss": 0.2665, + "step": 1546 + }, + { + "epoch": 0.48724409448818895, + "grad_norm": 59.07927703857422, + "learning_rate": 9.99757007874016e-06, + "loss": 0.8133, + "step": 1547 + }, + { + "epoch": 0.48755905511811026, + "grad_norm": 39.4317741394043, + "learning_rate": 9.997568503937009e-06, + "loss": 0.1684, + "step": 1548 + }, + { + "epoch": 0.4878740157480315, + "grad_norm": 61.61114501953125, + "learning_rate": 9.99756692913386e-06, + "loss": 0.5758, + "step": 1549 + }, + { + "epoch": 0.4881889763779528, + "grad_norm": 57.046142578125, + "learning_rate": 9.997565354330709e-06, + "loss": 0.4657, + "step": 1550 + }, + { + "epoch": 0.48850393700787403, + "grad_norm": 28.860565185546875, + "learning_rate": 9.99756377952756e-06, + "loss": 0.4151, + "step": 1551 + }, + { + "epoch": 0.4888188976377953, + "grad_norm": 92.83087158203125, + "learning_rate": 9.997562204724411e-06, + "loss": 0.7444, + "step": 1552 + }, + { + "epoch": 0.48913385826771655, + "grad_norm": 41.72923278808594, + "learning_rate": 9.99756062992126e-06, + "loss": 0.7972, + "step": 1553 + }, + { + "epoch": 0.4894488188976378, + "grad_norm": 43.05347442626953, + "learning_rate": 9.997559055118111e-06, + "loss": 0.1623, + "step": 1554 + }, + { + "epoch": 0.48976377952755906, + "grad_norm": 37.73609161376953, + "learning_rate": 9.99755748031496e-06, + "loss": 0.4308, + "step": 1555 + }, + { + "epoch": 0.4900787401574803, + "grad_norm": 22.235315322875977, + "learning_rate": 9.997555905511811e-06, + "loss": 0.2682, + "step": 1556 + }, + { + "epoch": 0.4903937007874016, + "grad_norm": 4.401269435882568, + "learning_rate": 9.997554330708662e-06, + "loss": 0.0164, + "step": 1557 + }, + { + "epoch": 0.49070866141732283, + "grad_norm": 44.11044692993164, + "learning_rate": 9.997552755905513e-06, + "loss": 0.3743, + "step": 1558 + }, + { + "epoch": 0.4910236220472441, + "grad_norm": 38.67685317993164, + "learning_rate": 9.997551181102363e-06, + "loss": 0.3507, + "step": 1559 + }, + { + "epoch": 0.49133858267716535, + "grad_norm": 20.494264602661133, + "learning_rate": 9.997549606299214e-06, + "loss": 0.3759, + "step": 1560 + }, + { + "epoch": 0.49133858267716535, + "eval_loss": 0.43347296118736267, + "eval_runtime": 339.139, + "eval_samples_per_second": 0.345, + "eval_steps_per_second": 0.345, + "step": 1560 + }, + { + "epoch": 0.4916535433070866, + "grad_norm": 21.03292465209961, + "learning_rate": 9.997548031496063e-06, + "loss": 0.2048, + "step": 1561 + }, + { + "epoch": 0.49196850393700786, + "grad_norm": 40.08341979980469, + "learning_rate": 9.997546456692914e-06, + "loss": 0.1279, + "step": 1562 + }, + { + "epoch": 0.4922834645669291, + "grad_norm": 31.835474014282227, + "learning_rate": 9.997544881889765e-06, + "loss": 0.26, + "step": 1563 + }, + { + "epoch": 0.4925984251968504, + "grad_norm": 26.725603103637695, + "learning_rate": 9.997543307086614e-06, + "loss": 0.2091, + "step": 1564 + }, + { + "epoch": 0.49291338582677163, + "grad_norm": 33.68645095825195, + "learning_rate": 9.997541732283465e-06, + "loss": 0.2875, + "step": 1565 + }, + { + "epoch": 0.4932283464566929, + "grad_norm": 56.92742156982422, + "learning_rate": 9.997540157480316e-06, + "loss": 1.1372, + "step": 1566 + }, + { + "epoch": 0.49354330708661415, + "grad_norm": 10.900672912597656, + "learning_rate": 9.997538582677167e-06, + "loss": 0.0537, + "step": 1567 + }, + { + "epoch": 0.49385826771653546, + "grad_norm": 20.076473236083984, + "learning_rate": 9.997537007874017e-06, + "loss": 0.1373, + "step": 1568 + }, + { + "epoch": 0.4941732283464567, + "grad_norm": 30.94587516784668, + "learning_rate": 9.997535433070868e-06, + "loss": 0.348, + "step": 1569 + }, + { + "epoch": 0.494488188976378, + "grad_norm": 79.50629425048828, + "learning_rate": 9.997533858267717e-06, + "loss": 0.9603, + "step": 1570 + }, + { + "epoch": 0.49480314960629923, + "grad_norm": 51.20693588256836, + "learning_rate": 9.997532283464568e-06, + "loss": 1.1457, + "step": 1571 + }, + { + "epoch": 0.4951181102362205, + "grad_norm": 34.06742477416992, + "learning_rate": 9.997530708661419e-06, + "loss": 0.4457, + "step": 1572 + }, + { + "epoch": 0.49543307086614174, + "grad_norm": 40.76358413696289, + "learning_rate": 9.997529133858268e-06, + "loss": 0.2998, + "step": 1573 + }, + { + "epoch": 0.495748031496063, + "grad_norm": 23.580713272094727, + "learning_rate": 9.997527559055119e-06, + "loss": 0.1263, + "step": 1574 + }, + { + "epoch": 0.49606299212598426, + "grad_norm": 6.127594470977783, + "learning_rate": 9.997525984251968e-06, + "loss": 0.0331, + "step": 1575 + }, + { + "epoch": 0.4963779527559055, + "grad_norm": 28.260459899902344, + "learning_rate": 9.99752440944882e-06, + "loss": 0.3719, + "step": 1576 + }, + { + "epoch": 0.4966929133858268, + "grad_norm": 68.38239288330078, + "learning_rate": 9.99752283464567e-06, + "loss": 0.6032, + "step": 1577 + }, + { + "epoch": 0.49700787401574803, + "grad_norm": 60.19215393066406, + "learning_rate": 9.997521259842521e-06, + "loss": 0.2563, + "step": 1578 + }, + { + "epoch": 0.4973228346456693, + "grad_norm": 40.08391571044922, + "learning_rate": 9.99751968503937e-06, + "loss": 0.2775, + "step": 1579 + }, + { + "epoch": 0.49763779527559054, + "grad_norm": 37.87879180908203, + "learning_rate": 9.997518110236222e-06, + "loss": 0.117, + "step": 1580 + }, + { + "epoch": 0.49763779527559054, + "eval_loss": 0.5100625157356262, + "eval_runtime": 338.9567, + "eval_samples_per_second": 0.345, + "eval_steps_per_second": 0.345, + "step": 1580 + }, + { + "epoch": 0.4979527559055118, + "grad_norm": 43.88140869140625, + "learning_rate": 9.997516535433071e-06, + "loss": 0.7505, + "step": 1581 + }, + { + "epoch": 0.49826771653543306, + "grad_norm": 10.613367080688477, + "learning_rate": 9.997514960629922e-06, + "loss": 0.0331, + "step": 1582 + }, + { + "epoch": 0.4985826771653543, + "grad_norm": 115.95816040039062, + "learning_rate": 9.997513385826773e-06, + "loss": 0.6034, + "step": 1583 + }, + { + "epoch": 0.4988976377952756, + "grad_norm": 52.247474670410156, + "learning_rate": 9.997511811023622e-06, + "loss": 0.7637, + "step": 1584 + }, + { + "epoch": 0.49921259842519683, + "grad_norm": 49.17156982421875, + "learning_rate": 9.997510236220473e-06, + "loss": 0.2622, + "step": 1585 + }, + { + "epoch": 0.4995275590551181, + "grad_norm": 52.29065704345703, + "learning_rate": 9.997508661417323e-06, + "loss": 0.2516, + "step": 1586 + }, + { + "epoch": 0.49984251968503934, + "grad_norm": 4.530742168426514, + "learning_rate": 9.997507086614175e-06, + "loss": 0.039, + "step": 1587 + }, + { + "epoch": 0.5001574803149607, + "grad_norm": 54.83158874511719, + "learning_rate": 9.997505511811024e-06, + "loss": 1.2184, + "step": 1588 + }, + { + "epoch": 0.5004724409448819, + "grad_norm": 80.9505615234375, + "learning_rate": 9.997503937007875e-06, + "loss": 0.6721, + "step": 1589 + }, + { + "epoch": 0.5007874015748032, + "grad_norm": 88.2104263305664, + "learning_rate": 9.997502362204725e-06, + "loss": 0.8541, + "step": 1590 + }, + { + "epoch": 0.5011023622047244, + "grad_norm": 54.88927459716797, + "learning_rate": 9.997500787401576e-06, + "loss": 0.7247, + "step": 1591 + }, + { + "epoch": 0.5014173228346457, + "grad_norm": 58.55228805541992, + "learning_rate": 9.997499212598427e-06, + "loss": 0.6949, + "step": 1592 + }, + { + "epoch": 0.5017322834645669, + "grad_norm": 70.87881469726562, + "learning_rate": 9.997497637795276e-06, + "loss": 0.5249, + "step": 1593 + }, + { + "epoch": 0.5020472440944882, + "grad_norm": 49.13249206542969, + "learning_rate": 9.997496062992127e-06, + "loss": 0.3263, + "step": 1594 + }, + { + "epoch": 0.5023622047244094, + "grad_norm": 44.89517593383789, + "learning_rate": 9.997494488188976e-06, + "loss": 0.2266, + "step": 1595 + }, + { + "epoch": 0.5026771653543307, + "grad_norm": 34.858665466308594, + "learning_rate": 9.997492913385827e-06, + "loss": 0.3243, + "step": 1596 + }, + { + "epoch": 0.5029921259842519, + "grad_norm": 48.156105041503906, + "learning_rate": 9.997491338582678e-06, + "loss": 0.6439, + "step": 1597 + }, + { + "epoch": 0.5033070866141732, + "grad_norm": 23.67864418029785, + "learning_rate": 9.99748976377953e-06, + "loss": 0.4013, + "step": 1598 + }, + { + "epoch": 0.5036220472440945, + "grad_norm": 33.883583068847656, + "learning_rate": 9.997488188976379e-06, + "loss": 0.3123, + "step": 1599 + }, + { + "epoch": 0.5039370078740157, + "grad_norm": 14.358415603637695, + "learning_rate": 9.99748661417323e-06, + "loss": 0.1317, + "step": 1600 + }, + { + "epoch": 0.5039370078740157, + "eval_loss": 0.46397241950035095, + "eval_runtime": 338.8895, + "eval_samples_per_second": 0.345, + "eval_steps_per_second": 0.345, + "step": 1600 + }, + { + "epoch": 0.504251968503937, + "grad_norm": 29.868349075317383, + "learning_rate": 9.997485039370079e-06, + "loss": 0.2516, + "step": 1601 + }, + { + "epoch": 0.5045669291338583, + "grad_norm": 21.991357803344727, + "learning_rate": 9.99748346456693e-06, + "loss": 0.2228, + "step": 1602 + }, + { + "epoch": 0.5048818897637796, + "grad_norm": 15.654339790344238, + "learning_rate": 9.997481889763781e-06, + "loss": 0.1345, + "step": 1603 + }, + { + "epoch": 0.5051968503937008, + "grad_norm": 15.310891151428223, + "learning_rate": 9.99748031496063e-06, + "loss": 0.1679, + "step": 1604 + }, + { + "epoch": 0.5055118110236221, + "grad_norm": 38.296146392822266, + "learning_rate": 9.997478740157481e-06, + "loss": 0.3338, + "step": 1605 + }, + { + "epoch": 0.5058267716535433, + "grad_norm": 43.396427154541016, + "learning_rate": 9.99747716535433e-06, + "loss": 0.4996, + "step": 1606 + }, + { + "epoch": 0.5061417322834646, + "grad_norm": 94.86878204345703, + "learning_rate": 9.997475590551181e-06, + "loss": 0.8225, + "step": 1607 + }, + { + "epoch": 0.5064566929133858, + "grad_norm": 52.50116729736328, + "learning_rate": 9.997474015748032e-06, + "loss": 0.3828, + "step": 1608 + }, + { + "epoch": 0.5067716535433071, + "grad_norm": 38.4481086730957, + "learning_rate": 9.997472440944883e-06, + "loss": 0.1611, + "step": 1609 + }, + { + "epoch": 0.5070866141732283, + "grad_norm": 49.7927131652832, + "learning_rate": 9.997470866141733e-06, + "loss": 1.0682, + "step": 1610 + }, + { + "epoch": 0.5074015748031496, + "grad_norm": 8.284367561340332, + "learning_rate": 9.997469291338584e-06, + "loss": 0.0571, + "step": 1611 + }, + { + "epoch": 0.5077165354330708, + "grad_norm": 24.673439025878906, + "learning_rate": 9.997467716535435e-06, + "loss": 0.1879, + "step": 1612 + }, + { + "epoch": 0.5080314960629921, + "grad_norm": 33.2513542175293, + "learning_rate": 9.997466141732284e-06, + "loss": 0.636, + "step": 1613 + }, + { + "epoch": 0.5083464566929133, + "grad_norm": 62.058738708496094, + "learning_rate": 9.997464566929135e-06, + "loss": 0.878, + "step": 1614 + }, + { + "epoch": 0.5086614173228347, + "grad_norm": 79.89865112304688, + "learning_rate": 9.997462992125984e-06, + "loss": 0.2207, + "step": 1615 + }, + { + "epoch": 0.5089763779527559, + "grad_norm": 37.931819915771484, + "learning_rate": 9.997461417322835e-06, + "loss": 0.1365, + "step": 1616 + }, + { + "epoch": 0.5092913385826772, + "grad_norm": 30.21257209777832, + "learning_rate": 9.997459842519686e-06, + "loss": 0.4815, + "step": 1617 + }, + { + "epoch": 0.5096062992125985, + "grad_norm": 35.088890075683594, + "learning_rate": 9.997458267716537e-06, + "loss": 0.2641, + "step": 1618 + }, + { + "epoch": 0.5099212598425197, + "grad_norm": 31.518869400024414, + "learning_rate": 9.997456692913387e-06, + "loss": 0.3096, + "step": 1619 + }, + { + "epoch": 0.510236220472441, + "grad_norm": 43.287391662597656, + "learning_rate": 9.997455118110238e-06, + "loss": 0.4844, + "step": 1620 + }, + { + "epoch": 0.510236220472441, + "eval_loss": 0.5515217781066895, + "eval_runtime": 339.8588, + "eval_samples_per_second": 0.344, + "eval_steps_per_second": 0.344, + "step": 1620 + }, + { + "epoch": 0.5105511811023622, + "grad_norm": 20.773025512695312, + "learning_rate": 9.997453543307087e-06, + "loss": 0.0673, + "step": 1621 + }, + { + "epoch": 0.5108661417322835, + "grad_norm": 22.52477264404297, + "learning_rate": 9.997451968503938e-06, + "loss": 0.2089, + "step": 1622 + }, + { + "epoch": 0.5111811023622047, + "grad_norm": 21.807056427001953, + "learning_rate": 9.997450393700789e-06, + "loss": 0.1542, + "step": 1623 + }, + { + "epoch": 0.511496062992126, + "grad_norm": 60.92094802856445, + "learning_rate": 9.997448818897638e-06, + "loss": 0.9874, + "step": 1624 + }, + { + "epoch": 0.5118110236220472, + "grad_norm": 100.76142120361328, + "learning_rate": 9.997447244094489e-06, + "loss": 0.3835, + "step": 1625 + }, + { + "epoch": 0.5121259842519685, + "grad_norm": 36.08515167236328, + "learning_rate": 9.997445669291338e-06, + "loss": 0.3638, + "step": 1626 + }, + { + "epoch": 0.5124409448818897, + "grad_norm": 36.939170837402344, + "learning_rate": 9.99744409448819e-06, + "loss": 0.7436, + "step": 1627 + }, + { + "epoch": 0.512755905511811, + "grad_norm": 48.17253494262695, + "learning_rate": 9.99744251968504e-06, + "loss": 0.4945, + "step": 1628 + }, + { + "epoch": 0.5130708661417323, + "grad_norm": 79.20890808105469, + "learning_rate": 9.997440944881891e-06, + "loss": 0.8642, + "step": 1629 + }, + { + "epoch": 0.5133858267716536, + "grad_norm": 19.723230361938477, + "learning_rate": 9.99743937007874e-06, + "loss": 0.1813, + "step": 1630 + }, + { + "epoch": 0.5137007874015748, + "grad_norm": 39.59589385986328, + "learning_rate": 9.997437795275592e-06, + "loss": 0.3752, + "step": 1631 + }, + { + "epoch": 0.5140157480314961, + "grad_norm": 77.0748062133789, + "learning_rate": 9.997436220472441e-06, + "loss": 0.6325, + "step": 1632 + }, + { + "epoch": 0.5143307086614173, + "grad_norm": 20.3349552154541, + "learning_rate": 9.997434645669292e-06, + "loss": 0.2886, + "step": 1633 + }, + { + "epoch": 0.5146456692913386, + "grad_norm": 46.71043014526367, + "learning_rate": 9.997433070866143e-06, + "loss": 0.9402, + "step": 1634 + }, + { + "epoch": 0.5149606299212598, + "grad_norm": 34.89626693725586, + "learning_rate": 9.997431496062992e-06, + "loss": 0.4485, + "step": 1635 + }, + { + "epoch": 0.5152755905511811, + "grad_norm": 28.397123336791992, + "learning_rate": 9.997429921259843e-06, + "loss": 0.3992, + "step": 1636 + }, + { + "epoch": 0.5155905511811023, + "grad_norm": 50.874427795410156, + "learning_rate": 9.997428346456694e-06, + "loss": 0.9479, + "step": 1637 + }, + { + "epoch": 0.5159055118110236, + "grad_norm": 38.96868896484375, + "learning_rate": 9.997426771653545e-06, + "loss": 0.5054, + "step": 1638 + }, + { + "epoch": 0.5162204724409449, + "grad_norm": 7.598008155822754, + "learning_rate": 9.997425196850395e-06, + "loss": 0.0787, + "step": 1639 + }, + { + "epoch": 0.5165354330708661, + "grad_norm": 24.391128540039062, + "learning_rate": 9.997423622047246e-06, + "loss": 0.1297, + "step": 1640 + }, + { + "epoch": 0.5165354330708661, + "eval_loss": 0.4850241541862488, + "eval_runtime": 337.9062, + "eval_samples_per_second": 0.346, + "eval_steps_per_second": 0.346, + "step": 1640 + }, + { + "epoch": 0.5168503937007874, + "grad_norm": 21.02195167541504, + "learning_rate": 9.997422047244095e-06, + "loss": 0.3174, + "step": 1641 + }, + { + "epoch": 0.5171653543307086, + "grad_norm": 19.174169540405273, + "learning_rate": 9.997420472440946e-06, + "loss": 0.4078, + "step": 1642 + }, + { + "epoch": 0.51748031496063, + "grad_norm": 35.037322998046875, + "learning_rate": 9.997418897637797e-06, + "loss": 0.3795, + "step": 1643 + }, + { + "epoch": 0.5177952755905512, + "grad_norm": 19.232954025268555, + "learning_rate": 9.997417322834646e-06, + "loss": 0.4354, + "step": 1644 + }, + { + "epoch": 0.5181102362204725, + "grad_norm": 41.107784271240234, + "learning_rate": 9.997415748031497e-06, + "loss": 0.1806, + "step": 1645 + }, + { + "epoch": 0.5184251968503937, + "grad_norm": 31.30357551574707, + "learning_rate": 9.997414173228346e-06, + "loss": 0.5762, + "step": 1646 + }, + { + "epoch": 0.518740157480315, + "grad_norm": 27.336881637573242, + "learning_rate": 9.997412598425197e-06, + "loss": 0.5801, + "step": 1647 + }, + { + "epoch": 0.5190551181102362, + "grad_norm": 43.55338668823242, + "learning_rate": 9.997411023622048e-06, + "loss": 0.9297, + "step": 1648 + }, + { + "epoch": 0.5193700787401575, + "grad_norm": 33.737937927246094, + "learning_rate": 9.9974094488189e-06, + "loss": 0.3568, + "step": 1649 + }, + { + "epoch": 0.5196850393700787, + "grad_norm": 24.022850036621094, + "learning_rate": 9.997407874015749e-06, + "loss": 0.1614, + "step": 1650 + }, + { + "epoch": 0.52, + "grad_norm": 27.58238410949707, + "learning_rate": 9.9974062992126e-06, + "loss": 0.3589, + "step": 1651 + }, + { + "epoch": 0.5203149606299212, + "grad_norm": 41.2049446105957, + "learning_rate": 9.997404724409449e-06, + "loss": 0.2244, + "step": 1652 + }, + { + "epoch": 0.5206299212598425, + "grad_norm": 19.164548873901367, + "learning_rate": 9.9974031496063e-06, + "loss": 0.3534, + "step": 1653 + }, + { + "epoch": 0.5209448818897637, + "grad_norm": 24.569271087646484, + "learning_rate": 9.997401574803151e-06, + "loss": 0.3207, + "step": 1654 + }, + { + "epoch": 0.521259842519685, + "grad_norm": 24.17620849609375, + "learning_rate": 9.9974e-06, + "loss": 0.2762, + "step": 1655 + }, + { + "epoch": 0.5215748031496062, + "grad_norm": 45.885250091552734, + "learning_rate": 9.997398425196851e-06, + "loss": 0.1903, + "step": 1656 + }, + { + "epoch": 0.5218897637795276, + "grad_norm": 10.778152465820312, + "learning_rate": 9.997396850393702e-06, + "loss": 0.0588, + "step": 1657 + }, + { + "epoch": 0.5222047244094489, + "grad_norm": 36.7594108581543, + "learning_rate": 9.997395275590553e-06, + "loss": 0.2227, + "step": 1658 + }, + { + "epoch": 0.5225196850393701, + "grad_norm": 13.830334663391113, + "learning_rate": 9.997393700787402e-06, + "loss": 0.2058, + "step": 1659 + }, + { + "epoch": 0.5228346456692914, + "grad_norm": 70.07479095458984, + "learning_rate": 9.997392125984253e-06, + "loss": 0.3415, + "step": 1660 + }, + { + "epoch": 0.5228346456692914, + "eval_loss": 0.4270039498806, + "eval_runtime": 338.5531, + "eval_samples_per_second": 0.346, + "eval_steps_per_second": 0.346, + "step": 1660 + }, + { + "epoch": 0.5231496062992126, + "grad_norm": 54.377689361572266, + "learning_rate": 9.997390551181103e-06, + "loss": 0.699, + "step": 1661 + }, + { + "epoch": 0.5234645669291339, + "grad_norm": 20.332332611083984, + "learning_rate": 9.997388976377954e-06, + "loss": 0.3042, + "step": 1662 + }, + { + "epoch": 0.5237795275590551, + "grad_norm": 29.605314254760742, + "learning_rate": 9.997387401574805e-06, + "loss": 0.616, + "step": 1663 + }, + { + "epoch": 0.5240944881889764, + "grad_norm": 26.555742263793945, + "learning_rate": 9.997385826771654e-06, + "loss": 0.2776, + "step": 1664 + }, + { + "epoch": 0.5244094488188976, + "grad_norm": 24.46210479736328, + "learning_rate": 9.997384251968505e-06, + "loss": 0.3006, + "step": 1665 + }, + { + "epoch": 0.5247244094488189, + "grad_norm": 135.4957275390625, + "learning_rate": 9.997382677165354e-06, + "loss": 0.4272, + "step": 1666 + }, + { + "epoch": 0.5250393700787401, + "grad_norm": 8.375167846679688, + "learning_rate": 9.997381102362205e-06, + "loss": 0.0964, + "step": 1667 + }, + { + "epoch": 0.5253543307086614, + "grad_norm": 17.8012752532959, + "learning_rate": 9.997379527559056e-06, + "loss": 0.1773, + "step": 1668 + }, + { + "epoch": 0.5256692913385826, + "grad_norm": 43.08966064453125, + "learning_rate": 9.997377952755907e-06, + "loss": 0.3974, + "step": 1669 + }, + { + "epoch": 0.525984251968504, + "grad_norm": 34.40397644042969, + "learning_rate": 9.997376377952757e-06, + "loss": 0.3196, + "step": 1670 + }, + { + "epoch": 0.5262992125984252, + "grad_norm": 16.06707000732422, + "learning_rate": 9.997374803149608e-06, + "loss": 0.1063, + "step": 1671 + }, + { + "epoch": 0.5266141732283465, + "grad_norm": 12.523601531982422, + "learning_rate": 9.997373228346457e-06, + "loss": 0.1203, + "step": 1672 + }, + { + "epoch": 0.5269291338582677, + "grad_norm": 43.67927169799805, + "learning_rate": 9.997371653543308e-06, + "loss": 0.285, + "step": 1673 + }, + { + "epoch": 0.527244094488189, + "grad_norm": 58.05824279785156, + "learning_rate": 9.997370078740159e-06, + "loss": 0.2903, + "step": 1674 + }, + { + "epoch": 0.5275590551181102, + "grad_norm": 52.33161163330078, + "learning_rate": 9.997368503937008e-06, + "loss": 0.3971, + "step": 1675 + }, + { + "epoch": 0.5278740157480315, + "grad_norm": 43.415462493896484, + "learning_rate": 9.997366929133859e-06, + "loss": 0.2529, + "step": 1676 + }, + { + "epoch": 0.5281889763779527, + "grad_norm": 37.16794967651367, + "learning_rate": 9.997365354330708e-06, + "loss": 0.6718, + "step": 1677 + }, + { + "epoch": 0.528503937007874, + "grad_norm": 19.802600860595703, + "learning_rate": 9.99736377952756e-06, + "loss": 0.1329, + "step": 1678 + }, + { + "epoch": 0.5288188976377953, + "grad_norm": 18.722341537475586, + "learning_rate": 9.99736220472441e-06, + "loss": 0.0449, + "step": 1679 + }, + { + "epoch": 0.5291338582677165, + "grad_norm": 30.56394386291504, + "learning_rate": 9.997360629921261e-06, + "loss": 0.0851, + "step": 1680 + }, + { + "epoch": 0.5291338582677165, + "eval_loss": 0.5506137013435364, + "eval_runtime": 340.4385, + "eval_samples_per_second": 0.344, + "eval_steps_per_second": 0.344, + "step": 1680 + }, + { + "epoch": 0.5294488188976378, + "grad_norm": 15.041778564453125, + "learning_rate": 9.99735905511811e-06, + "loss": 0.0869, + "step": 1681 + }, + { + "epoch": 0.529763779527559, + "grad_norm": 81.58389282226562, + "learning_rate": 9.997357480314962e-06, + "loss": 0.6074, + "step": 1682 + }, + { + "epoch": 0.5300787401574804, + "grad_norm": 99.90658569335938, + "learning_rate": 9.997355905511813e-06, + "loss": 0.2411, + "step": 1683 + }, + { + "epoch": 0.5303937007874016, + "grad_norm": 25.98404312133789, + "learning_rate": 9.997354330708662e-06, + "loss": 0.1055, + "step": 1684 + }, + { + "epoch": 0.5307086614173229, + "grad_norm": 34.40901184082031, + "learning_rate": 9.997352755905513e-06, + "loss": 0.4056, + "step": 1685 + }, + { + "epoch": 0.5310236220472441, + "grad_norm": 12.796487808227539, + "learning_rate": 9.997351181102362e-06, + "loss": 0.0303, + "step": 1686 + }, + { + "epoch": 0.5313385826771654, + "grad_norm": 14.561634063720703, + "learning_rate": 9.997349606299213e-06, + "loss": 0.0729, + "step": 1687 + }, + { + "epoch": 0.5316535433070866, + "grad_norm": 38.87438201904297, + "learning_rate": 9.997348031496064e-06, + "loss": 0.1667, + "step": 1688 + }, + { + "epoch": 0.5319685039370079, + "grad_norm": 39.482303619384766, + "learning_rate": 9.997346456692915e-06, + "loss": 0.0675, + "step": 1689 + }, + { + "epoch": 0.5322834645669291, + "grad_norm": 162.6251220703125, + "learning_rate": 9.997344881889765e-06, + "loss": 0.5522, + "step": 1690 + }, + { + "epoch": 0.5325984251968504, + "grad_norm": 10.552274703979492, + "learning_rate": 9.997343307086616e-06, + "loss": 0.036, + "step": 1691 + }, + { + "epoch": 0.5329133858267716, + "grad_norm": 118.2337646484375, + "learning_rate": 9.997341732283465e-06, + "loss": 0.1975, + "step": 1692 + }, + { + "epoch": 0.5332283464566929, + "grad_norm": 46.612266540527344, + "learning_rate": 9.997340157480316e-06, + "loss": 0.1823, + "step": 1693 + }, + { + "epoch": 0.5335433070866141, + "grad_norm": 93.66093444824219, + "learning_rate": 9.997338582677167e-06, + "loss": 1.0047, + "step": 1694 + }, + { + "epoch": 0.5338582677165354, + "grad_norm": 42.73204040527344, + "learning_rate": 9.997337007874016e-06, + "loss": 0.1202, + "step": 1695 + }, + { + "epoch": 0.5341732283464566, + "grad_norm": 116.17491912841797, + "learning_rate": 9.997335433070867e-06, + "loss": 0.5415, + "step": 1696 + }, + { + "epoch": 0.534488188976378, + "grad_norm": 38.49724197387695, + "learning_rate": 9.997333858267716e-06, + "loss": 0.7654, + "step": 1697 + }, + { + "epoch": 0.5348031496062993, + "grad_norm": 76.99092864990234, + "learning_rate": 9.997332283464567e-06, + "loss": 0.3025, + "step": 1698 + }, + { + "epoch": 0.5351181102362205, + "grad_norm": 102.11122131347656, + "learning_rate": 9.997330708661418e-06, + "loss": 0.6456, + "step": 1699 + }, + { + "epoch": 0.5354330708661418, + "grad_norm": 140.7232666015625, + "learning_rate": 9.99732913385827e-06, + "loss": 0.4218, + "step": 1700 + }, + { + "epoch": 0.5354330708661418, + "eval_loss": 0.5890030860900879, + "eval_runtime": 338.8108, + "eval_samples_per_second": 0.345, + "eval_steps_per_second": 0.345, + "step": 1700 + }, + { + "epoch": 0.535748031496063, + "grad_norm": 111.65464782714844, + "learning_rate": 9.997327559055119e-06, + "loss": 1.0181, + "step": 1701 + }, + { + "epoch": 0.5360629921259843, + "grad_norm": 37.92324447631836, + "learning_rate": 9.99732598425197e-06, + "loss": 0.7354, + "step": 1702 + }, + { + "epoch": 0.5363779527559055, + "grad_norm": 42.922149658203125, + "learning_rate": 9.997324409448819e-06, + "loss": 0.2447, + "step": 1703 + }, + { + "epoch": 0.5366929133858268, + "grad_norm": 45.183082580566406, + "learning_rate": 9.99732283464567e-06, + "loss": 0.337, + "step": 1704 + }, + { + "epoch": 0.537007874015748, + "grad_norm": 39.247962951660156, + "learning_rate": 9.997321259842521e-06, + "loss": 0.3811, + "step": 1705 + }, + { + "epoch": 0.5373228346456693, + "grad_norm": 128.93247985839844, + "learning_rate": 9.99731968503937e-06, + "loss": 0.5403, + "step": 1706 + }, + { + "epoch": 0.5376377952755905, + "grad_norm": 107.43035888671875, + "learning_rate": 9.997318110236221e-06, + "loss": 0.3911, + "step": 1707 + }, + { + "epoch": 0.5379527559055118, + "grad_norm": 43.16225051879883, + "learning_rate": 9.997316535433072e-06, + "loss": 0.2933, + "step": 1708 + }, + { + "epoch": 0.538267716535433, + "grad_norm": 87.62896728515625, + "learning_rate": 9.997314960629923e-06, + "loss": 0.4866, + "step": 1709 + }, + { + "epoch": 0.5385826771653544, + "grad_norm": 40.5244255065918, + "learning_rate": 9.997313385826772e-06, + "loss": 0.382, + "step": 1710 + }, + { + "epoch": 0.5388976377952756, + "grad_norm": 34.54233932495117, + "learning_rate": 9.997311811023623e-06, + "loss": 0.2301, + "step": 1711 + }, + { + "epoch": 0.5392125984251969, + "grad_norm": 95.98748016357422, + "learning_rate": 9.997310236220473e-06, + "loss": 0.8653, + "step": 1712 + }, + { + "epoch": 0.5395275590551181, + "grad_norm": 86.79911041259766, + "learning_rate": 9.997308661417324e-06, + "loss": 0.4673, + "step": 1713 + }, + { + "epoch": 0.5398425196850394, + "grad_norm": 13.791953086853027, + "learning_rate": 9.997307086614175e-06, + "loss": 0.0652, + "step": 1714 + }, + { + "epoch": 0.5401574803149606, + "grad_norm": 16.453601837158203, + "learning_rate": 9.997305511811024e-06, + "loss": 0.103, + "step": 1715 + }, + { + "epoch": 0.5404724409448819, + "grad_norm": 40.83580780029297, + "learning_rate": 9.997303937007875e-06, + "loss": 0.4618, + "step": 1716 + }, + { + "epoch": 0.5407874015748031, + "grad_norm": 57.49140548706055, + "learning_rate": 9.997302362204724e-06, + "loss": 0.2785, + "step": 1717 + }, + { + "epoch": 0.5411023622047244, + "grad_norm": 56.28849792480469, + "learning_rate": 9.997300787401575e-06, + "loss": 0.6193, + "step": 1718 + }, + { + "epoch": 0.5414173228346457, + "grad_norm": 37.70257568359375, + "learning_rate": 9.997299212598426e-06, + "loss": 0.2423, + "step": 1719 + }, + { + "epoch": 0.5417322834645669, + "grad_norm": 60.47585678100586, + "learning_rate": 9.997297637795277e-06, + "loss": 0.3203, + "step": 1720 + }, + { + "epoch": 0.5417322834645669, + "eval_loss": 0.4833095669746399, + "eval_runtime": 339.1777, + "eval_samples_per_second": 0.345, + "eval_steps_per_second": 0.345, + "step": 1720 + }, + { + "epoch": 0.5420472440944882, + "grad_norm": 45.25422286987305, + "learning_rate": 9.997296062992127e-06, + "loss": 0.4233, + "step": 1721 + }, + { + "epoch": 0.5423622047244094, + "grad_norm": 21.047258377075195, + "learning_rate": 9.997294488188978e-06, + "loss": 0.1222, + "step": 1722 + }, + { + "epoch": 0.5426771653543307, + "grad_norm": 41.04436492919922, + "learning_rate": 9.997292913385827e-06, + "loss": 0.1144, + "step": 1723 + }, + { + "epoch": 0.542992125984252, + "grad_norm": 9.362834930419922, + "learning_rate": 9.997291338582678e-06, + "loss": 0.0498, + "step": 1724 + }, + { + "epoch": 0.5433070866141733, + "grad_norm": 50.78684997558594, + "learning_rate": 9.997289763779529e-06, + "loss": 0.6966, + "step": 1725 + }, + { + "epoch": 0.5436220472440945, + "grad_norm": 33.23752212524414, + "learning_rate": 9.997288188976378e-06, + "loss": 0.1696, + "step": 1726 + }, + { + "epoch": 0.5439370078740158, + "grad_norm": 29.13566780090332, + "learning_rate": 9.99728661417323e-06, + "loss": 0.511, + "step": 1727 + }, + { + "epoch": 0.544251968503937, + "grad_norm": 36.36374282836914, + "learning_rate": 9.99728503937008e-06, + "loss": 0.3232, + "step": 1728 + }, + { + "epoch": 0.5445669291338583, + "grad_norm": 9.268793106079102, + "learning_rate": 9.997283464566931e-06, + "loss": 0.0484, + "step": 1729 + }, + { + "epoch": 0.5448818897637795, + "grad_norm": 10.723762512207031, + "learning_rate": 9.99728188976378e-06, + "loss": 0.0748, + "step": 1730 + }, + { + "epoch": 0.5451968503937008, + "grad_norm": 42.956424713134766, + "learning_rate": 9.997280314960631e-06, + "loss": 0.2678, + "step": 1731 + }, + { + "epoch": 0.545511811023622, + "grad_norm": 33.43272018432617, + "learning_rate": 9.99727874015748e-06, + "loss": 0.4772, + "step": 1732 + }, + { + "epoch": 0.5458267716535433, + "grad_norm": 59.14657974243164, + "learning_rate": 9.997277165354332e-06, + "loss": 0.2235, + "step": 1733 + }, + { + "epoch": 0.5461417322834645, + "grad_norm": 29.12813377380371, + "learning_rate": 9.997275590551183e-06, + "loss": 0.0931, + "step": 1734 + }, + { + "epoch": 0.5464566929133858, + "grad_norm": 17.903165817260742, + "learning_rate": 9.997274015748032e-06, + "loss": 0.4091, + "step": 1735 + }, + { + "epoch": 0.546771653543307, + "grad_norm": 72.72879028320312, + "learning_rate": 9.997272440944883e-06, + "loss": 0.3515, + "step": 1736 + }, + { + "epoch": 0.5470866141732283, + "grad_norm": 42.820655822753906, + "learning_rate": 9.997270866141732e-06, + "loss": 0.8093, + "step": 1737 + }, + { + "epoch": 0.5474015748031497, + "grad_norm": 34.83147048950195, + "learning_rate": 9.997269291338583e-06, + "loss": 0.0899, + "step": 1738 + }, + { + "epoch": 0.5477165354330709, + "grad_norm": 165.32769775390625, + "learning_rate": 9.997267716535434e-06, + "loss": 1.2147, + "step": 1739 + }, + { + "epoch": 0.5480314960629922, + "grad_norm": 54.89258575439453, + "learning_rate": 9.997266141732285e-06, + "loss": 0.3279, + "step": 1740 + }, + { + "epoch": 0.5480314960629922, + "eval_loss": 0.5086050033569336, + "eval_runtime": 340.7446, + "eval_samples_per_second": 0.343, + "eval_steps_per_second": 0.343, + "step": 1740 + }, + { + "epoch": 0.5483464566929134, + "grad_norm": 19.018238067626953, + "learning_rate": 9.997264566929135e-06, + "loss": 0.0628, + "step": 1741 + }, + { + "epoch": 0.5486614173228347, + "grad_norm": 44.377803802490234, + "learning_rate": 9.997262992125986e-06, + "loss": 0.1875, + "step": 1742 + }, + { + "epoch": 0.5489763779527559, + "grad_norm": 51.6882209777832, + "learning_rate": 9.997261417322835e-06, + "loss": 0.5064, + "step": 1743 + }, + { + "epoch": 0.5492913385826772, + "grad_norm": 6.151045799255371, + "learning_rate": 9.997259842519686e-06, + "loss": 0.0298, + "step": 1744 + }, + { + "epoch": 0.5496062992125984, + "grad_norm": 76.562255859375, + "learning_rate": 9.997258267716537e-06, + "loss": 0.6654, + "step": 1745 + }, + { + "epoch": 0.5499212598425197, + "grad_norm": 48.19951629638672, + "learning_rate": 9.997256692913386e-06, + "loss": 0.4798, + "step": 1746 + }, + { + "epoch": 0.5502362204724409, + "grad_norm": 68.81988525390625, + "learning_rate": 9.997255118110237e-06, + "loss": 0.4747, + "step": 1747 + }, + { + "epoch": 0.5505511811023622, + "grad_norm": 24.68579864501953, + "learning_rate": 9.997253543307086e-06, + "loss": 0.1881, + "step": 1748 + }, + { + "epoch": 0.5508661417322834, + "grad_norm": 21.575008392333984, + "learning_rate": 9.997251968503937e-06, + "loss": 0.3208, + "step": 1749 + }, + { + "epoch": 0.5511811023622047, + "grad_norm": 45.560203552246094, + "learning_rate": 9.997250393700788e-06, + "loss": 0.7353, + "step": 1750 + }, + { + "epoch": 0.551496062992126, + "grad_norm": 11.304056167602539, + "learning_rate": 9.99724881889764e-06, + "loss": 0.0611, + "step": 1751 + }, + { + "epoch": 0.5518110236220473, + "grad_norm": 53.74604797363281, + "learning_rate": 9.997247244094489e-06, + "loss": 0.3013, + "step": 1752 + }, + { + "epoch": 0.5521259842519685, + "grad_norm": 40.34061050415039, + "learning_rate": 9.99724566929134e-06, + "loss": 0.9412, + "step": 1753 + }, + { + "epoch": 0.5524409448818898, + "grad_norm": 38.01073455810547, + "learning_rate": 9.99724409448819e-06, + "loss": 0.1551, + "step": 1754 + }, + { + "epoch": 0.552755905511811, + "grad_norm": 64.29601287841797, + "learning_rate": 9.99724251968504e-06, + "loss": 1.8098, + "step": 1755 + }, + { + "epoch": 0.5530708661417323, + "grad_norm": 50.7859001159668, + "learning_rate": 9.997240944881891e-06, + "loss": 0.3809, + "step": 1756 + }, + { + "epoch": 0.5533858267716535, + "grad_norm": 23.059926986694336, + "learning_rate": 9.99723937007874e-06, + "loss": 0.1878, + "step": 1757 + }, + { + "epoch": 0.5537007874015748, + "grad_norm": 15.40706729888916, + "learning_rate": 9.997237795275591e-06, + "loss": 0.2192, + "step": 1758 + }, + { + "epoch": 0.5540157480314961, + "grad_norm": 50.74030685424805, + "learning_rate": 9.997236220472442e-06, + "loss": 0.8022, + "step": 1759 + }, + { + "epoch": 0.5543307086614173, + "grad_norm": 49.88266372680664, + "learning_rate": 9.997234645669293e-06, + "loss": 0.3615, + "step": 1760 + }, + { + "epoch": 0.5543307086614173, + "eval_loss": 0.4197517931461334, + "eval_runtime": 338.0619, + "eval_samples_per_second": 0.346, + "eval_steps_per_second": 0.346, + "step": 1760 + }, + { + "epoch": 0.5546456692913386, + "grad_norm": 33.29900360107422, + "learning_rate": 9.997233070866143e-06, + "loss": 0.4475, + "step": 1761 + }, + { + "epoch": 0.5549606299212598, + "grad_norm": 57.282649993896484, + "learning_rate": 9.997231496062994e-06, + "loss": 0.5688, + "step": 1762 + }, + { + "epoch": 0.5552755905511811, + "grad_norm": 16.435340881347656, + "learning_rate": 9.997229921259843e-06, + "loss": 0.2603, + "step": 1763 + }, + { + "epoch": 0.5555905511811023, + "grad_norm": 45.144737243652344, + "learning_rate": 9.997228346456694e-06, + "loss": 0.6854, + "step": 1764 + }, + { + "epoch": 0.5559055118110237, + "grad_norm": 42.861515045166016, + "learning_rate": 9.997226771653545e-06, + "loss": 0.6296, + "step": 1765 + }, + { + "epoch": 0.5562204724409449, + "grad_norm": 48.16584777832031, + "learning_rate": 9.997225196850394e-06, + "loss": 0.4873, + "step": 1766 + }, + { + "epoch": 0.5565354330708662, + "grad_norm": 61.42652893066406, + "learning_rate": 9.997223622047245e-06, + "loss": 0.7604, + "step": 1767 + }, + { + "epoch": 0.5568503937007874, + "grad_norm": 24.190889358520508, + "learning_rate": 9.997222047244094e-06, + "loss": 0.3546, + "step": 1768 + }, + { + "epoch": 0.5571653543307087, + "grad_norm": 19.951372146606445, + "learning_rate": 9.997220472440945e-06, + "loss": 0.4902, + "step": 1769 + }, + { + "epoch": 0.5574803149606299, + "grad_norm": 24.672101974487305, + "learning_rate": 9.997218897637796e-06, + "loss": 0.3603, + "step": 1770 + }, + { + "epoch": 0.5577952755905512, + "grad_norm": 16.915287017822266, + "learning_rate": 9.997217322834647e-06, + "loss": 0.1767, + "step": 1771 + }, + { + "epoch": 0.5581102362204724, + "grad_norm": 28.743501663208008, + "learning_rate": 9.997215748031497e-06, + "loss": 0.3357, + "step": 1772 + }, + { + "epoch": 0.5584251968503937, + "grad_norm": 48.64558029174805, + "learning_rate": 9.997214173228346e-06, + "loss": 0.3531, + "step": 1773 + }, + { + "epoch": 0.5587401574803149, + "grad_norm": 18.73015022277832, + "learning_rate": 9.997212598425197e-06, + "loss": 0.2041, + "step": 1774 + }, + { + "epoch": 0.5590551181102362, + "grad_norm": 27.524059295654297, + "learning_rate": 9.997211023622048e-06, + "loss": 0.2056, + "step": 1775 + }, + { + "epoch": 0.5593700787401574, + "grad_norm": 15.509268760681152, + "learning_rate": 9.997209448818899e-06, + "loss": 0.1131, + "step": 1776 + }, + { + "epoch": 0.5596850393700787, + "grad_norm": 39.2099609375, + "learning_rate": 9.997207874015748e-06, + "loss": 0.2837, + "step": 1777 + }, + { + "epoch": 0.56, + "grad_norm": 41.994815826416016, + "learning_rate": 9.9972062992126e-06, + "loss": 0.3212, + "step": 1778 + }, + { + "epoch": 0.5603149606299213, + "grad_norm": 50.88380432128906, + "learning_rate": 9.99720472440945e-06, + "loss": 0.476, + "step": 1779 + }, + { + "epoch": 0.5606299212598426, + "grad_norm": 23.539987564086914, + "learning_rate": 9.997203149606301e-06, + "loss": 0.1563, + "step": 1780 + }, + { + "epoch": 0.5606299212598426, + "eval_loss": 0.5017465353012085, + "eval_runtime": 339.1116, + "eval_samples_per_second": 0.345, + "eval_steps_per_second": 0.345, + "step": 1780 + }, + { + "epoch": 0.5609448818897638, + "grad_norm": 48.950191497802734, + "learning_rate": 9.99720157480315e-06, + "loss": 0.7028, + "step": 1781 + }, + { + "epoch": 0.5612598425196851, + "grad_norm": 15.981232643127441, + "learning_rate": 9.997200000000001e-06, + "loss": 0.0955, + "step": 1782 + }, + { + "epoch": 0.5615748031496063, + "grad_norm": 53.745574951171875, + "learning_rate": 9.99719842519685e-06, + "loss": 0.5677, + "step": 1783 + }, + { + "epoch": 0.5618897637795276, + "grad_norm": 57.129024505615234, + "learning_rate": 9.997196850393702e-06, + "loss": 0.5015, + "step": 1784 + }, + { + "epoch": 0.5622047244094488, + "grad_norm": 25.303375244140625, + "learning_rate": 9.997195275590553e-06, + "loss": 0.3141, + "step": 1785 + }, + { + "epoch": 0.5625196850393701, + "grad_norm": 29.012252807617188, + "learning_rate": 9.997193700787402e-06, + "loss": 0.3536, + "step": 1786 + }, + { + "epoch": 0.5628346456692913, + "grad_norm": 10.699189186096191, + "learning_rate": 9.997192125984253e-06, + "loss": 0.0627, + "step": 1787 + }, + { + "epoch": 0.5631496062992126, + "grad_norm": 8.34593391418457, + "learning_rate": 9.997190551181102e-06, + "loss": 0.0391, + "step": 1788 + }, + { + "epoch": 0.5634645669291338, + "grad_norm": 42.121437072753906, + "learning_rate": 9.997188976377953e-06, + "loss": 0.8368, + "step": 1789 + }, + { + "epoch": 0.5637795275590551, + "grad_norm": 35.72675704956055, + "learning_rate": 9.997187401574804e-06, + "loss": 0.5532, + "step": 1790 + }, + { + "epoch": 0.5640944881889763, + "grad_norm": 45.88935470581055, + "learning_rate": 9.997185826771655e-06, + "loss": 0.1764, + "step": 1791 + }, + { + "epoch": 0.5644094488188977, + "grad_norm": 32.38536071777344, + "learning_rate": 9.997184251968505e-06, + "loss": 0.1779, + "step": 1792 + }, + { + "epoch": 0.5647244094488189, + "grad_norm": 72.07355499267578, + "learning_rate": 9.997182677165354e-06, + "loss": 0.3102, + "step": 1793 + }, + { + "epoch": 0.5650393700787402, + "grad_norm": 82.99500274658203, + "learning_rate": 9.997181102362205e-06, + "loss": 0.6726, + "step": 1794 + }, + { + "epoch": 0.5653543307086614, + "grad_norm": 80.94422149658203, + "learning_rate": 9.997179527559056e-06, + "loss": 0.7582, + "step": 1795 + }, + { + "epoch": 0.5656692913385827, + "grad_norm": 4.303206443786621, + "learning_rate": 9.997177952755907e-06, + "loss": 0.0169, + "step": 1796 + }, + { + "epoch": 0.5659842519685039, + "grad_norm": 38.44367980957031, + "learning_rate": 9.997176377952756e-06, + "loss": 0.5267, + "step": 1797 + }, + { + "epoch": 0.5662992125984252, + "grad_norm": 48.196807861328125, + "learning_rate": 9.997174803149607e-06, + "loss": 0.6009, + "step": 1798 + }, + { + "epoch": 0.5666141732283465, + "grad_norm": 33.66616439819336, + "learning_rate": 9.997173228346458e-06, + "loss": 0.5003, + "step": 1799 + }, + { + "epoch": 0.5669291338582677, + "grad_norm": 34.213218688964844, + "learning_rate": 9.997171653543309e-06, + "loss": 0.2974, + "step": 1800 + }, + { + "epoch": 0.5669291338582677, + "eval_loss": 0.477321982383728, + "eval_runtime": 338.9113, + "eval_samples_per_second": 0.345, + "eval_steps_per_second": 0.345, + "step": 1800 + }, + { + "epoch": 0.567244094488189, + "grad_norm": 63.08566665649414, + "learning_rate": 9.997170078740158e-06, + "loss": 0.1859, + "step": 1801 + }, + { + "epoch": 0.5675590551181102, + "grad_norm": 61.78851318359375, + "learning_rate": 9.99716850393701e-06, + "loss": 0.9895, + "step": 1802 + }, + { + "epoch": 0.5678740157480315, + "grad_norm": 47.29201126098633, + "learning_rate": 9.997166929133859e-06, + "loss": 0.5153, + "step": 1803 + }, + { + "epoch": 0.5681889763779527, + "grad_norm": 33.21207046508789, + "learning_rate": 9.99716535433071e-06, + "loss": 0.3221, + "step": 1804 + }, + { + "epoch": 0.568503937007874, + "grad_norm": 20.20362663269043, + "learning_rate": 9.99716377952756e-06, + "loss": 0.0697, + "step": 1805 + }, + { + "epoch": 0.5688188976377953, + "grad_norm": 23.14605712890625, + "learning_rate": 9.99716220472441e-06, + "loss": 0.3096, + "step": 1806 + }, + { + "epoch": 0.5691338582677166, + "grad_norm": 58.990699768066406, + "learning_rate": 9.997160629921261e-06, + "loss": 0.6046, + "step": 1807 + }, + { + "epoch": 0.5694488188976378, + "grad_norm": 36.48517608642578, + "learning_rate": 9.99715905511811e-06, + "loss": 0.2886, + "step": 1808 + }, + { + "epoch": 0.5697637795275591, + "grad_norm": 56.9811897277832, + "learning_rate": 9.997157480314961e-06, + "loss": 0.4658, + "step": 1809 + }, + { + "epoch": 0.5700787401574803, + "grad_norm": 40.79648208618164, + "learning_rate": 9.997155905511812e-06, + "loss": 0.3177, + "step": 1810 + }, + { + "epoch": 0.5703937007874016, + "grad_norm": 31.444948196411133, + "learning_rate": 9.997154330708663e-06, + "loss": 0.3236, + "step": 1811 + }, + { + "epoch": 0.5707086614173228, + "grad_norm": 21.288089752197266, + "learning_rate": 9.997152755905513e-06, + "loss": 0.4159, + "step": 1812 + }, + { + "epoch": 0.5710236220472441, + "grad_norm": 58.0438117980957, + "learning_rate": 9.997151181102362e-06, + "loss": 0.6176, + "step": 1813 + }, + { + "epoch": 0.5713385826771653, + "grad_norm": 19.920330047607422, + "learning_rate": 9.997149606299213e-06, + "loss": 0.305, + "step": 1814 + }, + { + "epoch": 0.5716535433070866, + "grad_norm": 35.30345916748047, + "learning_rate": 9.997148031496064e-06, + "loss": 0.4661, + "step": 1815 + }, + { + "epoch": 0.5719685039370078, + "grad_norm": 16.50104522705078, + "learning_rate": 9.997146456692915e-06, + "loss": 0.2696, + "step": 1816 + }, + { + "epoch": 0.5722834645669291, + "grad_norm": 40.40153121948242, + "learning_rate": 9.997144881889764e-06, + "loss": 0.5205, + "step": 1817 + }, + { + "epoch": 0.5725984251968504, + "grad_norm": 12.835118293762207, + "learning_rate": 9.997143307086615e-06, + "loss": 0.1722, + "step": 1818 + }, + { + "epoch": 0.5729133858267716, + "grad_norm": 40.450950622558594, + "learning_rate": 9.997141732283464e-06, + "loss": 0.3829, + "step": 1819 + }, + { + "epoch": 0.573228346456693, + "grad_norm": 20.867347717285156, + "learning_rate": 9.997140157480315e-06, + "loss": 0.4121, + "step": 1820 + }, + { + "epoch": 0.573228346456693, + "eval_loss": 0.40022820234298706, + "eval_runtime": 337.3982, + "eval_samples_per_second": 0.347, + "eval_steps_per_second": 0.347, + "step": 1820 + }, + { + "epoch": 0.5735433070866142, + "grad_norm": 27.65648651123047, + "learning_rate": 9.997138582677166e-06, + "loss": 0.1822, + "step": 1821 + }, + { + "epoch": 0.5738582677165355, + "grad_norm": 57.512733459472656, + "learning_rate": 9.997137007874017e-06, + "loss": 0.8702, + "step": 1822 + }, + { + "epoch": 0.5741732283464567, + "grad_norm": 13.143925666809082, + "learning_rate": 9.997135433070867e-06, + "loss": 0.183, + "step": 1823 + }, + { + "epoch": 0.574488188976378, + "grad_norm": 49.15605926513672, + "learning_rate": 9.997133858267718e-06, + "loss": 0.3416, + "step": 1824 + }, + { + "epoch": 0.5748031496062992, + "grad_norm": 12.571438789367676, + "learning_rate": 9.997132283464569e-06, + "loss": 0.1638, + "step": 1825 + }, + { + "epoch": 0.5751181102362205, + "grad_norm": 66.83301544189453, + "learning_rate": 9.997130708661418e-06, + "loss": 0.6822, + "step": 1826 + }, + { + "epoch": 0.5754330708661417, + "grad_norm": 22.997888565063477, + "learning_rate": 9.997129133858269e-06, + "loss": 0.2096, + "step": 1827 + }, + { + "epoch": 0.575748031496063, + "grad_norm": 13.75336742401123, + "learning_rate": 9.997127559055118e-06, + "loss": 0.0664, + "step": 1828 + }, + { + "epoch": 0.5760629921259842, + "grad_norm": 60.55500411987305, + "learning_rate": 9.99712598425197e-06, + "loss": 0.8159, + "step": 1829 + }, + { + "epoch": 0.5763779527559055, + "grad_norm": 35.01063537597656, + "learning_rate": 9.99712440944882e-06, + "loss": 0.3943, + "step": 1830 + }, + { + "epoch": 0.5766929133858267, + "grad_norm": 31.06571388244629, + "learning_rate": 9.997122834645671e-06, + "loss": 0.4897, + "step": 1831 + }, + { + "epoch": 0.577007874015748, + "grad_norm": 78.39657592773438, + "learning_rate": 9.99712125984252e-06, + "loss": 0.5343, + "step": 1832 + }, + { + "epoch": 0.5773228346456692, + "grad_norm": 46.089942932128906, + "learning_rate": 9.99711968503937e-06, + "loss": 0.5686, + "step": 1833 + }, + { + "epoch": 0.5776377952755906, + "grad_norm": 37.687374114990234, + "learning_rate": 9.99711811023622e-06, + "loss": 0.6468, + "step": 1834 + }, + { + "epoch": 0.5779527559055118, + "grad_norm": 22.699844360351562, + "learning_rate": 9.997116535433072e-06, + "loss": 0.3065, + "step": 1835 + }, + { + "epoch": 0.5782677165354331, + "grad_norm": 89.35218048095703, + "learning_rate": 9.997114960629923e-06, + "loss": 0.4874, + "step": 1836 + }, + { + "epoch": 0.5785826771653543, + "grad_norm": 95.06536102294922, + "learning_rate": 9.997113385826772e-06, + "loss": 0.4179, + "step": 1837 + }, + { + "epoch": 0.5788976377952756, + "grad_norm": 44.81085205078125, + "learning_rate": 9.997111811023623e-06, + "loss": 0.3927, + "step": 1838 + }, + { + "epoch": 0.5792125984251969, + "grad_norm": 49.45285415649414, + "learning_rate": 9.997110236220472e-06, + "loss": 0.5742, + "step": 1839 + }, + { + "epoch": 0.5795275590551181, + "grad_norm": 35.893402099609375, + "learning_rate": 9.997108661417323e-06, + "loss": 0.4282, + "step": 1840 + }, + { + "epoch": 0.5795275590551181, + "eval_loss": 0.43516087532043457, + "eval_runtime": 337.9832, + "eval_samples_per_second": 0.346, + "eval_steps_per_second": 0.346, + "step": 1840 + }, + { + "epoch": 0.5798425196850394, + "grad_norm": 42.341896057128906, + "learning_rate": 9.997107086614174e-06, + "loss": 0.2899, + "step": 1841 + }, + { + "epoch": 0.5801574803149606, + "grad_norm": 28.99087142944336, + "learning_rate": 9.997105511811025e-06, + "loss": 0.1677, + "step": 1842 + }, + { + "epoch": 0.5804724409448819, + "grad_norm": 13.114380836486816, + "learning_rate": 9.997103937007875e-06, + "loss": 0.063, + "step": 1843 + }, + { + "epoch": 0.5807874015748031, + "grad_norm": 28.455913543701172, + "learning_rate": 9.997102362204724e-06, + "loss": 0.3202, + "step": 1844 + }, + { + "epoch": 0.5811023622047244, + "grad_norm": 65.52840423583984, + "learning_rate": 9.997100787401577e-06, + "loss": 0.5528, + "step": 1845 + }, + { + "epoch": 0.5814173228346456, + "grad_norm": 14.398350715637207, + "learning_rate": 9.997099212598426e-06, + "loss": 0.0874, + "step": 1846 + }, + { + "epoch": 0.581732283464567, + "grad_norm": 46.19868850708008, + "learning_rate": 9.997097637795277e-06, + "loss": 0.2662, + "step": 1847 + }, + { + "epoch": 0.5820472440944882, + "grad_norm": 13.150195121765137, + "learning_rate": 9.997096062992126e-06, + "loss": 0.1102, + "step": 1848 + }, + { + "epoch": 0.5823622047244095, + "grad_norm": 19.446687698364258, + "learning_rate": 9.997094488188977e-06, + "loss": 0.1399, + "step": 1849 + }, + { + "epoch": 0.5826771653543307, + "grad_norm": 112.38624572753906, + "learning_rate": 9.997092913385828e-06, + "loss": 0.3521, + "step": 1850 + }, + { + "epoch": 0.582992125984252, + "grad_norm": 14.854012489318848, + "learning_rate": 9.997091338582679e-06, + "loss": 0.1212, + "step": 1851 + }, + { + "epoch": 0.5833070866141732, + "grad_norm": 58.18556213378906, + "learning_rate": 9.997089763779528e-06, + "loss": 0.4609, + "step": 1852 + }, + { + "epoch": 0.5836220472440945, + "grad_norm": 8.105414390563965, + "learning_rate": 9.997088188976378e-06, + "loss": 0.047, + "step": 1853 + }, + { + "epoch": 0.5839370078740157, + "grad_norm": 78.19012451171875, + "learning_rate": 9.997086614173229e-06, + "loss": 0.3938, + "step": 1854 + }, + { + "epoch": 0.584251968503937, + "grad_norm": 60.095645904541016, + "learning_rate": 9.99708503937008e-06, + "loss": 1.0973, + "step": 1855 + }, + { + "epoch": 0.5845669291338582, + "grad_norm": 27.53265953063965, + "learning_rate": 9.99708346456693e-06, + "loss": 0.3794, + "step": 1856 + }, + { + "epoch": 0.5848818897637795, + "grad_norm": 90.6170654296875, + "learning_rate": 9.99708188976378e-06, + "loss": 0.7531, + "step": 1857 + }, + { + "epoch": 0.5851968503937008, + "grad_norm": 82.42227935791016, + "learning_rate": 9.997080314960631e-06, + "loss": 0.8938, + "step": 1858 + }, + { + "epoch": 0.585511811023622, + "grad_norm": 37.9282112121582, + "learning_rate": 9.99707874015748e-06, + "loss": 0.2363, + "step": 1859 + }, + { + "epoch": 0.5858267716535434, + "grad_norm": 17.56612777709961, + "learning_rate": 9.997077165354331e-06, + "loss": 0.3155, + "step": 1860 + }, + { + "epoch": 0.5858267716535434, + "eval_loss": 0.5003868937492371, + "eval_runtime": 368.9458, + "eval_samples_per_second": 0.317, + "eval_steps_per_second": 0.317, + "step": 1860 + }, + { + "epoch": 0.5861417322834646, + "grad_norm": 40.288368225097656, + "learning_rate": 9.997075590551182e-06, + "loss": 0.3474, + "step": 1861 + }, + { + "epoch": 0.5864566929133859, + "grad_norm": 29.836639404296875, + "learning_rate": 9.997074015748033e-06, + "loss": 0.3913, + "step": 1862 + }, + { + "epoch": 0.5867716535433071, + "grad_norm": 32.93700408935547, + "learning_rate": 9.997072440944883e-06, + "loss": 0.2077, + "step": 1863 + }, + { + "epoch": 0.5870866141732284, + "grad_norm": 44.143409729003906, + "learning_rate": 9.997070866141732e-06, + "loss": 0.359, + "step": 1864 + }, + { + "epoch": 0.5874015748031496, + "grad_norm": 18.514257431030273, + "learning_rate": 9.997069291338583e-06, + "loss": 0.2217, + "step": 1865 + }, + { + "epoch": 0.5877165354330709, + "grad_norm": 72.15349578857422, + "learning_rate": 9.997067716535434e-06, + "loss": 0.6858, + "step": 1866 + }, + { + "epoch": 0.5880314960629921, + "grad_norm": 53.466983795166016, + "learning_rate": 9.997066141732285e-06, + "loss": 0.4746, + "step": 1867 + }, + { + "epoch": 0.5883464566929134, + "grad_norm": 22.331130981445312, + "learning_rate": 9.997064566929134e-06, + "loss": 0.2087, + "step": 1868 + }, + { + "epoch": 0.5886614173228346, + "grad_norm": 49.12759017944336, + "learning_rate": 9.997062992125985e-06, + "loss": 0.8626, + "step": 1869 + }, + { + "epoch": 0.5889763779527559, + "grad_norm": 68.28533935546875, + "learning_rate": 9.997061417322836e-06, + "loss": 0.9698, + "step": 1870 + }, + { + "epoch": 0.5892913385826771, + "grad_norm": 61.11606979370117, + "learning_rate": 9.997059842519687e-06, + "loss": 0.6654, + "step": 1871 + }, + { + "epoch": 0.5896062992125984, + "grad_norm": 39.162593841552734, + "learning_rate": 9.997058267716536e-06, + "loss": 0.3398, + "step": 1872 + }, + { + "epoch": 0.5899212598425196, + "grad_norm": 34.68936538696289, + "learning_rate": 9.997056692913386e-06, + "loss": 0.4498, + "step": 1873 + }, + { + "epoch": 0.590236220472441, + "grad_norm": 31.508249282836914, + "learning_rate": 9.997055118110237e-06, + "loss": 0.5166, + "step": 1874 + }, + { + "epoch": 0.5905511811023622, + "grad_norm": 26.828365325927734, + "learning_rate": 9.997053543307088e-06, + "loss": 0.3993, + "step": 1875 + }, + { + "epoch": 0.5908661417322835, + "grad_norm": 31.714258193969727, + "learning_rate": 9.997051968503939e-06, + "loss": 0.4829, + "step": 1876 + }, + { + "epoch": 0.5911811023622047, + "grad_norm": 13.222908020019531, + "learning_rate": 9.997050393700788e-06, + "loss": 0.2407, + "step": 1877 + }, + { + "epoch": 0.591496062992126, + "grad_norm": 18.35155487060547, + "learning_rate": 9.997048818897639e-06, + "loss": 0.4279, + "step": 1878 + }, + { + "epoch": 0.5918110236220473, + "grad_norm": 11.358118057250977, + "learning_rate": 9.997047244094488e-06, + "loss": 0.208, + "step": 1879 + }, + { + "epoch": 0.5921259842519685, + "grad_norm": 26.622526168823242, + "learning_rate": 9.99704566929134e-06, + "loss": 0.4925, + "step": 1880 + }, + { + "epoch": 0.5921259842519685, + "eval_loss": 0.5159415602684021, + "eval_runtime": 351.9211, + "eval_samples_per_second": 0.332, + "eval_steps_per_second": 0.332, + "step": 1880 + }, + { + "epoch": 0.5924409448818898, + "grad_norm": 32.826812744140625, + "learning_rate": 9.99704409448819e-06, + "loss": 0.6125, + "step": 1881 + }, + { + "epoch": 0.592755905511811, + "grad_norm": 55.511314392089844, + "learning_rate": 9.997042519685041e-06, + "loss": 0.3093, + "step": 1882 + }, + { + "epoch": 0.5930708661417323, + "grad_norm": 35.73579788208008, + "learning_rate": 9.99704094488189e-06, + "loss": 0.7267, + "step": 1883 + }, + { + "epoch": 0.5933858267716535, + "grad_norm": 62.44449234008789, + "learning_rate": 9.99703937007874e-06, + "loss": 0.6222, + "step": 1884 + }, + { + "epoch": 0.5937007874015748, + "grad_norm": 33.10036087036133, + "learning_rate": 9.99703779527559e-06, + "loss": 0.9203, + "step": 1885 + }, + { + "epoch": 0.594015748031496, + "grad_norm": 12.248406410217285, + "learning_rate": 9.997036220472442e-06, + "loss": 0.1355, + "step": 1886 + }, + { + "epoch": 0.5943307086614174, + "grad_norm": 30.380659103393555, + "learning_rate": 9.997034645669293e-06, + "loss": 0.4853, + "step": 1887 + }, + { + "epoch": 0.5946456692913386, + "grad_norm": 32.66392135620117, + "learning_rate": 9.997033070866142e-06, + "loss": 0.6233, + "step": 1888 + }, + { + "epoch": 0.5949606299212599, + "grad_norm": 50.096702575683594, + "learning_rate": 9.997031496062993e-06, + "loss": 0.4143, + "step": 1889 + }, + { + "epoch": 0.5952755905511811, + "grad_norm": 10.48125171661377, + "learning_rate": 9.997029921259842e-06, + "loss": 0.1208, + "step": 1890 + }, + { + "epoch": 0.5955905511811024, + "grad_norm": 39.09800720214844, + "learning_rate": 9.997028346456693e-06, + "loss": 0.2144, + "step": 1891 + }, + { + "epoch": 0.5959055118110236, + "grad_norm": 18.518939971923828, + "learning_rate": 9.997026771653544e-06, + "loss": 0.3954, + "step": 1892 + }, + { + "epoch": 0.5962204724409449, + "grad_norm": 28.824628829956055, + "learning_rate": 9.997025196850394e-06, + "loss": 0.5161, + "step": 1893 + }, + { + "epoch": 0.5965354330708661, + "grad_norm": 36.205902099609375, + "learning_rate": 9.997023622047245e-06, + "loss": 0.3268, + "step": 1894 + }, + { + "epoch": 0.5968503937007874, + "grad_norm": 44.05073165893555, + "learning_rate": 9.997022047244096e-06, + "loss": 0.6844, + "step": 1895 + }, + { + "epoch": 0.5971653543307086, + "grad_norm": 32.883384704589844, + "learning_rate": 9.997020472440947e-06, + "loss": 0.2967, + "step": 1896 + }, + { + "epoch": 0.5974803149606299, + "grad_norm": 41.35356140136719, + "learning_rate": 9.997018897637796e-06, + "loss": 0.8327, + "step": 1897 + }, + { + "epoch": 0.5977952755905512, + "grad_norm": 30.476848602294922, + "learning_rate": 9.997017322834647e-06, + "loss": 0.384, + "step": 1898 + }, + { + "epoch": 0.5981102362204724, + "grad_norm": 20.232952117919922, + "learning_rate": 9.997015748031496e-06, + "loss": 0.4447, + "step": 1899 + }, + { + "epoch": 0.5984251968503937, + "grad_norm": 5.822659492492676, + "learning_rate": 9.997014173228347e-06, + "loss": 0.0435, + "step": 1900 + }, + { + "epoch": 0.5984251968503937, + "eval_loss": 0.49179500341415405, + "eval_runtime": 348.9961, + "eval_samples_per_second": 0.335, + "eval_steps_per_second": 0.335, + "step": 1900 + }, + { + "epoch": 0.598740157480315, + "grad_norm": 18.793500900268555, + "learning_rate": 9.997012598425198e-06, + "loss": 0.2095, + "step": 1901 + }, + { + "epoch": 0.5990551181102363, + "grad_norm": 17.714536666870117, + "learning_rate": 9.99701102362205e-06, + "loss": 0.222, + "step": 1902 + }, + { + "epoch": 0.5993700787401575, + "grad_norm": 35.780208587646484, + "learning_rate": 9.997009448818898e-06, + "loss": 0.2701, + "step": 1903 + }, + { + "epoch": 0.5996850393700788, + "grad_norm": 53.45842361450195, + "learning_rate": 9.997007874015748e-06, + "loss": 0.8297, + "step": 1904 + }, + { + "epoch": 0.6, + "grad_norm": 40.31769561767578, + "learning_rate": 9.997006299212599e-06, + "loss": 0.3894, + "step": 1905 + }, + { + "epoch": 0.6003149606299213, + "grad_norm": 24.99709129333496, + "learning_rate": 9.99700472440945e-06, + "loss": 0.3682, + "step": 1906 + }, + { + "epoch": 0.6006299212598425, + "grad_norm": 52.774383544921875, + "learning_rate": 9.9970031496063e-06, + "loss": 0.3242, + "step": 1907 + }, + { + "epoch": 0.6009448818897638, + "grad_norm": 19.92904281616211, + "learning_rate": 9.99700157480315e-06, + "loss": 0.2621, + "step": 1908 + }, + { + "epoch": 0.601259842519685, + "grad_norm": 38.63066864013672, + "learning_rate": 9.997000000000001e-06, + "loss": 0.2645, + "step": 1909 + }, + { + "epoch": 0.6015748031496063, + "grad_norm": 20.248315811157227, + "learning_rate": 9.99699842519685e-06, + "loss": 0.0956, + "step": 1910 + }, + { + "epoch": 0.6018897637795275, + "grad_norm": 162.4951934814453, + "learning_rate": 9.996996850393701e-06, + "loss": 0.4668, + "step": 1911 + }, + { + "epoch": 0.6022047244094488, + "grad_norm": 11.444823265075684, + "learning_rate": 9.996995275590552e-06, + "loss": 0.0933, + "step": 1912 + }, + { + "epoch": 0.60251968503937, + "grad_norm": 16.599746704101562, + "learning_rate": 9.996993700787403e-06, + "loss": 0.1043, + "step": 1913 + }, + { + "epoch": 0.6028346456692913, + "grad_norm": 66.4779281616211, + "learning_rate": 9.996992125984253e-06, + "loss": 0.818, + "step": 1914 + }, + { + "epoch": 0.6031496062992125, + "grad_norm": 26.693084716796875, + "learning_rate": 9.996990551181104e-06, + "loss": 0.572, + "step": 1915 + }, + { + "epoch": 0.6034645669291339, + "grad_norm": 71.67481994628906, + "learning_rate": 9.996988976377955e-06, + "loss": 0.1498, + "step": 1916 + }, + { + "epoch": 0.6037795275590552, + "grad_norm": 32.80756378173828, + "learning_rate": 9.996987401574804e-06, + "loss": 0.2654, + "step": 1917 + }, + { + "epoch": 0.6040944881889764, + "grad_norm": 73.54534149169922, + "learning_rate": 9.996985826771655e-06, + "loss": 0.8395, + "step": 1918 + }, + { + "epoch": 0.6044094488188977, + "grad_norm": 124.98379516601562, + "learning_rate": 9.996984251968504e-06, + "loss": 0.4122, + "step": 1919 + }, + { + "epoch": 0.6047244094488189, + "grad_norm": 58.07841873168945, + "learning_rate": 9.996982677165355e-06, + "loss": 0.4811, + "step": 1920 + }, + { + "epoch": 0.6047244094488189, + "eval_loss": 0.490536630153656, + "eval_runtime": 351.4081, + "eval_samples_per_second": 0.333, + "eval_steps_per_second": 0.333, + "step": 1920 + }, + { + "epoch": 0.6050393700787402, + "grad_norm": 47.089378356933594, + "learning_rate": 9.996981102362206e-06, + "loss": 0.5479, + "step": 1921 + }, + { + "epoch": 0.6053543307086614, + "grad_norm": 40.61234664916992, + "learning_rate": 9.996979527559057e-06, + "loss": 0.1702, + "step": 1922 + }, + { + "epoch": 0.6056692913385827, + "grad_norm": 76.10828399658203, + "learning_rate": 9.996977952755906e-06, + "loss": 1.0415, + "step": 1923 + }, + { + "epoch": 0.6059842519685039, + "grad_norm": 8.746352195739746, + "learning_rate": 9.996976377952756e-06, + "loss": 0.0595, + "step": 1924 + }, + { + "epoch": 0.6062992125984252, + "grad_norm": 49.07436752319336, + "learning_rate": 9.996974803149607e-06, + "loss": 0.8017, + "step": 1925 + }, + { + "epoch": 0.6066141732283464, + "grad_norm": 37.18422317504883, + "learning_rate": 9.996973228346458e-06, + "loss": 0.3798, + "step": 1926 + }, + { + "epoch": 0.6069291338582677, + "grad_norm": 41.602073669433594, + "learning_rate": 9.996971653543309e-06, + "loss": 0.4223, + "step": 1927 + }, + { + "epoch": 0.607244094488189, + "grad_norm": 46.16876220703125, + "learning_rate": 9.996970078740158e-06, + "loss": 0.5156, + "step": 1928 + }, + { + "epoch": 0.6075590551181103, + "grad_norm": 38.204036712646484, + "learning_rate": 9.996968503937009e-06, + "loss": 0.3032, + "step": 1929 + }, + { + "epoch": 0.6078740157480315, + "grad_norm": 7.078056335449219, + "learning_rate": 9.996966929133858e-06, + "loss": 0.0285, + "step": 1930 + }, + { + "epoch": 0.6081889763779528, + "grad_norm": 76.96788787841797, + "learning_rate": 9.99696535433071e-06, + "loss": 0.4164, + "step": 1931 + }, + { + "epoch": 0.608503937007874, + "grad_norm": 33.73112487792969, + "learning_rate": 9.99696377952756e-06, + "loss": 0.1971, + "step": 1932 + }, + { + "epoch": 0.6088188976377953, + "grad_norm": 50.70700454711914, + "learning_rate": 9.996962204724411e-06, + "loss": 0.6604, + "step": 1933 + }, + { + "epoch": 0.6091338582677165, + "grad_norm": 16.3695011138916, + "learning_rate": 9.99696062992126e-06, + "loss": 0.2403, + "step": 1934 + }, + { + "epoch": 0.6094488188976378, + "grad_norm": 32.59841537475586, + "learning_rate": 9.99695905511811e-06, + "loss": 0.3409, + "step": 1935 + }, + { + "epoch": 0.609763779527559, + "grad_norm": 33.59854507446289, + "learning_rate": 9.99695748031496e-06, + "loss": 0.2444, + "step": 1936 + }, + { + "epoch": 0.6100787401574803, + "grad_norm": 53.62559509277344, + "learning_rate": 9.996955905511812e-06, + "loss": 0.1804, + "step": 1937 + }, + { + "epoch": 0.6103937007874016, + "grad_norm": 41.21846389770508, + "learning_rate": 9.996954330708663e-06, + "loss": 0.8659, + "step": 1938 + }, + { + "epoch": 0.6107086614173228, + "grad_norm": 65.04766845703125, + "learning_rate": 9.996952755905512e-06, + "loss": 0.4908, + "step": 1939 + }, + { + "epoch": 0.6110236220472441, + "grad_norm": 63.45843505859375, + "learning_rate": 9.996951181102363e-06, + "loss": 0.8725, + "step": 1940 + }, + { + "epoch": 0.6110236220472441, + "eval_loss": 0.5175274014472961, + "eval_runtime": 349.2545, + "eval_samples_per_second": 0.335, + "eval_steps_per_second": 0.335, + "step": 1940 + }, + { + "epoch": 0.6113385826771653, + "grad_norm": 59.46821212768555, + "learning_rate": 9.996949606299214e-06, + "loss": 0.5103, + "step": 1941 + }, + { + "epoch": 0.6116535433070867, + "grad_norm": 18.964391708374023, + "learning_rate": 9.996948031496065e-06, + "loss": 0.3203, + "step": 1942 + }, + { + "epoch": 0.6119685039370079, + "grad_norm": 101.10069274902344, + "learning_rate": 9.996946456692914e-06, + "loss": 0.7592, + "step": 1943 + }, + { + "epoch": 0.6122834645669292, + "grad_norm": 36.87664794921875, + "learning_rate": 9.996944881889764e-06, + "loss": 0.5184, + "step": 1944 + }, + { + "epoch": 0.6125984251968504, + "grad_norm": 31.650327682495117, + "learning_rate": 9.996943307086615e-06, + "loss": 0.3825, + "step": 1945 + }, + { + "epoch": 0.6129133858267717, + "grad_norm": 23.641586303710938, + "learning_rate": 9.996941732283466e-06, + "loss": 0.1946, + "step": 1946 + }, + { + "epoch": 0.6132283464566929, + "grad_norm": 19.757226943969727, + "learning_rate": 9.996940157480317e-06, + "loss": 0.0762, + "step": 1947 + }, + { + "epoch": 0.6135433070866142, + "grad_norm": 10.93740177154541, + "learning_rate": 9.996938582677166e-06, + "loss": 0.0708, + "step": 1948 + }, + { + "epoch": 0.6138582677165354, + "grad_norm": 21.542688369750977, + "learning_rate": 9.996937007874017e-06, + "loss": 0.2724, + "step": 1949 + }, + { + "epoch": 0.6141732283464567, + "grad_norm": 27.340009689331055, + "learning_rate": 9.996935433070866e-06, + "loss": 0.0854, + "step": 1950 + }, + { + "epoch": 0.6144881889763779, + "grad_norm": 47.907955169677734, + "learning_rate": 9.996933858267717e-06, + "loss": 0.3816, + "step": 1951 + }, + { + "epoch": 0.6148031496062992, + "grad_norm": 65.70764923095703, + "learning_rate": 9.996932283464568e-06, + "loss": 0.2682, + "step": 1952 + }, + { + "epoch": 0.6151181102362204, + "grad_norm": 39.29658889770508, + "learning_rate": 9.99693070866142e-06, + "loss": 0.3791, + "step": 1953 + }, + { + "epoch": 0.6154330708661417, + "grad_norm": 7.722301006317139, + "learning_rate": 9.996929133858268e-06, + "loss": 0.0215, + "step": 1954 + }, + { + "epoch": 0.6157480314960629, + "grad_norm": 47.819461822509766, + "learning_rate": 9.996927559055118e-06, + "loss": 0.6154, + "step": 1955 + }, + { + "epoch": 0.6160629921259843, + "grad_norm": 32.92623519897461, + "learning_rate": 9.996925984251969e-06, + "loss": 0.1975, + "step": 1956 + }, + { + "epoch": 0.6163779527559056, + "grad_norm": 73.50929260253906, + "learning_rate": 9.99692440944882e-06, + "loss": 0.2781, + "step": 1957 + }, + { + "epoch": 0.6166929133858268, + "grad_norm": 26.083810806274414, + "learning_rate": 9.99692283464567e-06, + "loss": 0.3916, + "step": 1958 + }, + { + "epoch": 0.6170078740157481, + "grad_norm": 3.3253297805786133, + "learning_rate": 9.99692125984252e-06, + "loss": 0.0313, + "step": 1959 + }, + { + "epoch": 0.6173228346456693, + "grad_norm": 0.6226401329040527, + "learning_rate": 9.996919685039371e-06, + "loss": 0.0049, + "step": 1960 + }, + { + "epoch": 0.6173228346456693, + "eval_loss": 0.4793933033943176, + "eval_runtime": 341.1297, + "eval_samples_per_second": 0.343, + "eval_steps_per_second": 0.343, + "step": 1960 + }, + { + "epoch": 0.6176377952755906, + "grad_norm": 38.08949279785156, + "learning_rate": 9.99691811023622e-06, + "loss": 0.2089, + "step": 1961 + }, + { + "epoch": 0.6179527559055118, + "grad_norm": 96.51960754394531, + "learning_rate": 9.996916535433071e-06, + "loss": 0.6145, + "step": 1962 + }, + { + "epoch": 0.6182677165354331, + "grad_norm": 36.37595748901367, + "learning_rate": 9.996914960629922e-06, + "loss": 0.1043, + "step": 1963 + }, + { + "epoch": 0.6185826771653543, + "grad_norm": 112.60614013671875, + "learning_rate": 9.996913385826772e-06, + "loss": 1.0763, + "step": 1964 + }, + { + "epoch": 0.6188976377952756, + "grad_norm": 49.78593444824219, + "learning_rate": 9.996911811023623e-06, + "loss": 0.1965, + "step": 1965 + }, + { + "epoch": 0.6192125984251968, + "grad_norm": 58.12906265258789, + "learning_rate": 9.996910236220474e-06, + "loss": 0.3536, + "step": 1966 + }, + { + "epoch": 0.6195275590551181, + "grad_norm": 71.21797180175781, + "learning_rate": 9.996908661417325e-06, + "loss": 0.6605, + "step": 1967 + }, + { + "epoch": 0.6198425196850393, + "grad_norm": 104.464111328125, + "learning_rate": 9.996907086614174e-06, + "loss": 0.3808, + "step": 1968 + }, + { + "epoch": 0.6201574803149607, + "grad_norm": 79.76933288574219, + "learning_rate": 9.996905511811025e-06, + "loss": 0.5754, + "step": 1969 + }, + { + "epoch": 0.6204724409448819, + "grad_norm": 21.67342758178711, + "learning_rate": 9.996903937007874e-06, + "loss": 0.0917, + "step": 1970 + }, + { + "epoch": 0.6207874015748032, + "grad_norm": 34.137447357177734, + "learning_rate": 9.996902362204725e-06, + "loss": 0.5807, + "step": 1971 + }, + { + "epoch": 0.6211023622047244, + "grad_norm": 46.111595153808594, + "learning_rate": 9.996900787401576e-06, + "loss": 0.3738, + "step": 1972 + }, + { + "epoch": 0.6214173228346457, + "grad_norm": 45.28417205810547, + "learning_rate": 9.996899212598427e-06, + "loss": 0.7565, + "step": 1973 + }, + { + "epoch": 0.6217322834645669, + "grad_norm": 36.73102569580078, + "learning_rate": 9.996897637795276e-06, + "loss": 0.7887, + "step": 1974 + }, + { + "epoch": 0.6220472440944882, + "grad_norm": 69.32728576660156, + "learning_rate": 9.996896062992126e-06, + "loss": 0.4563, + "step": 1975 + }, + { + "epoch": 0.6223622047244094, + "grad_norm": 43.308555603027344, + "learning_rate": 9.996894488188977e-06, + "loss": 0.2911, + "step": 1976 + }, + { + "epoch": 0.6226771653543307, + "grad_norm": 40.517086029052734, + "learning_rate": 9.996892913385828e-06, + "loss": 0.3979, + "step": 1977 + }, + { + "epoch": 0.622992125984252, + "grad_norm": 68.62828826904297, + "learning_rate": 9.996891338582679e-06, + "loss": 0.3365, + "step": 1978 + }, + { + "epoch": 0.6233070866141732, + "grad_norm": 28.892871856689453, + "learning_rate": 9.996889763779528e-06, + "loss": 0.5079, + "step": 1979 + }, + { + "epoch": 0.6236220472440945, + "grad_norm": 22.088882446289062, + "learning_rate": 9.996888188976379e-06, + "loss": 0.3113, + "step": 1980 + }, + { + "epoch": 0.6236220472440945, + "eval_loss": 0.4940292537212372, + "eval_runtime": 352.4729, + "eval_samples_per_second": 0.332, + "eval_steps_per_second": 0.332, + "step": 1980 + }, + { + "epoch": 0.6239370078740157, + "grad_norm": 37.80302047729492, + "learning_rate": 9.996886614173228e-06, + "loss": 0.6656, + "step": 1981 + }, + { + "epoch": 0.624251968503937, + "grad_norm": 19.02584457397461, + "learning_rate": 9.99688503937008e-06, + "loss": 0.1548, + "step": 1982 + }, + { + "epoch": 0.6245669291338583, + "grad_norm": 10.896634101867676, + "learning_rate": 9.99688346456693e-06, + "loss": 0.1322, + "step": 1983 + }, + { + "epoch": 0.6248818897637796, + "grad_norm": 33.37139892578125, + "learning_rate": 9.99688188976378e-06, + "loss": 0.3109, + "step": 1984 + }, + { + "epoch": 0.6251968503937008, + "grad_norm": 20.222166061401367, + "learning_rate": 9.99688031496063e-06, + "loss": 0.3568, + "step": 1985 + }, + { + "epoch": 0.6255118110236221, + "grad_norm": 26.97902488708496, + "learning_rate": 9.996878740157482e-06, + "loss": 0.4469, + "step": 1986 + }, + { + "epoch": 0.6258267716535433, + "grad_norm": 3.6876025199890137, + "learning_rate": 9.996877165354333e-06, + "loss": 0.017, + "step": 1987 + }, + { + "epoch": 0.6261417322834646, + "grad_norm": 3.7643909454345703, + "learning_rate": 9.996875590551182e-06, + "loss": 0.0312, + "step": 1988 + }, + { + "epoch": 0.6264566929133858, + "grad_norm": 58.000083923339844, + "learning_rate": 9.996874015748033e-06, + "loss": 0.5242, + "step": 1989 + }, + { + "epoch": 0.6267716535433071, + "grad_norm": 57.707000732421875, + "learning_rate": 9.996872440944882e-06, + "loss": 0.537, + "step": 1990 + }, + { + "epoch": 0.6270866141732283, + "grad_norm": 23.221139907836914, + "learning_rate": 9.996870866141733e-06, + "loss": 0.1436, + "step": 1991 + }, + { + "epoch": 0.6274015748031496, + "grad_norm": 17.156478881835938, + "learning_rate": 9.996869291338584e-06, + "loss": 0.2777, + "step": 1992 + }, + { + "epoch": 0.6277165354330708, + "grad_norm": 81.11396789550781, + "learning_rate": 9.996867716535435e-06, + "loss": 0.6183, + "step": 1993 + }, + { + "epoch": 0.6280314960629921, + "grad_norm": 17.866575241088867, + "learning_rate": 9.996866141732284e-06, + "loss": 0.2052, + "step": 1994 + }, + { + "epoch": 0.6283464566929133, + "grad_norm": 49.33943557739258, + "learning_rate": 9.996864566929134e-06, + "loss": 0.2541, + "step": 1995 + }, + { + "epoch": 0.6286614173228346, + "grad_norm": 24.039709091186523, + "learning_rate": 9.996862992125985e-06, + "loss": 0.1116, + "step": 1996 + }, + { + "epoch": 0.628976377952756, + "grad_norm": 17.275371551513672, + "learning_rate": 9.996861417322836e-06, + "loss": 0.054, + "step": 1997 + }, + { + "epoch": 0.6292913385826772, + "grad_norm": 65.09687805175781, + "learning_rate": 9.996859842519687e-06, + "loss": 0.842, + "step": 1998 + }, + { + "epoch": 0.6296062992125985, + "grad_norm": 63.773319244384766, + "learning_rate": 9.996858267716536e-06, + "loss": 0.2476, + "step": 1999 + }, + { + "epoch": 0.6299212598425197, + "grad_norm": 101.19293212890625, + "learning_rate": 9.996856692913387e-06, + "loss": 1.0292, + "step": 2000 + }, + { + "epoch": 0.6299212598425197, + "eval_loss": 0.49186941981315613, + "eval_runtime": 611.4185, + "eval_samples_per_second": 0.191, + "eval_steps_per_second": 0.191, + "step": 2000 + } + ], + "logging_steps": 1, + "max_steps": 6350000, + "num_input_tokens_seen": 0, + "num_train_epochs": 2000, + "save_steps": 20, + "total_flos": 1.252556321381376e+18, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}