diff --git "a/checkpoint-574/trainer_state.json" "b/checkpoint-574/trainer_state.json" new file mode 100644--- /dev/null +++ "b/checkpoint-574/trainer_state.json" @@ -0,0 +1,4116 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.999129677980853, + "eval_steps": 72, + "global_step": 574, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0, + "eval_loss": 1.136976718902588, + "eval_runtime": 64.1168, + "eval_samples_per_second": 73.117, + "eval_steps_per_second": 9.14, + "step": 0 + }, + { + "epoch": 0.0017406440382941688, + "grad_norm": 1.859375, + "learning_rate": 0.0, + "loss": 1.1426, + "step": 1 + }, + { + "epoch": 0.0034812880765883376, + "grad_norm": 1.9765625, + "learning_rate": 1.7543859649122808e-07, + "loss": 1.1274, + "step": 2 + }, + { + "epoch": 0.005221932114882507, + "grad_norm": 1.828125, + "learning_rate": 3.5087719298245616e-07, + "loss": 1.1205, + "step": 3 + }, + { + "epoch": 0.006962576153176675, + "grad_norm": 1.8515625, + "learning_rate": 5.263157894736843e-07, + "loss": 1.1383, + "step": 4 + }, + { + "epoch": 0.008703220191470844, + "grad_norm": 1.796875, + "learning_rate": 7.017543859649123e-07, + "loss": 1.14, + "step": 5 + }, + { + "epoch": 0.010443864229765013, + "grad_norm": 1.734375, + "learning_rate": 8.771929824561404e-07, + "loss": 1.1031, + "step": 6 + }, + { + "epoch": 0.012184508268059183, + "grad_norm": 2.015625, + "learning_rate": 1.0526315789473685e-06, + "loss": 1.1794, + "step": 7 + }, + { + "epoch": 0.01392515230635335, + "grad_norm": 1.765625, + "learning_rate": 1.2280701754385965e-06, + "loss": 1.1766, + "step": 8 + }, + { + "epoch": 0.015665796344647518, + "grad_norm": 1.8046875, + "learning_rate": 1.4035087719298246e-06, + "loss": 1.1251, + "step": 9 + }, + { + "epoch": 0.017406440382941687, + "grad_norm": 1.6796875, + "learning_rate": 1.5789473684210526e-06, + "loss": 1.1622, + "step": 10 + }, + { + "epoch": 0.019147084421235857, + "grad_norm": 1.8671875, + "learning_rate": 1.7543859649122807e-06, + "loss": 1.0846, + "step": 11 + }, + { + "epoch": 0.020887728459530026, + "grad_norm": 1.765625, + "learning_rate": 1.929824561403509e-06, + "loss": 1.1644, + "step": 12 + }, + { + "epoch": 0.022628372497824196, + "grad_norm": 1.7265625, + "learning_rate": 2.105263157894737e-06, + "loss": 1.1361, + "step": 13 + }, + { + "epoch": 0.024369016536118365, + "grad_norm": 1.8671875, + "learning_rate": 2.280701754385965e-06, + "loss": 1.1783, + "step": 14 + }, + { + "epoch": 0.02610966057441253, + "grad_norm": 1.828125, + "learning_rate": 2.456140350877193e-06, + "loss": 1.127, + "step": 15 + }, + { + "epoch": 0.0278503046127067, + "grad_norm": 1.640625, + "learning_rate": 2.631578947368421e-06, + "loss": 1.1587, + "step": 16 + }, + { + "epoch": 0.02959094865100087, + "grad_norm": 1.671875, + "learning_rate": 2.8070175438596493e-06, + "loss": 1.1088, + "step": 17 + }, + { + "epoch": 0.031331592689295036, + "grad_norm": 1.59375, + "learning_rate": 2.9824561403508774e-06, + "loss": 1.174, + "step": 18 + }, + { + "epoch": 0.03307223672758921, + "grad_norm": 1.71875, + "learning_rate": 3.157894736842105e-06, + "loss": 1.1453, + "step": 19 + }, + { + "epoch": 0.034812880765883375, + "grad_norm": 1.8515625, + "learning_rate": 3.3333333333333333e-06, + "loss": 1.1922, + "step": 20 + }, + { + "epoch": 0.03655352480417755, + "grad_norm": 1.5390625, + "learning_rate": 3.5087719298245615e-06, + "loss": 1.1541, + "step": 21 + }, + { + "epoch": 0.038294168842471714, + "grad_norm": 1.5390625, + "learning_rate": 3.6842105263157896e-06, + "loss": 1.0909, + "step": 22 + }, + { + "epoch": 0.04003481288076589, + "grad_norm": 1.515625, + "learning_rate": 3.859649122807018e-06, + "loss": 1.1498, + "step": 23 + }, + { + "epoch": 0.04177545691906005, + "grad_norm": 1.5703125, + "learning_rate": 4.035087719298246e-06, + "loss": 1.1096, + "step": 24 + }, + { + "epoch": 0.04351610095735422, + "grad_norm": 1.453125, + "learning_rate": 4.210526315789474e-06, + "loss": 1.1152, + "step": 25 + }, + { + "epoch": 0.04525674499564839, + "grad_norm": 1.390625, + "learning_rate": 4.385964912280702e-06, + "loss": 1.1111, + "step": 26 + }, + { + "epoch": 0.04699738903394256, + "grad_norm": 1.296875, + "learning_rate": 4.56140350877193e-06, + "loss": 1.161, + "step": 27 + }, + { + "epoch": 0.04873803307223673, + "grad_norm": 1.390625, + "learning_rate": 4.736842105263158e-06, + "loss": 1.1557, + "step": 28 + }, + { + "epoch": 0.050478677110530897, + "grad_norm": 1.3125, + "learning_rate": 4.912280701754386e-06, + "loss": 1.1065, + "step": 29 + }, + { + "epoch": 0.05221932114882506, + "grad_norm": 1.28125, + "learning_rate": 5.087719298245615e-06, + "loss": 1.0493, + "step": 30 + }, + { + "epoch": 0.053959965187119235, + "grad_norm": 1.1953125, + "learning_rate": 5.263157894736842e-06, + "loss": 1.0652, + "step": 31 + }, + { + "epoch": 0.0557006092254134, + "grad_norm": 1.1484375, + "learning_rate": 5.438596491228071e-06, + "loss": 1.0389, + "step": 32 + }, + { + "epoch": 0.057441253263707574, + "grad_norm": 1.1015625, + "learning_rate": 5.6140350877192985e-06, + "loss": 1.0349, + "step": 33 + }, + { + "epoch": 0.05918189730200174, + "grad_norm": 1.0390625, + "learning_rate": 5.789473684210527e-06, + "loss": 1.1109, + "step": 34 + }, + { + "epoch": 0.060922541340295906, + "grad_norm": 1.03125, + "learning_rate": 5.964912280701755e-06, + "loss": 1.0541, + "step": 35 + }, + { + "epoch": 0.06266318537859007, + "grad_norm": 0.9453125, + "learning_rate": 6.140350877192983e-06, + "loss": 1.1176, + "step": 36 + }, + { + "epoch": 0.06440382941688425, + "grad_norm": 0.9140625, + "learning_rate": 6.31578947368421e-06, + "loss": 1.0774, + "step": 37 + }, + { + "epoch": 0.06614447345517842, + "grad_norm": 0.9453125, + "learning_rate": 6.491228070175439e-06, + "loss": 1.031, + "step": 38 + }, + { + "epoch": 0.06788511749347259, + "grad_norm": 0.85546875, + "learning_rate": 6.666666666666667e-06, + "loss": 1.0977, + "step": 39 + }, + { + "epoch": 0.06962576153176675, + "grad_norm": 0.8203125, + "learning_rate": 6.842105263157896e-06, + "loss": 1.0871, + "step": 40 + }, + { + "epoch": 0.07136640557006092, + "grad_norm": 0.83984375, + "learning_rate": 7.017543859649123e-06, + "loss": 1.0856, + "step": 41 + }, + { + "epoch": 0.0731070496083551, + "grad_norm": 0.82421875, + "learning_rate": 7.192982456140352e-06, + "loss": 1.0189, + "step": 42 + }, + { + "epoch": 0.07484769364664925, + "grad_norm": 0.76953125, + "learning_rate": 7.368421052631579e-06, + "loss": 1.0633, + "step": 43 + }, + { + "epoch": 0.07658833768494343, + "grad_norm": 0.7578125, + "learning_rate": 7.5438596491228074e-06, + "loss": 1.0582, + "step": 44 + }, + { + "epoch": 0.0783289817232376, + "grad_norm": 0.78125, + "learning_rate": 7.719298245614036e-06, + "loss": 1.0261, + "step": 45 + }, + { + "epoch": 0.08006962576153177, + "grad_norm": 0.66796875, + "learning_rate": 7.894736842105265e-06, + "loss": 0.9962, + "step": 46 + }, + { + "epoch": 0.08181026979982593, + "grad_norm": 0.6953125, + "learning_rate": 8.070175438596492e-06, + "loss": 1.0041, + "step": 47 + }, + { + "epoch": 0.0835509138381201, + "grad_norm": 0.6640625, + "learning_rate": 8.24561403508772e-06, + "loss": 1.0953, + "step": 48 + }, + { + "epoch": 0.08529155787641428, + "grad_norm": 0.671875, + "learning_rate": 8.421052631578948e-06, + "loss": 1.0397, + "step": 49 + }, + { + "epoch": 0.08703220191470844, + "grad_norm": 0.66015625, + "learning_rate": 8.596491228070176e-06, + "loss": 1.0366, + "step": 50 + }, + { + "epoch": 0.08877284595300261, + "grad_norm": 0.65625, + "learning_rate": 8.771929824561405e-06, + "loss": 1.005, + "step": 51 + }, + { + "epoch": 0.09051348999129678, + "grad_norm": 0.62109375, + "learning_rate": 8.947368421052632e-06, + "loss": 1.1106, + "step": 52 + }, + { + "epoch": 0.09225413402959094, + "grad_norm": 0.59375, + "learning_rate": 9.12280701754386e-06, + "loss": 1.002, + "step": 53 + }, + { + "epoch": 0.09399477806788512, + "grad_norm": 0.5859375, + "learning_rate": 9.298245614035088e-06, + "loss": 1.0599, + "step": 54 + }, + { + "epoch": 0.09573542210617929, + "grad_norm": 0.59765625, + "learning_rate": 9.473684210526315e-06, + "loss": 1.0736, + "step": 55 + }, + { + "epoch": 0.09747606614447346, + "grad_norm": 0.56640625, + "learning_rate": 9.649122807017545e-06, + "loss": 1.0179, + "step": 56 + }, + { + "epoch": 0.09921671018276762, + "grad_norm": 0.58203125, + "learning_rate": 9.824561403508772e-06, + "loss": 1.0308, + "step": 57 + }, + { + "epoch": 0.10095735422106179, + "grad_norm": 0.5625, + "learning_rate": 1e-05, + "loss": 1.0135, + "step": 58 + }, + { + "epoch": 0.10269799825935597, + "grad_norm": 0.5859375, + "learning_rate": 9.999979270446263e-06, + "loss": 0.9598, + "step": 59 + }, + { + "epoch": 0.10443864229765012, + "grad_norm": 0.5390625, + "learning_rate": 9.999917081956933e-06, + "loss": 1.0041, + "step": 60 + }, + { + "epoch": 0.1061792863359443, + "grad_norm": 0.53125, + "learning_rate": 9.999813435047668e-06, + "loss": 1.0726, + "step": 61 + }, + { + "epoch": 0.10791993037423847, + "grad_norm": 0.55078125, + "learning_rate": 9.99966833057789e-06, + "loss": 0.9915, + "step": 62 + }, + { + "epoch": 0.10966057441253264, + "grad_norm": 0.55078125, + "learning_rate": 9.999481769750779e-06, + "loss": 1.0266, + "step": 63 + }, + { + "epoch": 0.1114012184508268, + "grad_norm": 0.5234375, + "learning_rate": 9.999253754113263e-06, + "loss": 1.0524, + "step": 64 + }, + { + "epoch": 0.11314186248912098, + "grad_norm": 0.53125, + "learning_rate": 9.998984285556008e-06, + "loss": 1.0254, + "step": 65 + }, + { + "epoch": 0.11488250652741515, + "grad_norm": 0.53515625, + "learning_rate": 9.998673366313399e-06, + "loss": 1.0132, + "step": 66 + }, + { + "epoch": 0.11662315056570931, + "grad_norm": 0.5234375, + "learning_rate": 9.998320998963523e-06, + "loss": 0.9604, + "step": 67 + }, + { + "epoch": 0.11836379460400348, + "grad_norm": 0.54296875, + "learning_rate": 9.997927186428145e-06, + "loss": 1.0147, + "step": 68 + }, + { + "epoch": 0.12010443864229765, + "grad_norm": 0.52734375, + "learning_rate": 9.997491931972694e-06, + "loss": 0.998, + "step": 69 + }, + { + "epoch": 0.12184508268059181, + "grad_norm": 0.51953125, + "learning_rate": 9.997015239206216e-06, + "loss": 1.034, + "step": 70 + }, + { + "epoch": 0.12358572671888599, + "grad_norm": 0.53125, + "learning_rate": 9.996497112081365e-06, + "loss": 1.0103, + "step": 71 + }, + { + "epoch": 0.12532637075718014, + "grad_norm": 0.5078125, + "learning_rate": 9.99593755489436e-06, + "loss": 1.0053, + "step": 72 + }, + { + "epoch": 0.12532637075718014, + "eval_loss": 0.9893413186073303, + "eval_runtime": 60.3364, + "eval_samples_per_second": 77.698, + "eval_steps_per_second": 9.712, + "step": 72 + }, + { + "epoch": 0.12706701479547433, + "grad_norm": 0.50390625, + "learning_rate": 9.995336572284945e-06, + "loss": 0.9841, + "step": 73 + }, + { + "epoch": 0.1288076588337685, + "grad_norm": 0.53125, + "learning_rate": 9.994694169236366e-06, + "loss": 0.9298, + "step": 74 + }, + { + "epoch": 0.13054830287206268, + "grad_norm": 0.51171875, + "learning_rate": 9.99401035107531e-06, + "loss": 1.0449, + "step": 75 + }, + { + "epoch": 0.13228894691035684, + "grad_norm": 0.515625, + "learning_rate": 9.993285123471878e-06, + "loss": 1.0139, + "step": 76 + }, + { + "epoch": 0.134029590948651, + "grad_norm": 0.5078125, + "learning_rate": 9.992518492439526e-06, + "loss": 1.0129, + "step": 77 + }, + { + "epoch": 0.13577023498694518, + "grad_norm": 0.54296875, + "learning_rate": 9.991710464335022e-06, + "loss": 0.9696, + "step": 78 + }, + { + "epoch": 0.13751087902523934, + "grad_norm": 0.53125, + "learning_rate": 9.990861045858392e-06, + "loss": 1.0114, + "step": 79 + }, + { + "epoch": 0.1392515230635335, + "grad_norm": 0.515625, + "learning_rate": 9.989970244052861e-06, + "loss": 0.957, + "step": 80 + }, + { + "epoch": 0.1409921671018277, + "grad_norm": 0.5078125, + "learning_rate": 9.9890380663048e-06, + "loss": 0.9603, + "step": 81 + }, + { + "epoch": 0.14273281114012185, + "grad_norm": 0.52734375, + "learning_rate": 9.98806452034366e-06, + "loss": 1.0523, + "step": 82 + }, + { + "epoch": 0.144473455178416, + "grad_norm": 0.5078125, + "learning_rate": 9.987049614241907e-06, + "loss": 0.9804, + "step": 83 + }, + { + "epoch": 0.1462140992167102, + "grad_norm": 0.54296875, + "learning_rate": 9.985993356414965e-06, + "loss": 1.0307, + "step": 84 + }, + { + "epoch": 0.14795474325500435, + "grad_norm": 0.515625, + "learning_rate": 9.984895755621136e-06, + "loss": 0.9802, + "step": 85 + }, + { + "epoch": 0.1496953872932985, + "grad_norm": 0.494140625, + "learning_rate": 9.983756820961528e-06, + "loss": 0.9547, + "step": 86 + }, + { + "epoch": 0.1514360313315927, + "grad_norm": 0.5, + "learning_rate": 9.982576561879984e-06, + "loss": 1.014, + "step": 87 + }, + { + "epoch": 0.15317667536988686, + "grad_norm": 0.494140625, + "learning_rate": 9.981354988163002e-06, + "loss": 0.9708, + "step": 88 + }, + { + "epoch": 0.15491731940818101, + "grad_norm": 0.51171875, + "learning_rate": 9.980092109939651e-06, + "loss": 0.9717, + "step": 89 + }, + { + "epoch": 0.1566579634464752, + "grad_norm": 0.484375, + "learning_rate": 9.978787937681496e-06, + "loss": 0.9862, + "step": 90 + }, + { + "epoch": 0.15839860748476936, + "grad_norm": 0.498046875, + "learning_rate": 9.977442482202498e-06, + "loss": 0.9813, + "step": 91 + }, + { + "epoch": 0.16013925152306355, + "grad_norm": 0.494140625, + "learning_rate": 9.976055754658935e-06, + "loss": 0.9551, + "step": 92 + }, + { + "epoch": 0.1618798955613577, + "grad_norm": 0.51171875, + "learning_rate": 9.974627766549301e-06, + "loss": 0.9777, + "step": 93 + }, + { + "epoch": 0.16362053959965187, + "grad_norm": 0.50390625, + "learning_rate": 9.973158529714224e-06, + "loss": 0.9648, + "step": 94 + }, + { + "epoch": 0.16536118363794605, + "grad_norm": 0.48828125, + "learning_rate": 9.971648056336349e-06, + "loss": 0.964, + "step": 95 + }, + { + "epoch": 0.1671018276762402, + "grad_norm": 0.5078125, + "learning_rate": 9.97009635894025e-06, + "loss": 0.9635, + "step": 96 + }, + { + "epoch": 0.16884247171453437, + "grad_norm": 0.515625, + "learning_rate": 9.968503450392332e-06, + "loss": 1.0115, + "step": 97 + }, + { + "epoch": 0.17058311575282856, + "grad_norm": 0.51171875, + "learning_rate": 9.966869343900702e-06, + "loss": 0.9841, + "step": 98 + }, + { + "epoch": 0.17232375979112272, + "grad_norm": 0.484375, + "learning_rate": 9.965194053015083e-06, + "loss": 1.0241, + "step": 99 + }, + { + "epoch": 0.17406440382941687, + "grad_norm": 0.50390625, + "learning_rate": 9.963477591626687e-06, + "loss": 0.968, + "step": 100 + }, + { + "epoch": 0.17580504786771106, + "grad_norm": 0.50390625, + "learning_rate": 9.961719973968102e-06, + "loss": 0.946, + "step": 101 + }, + { + "epoch": 0.17754569190600522, + "grad_norm": 0.5234375, + "learning_rate": 9.959921214613187e-06, + "loss": 0.9284, + "step": 102 + }, + { + "epoch": 0.17928633594429938, + "grad_norm": 0.494140625, + "learning_rate": 9.958081328476926e-06, + "loss": 0.958, + "step": 103 + }, + { + "epoch": 0.18102697998259357, + "grad_norm": 0.486328125, + "learning_rate": 9.956200330815329e-06, + "loss": 0.9361, + "step": 104 + }, + { + "epoch": 0.18276762402088773, + "grad_norm": 0.5078125, + "learning_rate": 9.954278237225296e-06, + "loss": 1.0116, + "step": 105 + }, + { + "epoch": 0.18450826805918188, + "grad_norm": 0.5078125, + "learning_rate": 9.952315063644479e-06, + "loss": 1.0041, + "step": 106 + }, + { + "epoch": 0.18624891209747607, + "grad_norm": 0.5, + "learning_rate": 9.950310826351168e-06, + "loss": 0.9725, + "step": 107 + }, + { + "epoch": 0.18798955613577023, + "grad_norm": 0.498046875, + "learning_rate": 9.948265541964136e-06, + "loss": 0.9763, + "step": 108 + }, + { + "epoch": 0.18973020017406442, + "grad_norm": 0.5, + "learning_rate": 9.946179227442521e-06, + "loss": 0.9814, + "step": 109 + }, + { + "epoch": 0.19147084421235858, + "grad_norm": 0.50390625, + "learning_rate": 9.944051900085668e-06, + "loss": 1.0092, + "step": 110 + }, + { + "epoch": 0.19321148825065274, + "grad_norm": 0.515625, + "learning_rate": 9.941883577532993e-06, + "loss": 0.9836, + "step": 111 + }, + { + "epoch": 0.19495213228894692, + "grad_norm": 0.49609375, + "learning_rate": 9.939674277763845e-06, + "loss": 0.9598, + "step": 112 + }, + { + "epoch": 0.19669277632724108, + "grad_norm": 0.48828125, + "learning_rate": 9.937424019097337e-06, + "loss": 0.988, + "step": 113 + }, + { + "epoch": 0.19843342036553524, + "grad_norm": 0.498046875, + "learning_rate": 9.935132820192218e-06, + "loss": 0.9512, + "step": 114 + }, + { + "epoch": 0.20017406440382943, + "grad_norm": 0.498046875, + "learning_rate": 9.932800700046697e-06, + "loss": 0.9916, + "step": 115 + }, + { + "epoch": 0.20191470844212359, + "grad_norm": 0.51171875, + "learning_rate": 9.9304276779983e-06, + "loss": 0.9397, + "step": 116 + }, + { + "epoch": 0.20365535248041775, + "grad_norm": 0.48828125, + "learning_rate": 9.9280137737237e-06, + "loss": 0.9791, + "step": 117 + }, + { + "epoch": 0.20539599651871193, + "grad_norm": 0.515625, + "learning_rate": 9.925559007238564e-06, + "loss": 0.9231, + "step": 118 + }, + { + "epoch": 0.2071366405570061, + "grad_norm": 0.5, + "learning_rate": 9.923063398897372e-06, + "loss": 0.9854, + "step": 119 + }, + { + "epoch": 0.20887728459530025, + "grad_norm": 0.50390625, + "learning_rate": 9.920526969393267e-06, + "loss": 1.0411, + "step": 120 + }, + { + "epoch": 0.21061792863359444, + "grad_norm": 0.5078125, + "learning_rate": 9.917949739757869e-06, + "loss": 1.0254, + "step": 121 + }, + { + "epoch": 0.2123585726718886, + "grad_norm": 0.5, + "learning_rate": 9.915331731361104e-06, + "loss": 0.9208, + "step": 122 + }, + { + "epoch": 0.21409921671018275, + "grad_norm": 0.5078125, + "learning_rate": 9.912672965911034e-06, + "loss": 0.9195, + "step": 123 + }, + { + "epoch": 0.21583986074847694, + "grad_norm": 0.48046875, + "learning_rate": 9.909973465453666e-06, + "loss": 0.9938, + "step": 124 + }, + { + "epoch": 0.2175805047867711, + "grad_norm": 0.5078125, + "learning_rate": 9.907233252372775e-06, + "loss": 0.904, + "step": 125 + }, + { + "epoch": 0.2193211488250653, + "grad_norm": 0.48828125, + "learning_rate": 9.904452349389717e-06, + "loss": 0.9882, + "step": 126 + }, + { + "epoch": 0.22106179286335945, + "grad_norm": 0.50390625, + "learning_rate": 9.901630779563247e-06, + "loss": 0.9688, + "step": 127 + }, + { + "epoch": 0.2228024369016536, + "grad_norm": 0.51953125, + "learning_rate": 9.898768566289316e-06, + "loss": 1.0522, + "step": 128 + }, + { + "epoch": 0.2245430809399478, + "grad_norm": 0.48828125, + "learning_rate": 9.895865733300887e-06, + "loss": 0.9551, + "step": 129 + }, + { + "epoch": 0.22628372497824195, + "grad_norm": 0.498046875, + "learning_rate": 9.89292230466773e-06, + "loss": 0.9623, + "step": 130 + }, + { + "epoch": 0.2280243690165361, + "grad_norm": 0.515625, + "learning_rate": 9.889938304796236e-06, + "loss": 0.9702, + "step": 131 + }, + { + "epoch": 0.2297650130548303, + "grad_norm": 0.5078125, + "learning_rate": 9.886913758429194e-06, + "loss": 0.9679, + "step": 132 + }, + { + "epoch": 0.23150565709312446, + "grad_norm": 0.466796875, + "learning_rate": 9.883848690645601e-06, + "loss": 0.9719, + "step": 133 + }, + { + "epoch": 0.23324630113141862, + "grad_norm": 0.4921875, + "learning_rate": 9.880743126860458e-06, + "loss": 0.9717, + "step": 134 + }, + { + "epoch": 0.2349869451697128, + "grad_norm": 0.4921875, + "learning_rate": 9.87759709282454e-06, + "loss": 0.9365, + "step": 135 + }, + { + "epoch": 0.23672758920800696, + "grad_norm": 0.478515625, + "learning_rate": 9.874410614624202e-06, + "loss": 0.9254, + "step": 136 + }, + { + "epoch": 0.23846823324630112, + "grad_norm": 0.498046875, + "learning_rate": 9.871183718681153e-06, + "loss": 1.0045, + "step": 137 + }, + { + "epoch": 0.2402088772845953, + "grad_norm": 0.51171875, + "learning_rate": 9.867916431752237e-06, + "loss": 0.9693, + "step": 138 + }, + { + "epoch": 0.24194952132288947, + "grad_norm": 0.474609375, + "learning_rate": 9.864608780929218e-06, + "loss": 0.9981, + "step": 139 + }, + { + "epoch": 0.24369016536118362, + "grad_norm": 0.4921875, + "learning_rate": 9.861260793638539e-06, + "loss": 0.9569, + "step": 140 + }, + { + "epoch": 0.2454308093994778, + "grad_norm": 0.490234375, + "learning_rate": 9.857872497641117e-06, + "loss": 1.0422, + "step": 141 + }, + { + "epoch": 0.24717145343777197, + "grad_norm": 0.49609375, + "learning_rate": 9.854443921032098e-06, + "loss": 1.0408, + "step": 142 + }, + { + "epoch": 0.24891209747606616, + "grad_norm": 0.484375, + "learning_rate": 9.850975092240625e-06, + "loss": 0.8893, + "step": 143 + }, + { + "epoch": 0.2506527415143603, + "grad_norm": 0.5078125, + "learning_rate": 9.84746604002961e-06, + "loss": 0.9679, + "step": 144 + }, + { + "epoch": 0.2506527415143603, + "eval_loss": 0.9576423168182373, + "eval_runtime": 59.1215, + "eval_samples_per_second": 79.294, + "eval_steps_per_second": 9.912, + "step": 144 + }, + { + "epoch": 0.2523933855526545, + "grad_norm": 0.5, + "learning_rate": 9.843916793495487e-06, + "loss": 0.9071, + "step": 145 + }, + { + "epoch": 0.25413402959094866, + "grad_norm": 0.498046875, + "learning_rate": 9.840327382067972e-06, + "loss": 0.9496, + "step": 146 + }, + { + "epoch": 0.2558746736292428, + "grad_norm": 0.5078125, + "learning_rate": 9.836697835509827e-06, + "loss": 0.9864, + "step": 147 + }, + { + "epoch": 0.257615317667537, + "grad_norm": 0.5, + "learning_rate": 9.833028183916601e-06, + "loss": 1.0082, + "step": 148 + }, + { + "epoch": 0.25935596170583114, + "grad_norm": 0.494140625, + "learning_rate": 9.829318457716395e-06, + "loss": 0.9591, + "step": 149 + }, + { + "epoch": 0.26109660574412535, + "grad_norm": 0.515625, + "learning_rate": 9.82556868766959e-06, + "loss": 1.0204, + "step": 150 + }, + { + "epoch": 0.2628372497824195, + "grad_norm": 0.4921875, + "learning_rate": 9.821778904868616e-06, + "loss": 0.9805, + "step": 151 + }, + { + "epoch": 0.26457789382071367, + "grad_norm": 0.5078125, + "learning_rate": 9.817949140737672e-06, + "loss": 0.9761, + "step": 152 + }, + { + "epoch": 0.26631853785900783, + "grad_norm": 0.50390625, + "learning_rate": 9.81407942703248e-06, + "loss": 0.9789, + "step": 153 + }, + { + "epoch": 0.268059181897302, + "grad_norm": 0.49609375, + "learning_rate": 9.810169795840012e-06, + "loss": 0.952, + "step": 154 + }, + { + "epoch": 0.26979982593559615, + "grad_norm": 0.484375, + "learning_rate": 9.806220279578236e-06, + "loss": 0.9431, + "step": 155 + }, + { + "epoch": 0.27154046997389036, + "grad_norm": 0.5078125, + "learning_rate": 9.802230910995833e-06, + "loss": 1.0015, + "step": 156 + }, + { + "epoch": 0.2732811140121845, + "grad_norm": 0.49609375, + "learning_rate": 9.798201723171938e-06, + "loss": 0.9513, + "step": 157 + }, + { + "epoch": 0.2750217580504787, + "grad_norm": 0.48828125, + "learning_rate": 9.794132749515854e-06, + "loss": 0.9454, + "step": 158 + }, + { + "epoch": 0.27676240208877284, + "grad_norm": 0.494140625, + "learning_rate": 9.790024023766789e-06, + "loss": 0.9581, + "step": 159 + }, + { + "epoch": 0.278503046127067, + "grad_norm": 0.482421875, + "learning_rate": 9.785875579993558e-06, + "loss": 0.9874, + "step": 160 + }, + { + "epoch": 0.28024369016536116, + "grad_norm": 0.49609375, + "learning_rate": 9.781687452594318e-06, + "loss": 0.9417, + "step": 161 + }, + { + "epoch": 0.2819843342036554, + "grad_norm": 0.478515625, + "learning_rate": 9.777459676296276e-06, + "loss": 0.9589, + "step": 162 + }, + { + "epoch": 0.28372497824194953, + "grad_norm": 0.515625, + "learning_rate": 9.773192286155395e-06, + "loss": 0.9851, + "step": 163 + }, + { + "epoch": 0.2854656222802437, + "grad_norm": 0.49609375, + "learning_rate": 9.768885317556116e-06, + "loss": 0.98, + "step": 164 + }, + { + "epoch": 0.28720626631853785, + "grad_norm": 0.5078125, + "learning_rate": 9.764538806211052e-06, + "loss": 0.9651, + "step": 165 + }, + { + "epoch": 0.288946910356832, + "grad_norm": 0.50390625, + "learning_rate": 9.760152788160697e-06, + "loss": 0.9407, + "step": 166 + }, + { + "epoch": 0.2906875543951262, + "grad_norm": 0.46875, + "learning_rate": 9.755727299773135e-06, + "loss": 0.9553, + "step": 167 + }, + { + "epoch": 0.2924281984334204, + "grad_norm": 0.470703125, + "learning_rate": 9.75126237774372e-06, + "loss": 0.9823, + "step": 168 + }, + { + "epoch": 0.29416884247171454, + "grad_norm": 0.5, + "learning_rate": 9.746758059094791e-06, + "loss": 0.9832, + "step": 169 + }, + { + "epoch": 0.2959094865100087, + "grad_norm": 0.50390625, + "learning_rate": 9.742214381175355e-06, + "loss": 0.8976, + "step": 170 + }, + { + "epoch": 0.29765013054830286, + "grad_norm": 0.48828125, + "learning_rate": 9.737631381660777e-06, + "loss": 0.9331, + "step": 171 + }, + { + "epoch": 0.299390774586597, + "grad_norm": 0.50390625, + "learning_rate": 9.733009098552473e-06, + "loss": 0.9666, + "step": 172 + }, + { + "epoch": 0.30113141862489123, + "grad_norm": 0.490234375, + "learning_rate": 9.728347570177587e-06, + "loss": 0.9781, + "step": 173 + }, + { + "epoch": 0.3028720626631854, + "grad_norm": 0.484375, + "learning_rate": 9.723646835188681e-06, + "loss": 0.969, + "step": 174 + }, + { + "epoch": 0.30461270670147955, + "grad_norm": 0.51171875, + "learning_rate": 9.71890693256341e-06, + "loss": 0.9481, + "step": 175 + }, + { + "epoch": 0.3063533507397737, + "grad_norm": 0.49609375, + "learning_rate": 9.7141279016042e-06, + "loss": 1.0301, + "step": 176 + }, + { + "epoch": 0.30809399477806787, + "grad_norm": 0.48828125, + "learning_rate": 9.709309781937925e-06, + "loss": 1.023, + "step": 177 + }, + { + "epoch": 0.30983463881636203, + "grad_norm": 0.490234375, + "learning_rate": 9.704452613515571e-06, + "loss": 0.9598, + "step": 178 + }, + { + "epoch": 0.31157528285465624, + "grad_norm": 0.482421875, + "learning_rate": 9.699556436611912e-06, + "loss": 0.9699, + "step": 179 + }, + { + "epoch": 0.3133159268929504, + "grad_norm": 0.498046875, + "learning_rate": 9.694621291825174e-06, + "loss": 0.903, + "step": 180 + }, + { + "epoch": 0.31505657093124456, + "grad_norm": 0.515625, + "learning_rate": 9.689647220076696e-06, + "loss": 0.9581, + "step": 181 + }, + { + "epoch": 0.3167972149695387, + "grad_norm": 0.490234375, + "learning_rate": 9.684634262610593e-06, + "loss": 0.939, + "step": 182 + }, + { + "epoch": 0.3185378590078329, + "grad_norm": 0.498046875, + "learning_rate": 9.679582460993413e-06, + "loss": 0.9363, + "step": 183 + }, + { + "epoch": 0.3202785030461271, + "grad_norm": 0.482421875, + "learning_rate": 9.674491857113792e-06, + "loss": 1.0215, + "step": 184 + }, + { + "epoch": 0.32201914708442125, + "grad_norm": 0.48828125, + "learning_rate": 9.669362493182112e-06, + "loss": 0.9464, + "step": 185 + }, + { + "epoch": 0.3237597911227154, + "grad_norm": 0.482421875, + "learning_rate": 9.66419441173014e-06, + "loss": 0.8955, + "step": 186 + }, + { + "epoch": 0.32550043516100957, + "grad_norm": 0.494140625, + "learning_rate": 9.658987655610687e-06, + "loss": 0.9503, + "step": 187 + }, + { + "epoch": 0.32724107919930373, + "grad_norm": 0.474609375, + "learning_rate": 9.653742267997245e-06, + "loss": 0.9808, + "step": 188 + }, + { + "epoch": 0.3289817232375979, + "grad_norm": 0.48046875, + "learning_rate": 9.648458292383631e-06, + "loss": 1.0155, + "step": 189 + }, + { + "epoch": 0.3307223672758921, + "grad_norm": 0.5078125, + "learning_rate": 9.643135772583627e-06, + "loss": 0.929, + "step": 190 + }, + { + "epoch": 0.33246301131418626, + "grad_norm": 0.5, + "learning_rate": 9.63777475273062e-06, + "loss": 0.9214, + "step": 191 + }, + { + "epoch": 0.3342036553524804, + "grad_norm": 0.47265625, + "learning_rate": 9.632375277277226e-06, + "loss": 0.9762, + "step": 192 + }, + { + "epoch": 0.3359442993907746, + "grad_norm": 0.474609375, + "learning_rate": 9.626937390994932e-06, + "loss": 0.9734, + "step": 193 + }, + { + "epoch": 0.33768494342906874, + "grad_norm": 0.498046875, + "learning_rate": 9.621461138973725e-06, + "loss": 0.9599, + "step": 194 + }, + { + "epoch": 0.3394255874673629, + "grad_norm": 0.50390625, + "learning_rate": 9.615946566621704e-06, + "loss": 0.9448, + "step": 195 + }, + { + "epoch": 0.3411662315056571, + "grad_norm": 0.494140625, + "learning_rate": 9.61039371966472e-06, + "loss": 0.9721, + "step": 196 + }, + { + "epoch": 0.3429068755439513, + "grad_norm": 0.48828125, + "learning_rate": 9.60480264414599e-06, + "loss": 0.97, + "step": 197 + }, + { + "epoch": 0.34464751958224543, + "grad_norm": 0.49609375, + "learning_rate": 9.599173386425711e-06, + "loss": 0.9964, + "step": 198 + }, + { + "epoch": 0.3463881636205396, + "grad_norm": 0.486328125, + "learning_rate": 9.593505993180687e-06, + "loss": 0.9715, + "step": 199 + }, + { + "epoch": 0.34812880765883375, + "grad_norm": 0.4921875, + "learning_rate": 9.587800511403931e-06, + "loss": 0.9655, + "step": 200 + }, + { + "epoch": 0.34986945169712796, + "grad_norm": 0.5078125, + "learning_rate": 9.582056988404276e-06, + "loss": 0.9606, + "step": 201 + }, + { + "epoch": 0.3516100957354221, + "grad_norm": 0.50390625, + "learning_rate": 9.576275471805993e-06, + "loss": 0.9166, + "step": 202 + }, + { + "epoch": 0.3533507397737163, + "grad_norm": 0.494140625, + "learning_rate": 9.570456009548383e-06, + "loss": 0.9063, + "step": 203 + }, + { + "epoch": 0.35509138381201044, + "grad_norm": 0.484375, + "learning_rate": 9.564598649885391e-06, + "loss": 0.9295, + "step": 204 + }, + { + "epoch": 0.3568320278503046, + "grad_norm": 0.48046875, + "learning_rate": 9.558703441385195e-06, + "loss": 0.9933, + "step": 205 + }, + { + "epoch": 0.35857267188859876, + "grad_norm": 0.48046875, + "learning_rate": 9.552770432929812e-06, + "loss": 0.9572, + "step": 206 + }, + { + "epoch": 0.360313315926893, + "grad_norm": 0.482421875, + "learning_rate": 9.54679967371469e-06, + "loss": 0.9484, + "step": 207 + }, + { + "epoch": 0.36205395996518713, + "grad_norm": 0.47265625, + "learning_rate": 9.540791213248299e-06, + "loss": 0.9266, + "step": 208 + }, + { + "epoch": 0.3637946040034813, + "grad_norm": 0.5, + "learning_rate": 9.534745101351719e-06, + "loss": 0.9526, + "step": 209 + }, + { + "epoch": 0.36553524804177545, + "grad_norm": 0.5, + "learning_rate": 9.528661388158234e-06, + "loss": 1.0046, + "step": 210 + }, + { + "epoch": 0.3672758920800696, + "grad_norm": 0.51171875, + "learning_rate": 9.522540124112902e-06, + "loss": 0.9757, + "step": 211 + }, + { + "epoch": 0.36901653611836377, + "grad_norm": 0.48046875, + "learning_rate": 9.516381359972157e-06, + "loss": 1.0195, + "step": 212 + }, + { + "epoch": 0.370757180156658, + "grad_norm": 0.490234375, + "learning_rate": 9.51018514680337e-06, + "loss": 0.971, + "step": 213 + }, + { + "epoch": 0.37249782419495214, + "grad_norm": 0.5, + "learning_rate": 9.503951535984434e-06, + "loss": 0.9336, + "step": 214 + }, + { + "epoch": 0.3742384682332463, + "grad_norm": 0.494140625, + "learning_rate": 9.49768057920334e-06, + "loss": 0.9887, + "step": 215 + }, + { + "epoch": 0.37597911227154046, + "grad_norm": 0.486328125, + "learning_rate": 9.491372328457737e-06, + "loss": 0.966, + "step": 216 + }, + { + "epoch": 0.37597911227154046, + "eval_loss": 0.9439952373504639, + "eval_runtime": 59.4647, + "eval_samples_per_second": 78.837, + "eval_steps_per_second": 9.855, + "step": 216 + }, + { + "epoch": 0.3777197563098346, + "grad_norm": 0.494140625, + "learning_rate": 9.485026836054519e-06, + "loss": 0.9771, + "step": 217 + }, + { + "epoch": 0.37946040034812883, + "grad_norm": 0.498046875, + "learning_rate": 9.478644154609372e-06, + "loss": 0.9988, + "step": 218 + }, + { + "epoch": 0.381201044386423, + "grad_norm": 0.470703125, + "learning_rate": 9.472224337046357e-06, + "loss": 0.9875, + "step": 219 + }, + { + "epoch": 0.38294168842471715, + "grad_norm": 0.482421875, + "learning_rate": 9.46576743659745e-06, + "loss": 0.9664, + "step": 220 + }, + { + "epoch": 0.3846823324630113, + "grad_norm": 0.478515625, + "learning_rate": 9.45927350680212e-06, + "loss": 0.9693, + "step": 221 + }, + { + "epoch": 0.38642297650130547, + "grad_norm": 0.48828125, + "learning_rate": 9.452742601506873e-06, + "loss": 0.9137, + "step": 222 + }, + { + "epoch": 0.38816362053959963, + "grad_norm": 0.5078125, + "learning_rate": 9.446174774864808e-06, + "loss": 0.9047, + "step": 223 + }, + { + "epoch": 0.38990426457789384, + "grad_norm": 0.486328125, + "learning_rate": 9.439570081335173e-06, + "loss": 0.9926, + "step": 224 + }, + { + "epoch": 0.391644908616188, + "grad_norm": 0.478515625, + "learning_rate": 9.432928575682908e-06, + "loss": 0.9693, + "step": 225 + }, + { + "epoch": 0.39338555265448216, + "grad_norm": 0.50390625, + "learning_rate": 9.426250312978191e-06, + "loss": 0.9106, + "step": 226 + }, + { + "epoch": 0.3951261966927763, + "grad_norm": 0.5, + "learning_rate": 9.419535348595985e-06, + "loss": 0.9639, + "step": 227 + }, + { + "epoch": 0.3968668407310705, + "grad_norm": 0.49609375, + "learning_rate": 9.412783738215576e-06, + "loss": 0.9338, + "step": 228 + }, + { + "epoch": 0.39860748476936464, + "grad_norm": 0.48828125, + "learning_rate": 9.405995537820111e-06, + "loss": 1.0216, + "step": 229 + }, + { + "epoch": 0.40034812880765885, + "grad_norm": 0.498046875, + "learning_rate": 9.399170803696139e-06, + "loss": 0.942, + "step": 230 + }, + { + "epoch": 0.402088772845953, + "grad_norm": 0.486328125, + "learning_rate": 9.392309592433134e-06, + "loss": 0.9184, + "step": 231 + }, + { + "epoch": 0.40382941688424717, + "grad_norm": 0.5234375, + "learning_rate": 9.385411960923036e-06, + "loss": 0.9178, + "step": 232 + }, + { + "epoch": 0.40557006092254133, + "grad_norm": 0.4921875, + "learning_rate": 9.378477966359773e-06, + "loss": 0.9303, + "step": 233 + }, + { + "epoch": 0.4073107049608355, + "grad_norm": 0.5, + "learning_rate": 9.371507666238793e-06, + "loss": 0.9563, + "step": 234 + }, + { + "epoch": 0.4090513489991297, + "grad_norm": 0.47265625, + "learning_rate": 9.364501118356579e-06, + "loss": 0.92, + "step": 235 + }, + { + "epoch": 0.41079199303742386, + "grad_norm": 0.4765625, + "learning_rate": 9.357458380810175e-06, + "loss": 0.9532, + "step": 236 + }, + { + "epoch": 0.412532637075718, + "grad_norm": 0.482421875, + "learning_rate": 9.350379511996706e-06, + "loss": 0.9604, + "step": 237 + }, + { + "epoch": 0.4142732811140122, + "grad_norm": 0.48828125, + "learning_rate": 9.343264570612883e-06, + "loss": 0.9415, + "step": 238 + }, + { + "epoch": 0.41601392515230634, + "grad_norm": 0.48828125, + "learning_rate": 9.336113615654535e-06, + "loss": 0.9752, + "step": 239 + }, + { + "epoch": 0.4177545691906005, + "grad_norm": 0.484375, + "learning_rate": 9.328926706416102e-06, + "loss": 0.9517, + "step": 240 + }, + { + "epoch": 0.4194952132288947, + "grad_norm": 0.484375, + "learning_rate": 9.321703902490152e-06, + "loss": 0.9245, + "step": 241 + }, + { + "epoch": 0.4212358572671889, + "grad_norm": 0.50390625, + "learning_rate": 9.314445263766888e-06, + "loss": 0.9341, + "step": 242 + }, + { + "epoch": 0.42297650130548303, + "grad_norm": 0.47265625, + "learning_rate": 9.307150850433643e-06, + "loss": 0.9399, + "step": 243 + }, + { + "epoch": 0.4247171453437772, + "grad_norm": 0.48828125, + "learning_rate": 9.299820722974396e-06, + "loss": 0.9865, + "step": 244 + }, + { + "epoch": 0.42645778938207135, + "grad_norm": 0.5078125, + "learning_rate": 9.29245494216925e-06, + "loss": 0.9538, + "step": 245 + }, + { + "epoch": 0.4281984334203655, + "grad_norm": 0.5, + "learning_rate": 9.285053569093948e-06, + "loss": 1.0095, + "step": 246 + }, + { + "epoch": 0.4299390774586597, + "grad_norm": 0.494140625, + "learning_rate": 9.277616665119352e-06, + "loss": 0.9691, + "step": 247 + }, + { + "epoch": 0.4316797214969539, + "grad_norm": 0.4921875, + "learning_rate": 9.27014429191094e-06, + "loss": 0.9854, + "step": 248 + }, + { + "epoch": 0.43342036553524804, + "grad_norm": 0.498046875, + "learning_rate": 9.262636511428304e-06, + "loss": 0.9179, + "step": 249 + }, + { + "epoch": 0.4351610095735422, + "grad_norm": 0.490234375, + "learning_rate": 9.255093385924616e-06, + "loss": 0.9388, + "step": 250 + }, + { + "epoch": 0.43690165361183636, + "grad_norm": 0.51953125, + "learning_rate": 9.247514977946124e-06, + "loss": 0.9788, + "step": 251 + }, + { + "epoch": 0.4386422976501306, + "grad_norm": 0.515625, + "learning_rate": 9.239901350331635e-06, + "loss": 0.9301, + "step": 252 + }, + { + "epoch": 0.44038294168842473, + "grad_norm": 0.474609375, + "learning_rate": 9.232252566211993e-06, + "loss": 0.9656, + "step": 253 + }, + { + "epoch": 0.4421235857267189, + "grad_norm": 0.4921875, + "learning_rate": 9.224568689009548e-06, + "loss": 1.0119, + "step": 254 + }, + { + "epoch": 0.44386422976501305, + "grad_norm": 0.49609375, + "learning_rate": 9.216849782437637e-06, + "loss": 0.874, + "step": 255 + }, + { + "epoch": 0.4456048738033072, + "grad_norm": 0.5, + "learning_rate": 9.20909591050006e-06, + "loss": 0.9191, + "step": 256 + }, + { + "epoch": 0.44734551784160137, + "grad_norm": 0.50390625, + "learning_rate": 9.201307137490536e-06, + "loss": 0.9017, + "step": 257 + }, + { + "epoch": 0.4490861618798956, + "grad_norm": 0.4921875, + "learning_rate": 9.19348352799218e-06, + "loss": 0.9363, + "step": 258 + }, + { + "epoch": 0.45082680591818974, + "grad_norm": 0.494140625, + "learning_rate": 9.185625146876966e-06, + "loss": 0.9921, + "step": 259 + }, + { + "epoch": 0.4525674499564839, + "grad_norm": 0.482421875, + "learning_rate": 9.177732059305187e-06, + "loss": 0.9358, + "step": 260 + }, + { + "epoch": 0.45430809399477806, + "grad_norm": 0.494140625, + "learning_rate": 9.169804330724916e-06, + "loss": 0.9257, + "step": 261 + }, + { + "epoch": 0.4560487380330722, + "grad_norm": 0.49609375, + "learning_rate": 9.161842026871465e-06, + "loss": 0.9201, + "step": 262 + }, + { + "epoch": 0.4577893820713664, + "grad_norm": 0.498046875, + "learning_rate": 9.153845213766837e-06, + "loss": 0.9212, + "step": 263 + }, + { + "epoch": 0.4595300261096606, + "grad_norm": 0.5, + "learning_rate": 9.145813957719174e-06, + "loss": 0.9735, + "step": 264 + }, + { + "epoch": 0.46127067014795475, + "grad_norm": 0.5078125, + "learning_rate": 9.137748325322223e-06, + "loss": 0.9585, + "step": 265 + }, + { + "epoch": 0.4630113141862489, + "grad_norm": 0.50390625, + "learning_rate": 9.129648383454764e-06, + "loss": 0.9781, + "step": 266 + }, + { + "epoch": 0.46475195822454307, + "grad_norm": 0.486328125, + "learning_rate": 9.121514199280072e-06, + "loss": 0.9759, + "step": 267 + }, + { + "epoch": 0.46649260226283723, + "grad_norm": 0.5078125, + "learning_rate": 9.113345840245348e-06, + "loss": 0.9688, + "step": 268 + }, + { + "epoch": 0.46823324630113144, + "grad_norm": 0.49609375, + "learning_rate": 9.105143374081167e-06, + "loss": 0.9092, + "step": 269 + }, + { + "epoch": 0.4699738903394256, + "grad_norm": 0.48046875, + "learning_rate": 9.096906868800917e-06, + "loss": 0.9357, + "step": 270 + }, + { + "epoch": 0.47171453437771976, + "grad_norm": 0.4921875, + "learning_rate": 9.088636392700227e-06, + "loss": 1.0134, + "step": 271 + }, + { + "epoch": 0.4734551784160139, + "grad_norm": 0.49609375, + "learning_rate": 9.08033201435641e-06, + "loss": 0.9494, + "step": 272 + }, + { + "epoch": 0.4751958224543081, + "grad_norm": 0.48046875, + "learning_rate": 9.071993802627887e-06, + "loss": 0.9446, + "step": 273 + }, + { + "epoch": 0.47693646649260224, + "grad_norm": 0.494140625, + "learning_rate": 9.063621826653624e-06, + "loss": 0.8926, + "step": 274 + }, + { + "epoch": 0.47867711053089645, + "grad_norm": 0.484375, + "learning_rate": 9.055216155852548e-06, + "loss": 0.9216, + "step": 275 + }, + { + "epoch": 0.4804177545691906, + "grad_norm": 0.470703125, + "learning_rate": 9.046776859922983e-06, + "loss": 0.9442, + "step": 276 + }, + { + "epoch": 0.4821583986074848, + "grad_norm": 0.484375, + "learning_rate": 9.038304008842064e-06, + "loss": 0.9456, + "step": 277 + }, + { + "epoch": 0.48389904264577893, + "grad_norm": 0.482421875, + "learning_rate": 9.02979767286516e-06, + "loss": 0.909, + "step": 278 + }, + { + "epoch": 0.4856396866840731, + "grad_norm": 0.50390625, + "learning_rate": 9.021257922525289e-06, + "loss": 0.9597, + "step": 279 + }, + { + "epoch": 0.48738033072236725, + "grad_norm": 0.5, + "learning_rate": 9.012684828632538e-06, + "loss": 0.9646, + "step": 280 + }, + { + "epoch": 0.48912097476066146, + "grad_norm": 0.50390625, + "learning_rate": 9.004078462273471e-06, + "loss": 0.9679, + "step": 281 + }, + { + "epoch": 0.4908616187989556, + "grad_norm": 0.5078125, + "learning_rate": 8.995438894810541e-06, + "loss": 0.952, + "step": 282 + }, + { + "epoch": 0.4926022628372498, + "grad_norm": 0.484375, + "learning_rate": 8.9867661978815e-06, + "loss": 0.9792, + "step": 283 + }, + { + "epoch": 0.49434290687554394, + "grad_norm": 0.5, + "learning_rate": 8.978060443398802e-06, + "loss": 0.939, + "step": 284 + }, + { + "epoch": 0.4960835509138381, + "grad_norm": 0.486328125, + "learning_rate": 8.96932170354901e-06, + "loss": 0.9974, + "step": 285 + }, + { + "epoch": 0.4978241949521323, + "grad_norm": 0.48828125, + "learning_rate": 8.960550050792194e-06, + "loss": 0.9947, + "step": 286 + }, + { + "epoch": 0.4995648389904265, + "grad_norm": 0.50390625, + "learning_rate": 8.951745557861333e-06, + "loss": 0.9139, + "step": 287 + }, + { + "epoch": 0.5013054830287206, + "grad_norm": 0.49609375, + "learning_rate": 8.942908297761712e-06, + "loss": 0.9397, + "step": 288 + }, + { + "epoch": 0.5013054830287206, + "eval_loss": 0.9357889294624329, + "eval_runtime": 60.8471, + "eval_samples_per_second": 77.046, + "eval_steps_per_second": 9.631, + "step": 288 + }, + { + "epoch": 0.5030461270670148, + "grad_norm": 0.482421875, + "learning_rate": 8.934038343770312e-06, + "loss": 0.9656, + "step": 289 + }, + { + "epoch": 0.504786771105309, + "grad_norm": 0.5078125, + "learning_rate": 8.925135769435211e-06, + "loss": 0.9896, + "step": 290 + }, + { + "epoch": 0.5065274151436031, + "grad_norm": 0.5, + "learning_rate": 8.916200648574964e-06, + "loss": 0.905, + "step": 291 + }, + { + "epoch": 0.5082680591818973, + "grad_norm": 0.490234375, + "learning_rate": 8.907233055277999e-06, + "loss": 0.9309, + "step": 292 + }, + { + "epoch": 0.5100087032201914, + "grad_norm": 0.48046875, + "learning_rate": 8.898233063902e-06, + "loss": 0.9796, + "step": 293 + }, + { + "epoch": 0.5117493472584856, + "grad_norm": 0.5234375, + "learning_rate": 8.889200749073285e-06, + "loss": 0.9335, + "step": 294 + }, + { + "epoch": 0.5134899912967799, + "grad_norm": 0.498046875, + "learning_rate": 8.880136185686202e-06, + "loss": 0.9292, + "step": 295 + }, + { + "epoch": 0.515230635335074, + "grad_norm": 0.478515625, + "learning_rate": 8.871039448902488e-06, + "loss": 0.9116, + "step": 296 + }, + { + "epoch": 0.5169712793733682, + "grad_norm": 0.47265625, + "learning_rate": 8.861910614150662e-06, + "loss": 0.9315, + "step": 297 + }, + { + "epoch": 0.5187119234116623, + "grad_norm": 0.478515625, + "learning_rate": 8.852749757125392e-06, + "loss": 0.9283, + "step": 298 + }, + { + "epoch": 0.5204525674499565, + "grad_norm": 0.5078125, + "learning_rate": 8.843556953786872e-06, + "loss": 0.952, + "step": 299 + }, + { + "epoch": 0.5221932114882507, + "grad_norm": 0.48828125, + "learning_rate": 8.834332280360181e-06, + "loss": 0.9999, + "step": 300 + }, + { + "epoch": 0.5239338555265448, + "grad_norm": 0.498046875, + "learning_rate": 8.82507581333467e-06, + "loss": 0.9453, + "step": 301 + }, + { + "epoch": 0.525674499564839, + "grad_norm": 0.482421875, + "learning_rate": 8.815787629463306e-06, + "loss": 0.8678, + "step": 302 + }, + { + "epoch": 0.5274151436031331, + "grad_norm": 0.5, + "learning_rate": 8.806467805762056e-06, + "loss": 0.9878, + "step": 303 + }, + { + "epoch": 0.5291557876414273, + "grad_norm": 0.49609375, + "learning_rate": 8.797116419509232e-06, + "loss": 0.8964, + "step": 304 + }, + { + "epoch": 0.5308964316797214, + "grad_norm": 0.474609375, + "learning_rate": 8.78773354824486e-06, + "loss": 0.9584, + "step": 305 + }, + { + "epoch": 0.5326370757180157, + "grad_norm": 0.482421875, + "learning_rate": 8.778319269770033e-06, + "loss": 0.9715, + "step": 306 + }, + { + "epoch": 0.5343777197563099, + "grad_norm": 0.484375, + "learning_rate": 8.768873662146271e-06, + "loss": 0.9034, + "step": 307 + }, + { + "epoch": 0.536118363794604, + "grad_norm": 0.494140625, + "learning_rate": 8.759396803694863e-06, + "loss": 0.9189, + "step": 308 + }, + { + "epoch": 0.5378590078328982, + "grad_norm": 0.484375, + "learning_rate": 8.749888772996226e-06, + "loss": 1.0066, + "step": 309 + }, + { + "epoch": 0.5395996518711923, + "grad_norm": 0.47265625, + "learning_rate": 8.74034964888926e-06, + "loss": 1.0147, + "step": 310 + }, + { + "epoch": 0.5413402959094865, + "grad_norm": 0.484375, + "learning_rate": 8.730779510470672e-06, + "loss": 0.9504, + "step": 311 + }, + { + "epoch": 0.5430809399477807, + "grad_norm": 0.482421875, + "learning_rate": 8.721178437094346e-06, + "loss": 0.9239, + "step": 312 + }, + { + "epoch": 0.5448215839860748, + "grad_norm": 0.5078125, + "learning_rate": 8.711546508370666e-06, + "loss": 0.9145, + "step": 313 + }, + { + "epoch": 0.546562228024369, + "grad_norm": 0.498046875, + "learning_rate": 8.701883804165867e-06, + "loss": 0.897, + "step": 314 + }, + { + "epoch": 0.5483028720626631, + "grad_norm": 0.490234375, + "learning_rate": 8.692190404601368e-06, + "loss": 0.8796, + "step": 315 + }, + { + "epoch": 0.5500435161009574, + "grad_norm": 0.53515625, + "learning_rate": 8.682466390053106e-06, + "loss": 1.03, + "step": 316 + }, + { + "epoch": 0.5517841601392516, + "grad_norm": 0.482421875, + "learning_rate": 8.672711841150877e-06, + "loss": 0.9676, + "step": 317 + }, + { + "epoch": 0.5535248041775457, + "grad_norm": 0.490234375, + "learning_rate": 8.662926838777657e-06, + "loss": 1.0098, + "step": 318 + }, + { + "epoch": 0.5552654482158399, + "grad_norm": 0.5078125, + "learning_rate": 8.653111464068937e-06, + "loss": 0.9344, + "step": 319 + }, + { + "epoch": 0.557006092254134, + "grad_norm": 0.51171875, + "learning_rate": 8.643265798412057e-06, + "loss": 0.9555, + "step": 320 + }, + { + "epoch": 0.5587467362924282, + "grad_norm": 0.5, + "learning_rate": 8.633389923445515e-06, + "loss": 0.8845, + "step": 321 + }, + { + "epoch": 0.5604873803307223, + "grad_norm": 0.48046875, + "learning_rate": 8.623483921058304e-06, + "loss": 0.9403, + "step": 322 + }, + { + "epoch": 0.5622280243690165, + "grad_norm": 0.482421875, + "learning_rate": 8.613547873389228e-06, + "loss": 0.9654, + "step": 323 + }, + { + "epoch": 0.5639686684073107, + "grad_norm": 0.490234375, + "learning_rate": 8.603581862826222e-06, + "loss": 1.0108, + "step": 324 + }, + { + "epoch": 0.5657093124456049, + "grad_norm": 0.4921875, + "learning_rate": 8.593585972005665e-06, + "loss": 0.9708, + "step": 325 + }, + { + "epoch": 0.5674499564838991, + "grad_norm": 0.484375, + "learning_rate": 8.5835602838117e-06, + "loss": 0.9333, + "step": 326 + }, + { + "epoch": 0.5691906005221932, + "grad_norm": 0.478515625, + "learning_rate": 8.573504881375543e-06, + "loss": 0.9067, + "step": 327 + }, + { + "epoch": 0.5709312445604874, + "grad_norm": 0.486328125, + "learning_rate": 8.563419848074798e-06, + "loss": 0.9388, + "step": 328 + }, + { + "epoch": 0.5726718885987816, + "grad_norm": 0.4921875, + "learning_rate": 8.55330526753276e-06, + "loss": 0.9246, + "step": 329 + }, + { + "epoch": 0.5744125326370757, + "grad_norm": 0.498046875, + "learning_rate": 8.543161223617724e-06, + "loss": 0.9222, + "step": 330 + }, + { + "epoch": 0.5761531766753699, + "grad_norm": 0.490234375, + "learning_rate": 8.532987800442292e-06, + "loss": 1.0001, + "step": 331 + }, + { + "epoch": 0.577893820713664, + "grad_norm": 0.498046875, + "learning_rate": 8.522785082362675e-06, + "loss": 1.0003, + "step": 332 + }, + { + "epoch": 0.5796344647519582, + "grad_norm": 0.498046875, + "learning_rate": 8.512553153977988e-06, + "loss": 0.9103, + "step": 333 + }, + { + "epoch": 0.5813751087902524, + "grad_norm": 0.498046875, + "learning_rate": 8.502292100129553e-06, + "loss": 0.9895, + "step": 334 + }, + { + "epoch": 0.5831157528285466, + "grad_norm": 0.4921875, + "learning_rate": 8.492002005900201e-06, + "loss": 0.9981, + "step": 335 + }, + { + "epoch": 0.5848563968668408, + "grad_norm": 0.4765625, + "learning_rate": 8.481682956613555e-06, + "loss": 0.8758, + "step": 336 + }, + { + "epoch": 0.5865970409051349, + "grad_norm": 0.49609375, + "learning_rate": 8.471335037833328e-06, + "loss": 0.9845, + "step": 337 + }, + { + "epoch": 0.5883376849434291, + "grad_norm": 0.474609375, + "learning_rate": 8.460958335362617e-06, + "loss": 0.9974, + "step": 338 + }, + { + "epoch": 0.5900783289817232, + "grad_norm": 0.4921875, + "learning_rate": 8.450552935243186e-06, + "loss": 0.9725, + "step": 339 + }, + { + "epoch": 0.5918189730200174, + "grad_norm": 0.48046875, + "learning_rate": 8.440118923754757e-06, + "loss": 0.9362, + "step": 340 + }, + { + "epoch": 0.5935596170583116, + "grad_norm": 0.47265625, + "learning_rate": 8.429656387414289e-06, + "loss": 0.9659, + "step": 341 + }, + { + "epoch": 0.5953002610966057, + "grad_norm": 0.5078125, + "learning_rate": 8.419165412975265e-06, + "loss": 0.9346, + "step": 342 + }, + { + "epoch": 0.5970409051348999, + "grad_norm": 0.48828125, + "learning_rate": 8.408646087426975e-06, + "loss": 0.8759, + "step": 343 + }, + { + "epoch": 0.598781549173194, + "grad_norm": 0.5, + "learning_rate": 8.398098497993785e-06, + "loss": 0.9359, + "step": 344 + }, + { + "epoch": 0.6005221932114883, + "grad_norm": 0.5, + "learning_rate": 8.387522732134428e-06, + "loss": 0.9171, + "step": 345 + }, + { + "epoch": 0.6022628372497825, + "grad_norm": 0.49609375, + "learning_rate": 8.376918877541263e-06, + "loss": 0.919, + "step": 346 + }, + { + "epoch": 0.6040034812880766, + "grad_norm": 0.5234375, + "learning_rate": 8.36628702213956e-06, + "loss": 0.9211, + "step": 347 + }, + { + "epoch": 0.6057441253263708, + "grad_norm": 0.494140625, + "learning_rate": 8.355627254086771e-06, + "loss": 0.9191, + "step": 348 + }, + { + "epoch": 0.6074847693646649, + "grad_norm": 0.49609375, + "learning_rate": 8.344939661771784e-06, + "loss": 0.926, + "step": 349 + }, + { + "epoch": 0.6092254134029591, + "grad_norm": 0.5, + "learning_rate": 8.334224333814209e-06, + "loss": 0.9088, + "step": 350 + }, + { + "epoch": 0.6109660574412533, + "grad_norm": 0.50390625, + "learning_rate": 8.323481359063631e-06, + "loss": 0.9455, + "step": 351 + }, + { + "epoch": 0.6127067014795474, + "grad_norm": 0.50390625, + "learning_rate": 8.312710826598884e-06, + "loss": 0.8873, + "step": 352 + }, + { + "epoch": 0.6144473455178416, + "grad_norm": 0.49609375, + "learning_rate": 8.301912825727294e-06, + "loss": 0.9562, + "step": 353 + }, + { + "epoch": 0.6161879895561357, + "grad_norm": 0.48828125, + "learning_rate": 8.29108744598396e-06, + "loss": 0.9324, + "step": 354 + }, + { + "epoch": 0.61792863359443, + "grad_norm": 0.51171875, + "learning_rate": 8.280234777131e-06, + "loss": 0.9037, + "step": 355 + }, + { + "epoch": 0.6196692776327241, + "grad_norm": 0.5078125, + "learning_rate": 8.269354909156803e-06, + "loss": 0.9127, + "step": 356 + }, + { + "epoch": 0.6214099216710183, + "grad_norm": 0.494140625, + "learning_rate": 8.258447932275296e-06, + "loss": 0.9003, + "step": 357 + }, + { + "epoch": 0.6231505657093125, + "grad_norm": 0.494140625, + "learning_rate": 8.247513936925182e-06, + "loss": 0.9871, + "step": 358 + }, + { + "epoch": 0.6248912097476066, + "grad_norm": 0.50390625, + "learning_rate": 8.236553013769198e-06, + "loss": 0.8732, + "step": 359 + }, + { + "epoch": 0.6266318537859008, + "grad_norm": 0.486328125, + "learning_rate": 8.225565253693365e-06, + "loss": 0.9563, + "step": 360 + }, + { + "epoch": 0.6266318537859008, + "eval_loss": 0.9300395846366882, + "eval_runtime": 59.3628, + "eval_samples_per_second": 78.972, + "eval_steps_per_second": 9.871, + "step": 360 + }, + { + "epoch": 0.6283724978241949, + "grad_norm": 0.5078125, + "learning_rate": 8.214550747806227e-06, + "loss": 0.9173, + "step": 361 + }, + { + "epoch": 0.6301131418624891, + "grad_norm": 0.494140625, + "learning_rate": 8.2035095874381e-06, + "loss": 0.9297, + "step": 362 + }, + { + "epoch": 0.6318537859007833, + "grad_norm": 0.50390625, + "learning_rate": 8.192441864140314e-06, + "loss": 0.9582, + "step": 363 + }, + { + "epoch": 0.6335944299390774, + "grad_norm": 0.50390625, + "learning_rate": 8.181347669684456e-06, + "loss": 0.9176, + "step": 364 + }, + { + "epoch": 0.6353350739773717, + "grad_norm": 0.494140625, + "learning_rate": 8.170227096061607e-06, + "loss": 0.9181, + "step": 365 + }, + { + "epoch": 0.6370757180156658, + "grad_norm": 0.486328125, + "learning_rate": 8.15908023548158e-06, + "loss": 0.9708, + "step": 366 + }, + { + "epoch": 0.63881636205396, + "grad_norm": 0.48046875, + "learning_rate": 8.147907180372147e-06, + "loss": 0.9683, + "step": 367 + }, + { + "epoch": 0.6405570060922542, + "grad_norm": 0.4765625, + "learning_rate": 8.136708023378292e-06, + "loss": 0.9728, + "step": 368 + }, + { + "epoch": 0.6422976501305483, + "grad_norm": 0.498046875, + "learning_rate": 8.125482857361426e-06, + "loss": 0.9696, + "step": 369 + }, + { + "epoch": 0.6440382941688425, + "grad_norm": 0.49609375, + "learning_rate": 8.114231775398618e-06, + "loss": 0.9189, + "step": 370 + }, + { + "epoch": 0.6457789382071366, + "grad_norm": 0.486328125, + "learning_rate": 8.102954870781831e-06, + "loss": 0.9361, + "step": 371 + }, + { + "epoch": 0.6475195822454308, + "grad_norm": 0.490234375, + "learning_rate": 8.091652237017152e-06, + "loss": 0.9612, + "step": 372 + }, + { + "epoch": 0.6492602262837249, + "grad_norm": 0.498046875, + "learning_rate": 8.080323967823993e-06, + "loss": 0.9442, + "step": 373 + }, + { + "epoch": 0.6510008703220191, + "grad_norm": 0.48828125, + "learning_rate": 8.068970157134349e-06, + "loss": 0.9346, + "step": 374 + }, + { + "epoch": 0.6527415143603134, + "grad_norm": 0.486328125, + "learning_rate": 8.057590899091985e-06, + "loss": 0.8999, + "step": 375 + }, + { + "epoch": 0.6544821583986075, + "grad_norm": 0.498046875, + "learning_rate": 8.046186288051681e-06, + "loss": 0.9674, + "step": 376 + }, + { + "epoch": 0.6562228024369017, + "grad_norm": 0.48046875, + "learning_rate": 8.034756418578434e-06, + "loss": 0.9364, + "step": 377 + }, + { + "epoch": 0.6579634464751958, + "grad_norm": 0.484375, + "learning_rate": 8.023301385446682e-06, + "loss": 0.9341, + "step": 378 + }, + { + "epoch": 0.65970409051349, + "grad_norm": 0.4921875, + "learning_rate": 8.011821283639515e-06, + "loss": 0.8986, + "step": 379 + }, + { + "epoch": 0.6614447345517842, + "grad_norm": 0.515625, + "learning_rate": 8.000316208347891e-06, + "loss": 0.9266, + "step": 380 + }, + { + "epoch": 0.6631853785900783, + "grad_norm": 0.5078125, + "learning_rate": 7.988786254969837e-06, + "loss": 0.8911, + "step": 381 + }, + { + "epoch": 0.6649260226283725, + "grad_norm": 0.5078125, + "learning_rate": 7.977231519109665e-06, + "loss": 0.9221, + "step": 382 + }, + { + "epoch": 0.6666666666666666, + "grad_norm": 0.466796875, + "learning_rate": 7.965652096577188e-06, + "loss": 0.9635, + "step": 383 + }, + { + "epoch": 0.6684073107049608, + "grad_norm": 0.48828125, + "learning_rate": 7.954048083386909e-06, + "loss": 0.9023, + "step": 384 + }, + { + "epoch": 0.6701479547432551, + "grad_norm": 0.48828125, + "learning_rate": 7.942419575757235e-06, + "loss": 0.9277, + "step": 385 + }, + { + "epoch": 0.6718885987815492, + "grad_norm": 0.486328125, + "learning_rate": 7.930766670109675e-06, + "loss": 0.9826, + "step": 386 + }, + { + "epoch": 0.6736292428198434, + "grad_norm": 0.490234375, + "learning_rate": 7.919089463068038e-06, + "loss": 0.9223, + "step": 387 + }, + { + "epoch": 0.6753698868581375, + "grad_norm": 0.51953125, + "learning_rate": 7.907388051457647e-06, + "loss": 1.0468, + "step": 388 + }, + { + "epoch": 0.6771105308964317, + "grad_norm": 0.4921875, + "learning_rate": 7.895662532304516e-06, + "loss": 0.9233, + "step": 389 + }, + { + "epoch": 0.6788511749347258, + "grad_norm": 0.478515625, + "learning_rate": 7.883913002834555e-06, + "loss": 0.9756, + "step": 390 + }, + { + "epoch": 0.68059181897302, + "grad_norm": 0.494140625, + "learning_rate": 7.872139560472767e-06, + "loss": 0.9, + "step": 391 + }, + { + "epoch": 0.6823324630113142, + "grad_norm": 0.51953125, + "learning_rate": 7.860342302842434e-06, + "loss": 0.894, + "step": 392 + }, + { + "epoch": 0.6840731070496083, + "grad_norm": 0.5, + "learning_rate": 7.848521327764309e-06, + "loss": 0.9369, + "step": 393 + }, + { + "epoch": 0.6858137510879025, + "grad_norm": 0.5, + "learning_rate": 7.836676733255809e-06, + "loss": 0.9416, + "step": 394 + }, + { + "epoch": 0.6875543951261966, + "grad_norm": 0.494140625, + "learning_rate": 7.824808617530197e-06, + "loss": 0.8992, + "step": 395 + }, + { + "epoch": 0.6892950391644909, + "grad_norm": 0.486328125, + "learning_rate": 7.812917078995769e-06, + "loss": 0.9457, + "step": 396 + }, + { + "epoch": 0.6910356832027851, + "grad_norm": 0.4765625, + "learning_rate": 7.801002216255042e-06, + "loss": 0.9484, + "step": 397 + }, + { + "epoch": 0.6927763272410792, + "grad_norm": 0.50390625, + "learning_rate": 7.78906412810393e-06, + "loss": 0.9297, + "step": 398 + }, + { + "epoch": 0.6945169712793734, + "grad_norm": 0.498046875, + "learning_rate": 7.777102913530927e-06, + "loss": 0.951, + "step": 399 + }, + { + "epoch": 0.6962576153176675, + "grad_norm": 0.4921875, + "learning_rate": 7.76511867171629e-06, + "loss": 0.9091, + "step": 400 + }, + { + "epoch": 0.6979982593559617, + "grad_norm": 0.49609375, + "learning_rate": 7.753111502031214e-06, + "loss": 0.9041, + "step": 401 + }, + { + "epoch": 0.6997389033942559, + "grad_norm": 0.50390625, + "learning_rate": 7.741081504037009e-06, + "loss": 0.9463, + "step": 402 + }, + { + "epoch": 0.70147954743255, + "grad_norm": 0.486328125, + "learning_rate": 7.729028777484266e-06, + "loss": 0.9494, + "step": 403 + }, + { + "epoch": 0.7032201914708442, + "grad_norm": 0.482421875, + "learning_rate": 7.716953422312044e-06, + "loss": 0.9266, + "step": 404 + }, + { + "epoch": 0.7049608355091384, + "grad_norm": 0.5078125, + "learning_rate": 7.704855538647033e-06, + "loss": 0.9731, + "step": 405 + }, + { + "epoch": 0.7067014795474326, + "grad_norm": 0.490234375, + "learning_rate": 7.692735226802729e-06, + "loss": 0.961, + "step": 406 + }, + { + "epoch": 0.7084421235857267, + "grad_norm": 0.482421875, + "learning_rate": 7.680592587278585e-06, + "loss": 0.9506, + "step": 407 + }, + { + "epoch": 0.7101827676240209, + "grad_norm": 0.5, + "learning_rate": 7.668427720759207e-06, + "loss": 0.9084, + "step": 408 + }, + { + "epoch": 0.7119234116623151, + "grad_norm": 0.5, + "learning_rate": 7.656240728113493e-06, + "loss": 0.9147, + "step": 409 + }, + { + "epoch": 0.7136640557006092, + "grad_norm": 0.4921875, + "learning_rate": 7.644031710393815e-06, + "loss": 0.9498, + "step": 410 + }, + { + "epoch": 0.7154046997389034, + "grad_norm": 0.490234375, + "learning_rate": 7.631800768835167e-06, + "loss": 0.9145, + "step": 411 + }, + { + "epoch": 0.7171453437771975, + "grad_norm": 0.498046875, + "learning_rate": 7.619548004854332e-06, + "loss": 0.8884, + "step": 412 + }, + { + "epoch": 0.7188859878154917, + "grad_norm": 0.48828125, + "learning_rate": 7.607273520049041e-06, + "loss": 0.9777, + "step": 413 + }, + { + "epoch": 0.720626631853786, + "grad_norm": 0.478515625, + "learning_rate": 7.594977416197134e-06, + "loss": 0.9802, + "step": 414 + }, + { + "epoch": 0.72236727589208, + "grad_norm": 0.49609375, + "learning_rate": 7.582659795255707e-06, + "loss": 0.9484, + "step": 415 + }, + { + "epoch": 0.7241079199303743, + "grad_norm": 0.5078125, + "learning_rate": 7.570320759360273e-06, + "loss": 0.9231, + "step": 416 + }, + { + "epoch": 0.7258485639686684, + "grad_norm": 0.498046875, + "learning_rate": 7.557960410823917e-06, + "loss": 0.9416, + "step": 417 + }, + { + "epoch": 0.7275892080069626, + "grad_norm": 0.4921875, + "learning_rate": 7.545578852136443e-06, + "loss": 1.0058, + "step": 418 + }, + { + "epoch": 0.7293298520452568, + "grad_norm": 0.5, + "learning_rate": 7.533176185963523e-06, + "loss": 0.9426, + "step": 419 + }, + { + "epoch": 0.7310704960835509, + "grad_norm": 0.498046875, + "learning_rate": 7.520752515145855e-06, + "loss": 0.949, + "step": 420 + }, + { + "epoch": 0.7328111401218451, + "grad_norm": 0.484375, + "learning_rate": 7.508307942698296e-06, + "loss": 0.9727, + "step": 421 + }, + { + "epoch": 0.7345517841601392, + "grad_norm": 0.515625, + "learning_rate": 7.495842571809021e-06, + "loss": 0.8763, + "step": 422 + }, + { + "epoch": 0.7362924281984334, + "grad_norm": 0.5, + "learning_rate": 7.4833565058386595e-06, + "loss": 0.8914, + "step": 423 + }, + { + "epoch": 0.7380330722367275, + "grad_norm": 0.49609375, + "learning_rate": 7.470849848319443e-06, + "loss": 0.9157, + "step": 424 + }, + { + "epoch": 0.7397737162750218, + "grad_norm": 0.490234375, + "learning_rate": 7.458322702954342e-06, + "loss": 0.8703, + "step": 425 + }, + { + "epoch": 0.741514360313316, + "grad_norm": 0.48828125, + "learning_rate": 7.44577517361621e-06, + "loss": 0.9106, + "step": 426 + }, + { + "epoch": 0.7432550043516101, + "grad_norm": 0.5078125, + "learning_rate": 7.4332073643469196e-06, + "loss": 0.8928, + "step": 427 + }, + { + "epoch": 0.7449956483899043, + "grad_norm": 0.5234375, + "learning_rate": 7.420619379356504e-06, + "loss": 0.9402, + "step": 428 + }, + { + "epoch": 0.7467362924281984, + "grad_norm": 0.490234375, + "learning_rate": 7.408011323022286e-06, + "loss": 0.8997, + "step": 429 + }, + { + "epoch": 0.7484769364664926, + "grad_norm": 0.48828125, + "learning_rate": 7.395383299888019e-06, + "loss": 0.9831, + "step": 430 + }, + { + "epoch": 0.7502175805047868, + "grad_norm": 0.494140625, + "learning_rate": 7.382735414663017e-06, + "loss": 1.04, + "step": 431 + }, + { + "epoch": 0.7519582245430809, + "grad_norm": 0.49609375, + "learning_rate": 7.370067772221285e-06, + "loss": 0.9034, + "step": 432 + }, + { + "epoch": 0.7519582245430809, + "eval_loss": 0.9259106516838074, + "eval_runtime": 59.5681, + "eval_samples_per_second": 78.7, + "eval_steps_per_second": 9.837, + "step": 432 + }, + { + "epoch": 0.7536988685813751, + "grad_norm": 0.4921875, + "learning_rate": 7.357380477600654e-06, + "loss": 0.9201, + "step": 433 + }, + { + "epoch": 0.7554395126196692, + "grad_norm": 0.5078125, + "learning_rate": 7.3446736360019065e-06, + "loss": 0.9291, + "step": 434 + }, + { + "epoch": 0.7571801566579635, + "grad_norm": 0.47265625, + "learning_rate": 7.331947352787905e-06, + "loss": 0.951, + "step": 435 + }, + { + "epoch": 0.7589208006962577, + "grad_norm": 0.49609375, + "learning_rate": 7.319201733482715e-06, + "loss": 0.9208, + "step": 436 + }, + { + "epoch": 0.7606614447345518, + "grad_norm": 0.490234375, + "learning_rate": 7.3064368837707425e-06, + "loss": 0.9116, + "step": 437 + }, + { + "epoch": 0.762402088772846, + "grad_norm": 0.50390625, + "learning_rate": 7.2936529094958365e-06, + "loss": 0.9165, + "step": 438 + }, + { + "epoch": 0.7641427328111401, + "grad_norm": 0.494140625, + "learning_rate": 7.280849916660434e-06, + "loss": 0.9513, + "step": 439 + }, + { + "epoch": 0.7658833768494343, + "grad_norm": 0.48828125, + "learning_rate": 7.268028011424664e-06, + "loss": 0.9725, + "step": 440 + }, + { + "epoch": 0.7676240208877284, + "grad_norm": 0.494140625, + "learning_rate": 7.255187300105477e-06, + "loss": 0.9328, + "step": 441 + }, + { + "epoch": 0.7693646649260226, + "grad_norm": 0.50390625, + "learning_rate": 7.24232788917576e-06, + "loss": 0.8776, + "step": 442 + }, + { + "epoch": 0.7711053089643168, + "grad_norm": 0.5078125, + "learning_rate": 7.229449885263451e-06, + "loss": 0.9432, + "step": 443 + }, + { + "epoch": 0.7728459530026109, + "grad_norm": 0.478515625, + "learning_rate": 7.21655339515066e-06, + "loss": 0.8951, + "step": 444 + }, + { + "epoch": 0.7745865970409052, + "grad_norm": 0.4921875, + "learning_rate": 7.203638525772783e-06, + "loss": 0.961, + "step": 445 + }, + { + "epoch": 0.7763272410791993, + "grad_norm": 0.48828125, + "learning_rate": 7.1907053842176075e-06, + "loss": 0.9498, + "step": 446 + }, + { + "epoch": 0.7780678851174935, + "grad_norm": 0.498046875, + "learning_rate": 7.17775407772444e-06, + "loss": 0.9638, + "step": 447 + }, + { + "epoch": 0.7798085291557877, + "grad_norm": 0.490234375, + "learning_rate": 7.164784713683197e-06, + "loss": 0.9516, + "step": 448 + }, + { + "epoch": 0.7815491731940818, + "grad_norm": 0.466796875, + "learning_rate": 7.1517973996335335e-06, + "loss": 0.9108, + "step": 449 + }, + { + "epoch": 0.783289817232376, + "grad_norm": 0.50390625, + "learning_rate": 7.138792243263936e-06, + "loss": 0.9089, + "step": 450 + }, + { + "epoch": 0.7850304612706701, + "grad_norm": 0.50390625, + "learning_rate": 7.125769352410845e-06, + "loss": 0.9667, + "step": 451 + }, + { + "epoch": 0.7867711053089643, + "grad_norm": 0.484375, + "learning_rate": 7.112728835057742e-06, + "loss": 0.9458, + "step": 452 + }, + { + "epoch": 0.7885117493472585, + "grad_norm": 0.478515625, + "learning_rate": 7.099670799334269e-06, + "loss": 0.899, + "step": 453 + }, + { + "epoch": 0.7902523933855526, + "grad_norm": 0.498046875, + "learning_rate": 7.08659535351533e-06, + "loss": 0.9031, + "step": 454 + }, + { + "epoch": 0.7919930374238469, + "grad_norm": 0.49609375, + "learning_rate": 7.073502606020187e-06, + "loss": 0.9771, + "step": 455 + }, + { + "epoch": 0.793733681462141, + "grad_norm": 0.52734375, + "learning_rate": 7.060392665411564e-06, + "loss": 0.9113, + "step": 456 + }, + { + "epoch": 0.7954743255004352, + "grad_norm": 0.498046875, + "learning_rate": 7.0472656403947505e-06, + "loss": 1.0172, + "step": 457 + }, + { + "epoch": 0.7972149695387293, + "grad_norm": 0.4765625, + "learning_rate": 7.034121639816691e-06, + "loss": 0.9282, + "step": 458 + }, + { + "epoch": 0.7989556135770235, + "grad_norm": 0.50390625, + "learning_rate": 7.020960772665096e-06, + "loss": 0.896, + "step": 459 + }, + { + "epoch": 0.8006962576153177, + "grad_norm": 0.490234375, + "learning_rate": 7.007783148067524e-06, + "loss": 0.881, + "step": 460 + }, + { + "epoch": 0.8024369016536118, + "grad_norm": 0.5078125, + "learning_rate": 6.994588875290488e-06, + "loss": 0.9155, + "step": 461 + }, + { + "epoch": 0.804177545691906, + "grad_norm": 0.5078125, + "learning_rate": 6.9813780637385385e-06, + "loss": 0.9012, + "step": 462 + }, + { + "epoch": 0.8059181897302001, + "grad_norm": 0.50390625, + "learning_rate": 6.968150822953372e-06, + "loss": 0.9085, + "step": 463 + }, + { + "epoch": 0.8076588337684943, + "grad_norm": 0.498046875, + "learning_rate": 6.954907262612906e-06, + "loss": 0.9818, + "step": 464 + }, + { + "epoch": 0.8093994778067886, + "grad_norm": 0.484375, + "learning_rate": 6.941647492530378e-06, + "loss": 0.9717, + "step": 465 + }, + { + "epoch": 0.8111401218450827, + "grad_norm": 0.5, + "learning_rate": 6.928371622653434e-06, + "loss": 0.9369, + "step": 466 + }, + { + "epoch": 0.8128807658833769, + "grad_norm": 0.486328125, + "learning_rate": 6.91507976306322e-06, + "loss": 0.8943, + "step": 467 + }, + { + "epoch": 0.814621409921671, + "grad_norm": 0.498046875, + "learning_rate": 6.901772023973459e-06, + "loss": 0.9113, + "step": 468 + }, + { + "epoch": 0.8163620539599652, + "grad_norm": 0.49609375, + "learning_rate": 6.888448515729552e-06, + "loss": 0.8995, + "step": 469 + }, + { + "epoch": 0.8181026979982594, + "grad_norm": 0.486328125, + "learning_rate": 6.8751093488076485e-06, + "loss": 0.9159, + "step": 470 + }, + { + "epoch": 0.8198433420365535, + "grad_norm": 0.498046875, + "learning_rate": 6.86175463381374e-06, + "loss": 0.9388, + "step": 471 + }, + { + "epoch": 0.8215839860748477, + "grad_norm": 0.49609375, + "learning_rate": 6.8483844814827405e-06, + "loss": 0.9524, + "step": 472 + }, + { + "epoch": 0.8233246301131418, + "grad_norm": 0.5078125, + "learning_rate": 6.8349990026775656e-06, + "loss": 0.9458, + "step": 473 + }, + { + "epoch": 0.825065274151436, + "grad_norm": 0.5, + "learning_rate": 6.821598308388217e-06, + "loss": 0.9258, + "step": 474 + }, + { + "epoch": 0.8268059181897301, + "grad_norm": 0.482421875, + "learning_rate": 6.8081825097308584e-06, + "loss": 0.9159, + "step": 475 + }, + { + "epoch": 0.8285465622280244, + "grad_norm": 0.5, + "learning_rate": 6.794751717946897e-06, + "loss": 0.9029, + "step": 476 + }, + { + "epoch": 0.8302872062663186, + "grad_norm": 0.484375, + "learning_rate": 6.781306044402064e-06, + "loss": 0.8852, + "step": 477 + }, + { + "epoch": 0.8320278503046127, + "grad_norm": 0.482421875, + "learning_rate": 6.767845600585479e-06, + "loss": 0.945, + "step": 478 + }, + { + "epoch": 0.8337684943429069, + "grad_norm": 0.474609375, + "learning_rate": 6.754370498108747e-06, + "loss": 0.9776, + "step": 479 + }, + { + "epoch": 0.835509138381201, + "grad_norm": 0.490234375, + "learning_rate": 6.740880848705005e-06, + "loss": 0.9229, + "step": 480 + }, + { + "epoch": 0.8372497824194952, + "grad_norm": 0.50390625, + "learning_rate": 6.72737676422802e-06, + "loss": 0.9686, + "step": 481 + }, + { + "epoch": 0.8389904264577894, + "grad_norm": 0.484375, + "learning_rate": 6.713858356651253e-06, + "loss": 0.9459, + "step": 482 + }, + { + "epoch": 0.8407310704960835, + "grad_norm": 0.486328125, + "learning_rate": 6.700325738066923e-06, + "loss": 0.9472, + "step": 483 + }, + { + "epoch": 0.8424717145343777, + "grad_norm": 0.5, + "learning_rate": 6.686779020685089e-06, + "loss": 0.9417, + "step": 484 + }, + { + "epoch": 0.8442123585726719, + "grad_norm": 0.490234375, + "learning_rate": 6.6732183168327146e-06, + "loss": 0.9666, + "step": 485 + }, + { + "epoch": 0.8459530026109661, + "grad_norm": 0.490234375, + "learning_rate": 6.659643738952732e-06, + "loss": 0.9368, + "step": 486 + }, + { + "epoch": 0.8476936466492603, + "grad_norm": 0.484375, + "learning_rate": 6.646055399603122e-06, + "loss": 0.9466, + "step": 487 + }, + { + "epoch": 0.8494342906875544, + "grad_norm": 0.5, + "learning_rate": 6.6324534114559656e-06, + "loss": 0.9159, + "step": 488 + }, + { + "epoch": 0.8511749347258486, + "grad_norm": 0.5078125, + "learning_rate": 6.618837887296523e-06, + "loss": 0.9067, + "step": 489 + }, + { + "epoch": 0.8529155787641427, + "grad_norm": 0.48046875, + "learning_rate": 6.605208940022289e-06, + "loss": 0.9608, + "step": 490 + }, + { + "epoch": 0.8546562228024369, + "grad_norm": 0.498046875, + "learning_rate": 6.591566682642061e-06, + "loss": 0.9663, + "step": 491 + }, + { + "epoch": 0.856396866840731, + "grad_norm": 0.5234375, + "learning_rate": 6.5779112282750035e-06, + "loss": 0.8931, + "step": 492 + }, + { + "epoch": 0.8581375108790252, + "grad_norm": 0.498046875, + "learning_rate": 6.564242690149705e-06, + "loss": 0.9383, + "step": 493 + }, + { + "epoch": 0.8598781549173194, + "grad_norm": 0.484375, + "learning_rate": 6.550561181603244e-06, + "loss": 0.9343, + "step": 494 + }, + { + "epoch": 0.8616187989556136, + "grad_norm": 0.486328125, + "learning_rate": 6.536866816080247e-06, + "loss": 0.9999, + "step": 495 + }, + { + "epoch": 0.8633594429939078, + "grad_norm": 0.4765625, + "learning_rate": 6.523159707131951e-06, + "loss": 0.9955, + "step": 496 + }, + { + "epoch": 0.8651000870322019, + "grad_norm": 0.50390625, + "learning_rate": 6.509439968415252e-06, + "loss": 0.8897, + "step": 497 + }, + { + "epoch": 0.8668407310704961, + "grad_norm": 0.498046875, + "learning_rate": 6.4957077136917776e-06, + "loss": 0.9065, + "step": 498 + }, + { + "epoch": 0.8685813751087903, + "grad_norm": 0.4921875, + "learning_rate": 6.481963056826932e-06, + "loss": 0.9224, + "step": 499 + }, + { + "epoch": 0.8703220191470844, + "grad_norm": 0.50390625, + "learning_rate": 6.468206111788957e-06, + "loss": 0.9638, + "step": 500 + }, + { + "epoch": 0.8720626631853786, + "grad_norm": 0.5, + "learning_rate": 6.454436992647984e-06, + "loss": 0.9232, + "step": 501 + }, + { + "epoch": 0.8738033072236727, + "grad_norm": 0.51171875, + "learning_rate": 6.440655813575093e-06, + "loss": 0.9412, + "step": 502 + }, + { + "epoch": 0.8755439512619669, + "grad_norm": 0.5078125, + "learning_rate": 6.426862688841359e-06, + "loss": 0.9051, + "step": 503 + }, + { + "epoch": 0.8772845953002611, + "grad_norm": 0.5078125, + "learning_rate": 6.413057732816911e-06, + "loss": 0.9214, + "step": 504 + }, + { + "epoch": 0.8772845953002611, + "eval_loss": 0.9230473637580872, + "eval_runtime": 59.2532, + "eval_samples_per_second": 79.118, + "eval_steps_per_second": 9.89, + "step": 504 + }, + { + "epoch": 0.8790252393385553, + "grad_norm": 0.49609375, + "learning_rate": 6.3992410599699786e-06, + "loss": 0.908, + "step": 505 + }, + { + "epoch": 0.8807658833768495, + "grad_norm": 0.48828125, + "learning_rate": 6.385412784865948e-06, + "loss": 0.9613, + "step": 506 + }, + { + "epoch": 0.8825065274151436, + "grad_norm": 0.498046875, + "learning_rate": 6.371573022166409e-06, + "loss": 0.9061, + "step": 507 + }, + { + "epoch": 0.8842471714534378, + "grad_norm": 0.5, + "learning_rate": 6.357721886628201e-06, + "loss": 0.9117, + "step": 508 + }, + { + "epoch": 0.8859878154917319, + "grad_norm": 0.48046875, + "learning_rate": 6.34385949310247e-06, + "loss": 0.9219, + "step": 509 + }, + { + "epoch": 0.8877284595300261, + "grad_norm": 0.5078125, + "learning_rate": 6.329985956533708e-06, + "loss": 0.976, + "step": 510 + }, + { + "epoch": 0.8894691035683203, + "grad_norm": 0.515625, + "learning_rate": 6.3161013919588e-06, + "loss": 0.9131, + "step": 511 + }, + { + "epoch": 0.8912097476066144, + "grad_norm": 0.5, + "learning_rate": 6.302205914506083e-06, + "loss": 0.908, + "step": 512 + }, + { + "epoch": 0.8929503916449086, + "grad_norm": 0.4921875, + "learning_rate": 6.2882996393943706e-06, + "loss": 0.9786, + "step": 513 + }, + { + "epoch": 0.8946910356832027, + "grad_norm": 0.48828125, + "learning_rate": 6.274382681932019e-06, + "loss": 0.8728, + "step": 514 + }, + { + "epoch": 0.896431679721497, + "grad_norm": 0.482421875, + "learning_rate": 6.2604551575159476e-06, + "loss": 0.9491, + "step": 515 + }, + { + "epoch": 0.8981723237597912, + "grad_norm": 0.49609375, + "learning_rate": 6.24651718163071e-06, + "loss": 0.995, + "step": 516 + }, + { + "epoch": 0.8999129677980853, + "grad_norm": 0.482421875, + "learning_rate": 6.2325688698475106e-06, + "loss": 0.9574, + "step": 517 + }, + { + "epoch": 0.9016536118363795, + "grad_norm": 0.486328125, + "learning_rate": 6.218610337823262e-06, + "loss": 0.9004, + "step": 518 + }, + { + "epoch": 0.9033942558746736, + "grad_norm": 0.48046875, + "learning_rate": 6.2046417012996195e-06, + "loss": 0.9165, + "step": 519 + }, + { + "epoch": 0.9051348999129678, + "grad_norm": 0.498046875, + "learning_rate": 6.1906630761020245e-06, + "loss": 0.9534, + "step": 520 + }, + { + "epoch": 0.906875543951262, + "grad_norm": 0.5, + "learning_rate": 6.17667457813874e-06, + "loss": 0.8883, + "step": 521 + }, + { + "epoch": 0.9086161879895561, + "grad_norm": 0.51171875, + "learning_rate": 6.162676323399898e-06, + "loss": 0.919, + "step": 522 + }, + { + "epoch": 0.9103568320278503, + "grad_norm": 0.4921875, + "learning_rate": 6.148668427956523e-06, + "loss": 0.9187, + "step": 523 + }, + { + "epoch": 0.9120974760661444, + "grad_norm": 0.494140625, + "learning_rate": 6.134651007959586e-06, + "loss": 0.8984, + "step": 524 + }, + { + "epoch": 0.9138381201044387, + "grad_norm": 0.4921875, + "learning_rate": 6.120624179639032e-06, + "loss": 0.9585, + "step": 525 + }, + { + "epoch": 0.9155787641427328, + "grad_norm": 0.49609375, + "learning_rate": 6.106588059302818e-06, + "loss": 0.9848, + "step": 526 + }, + { + "epoch": 0.917319408181027, + "grad_norm": 0.4921875, + "learning_rate": 6.092542763335947e-06, + "loss": 0.9542, + "step": 527 + }, + { + "epoch": 0.9190600522193212, + "grad_norm": 0.5, + "learning_rate": 6.0784884081995065e-06, + "loss": 0.9476, + "step": 528 + }, + { + "epoch": 0.9208006962576153, + "grad_norm": 0.498046875, + "learning_rate": 6.0644251104296995e-06, + "loss": 0.9425, + "step": 529 + }, + { + "epoch": 0.9225413402959095, + "grad_norm": 0.5, + "learning_rate": 6.0503529866368824e-06, + "loss": 0.9532, + "step": 530 + }, + { + "epoch": 0.9242819843342036, + "grad_norm": 0.49609375, + "learning_rate": 6.036272153504592e-06, + "loss": 0.9243, + "step": 531 + }, + { + "epoch": 0.9260226283724978, + "grad_norm": 0.490234375, + "learning_rate": 6.022182727788586e-06, + "loss": 0.9452, + "step": 532 + }, + { + "epoch": 0.927763272410792, + "grad_norm": 0.5234375, + "learning_rate": 6.008084826315863e-06, + "loss": 0.9243, + "step": 533 + }, + { + "epoch": 0.9295039164490861, + "grad_norm": 0.498046875, + "learning_rate": 5.993978565983709e-06, + "loss": 0.8917, + "step": 534 + }, + { + "epoch": 0.9312445604873804, + "grad_norm": 0.4921875, + "learning_rate": 5.979864063758717e-06, + "loss": 0.9562, + "step": 535 + }, + { + "epoch": 0.9329852045256745, + "grad_norm": 0.478515625, + "learning_rate": 5.965741436675816e-06, + "loss": 0.9904, + "step": 536 + }, + { + "epoch": 0.9347258485639687, + "grad_norm": 0.5, + "learning_rate": 5.9516108018373145e-06, + "loss": 0.8953, + "step": 537 + }, + { + "epoch": 0.9364664926022629, + "grad_norm": 0.494140625, + "learning_rate": 5.937472276411909e-06, + "loss": 0.8674, + "step": 538 + }, + { + "epoch": 0.938207136640557, + "grad_norm": 0.498046875, + "learning_rate": 5.923325977633732e-06, + "loss": 1.0281, + "step": 539 + }, + { + "epoch": 0.9399477806788512, + "grad_norm": 0.494140625, + "learning_rate": 5.909172022801364e-06, + "loss": 0.9278, + "step": 540 + }, + { + "epoch": 0.9416884247171453, + "grad_norm": 0.486328125, + "learning_rate": 5.8950105292768754e-06, + "loss": 0.8704, + "step": 541 + }, + { + "epoch": 0.9434290687554395, + "grad_norm": 0.515625, + "learning_rate": 5.880841614484841e-06, + "loss": 0.9561, + "step": 542 + }, + { + "epoch": 0.9451697127937336, + "grad_norm": 0.47265625, + "learning_rate": 5.866665395911375e-06, + "loss": 0.9407, + "step": 543 + }, + { + "epoch": 0.9469103568320278, + "grad_norm": 0.484375, + "learning_rate": 5.852481991103149e-06, + "loss": 0.8833, + "step": 544 + }, + { + "epoch": 0.9486510008703221, + "grad_norm": 0.5, + "learning_rate": 5.838291517666427e-06, + "loss": 0.993, + "step": 545 + }, + { + "epoch": 0.9503916449086162, + "grad_norm": 0.494140625, + "learning_rate": 5.824094093266077e-06, + "loss": 0.9231, + "step": 546 + }, + { + "epoch": 0.9521322889469104, + "grad_norm": 0.490234375, + "learning_rate": 5.809889835624611e-06, + "loss": 0.9512, + "step": 547 + }, + { + "epoch": 0.9538729329852045, + "grad_norm": 0.494140625, + "learning_rate": 5.795678862521197e-06, + "loss": 0.9175, + "step": 548 + }, + { + "epoch": 0.9556135770234987, + "grad_norm": 0.486328125, + "learning_rate": 5.781461291790687e-06, + "loss": 0.8922, + "step": 549 + }, + { + "epoch": 0.9573542210617929, + "grad_norm": 0.494140625, + "learning_rate": 5.767237241322641e-06, + "loss": 0.9701, + "step": 550 + }, + { + "epoch": 0.959094865100087, + "grad_norm": 0.49609375, + "learning_rate": 5.753006829060343e-06, + "loss": 0.9464, + "step": 551 + }, + { + "epoch": 0.9608355091383812, + "grad_norm": 0.5, + "learning_rate": 5.738770172999835e-06, + "loss": 0.9335, + "step": 552 + }, + { + "epoch": 0.9625761531766753, + "grad_norm": 0.458984375, + "learning_rate": 5.724527391188927e-06, + "loss": 0.9689, + "step": 553 + }, + { + "epoch": 0.9643167972149695, + "grad_norm": 0.50390625, + "learning_rate": 5.710278601726222e-06, + "loss": 0.8756, + "step": 554 + }, + { + "epoch": 0.9660574412532638, + "grad_norm": 0.484375, + "learning_rate": 5.696023922760141e-06, + "loss": 0.9459, + "step": 555 + }, + { + "epoch": 0.9677980852915579, + "grad_norm": 0.50390625, + "learning_rate": 5.681763472487933e-06, + "loss": 0.9305, + "step": 556 + }, + { + "epoch": 0.9695387293298521, + "grad_norm": 0.50390625, + "learning_rate": 5.667497369154712e-06, + "loss": 0.8707, + "step": 557 + }, + { + "epoch": 0.9712793733681462, + "grad_norm": 0.47265625, + "learning_rate": 5.6532257310524565e-06, + "loss": 0.8977, + "step": 558 + }, + { + "epoch": 0.9730200174064404, + "grad_norm": 0.494140625, + "learning_rate": 5.638948676519043e-06, + "loss": 0.8735, + "step": 559 + }, + { + "epoch": 0.9747606614447345, + "grad_norm": 0.50390625, + "learning_rate": 5.624666323937257e-06, + "loss": 0.9788, + "step": 560 + }, + { + "epoch": 0.9765013054830287, + "grad_norm": 0.49609375, + "learning_rate": 5.610378791733821e-06, + "loss": 0.9348, + "step": 561 + }, + { + "epoch": 0.9782419495213229, + "grad_norm": 0.48046875, + "learning_rate": 5.596086198378399e-06, + "loss": 0.9258, + "step": 562 + }, + { + "epoch": 0.979982593559617, + "grad_norm": 0.490234375, + "learning_rate": 5.5817886623826245e-06, + "loss": 0.9184, + "step": 563 + }, + { + "epoch": 0.9817232375979112, + "grad_norm": 0.490234375, + "learning_rate": 5.567486302299112e-06, + "loss": 0.9439, + "step": 564 + }, + { + "epoch": 0.9834638816362054, + "grad_norm": 0.474609375, + "learning_rate": 5.553179236720482e-06, + "loss": 0.9887, + "step": 565 + }, + { + "epoch": 0.9852045256744996, + "grad_norm": 0.48828125, + "learning_rate": 5.5388675842783644e-06, + "loss": 0.9439, + "step": 566 + }, + { + "epoch": 0.9869451697127938, + "grad_norm": 0.490234375, + "learning_rate": 5.524551463642429e-06, + "loss": 0.971, + "step": 567 + }, + { + "epoch": 0.9886858137510879, + "grad_norm": 0.486328125, + "learning_rate": 5.510230993519391e-06, + "loss": 0.9583, + "step": 568 + }, + { + "epoch": 0.9904264577893821, + "grad_norm": 0.50390625, + "learning_rate": 5.495906292652035e-06, + "loss": 0.9788, + "step": 569 + }, + { + "epoch": 0.9921671018276762, + "grad_norm": 0.4921875, + "learning_rate": 5.48157747981822e-06, + "loss": 0.9634, + "step": 570 + }, + { + "epoch": 0.9939077458659704, + "grad_norm": 0.5078125, + "learning_rate": 5.467244673829908e-06, + "loss": 0.8834, + "step": 571 + }, + { + "epoch": 0.9956483899042646, + "grad_norm": 0.49609375, + "learning_rate": 5.452907993532164e-06, + "loss": 0.9113, + "step": 572 + }, + { + "epoch": 0.9973890339425587, + "grad_norm": 0.50390625, + "learning_rate": 5.438567557802186e-06, + "loss": 0.8828, + "step": 573 + }, + { + "epoch": 0.999129677980853, + "grad_norm": 0.48828125, + "learning_rate": 5.424223485548303e-06, + "loss": 0.9363, + "step": 574 + } + ], + "logging_steps": 1, + "max_steps": 1148, + "num_input_tokens_seen": 0, + "num_train_epochs": 2, + "save_steps": 574, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 3.613017336006574e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +}