{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.1, "eval_steps": 500, "global_step": 500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0002, "grad_norm": 5.698113441467285, "learning_rate": 5e-06, "loss": 12.2509, "mean_token_accuracy": 0.2450142428278923, "num_tokens": 16438.0, "step": 1 }, { "epoch": 0.0004, "grad_norm": 6.063949108123779, "learning_rate": 4.999000000000001e-06, "loss": 13.3748, "mean_token_accuracy": 0.2325708046555519, "num_tokens": 31368.0, "step": 2 }, { "epoch": 0.0006, "grad_norm": 8.026069641113281, "learning_rate": 4.998e-06, "loss": 12.5454, "mean_token_accuracy": 0.2398785501718521, "num_tokens": 58542.0, "step": 3 }, { "epoch": 0.0008, "grad_norm": 6.758559703826904, "learning_rate": 4.997000000000001e-06, "loss": 13.9155, "mean_token_accuracy": 0.21041666716337204, "num_tokens": 77844.0, "step": 4 }, { "epoch": 0.001, "grad_norm": 6.786197185516357, "learning_rate": 4.996e-06, "loss": 13.2561, "mean_token_accuracy": 0.1967741921544075, "num_tokens": 89661.0, "step": 5 }, { "epoch": 0.0012, "grad_norm": 9.514345169067383, "learning_rate": 4.9950000000000005e-06, "loss": 12.3378, "mean_token_accuracy": 0.20638945698738098, "num_tokens": 104612.0, "step": 6 }, { "epoch": 0.0014, "grad_norm": 7.438795566558838, "learning_rate": 4.994000000000001e-06, "loss": 12.0814, "mean_token_accuracy": 0.2916666716337204, "num_tokens": 114269.0, "step": 7 }, { "epoch": 0.0016, "grad_norm": 15.847832679748535, "learning_rate": 4.993e-06, "loss": 13.2761, "mean_token_accuracy": 0.2450142428278923, "num_tokens": 128762.0, "step": 8 }, { "epoch": 0.0018, "grad_norm": 7.620685577392578, "learning_rate": 4.992e-06, "loss": 13.1792, "mean_token_accuracy": 0.2165178582072258, "num_tokens": 148453.0, "step": 9 }, { "epoch": 0.002, "grad_norm": 8.26969051361084, "learning_rate": 4.991e-06, "loss": 12.853, "mean_token_accuracy": 0.21405228972434998, "num_tokens": 167693.0, "step": 10 }, { "epoch": 0.0022, "grad_norm": 7.328786373138428, "learning_rate": 4.9900000000000005e-06, "loss": 12.9578, "mean_token_accuracy": 0.2115987464785576, "num_tokens": 182157.0, "step": 11 }, { "epoch": 0.0024, "grad_norm": 7.058177947998047, "learning_rate": 4.989000000000001e-06, "loss": 12.7711, "mean_token_accuracy": 0.2002224698662758, "num_tokens": 201457.0, "step": 12 }, { "epoch": 0.0026, "grad_norm": 6.48744010925293, "learning_rate": 4.988e-06, "loss": 13.006, "mean_token_accuracy": 0.23790322244167328, "num_tokens": 215922.0, "step": 13 }, { "epoch": 0.0028, "grad_norm": 6.006223201751709, "learning_rate": 4.987e-06, "loss": 10.5975, "mean_token_accuracy": 0.2096899226307869, "num_tokens": 225597.0, "step": 14 }, { "epoch": 0.003, "grad_norm": 6.274689674377441, "learning_rate": 4.986e-06, "loss": 11.2766, "mean_token_accuracy": 0.26851852238178253, "num_tokens": 240100.0, "step": 15 }, { "epoch": 0.0032, "grad_norm": 8.211908340454102, "learning_rate": 4.9850000000000006e-06, "loss": 13.7091, "mean_token_accuracy": 0.20202020555734634, "num_tokens": 259400.0, "step": 16 }, { "epoch": 0.0034, "grad_norm": 6.356493949890137, "learning_rate": 4.984000000000001e-06, "loss": 12.043, "mean_token_accuracy": 0.20519480854272842, "num_tokens": 278628.0, "step": 17 }, { "epoch": 0.0036, "grad_norm": 6.438048839569092, "learning_rate": 4.983e-06, "loss": 11.5484, "mean_token_accuracy": 0.2343137264251709, "num_tokens": 303856.0, "step": 18 }, { "epoch": 0.0038, "grad_norm": 7.152822494506836, "learning_rate": 4.982e-06, "loss": 13.4607, "mean_token_accuracy": 0.17500000447034836, "num_tokens": 318686.0, "step": 19 }, { "epoch": 0.004, "grad_norm": 6.464632034301758, "learning_rate": 4.981e-06, "loss": 11.7533, "mean_token_accuracy": 0.21959459781646729, "num_tokens": 337999.0, "step": 20 }, { "epoch": 0.0042, "grad_norm": 7.051283836364746, "learning_rate": 4.980000000000001e-06, "loss": 12.9788, "mean_token_accuracy": 0.22011494636535645, "num_tokens": 356602.0, "step": 21 }, { "epoch": 0.0044, "grad_norm": 9.595747947692871, "learning_rate": 4.979e-06, "loss": 10.9347, "mean_token_accuracy": 0.2645348906517029, "num_tokens": 373201.0, "step": 22 }, { "epoch": 0.0046, "grad_norm": 7.404125213623047, "learning_rate": 4.9780000000000005e-06, "loss": 12.5181, "mean_token_accuracy": 0.2666666731238365, "num_tokens": 387660.0, "step": 23 }, { "epoch": 0.0048, "grad_norm": 6.656332969665527, "learning_rate": 4.977e-06, "loss": 11.5566, "mean_token_accuracy": 0.209001787006855, "num_tokens": 402129.0, "step": 24 }, { "epoch": 0.005, "grad_norm": 6.866989612579346, "learning_rate": 4.976e-06, "loss": 12.4797, "mean_token_accuracy": 0.2379310354590416, "num_tokens": 421133.0, "step": 25 }, { "epoch": 0.0052, "grad_norm": 6.77735710144043, "learning_rate": 4.975000000000001e-06, "loss": 13.6256, "mean_token_accuracy": 0.22177419066429138, "num_tokens": 440531.0, "step": 26 }, { "epoch": 0.0054, "grad_norm": 8.206353187561035, "learning_rate": 4.974e-06, "loss": 13.0667, "mean_token_accuracy": 0.21791187673807144, "num_tokens": 459798.0, "step": 27 }, { "epoch": 0.0056, "grad_norm": 7.40612268447876, "learning_rate": 4.9730000000000005e-06, "loss": 13.5884, "mean_token_accuracy": 0.19052419066429138, "num_tokens": 478587.0, "step": 28 }, { "epoch": 0.0058, "grad_norm": 6.71999454498291, "learning_rate": 4.972e-06, "loss": 12.0322, "mean_token_accuracy": 0.24526315927505493, "num_tokens": 492356.0, "step": 29 }, { "epoch": 0.006, "grad_norm": 8.108094215393066, "learning_rate": 4.971e-06, "loss": 12.4858, "mean_token_accuracy": 0.2566666677594185, "num_tokens": 507242.0, "step": 30 }, { "epoch": 0.0062, "grad_norm": 7.7994513511657715, "learning_rate": 4.970000000000001e-06, "loss": 13.0762, "mean_token_accuracy": 0.24568965286016464, "num_tokens": 525987.0, "step": 31 }, { "epoch": 0.0064, "grad_norm": 7.258217811584473, "learning_rate": 4.969e-06, "loss": 12.7757, "mean_token_accuracy": 0.24137930572032928, "num_tokens": 540447.0, "step": 32 }, { "epoch": 0.0066, "grad_norm": 14.746047973632812, "learning_rate": 4.9680000000000005e-06, "loss": 14.4335, "mean_token_accuracy": 0.22649572789669037, "num_tokens": 560131.0, "step": 33 }, { "epoch": 0.0068, "grad_norm": 5.289712429046631, "learning_rate": 4.967e-06, "loss": 9.6693, "mean_token_accuracy": 0.2875000089406967, "num_tokens": 579313.0, "step": 34 }, { "epoch": 0.007, "grad_norm": 7.960392951965332, "learning_rate": 4.966e-06, "loss": 13.1738, "mean_token_accuracy": 0.25833334028720856, "num_tokens": 598611.0, "step": 35 }, { "epoch": 0.0072, "grad_norm": 8.295417785644531, "learning_rate": 4.965000000000001e-06, "loss": 11.8889, "mean_token_accuracy": 0.24144145101308823, "num_tokens": 615588.0, "step": 36 }, { "epoch": 0.0074, "grad_norm": 7.051126480102539, "learning_rate": 4.964e-06, "loss": 12.1364, "mean_token_accuracy": 0.23590733855962753, "num_tokens": 636032.0, "step": 37 }, { "epoch": 0.0076, "grad_norm": 7.895816326141357, "learning_rate": 4.963000000000001e-06, "loss": 12.8971, "mean_token_accuracy": 0.18571428954601288, "num_tokens": 655299.0, "step": 38 }, { "epoch": 0.0078, "grad_norm": 7.544738292694092, "learning_rate": 4.962e-06, "loss": 13.0581, "mean_token_accuracy": 0.20937500149011612, "num_tokens": 674357.0, "step": 39 }, { "epoch": 0.008, "grad_norm": 7.548039436340332, "learning_rate": 4.9610000000000004e-06, "loss": 13.2158, "mean_token_accuracy": 0.20892494916915894, "num_tokens": 693698.0, "step": 40 }, { "epoch": 0.0082, "grad_norm": 7.687658309936523, "learning_rate": 4.960000000000001e-06, "loss": 12.8524, "mean_token_accuracy": 0.28735632449388504, "num_tokens": 703433.0, "step": 41 }, { "epoch": 0.0084, "grad_norm": 8.011468887329102, "learning_rate": 4.959e-06, "loss": 13.5655, "mean_token_accuracy": 0.21635150164365768, "num_tokens": 722733.0, "step": 42 }, { "epoch": 0.0086, "grad_norm": 11.084840774536133, "learning_rate": 4.958000000000001e-06, "loss": 12.7054, "mean_token_accuracy": 0.21746384352445602, "num_tokens": 738121.0, "step": 43 }, { "epoch": 0.0088, "grad_norm": 9.436426162719727, "learning_rate": 4.957e-06, "loss": 13.5213, "mean_token_accuracy": 0.22783251106739044, "num_tokens": 757723.0, "step": 44 }, { "epoch": 0.009, "grad_norm": 8.382990837097168, "learning_rate": 4.9560000000000005e-06, "loss": 12.6328, "mean_token_accuracy": 0.2567741870880127, "num_tokens": 772646.0, "step": 45 }, { "epoch": 0.0092, "grad_norm": 8.272336959838867, "learning_rate": 4.955e-06, "loss": 13.0365, "mean_token_accuracy": 0.2611111178994179, "num_tokens": 792334.0, "step": 46 }, { "epoch": 0.0094, "grad_norm": 10.347405433654785, "learning_rate": 4.954e-06, "loss": 12.8164, "mean_token_accuracy": 0.22616633027791977, "num_tokens": 811713.0, "step": 47 }, { "epoch": 0.0096, "grad_norm": 7.27515983581543, "learning_rate": 4.953000000000001e-06, "loss": 11.0123, "mean_token_accuracy": 0.23790322244167328, "num_tokens": 832998.0, "step": 48 }, { "epoch": 0.0098, "grad_norm": 8.973237991333008, "learning_rate": 4.952e-06, "loss": 12.9664, "mean_token_accuracy": 0.19805195182561874, "num_tokens": 861717.0, "step": 49 }, { "epoch": 0.01, "grad_norm": 8.738320350646973, "learning_rate": 4.9510000000000005e-06, "loss": 12.5728, "mean_token_accuracy": 0.24344827979803085, "num_tokens": 875736.0, "step": 50 }, { "epoch": 0.0102, "grad_norm": 8.51733112335205, "learning_rate": 4.95e-06, "loss": 12.7596, "mean_token_accuracy": 0.22380952537059784, "num_tokens": 894386.0, "step": 51 }, { "epoch": 0.0104, "grad_norm": 8.787413597106934, "learning_rate": 4.949e-06, "loss": 12.4493, "mean_token_accuracy": 0.268075630068779, "num_tokens": 908848.0, "step": 52 }, { "epoch": 0.0106, "grad_norm": 7.849542617797852, "learning_rate": 4.948000000000001e-06, "loss": 12.7514, "mean_token_accuracy": 0.23790322244167328, "num_tokens": 923309.0, "step": 53 }, { "epoch": 0.0108, "grad_norm": 8.378942489624023, "learning_rate": 4.947e-06, "loss": 11.4953, "mean_token_accuracy": 0.2290322557091713, "num_tokens": 938298.0, "step": 54 }, { "epoch": 0.011, "grad_norm": 8.311882972717285, "learning_rate": 4.946000000000001e-06, "loss": 12.0904, "mean_token_accuracy": 0.24014336615800858, "num_tokens": 948288.0, "step": 55 }, { "epoch": 0.0112, "grad_norm": 9.599881172180176, "learning_rate": 4.945e-06, "loss": 14.3569, "mean_token_accuracy": 0.22649572789669037, "num_tokens": 967543.0, "step": 56 }, { "epoch": 0.0114, "grad_norm": 8.84776496887207, "learning_rate": 4.9440000000000004e-06, "loss": 12.2268, "mean_token_accuracy": 0.28287841379642487, "num_tokens": 981726.0, "step": 57 }, { "epoch": 0.0116, "grad_norm": 11.259871482849121, "learning_rate": 4.943000000000001e-06, "loss": 12.6356, "mean_token_accuracy": 0.22177419066429138, "num_tokens": 996225.0, "step": 58 }, { "epoch": 0.0118, "grad_norm": 10.529711723327637, "learning_rate": 4.942e-06, "loss": 11.9829, "mean_token_accuracy": 0.26986077427864075, "num_tokens": 1015573.0, "step": 59 }, { "epoch": 0.012, "grad_norm": 8.90577220916748, "learning_rate": 4.941000000000001e-06, "loss": 11.4895, "mean_token_accuracy": 0.2374911978840828, "num_tokens": 1034198.0, "step": 60 }, { "epoch": 0.0122, "grad_norm": 7.851955413818359, "learning_rate": 4.94e-06, "loss": 12.0083, "mean_token_accuracy": 0.22478991746902466, "num_tokens": 1045419.0, "step": 61 }, { "epoch": 0.0124, "grad_norm": 9.812698364257812, "learning_rate": 4.9390000000000005e-06, "loss": 12.9777, "mean_token_accuracy": 0.2379310354590416, "num_tokens": 1064723.0, "step": 62 }, { "epoch": 0.0126, "grad_norm": 8.35107707977295, "learning_rate": 4.938000000000001e-06, "loss": 11.3187, "mean_token_accuracy": 0.24358975142240524, "num_tokens": 1079171.0, "step": 63 }, { "epoch": 0.0128, "grad_norm": 7.16640567779541, "learning_rate": 4.937e-06, "loss": 10.2182, "mean_token_accuracy": 0.34151194989681244, "num_tokens": 1093695.0, "step": 64 }, { "epoch": 0.013, "grad_norm": 10.18386459350586, "learning_rate": 4.936e-06, "loss": 12.2175, "mean_token_accuracy": 0.25961539149284363, "num_tokens": 1108580.0, "step": 65 }, { "epoch": 0.0132, "grad_norm": 8.232446670532227, "learning_rate": 4.935e-06, "loss": 11.8182, "mean_token_accuracy": 0.27314814925193787, "num_tokens": 1123098.0, "step": 66 }, { "epoch": 0.0134, "grad_norm": 8.809263229370117, "learning_rate": 4.9340000000000005e-06, "loss": 10.9148, "mean_token_accuracy": 0.20927418768405914, "num_tokens": 1137548.0, "step": 67 }, { "epoch": 0.0136, "grad_norm": 9.865777015686035, "learning_rate": 4.933000000000001e-06, "loss": 12.6294, "mean_token_accuracy": 0.22962962836027145, "num_tokens": 1156845.0, "step": 68 }, { "epoch": 0.0138, "grad_norm": 7.259024620056152, "learning_rate": 4.932e-06, "loss": 9.7717, "mean_token_accuracy": 0.26875001192092896, "num_tokens": 1171363.0, "step": 69 }, { "epoch": 0.014, "grad_norm": 11.756244659423828, "learning_rate": 4.931e-06, "loss": 12.6686, "mean_token_accuracy": 0.28285714983940125, "num_tokens": 1185814.0, "step": 70 }, { "epoch": 0.0142, "grad_norm": 9.128395080566406, "learning_rate": 4.93e-06, "loss": 11.1826, "mean_token_accuracy": 0.21765056997537613, "num_tokens": 1200437.0, "step": 71 }, { "epoch": 0.0144, "grad_norm": 8.472599029541016, "learning_rate": 4.929000000000001e-06, "loss": 11.6617, "mean_token_accuracy": 0.2060810774564743, "num_tokens": 1219264.0, "step": 72 }, { "epoch": 0.0146, "grad_norm": 11.549400329589844, "learning_rate": 4.928000000000001e-06, "loss": 12.9588, "mean_token_accuracy": 0.25, "num_tokens": 1238560.0, "step": 73 }, { "epoch": 0.0148, "grad_norm": 11.084433555603027, "learning_rate": 4.9270000000000004e-06, "loss": 10.4111, "mean_token_accuracy": 0.21954887360334396, "num_tokens": 1247376.0, "step": 74 }, { "epoch": 0.015, "grad_norm": 10.906563758850098, "learning_rate": 4.926e-06, "loss": 12.5231, "mean_token_accuracy": 0.23307790607213974, "num_tokens": 1266611.0, "step": 75 }, { "epoch": 0.0152, "grad_norm": 9.466647148132324, "learning_rate": 4.925e-06, "loss": 11.4363, "mean_token_accuracy": 0.24193547666072845, "num_tokens": 1281132.0, "step": 76 }, { "epoch": 0.0154, "grad_norm": 11.185935020446777, "learning_rate": 4.924000000000001e-06, "loss": 12.6383, "mean_token_accuracy": 0.22685185074806213, "num_tokens": 1301049.0, "step": 77 }, { "epoch": 0.0156, "grad_norm": 11.0143461227417, "learning_rate": 4.923000000000001e-06, "loss": 12.479, "mean_token_accuracy": 0.22962962836027145, "num_tokens": 1322278.0, "step": 78 }, { "epoch": 0.0158, "grad_norm": 12.330080032348633, "learning_rate": 4.9220000000000005e-06, "loss": 12.2351, "mean_token_accuracy": 0.2675606608390808, "num_tokens": 1341536.0, "step": 79 }, { "epoch": 0.016, "grad_norm": 10.486513137817383, "learning_rate": 4.921e-06, "loss": 11.6896, "mean_token_accuracy": 0.25356507301330566, "num_tokens": 1355885.0, "step": 80 }, { "epoch": 0.0162, "grad_norm": 12.150262832641602, "learning_rate": 4.92e-06, "loss": 12.5725, "mean_token_accuracy": 0.2321428582072258, "num_tokens": 1375181.0, "step": 81 }, { "epoch": 0.0164, "grad_norm": 11.387964248657227, "learning_rate": 4.919000000000001e-06, "loss": 11.5796, "mean_token_accuracy": 0.29461538791656494, "num_tokens": 1384853.0, "step": 82 }, { "epoch": 0.0166, "grad_norm": 9.878727912902832, "learning_rate": 4.918e-06, "loss": 11.7534, "mean_token_accuracy": 0.24473684281110764, "num_tokens": 1401178.0, "step": 83 }, { "epoch": 0.0168, "grad_norm": 9.827662467956543, "learning_rate": 4.9170000000000005e-06, "loss": 12.0345, "mean_token_accuracy": 0.18034055829048157, "num_tokens": 1419976.0, "step": 84 }, { "epoch": 0.017, "grad_norm": 10.327628135681152, "learning_rate": 4.916e-06, "loss": 11.1049, "mean_token_accuracy": 0.3095238208770752, "num_tokens": 1429652.0, "step": 85 }, { "epoch": 0.0172, "grad_norm": 10.67590045928955, "learning_rate": 4.915e-06, "loss": 10.7087, "mean_token_accuracy": 0.2750582844018936, "num_tokens": 1444113.0, "step": 86 }, { "epoch": 0.0174, "grad_norm": 20.292959213256836, "learning_rate": 4.914000000000001e-06, "loss": 10.797, "mean_token_accuracy": 0.32500000298023224, "num_tokens": 1453773.0, "step": 87 }, { "epoch": 0.0176, "grad_norm": 8.3117036819458, "learning_rate": 4.913e-06, "loss": 9.394, "mean_token_accuracy": 0.31915584206581116, "num_tokens": 1475024.0, "step": 88 }, { "epoch": 0.0178, "grad_norm": 11.334404945373535, "learning_rate": 4.9120000000000006e-06, "loss": 11.0718, "mean_token_accuracy": 0.2678571492433548, "num_tokens": 1489489.0, "step": 89 }, { "epoch": 0.018, "grad_norm": 15.36023235321045, "learning_rate": 4.911e-06, "loss": 11.7992, "mean_token_accuracy": 0.28607918322086334, "num_tokens": 1508816.0, "step": 90 }, { "epoch": 0.0182, "grad_norm": 11.725635528564453, "learning_rate": 4.9100000000000004e-06, "loss": 11.9207, "mean_token_accuracy": 0.26123301684856415, "num_tokens": 1530050.0, "step": 91 }, { "epoch": 0.0184, "grad_norm": 14.75428295135498, "learning_rate": 4.909000000000001e-06, "loss": 11.0294, "mean_token_accuracy": 0.21236559003591537, "num_tokens": 1544713.0, "step": 92 }, { "epoch": 0.0186, "grad_norm": 10.447436332702637, "learning_rate": 4.908e-06, "loss": 9.7536, "mean_token_accuracy": 0.24568965286016464, "num_tokens": 1559226.0, "step": 93 }, { "epoch": 0.0188, "grad_norm": 14.036280632019043, "learning_rate": 4.907000000000001e-06, "loss": 12.2065, "mean_token_accuracy": 0.2771739065647125, "num_tokens": 1568898.0, "step": 94 }, { "epoch": 0.019, "grad_norm": 13.364744186401367, "learning_rate": 4.906e-06, "loss": 13.168, "mean_token_accuracy": 0.2321428582072258, "num_tokens": 1588133.0, "step": 95 }, { "epoch": 0.0192, "grad_norm": 12.441611289978027, "learning_rate": 4.9050000000000005e-06, "loss": 11.6806, "mean_token_accuracy": 0.21875, "num_tokens": 1607218.0, "step": 96 }, { "epoch": 0.0194, "grad_norm": 11.559666633605957, "learning_rate": 4.904000000000001e-06, "loss": 12.2055, "mean_token_accuracy": 0.23010753095149994, "num_tokens": 1626557.0, "step": 97 }, { "epoch": 0.0196, "grad_norm": 11.131147384643555, "learning_rate": 4.903e-06, "loss": 11.4201, "mean_token_accuracy": 0.2931034490466118, "num_tokens": 1641017.0, "step": 98 }, { "epoch": 0.0198, "grad_norm": 12.186894416809082, "learning_rate": 4.902000000000001e-06, "loss": 12.1568, "mean_token_accuracy": 0.23180076479911804, "num_tokens": 1655475.0, "step": 99 }, { "epoch": 0.02, "grad_norm": 11.864778518676758, "learning_rate": 4.901e-06, "loss": 12.4851, "mean_token_accuracy": 0.21008403599262238, "num_tokens": 1674815.0, "step": 100 }, { "epoch": 0.0202, "grad_norm": 14.35185718536377, "learning_rate": 4.9000000000000005e-06, "loss": 11.1814, "mean_token_accuracy": 0.2612612694501877, "num_tokens": 1686681.0, "step": 101 }, { "epoch": 0.0204, "grad_norm": 18.715627670288086, "learning_rate": 4.899e-06, "loss": 11.632, "mean_token_accuracy": 0.30199430882930756, "num_tokens": 1701565.0, "step": 102 }, { "epoch": 0.0206, "grad_norm": 12.892010688781738, "learning_rate": 4.898e-06, "loss": 12.3157, "mean_token_accuracy": 0.23806367069482803, "num_tokens": 1715938.0, "step": 103 }, { "epoch": 0.0208, "grad_norm": 13.078173637390137, "learning_rate": 4.897000000000001e-06, "loss": 11.4954, "mean_token_accuracy": 0.22177419066429138, "num_tokens": 1734685.0, "step": 104 }, { "epoch": 0.021, "grad_norm": 14.580618858337402, "learning_rate": 4.896e-06, "loss": 10.4619, "mean_token_accuracy": 0.2797202914953232, "num_tokens": 1757899.0, "step": 105 }, { "epoch": 0.0212, "grad_norm": 13.174345016479492, "learning_rate": 4.8950000000000006e-06, "loss": 10.5835, "mean_token_accuracy": 0.27546295523643494, "num_tokens": 1778410.0, "step": 106 }, { "epoch": 0.0214, "grad_norm": 13.970446586608887, "learning_rate": 4.894e-06, "loss": 11.1265, "mean_token_accuracy": 0.2693749964237213, "num_tokens": 1791819.0, "step": 107 }, { "epoch": 0.0216, "grad_norm": 11.994514465332031, "learning_rate": 4.893e-06, "loss": 11.3511, "mean_token_accuracy": 0.25462962687015533, "num_tokens": 1811118.0, "step": 108 }, { "epoch": 0.0218, "grad_norm": 13.328775405883789, "learning_rate": 4.892000000000001e-06, "loss": 11.9787, "mean_token_accuracy": 0.23790322244167328, "num_tokens": 1825598.0, "step": 109 }, { "epoch": 0.022, "grad_norm": 11.813949584960938, "learning_rate": 4.891e-06, "loss": 10.614, "mean_token_accuracy": 0.25833334028720856, "num_tokens": 1840096.0, "step": 110 }, { "epoch": 0.0222, "grad_norm": 17.409263610839844, "learning_rate": 4.890000000000001e-06, "loss": 10.5087, "mean_token_accuracy": 0.32692308723926544, "num_tokens": 1849750.0, "step": 111 }, { "epoch": 0.0224, "grad_norm": 19.462722778320312, "learning_rate": 4.889e-06, "loss": 10.2288, "mean_token_accuracy": 0.25729166716337204, "num_tokens": 1859433.0, "step": 112 }, { "epoch": 0.0226, "grad_norm": 11.595419883728027, "learning_rate": 4.8880000000000005e-06, "loss": 11.0996, "mean_token_accuracy": 0.2807881832122803, "num_tokens": 1873892.0, "step": 113 }, { "epoch": 0.0228, "grad_norm": 14.413046836853027, "learning_rate": 4.887000000000001e-06, "loss": 11.6535, "mean_token_accuracy": 0.2341153472661972, "num_tokens": 1893196.0, "step": 114 }, { "epoch": 0.023, "grad_norm": 12.218829154968262, "learning_rate": 4.886e-06, "loss": 10.9169, "mean_token_accuracy": 0.2557603716850281, "num_tokens": 1907676.0, "step": 115 }, { "epoch": 0.0232, "grad_norm": 16.51720428466797, "learning_rate": 4.885000000000001e-06, "loss": 9.8406, "mean_token_accuracy": 0.25968992710113525, "num_tokens": 1921596.0, "step": 116 }, { "epoch": 0.0234, "grad_norm": 10.608232498168945, "learning_rate": 4.884e-06, "loss": 9.9809, "mean_token_accuracy": 0.22068965435028076, "num_tokens": 1940247.0, "step": 117 }, { "epoch": 0.0236, "grad_norm": 14.17568302154541, "learning_rate": 4.8830000000000005e-06, "loss": 11.3908, "mean_token_accuracy": 0.23790322244167328, "num_tokens": 1959268.0, "step": 118 }, { "epoch": 0.0238, "grad_norm": 12.197131156921387, "learning_rate": 4.882000000000001e-06, "loss": 10.9438, "mean_token_accuracy": 0.2538699731230736, "num_tokens": 1978218.0, "step": 119 }, { "epoch": 0.024, "grad_norm": 15.261062622070312, "learning_rate": 4.881e-06, "loss": 11.0329, "mean_token_accuracy": 0.24621212482452393, "num_tokens": 1997519.0, "step": 120 }, { "epoch": 0.0242, "grad_norm": 13.116536140441895, "learning_rate": 4.880000000000001e-06, "loss": 11.9918, "mean_token_accuracy": 0.24137930572032928, "num_tokens": 2016817.0, "step": 121 }, { "epoch": 0.0244, "grad_norm": 18.174373626708984, "learning_rate": 4.879e-06, "loss": 11.0414, "mean_token_accuracy": 0.25833334028720856, "num_tokens": 2035875.0, "step": 122 }, { "epoch": 0.0246, "grad_norm": 17.258121490478516, "learning_rate": 4.8780000000000006e-06, "loss": 11.1173, "mean_token_accuracy": 0.3010057359933853, "num_tokens": 2050330.0, "step": 123 }, { "epoch": 0.0248, "grad_norm": 12.636884689331055, "learning_rate": 4.877000000000001e-06, "loss": 11.3399, "mean_token_accuracy": 0.2879464328289032, "num_tokens": 2070156.0, "step": 124 }, { "epoch": 0.025, "grad_norm": 18.457618713378906, "learning_rate": 4.876e-06, "loss": 11.5595, "mean_token_accuracy": 0.28418803960084915, "num_tokens": 2084771.0, "step": 125 }, { "epoch": 0.0252, "grad_norm": 14.281397819519043, "learning_rate": 4.875e-06, "loss": 10.9901, "mean_token_accuracy": 0.28140393644571304, "num_tokens": 2104228.0, "step": 126 }, { "epoch": 0.0254, "grad_norm": 13.08484935760498, "learning_rate": 4.874e-06, "loss": 9.4802, "mean_token_accuracy": 0.2637759745121002, "num_tokens": 2118877.0, "step": 127 }, { "epoch": 0.0256, "grad_norm": 11.949925422668457, "learning_rate": 4.873000000000001e-06, "loss": 10.3703, "mean_token_accuracy": 0.2619825750589371, "num_tokens": 2133298.0, "step": 128 }, { "epoch": 0.0258, "grad_norm": 14.950297355651855, "learning_rate": 4.872000000000001e-06, "loss": 10.6116, "mean_token_accuracy": 0.33000001311302185, "num_tokens": 2142955.0, "step": 129 }, { "epoch": 0.026, "grad_norm": 16.433286666870117, "learning_rate": 4.8710000000000005e-06, "loss": 9.4482, "mean_token_accuracy": 0.3333333432674408, "num_tokens": 2159447.0, "step": 130 }, { "epoch": 0.0262, "grad_norm": 12.467981338500977, "learning_rate": 4.87e-06, "loss": 10.2354, "mean_token_accuracy": 0.33796295523643494, "num_tokens": 2178352.0, "step": 131 }, { "epoch": 0.0264, "grad_norm": 11.493000030517578, "learning_rate": 4.869e-06, "loss": 9.0597, "mean_token_accuracy": 0.2736175060272217, "num_tokens": 2199174.0, "step": 132 }, { "epoch": 0.0266, "grad_norm": 13.966115951538086, "learning_rate": 4.868000000000001e-06, "loss": 10.3177, "mean_token_accuracy": 0.3325917571783066, "num_tokens": 2213636.0, "step": 133 }, { "epoch": 0.0268, "grad_norm": 13.971321105957031, "learning_rate": 4.867000000000001e-06, "loss": 10.2564, "mean_token_accuracy": 0.31481482088565826, "num_tokens": 2228050.0, "step": 134 }, { "epoch": 0.027, "grad_norm": 16.246124267578125, "learning_rate": 4.8660000000000005e-06, "loss": 10.7549, "mean_token_accuracy": 0.30943846702575684, "num_tokens": 2247348.0, "step": 135 }, { "epoch": 0.0272, "grad_norm": 17.702425003051758, "learning_rate": 4.865e-06, "loss": 10.1729, "mean_token_accuracy": 0.3575989753007889, "num_tokens": 2261806.0, "step": 136 }, { "epoch": 0.0274, "grad_norm": 13.732104301452637, "learning_rate": 4.864e-06, "loss": 10.9004, "mean_token_accuracy": 0.28607918322086334, "num_tokens": 2281224.0, "step": 137 }, { "epoch": 0.0276, "grad_norm": 17.568925857543945, "learning_rate": 4.863000000000001e-06, "loss": 10.1658, "mean_token_accuracy": 0.2857142984867096, "num_tokens": 2295720.0, "step": 138 }, { "epoch": 0.0278, "grad_norm": 13.424271583557129, "learning_rate": 4.862e-06, "loss": 10.3543, "mean_token_accuracy": 0.28928571939468384, "num_tokens": 2310345.0, "step": 139 }, { "epoch": 0.028, "grad_norm": 14.413524627685547, "learning_rate": 4.8610000000000006e-06, "loss": 11.3822, "mean_token_accuracy": 0.22828783839941025, "num_tokens": 2347736.0, "step": 140 }, { "epoch": 0.0282, "grad_norm": 22.185630798339844, "learning_rate": 4.86e-06, "loss": 10.6185, "mean_token_accuracy": 0.2911111190915108, "num_tokens": 2361685.0, "step": 141 }, { "epoch": 0.0284, "grad_norm": 12.253998756408691, "learning_rate": 4.859e-06, "loss": 9.2399, "mean_token_accuracy": 0.288621261715889, "num_tokens": 2376158.0, "step": 142 }, { "epoch": 0.0286, "grad_norm": 16.229686737060547, "learning_rate": 4.858000000000001e-06, "loss": 10.3377, "mean_token_accuracy": 0.22227822244167328, "num_tokens": 2391501.0, "step": 143 }, { "epoch": 0.0288, "grad_norm": 9.664397239685059, "learning_rate": 4.857e-06, "loss": 9.3351, "mean_token_accuracy": 0.26456456631422043, "num_tokens": 2411105.0, "step": 144 }, { "epoch": 0.029, "grad_norm": 12.8119478225708, "learning_rate": 4.856e-06, "loss": 9.87, "mean_token_accuracy": 0.24166666716337204, "num_tokens": 2431185.0, "step": 145 }, { "epoch": 0.0292, "grad_norm": 10.490764617919922, "learning_rate": 4.855e-06, "loss": 8.7889, "mean_token_accuracy": 0.36666667461395264, "num_tokens": 2452470.0, "step": 146 }, { "epoch": 0.0294, "grad_norm": 13.65211296081543, "learning_rate": 4.8540000000000005e-06, "loss": 10.4286, "mean_token_accuracy": 0.2718253955245018, "num_tokens": 2472396.0, "step": 147 }, { "epoch": 0.0296, "grad_norm": 18.866209030151367, "learning_rate": 4.853000000000001e-06, "loss": 10.2372, "mean_token_accuracy": 0.38141025602817535, "num_tokens": 2486886.0, "step": 148 }, { "epoch": 0.0298, "grad_norm": 14.852785110473633, "learning_rate": 4.852e-06, "loss": 10.0487, "mean_token_accuracy": 0.30820105969905853, "num_tokens": 2506143.0, "step": 149 }, { "epoch": 0.03, "grad_norm": 13.972378730773926, "learning_rate": 4.851e-06, "loss": 10.6291, "mean_token_accuracy": 0.25820106267929077, "num_tokens": 2520401.0, "step": 150 }, { "epoch": 0.0302, "grad_norm": 14.618459701538086, "learning_rate": 4.85e-06, "loss": 10.7579, "mean_token_accuracy": 0.23571428656578064, "num_tokens": 2539135.0, "step": 151 }, { "epoch": 0.0304, "grad_norm": 15.176739692687988, "learning_rate": 4.8490000000000005e-06, "loss": 9.6595, "mean_token_accuracy": 0.2510339096188545, "num_tokens": 2558562.0, "step": 152 }, { "epoch": 0.0306, "grad_norm": 16.972919464111328, "learning_rate": 4.848000000000001e-06, "loss": 10.8999, "mean_token_accuracy": 0.25925925374031067, "num_tokens": 2577856.0, "step": 153 }, { "epoch": 0.0308, "grad_norm": 16.451147079467773, "learning_rate": 4.847e-06, "loss": 8.5505, "mean_token_accuracy": 0.36249999701976776, "num_tokens": 2592246.0, "step": 154 }, { "epoch": 0.031, "grad_norm": 24.95278549194336, "learning_rate": 4.846e-06, "loss": 10.2367, "mean_token_accuracy": 0.28735632449388504, "num_tokens": 2607133.0, "step": 155 }, { "epoch": 0.0312, "grad_norm": 15.770346641540527, "learning_rate": 4.845e-06, "loss": 9.6085, "mean_token_accuracy": 0.27272728085517883, "num_tokens": 2621734.0, "step": 156 }, { "epoch": 0.0314, "grad_norm": 12.314064025878906, "learning_rate": 4.8440000000000005e-06, "loss": 9.816, "mean_token_accuracy": 0.2540322542190552, "num_tokens": 2641357.0, "step": 157 }, { "epoch": 0.0316, "grad_norm": 13.379799842834473, "learning_rate": 4.843000000000001e-06, "loss": 10.378, "mean_token_accuracy": 0.32804232835769653, "num_tokens": 2660177.0, "step": 158 }, { "epoch": 0.0318, "grad_norm": 17.458240509033203, "learning_rate": 4.842e-06, "loss": 9.3842, "mean_token_accuracy": 0.2557164579629898, "num_tokens": 2679528.0, "step": 159 }, { "epoch": 0.032, "grad_norm": 25.741785049438477, "learning_rate": 4.841e-06, "loss": 9.0768, "mean_token_accuracy": 0.3270474076271057, "num_tokens": 2693652.0, "step": 160 }, { "epoch": 0.0322, "grad_norm": 13.557204246520996, "learning_rate": 4.84e-06, "loss": 9.6471, "mean_token_accuracy": 0.28735632449388504, "num_tokens": 2713320.0, "step": 161 }, { "epoch": 0.0324, "grad_norm": 14.830061912536621, "learning_rate": 4.839000000000001e-06, "loss": 11.3496, "mean_token_accuracy": 0.24049513787031174, "num_tokens": 2733424.0, "step": 162 }, { "epoch": 0.0326, "grad_norm": 12.371265411376953, "learning_rate": 4.838e-06, "loss": 8.6217, "mean_token_accuracy": 0.25988225638866425, "num_tokens": 2752613.0, "step": 163 }, { "epoch": 0.0328, "grad_norm": 12.722640037536621, "learning_rate": 4.8370000000000004e-06, "loss": 9.2979, "mean_token_accuracy": 0.2337121218442917, "num_tokens": 2773896.0, "step": 164 }, { "epoch": 0.033, "grad_norm": 12.796667098999023, "learning_rate": 4.836e-06, "loss": 9.9112, "mean_token_accuracy": 0.3337438404560089, "num_tokens": 2793334.0, "step": 165 }, { "epoch": 0.0332, "grad_norm": 15.983271598815918, "learning_rate": 4.835e-06, "loss": 9.5066, "mean_token_accuracy": 0.29256465286016464, "num_tokens": 2807755.0, "step": 166 }, { "epoch": 0.0334, "grad_norm": 15.554715156555176, "learning_rate": 4.834000000000001e-06, "loss": 10.5127, "mean_token_accuracy": 0.2986453175544739, "num_tokens": 2827193.0, "step": 167 }, { "epoch": 0.0336, "grad_norm": 14.6381196975708, "learning_rate": 4.833e-06, "loss": 9.7383, "mean_token_accuracy": 0.2586618810892105, "num_tokens": 2846449.0, "step": 168 }, { "epoch": 0.0338, "grad_norm": 16.013647079467773, "learning_rate": 4.8320000000000005e-06, "loss": 9.0009, "mean_token_accuracy": 0.32356322556734085, "num_tokens": 2860910.0, "step": 169 }, { "epoch": 0.034, "grad_norm": 13.708538055419922, "learning_rate": 4.831e-06, "loss": 9.4298, "mean_token_accuracy": 0.28114478290081024, "num_tokens": 2880172.0, "step": 170 }, { "epoch": 0.0342, "grad_norm": 14.314607620239258, "learning_rate": 4.83e-06, "loss": 9.2596, "mean_token_accuracy": 0.3014460504055023, "num_tokens": 2899434.0, "step": 171 }, { "epoch": 0.0344, "grad_norm": 12.27084732055664, "learning_rate": 4.829000000000001e-06, "loss": 9.5638, "mean_token_accuracy": 0.33095238357782364, "num_tokens": 2913932.0, "step": 172 }, { "epoch": 0.0346, "grad_norm": 14.023222923278809, "learning_rate": 4.828e-06, "loss": 9.689, "mean_token_accuracy": 0.31680162250995636, "num_tokens": 2928356.0, "step": 173 }, { "epoch": 0.0348, "grad_norm": 14.490949630737305, "learning_rate": 4.8270000000000005e-06, "loss": 8.8018, "mean_token_accuracy": 0.34068627655506134, "num_tokens": 2940589.0, "step": 174 }, { "epoch": 0.035, "grad_norm": 17.97809410095215, "learning_rate": 4.826e-06, "loss": 9.9556, "mean_token_accuracy": 0.3452381044626236, "num_tokens": 2959486.0, "step": 175 }, { "epoch": 0.0352, "grad_norm": 13.302875518798828, "learning_rate": 4.825e-06, "loss": 9.4104, "mean_token_accuracy": 0.23885918408632278, "num_tokens": 2974195.0, "step": 176 }, { "epoch": 0.0354, "grad_norm": 12.792606353759766, "learning_rate": 4.824000000000001e-06, "loss": 8.2518, "mean_token_accuracy": 0.42592592537403107, "num_tokens": 2984011.0, "step": 177 }, { "epoch": 0.0356, "grad_norm": 18.23525619506836, "learning_rate": 4.823e-06, "loss": 8.8069, "mean_token_accuracy": 0.40079365670681, "num_tokens": 2998487.0, "step": 178 }, { "epoch": 0.0358, "grad_norm": 14.206355094909668, "learning_rate": 4.822000000000001e-06, "loss": 9.1892, "mean_token_accuracy": 0.2290322557091713, "num_tokens": 3019773.0, "step": 179 }, { "epoch": 0.036, "grad_norm": 12.26903247833252, "learning_rate": 4.821e-06, "loss": 8.4939, "mean_token_accuracy": 0.28287841379642487, "num_tokens": 3040775.0, "step": 180 }, { "epoch": 0.0362, "grad_norm": 15.23544979095459, "learning_rate": 4.8200000000000004e-06, "loss": 8.9121, "mean_token_accuracy": 0.3285440653562546, "num_tokens": 3060080.0, "step": 181 }, { "epoch": 0.0364, "grad_norm": 57.132049560546875, "learning_rate": 4.819e-06, "loss": 8.9738, "mean_token_accuracy": 0.3175750821828842, "num_tokens": 3079332.0, "step": 182 }, { "epoch": 0.0366, "grad_norm": 13.691211700439453, "learning_rate": 4.818e-06, "loss": 8.0025, "mean_token_accuracy": 0.34656085073947906, "num_tokens": 3088989.0, "step": 183 }, { "epoch": 0.0368, "grad_norm": 15.762035369873047, "learning_rate": 4.817000000000001e-06, "loss": 9.4602, "mean_token_accuracy": 0.34457671642303467, "num_tokens": 3107732.0, "step": 184 }, { "epoch": 0.037, "grad_norm": 17.034019470214844, "learning_rate": 4.816e-06, "loss": 9.4491, "mean_token_accuracy": 0.34666667878627777, "num_tokens": 3122423.0, "step": 185 }, { "epoch": 0.0372, "grad_norm": 12.461385726928711, "learning_rate": 4.8150000000000005e-06, "loss": 9.0659, "mean_token_accuracy": 0.2838345915079117, "num_tokens": 3141889.0, "step": 186 }, { "epoch": 0.0374, "grad_norm": 13.194416046142578, "learning_rate": 4.814e-06, "loss": 8.8107, "mean_token_accuracy": 0.36685824394226074, "num_tokens": 3161194.0, "step": 187 }, { "epoch": 0.0376, "grad_norm": 14.799727439880371, "learning_rate": 4.813e-06, "loss": 9.7196, "mean_token_accuracy": 0.377616748213768, "num_tokens": 3175646.0, "step": 188 }, { "epoch": 0.0378, "grad_norm": 15.303200721740723, "learning_rate": 4.812000000000001e-06, "loss": 10.0029, "mean_token_accuracy": 0.38161374628543854, "num_tokens": 3194941.0, "step": 189 }, { "epoch": 0.038, "grad_norm": 13.99432373046875, "learning_rate": 4.811000000000001e-06, "loss": 8.4004, "mean_token_accuracy": 0.37096773087978363, "num_tokens": 3214205.0, "step": 190 }, { "epoch": 0.0382, "grad_norm": 12.163537979125977, "learning_rate": 4.8100000000000005e-06, "loss": 9.0058, "mean_token_accuracy": 0.35395538806915283, "num_tokens": 3235063.0, "step": 191 }, { "epoch": 0.0384, "grad_norm": 12.369959831237793, "learning_rate": 4.809e-06, "loss": 9.0358, "mean_token_accuracy": 0.4807407408952713, "num_tokens": 3249517.0, "step": 192 }, { "epoch": 0.0386, "grad_norm": 21.97284698486328, "learning_rate": 4.808e-06, "loss": 8.7725, "mean_token_accuracy": 0.41179338097572327, "num_tokens": 3268738.0, "step": 193 }, { "epoch": 0.0388, "grad_norm": 13.040814399719238, "learning_rate": 4.807000000000001e-06, "loss": 9.4628, "mean_token_accuracy": 0.424450546503067, "num_tokens": 3288032.0, "step": 194 }, { "epoch": 0.039, "grad_norm": 14.395992279052734, "learning_rate": 4.806000000000001e-06, "loss": 9.1626, "mean_token_accuracy": 0.454365074634552, "num_tokens": 3306813.0, "step": 195 }, { "epoch": 0.0392, "grad_norm": 11.64809799194336, "learning_rate": 4.805000000000001e-06, "loss": 8.7527, "mean_token_accuracy": 0.3843159079551697, "num_tokens": 3325561.0, "step": 196 }, { "epoch": 0.0394, "grad_norm": 25.551607131958008, "learning_rate": 4.804e-06, "loss": 8.5253, "mean_token_accuracy": 0.3452381044626236, "num_tokens": 3339397.0, "step": 197 }, { "epoch": 0.0396, "grad_norm": 12.025030136108398, "learning_rate": 4.8030000000000004e-06, "loss": 7.7581, "mean_token_accuracy": 0.5191570967435837, "num_tokens": 3349055.0, "step": 198 }, { "epoch": 0.0398, "grad_norm": 10.044900894165039, "learning_rate": 4.802000000000001e-06, "loss": 9.0733, "mean_token_accuracy": 0.4278416335582733, "num_tokens": 3368313.0, "step": 199 }, { "epoch": 0.04, "grad_norm": 10.911112785339355, "learning_rate": 4.801e-06, "loss": 8.6906, "mean_token_accuracy": 0.5000000149011612, "num_tokens": 3383810.0, "step": 200 }, { "epoch": 0.0402, "grad_norm": 10.119377136230469, "learning_rate": 4.800000000000001e-06, "loss": 8.8941, "mean_token_accuracy": 0.4000000059604645, "num_tokens": 3398244.0, "step": 201 }, { "epoch": 0.0404, "grad_norm": 10.526436805725098, "learning_rate": 4.799e-06, "loss": 7.8408, "mean_token_accuracy": 0.39772726595401764, "num_tokens": 3414686.0, "step": 202 }, { "epoch": 0.0406, "grad_norm": 10.947959899902344, "learning_rate": 4.7980000000000005e-06, "loss": 8.4085, "mean_token_accuracy": 0.40992647409439087, "num_tokens": 3433954.0, "step": 203 }, { "epoch": 0.0408, "grad_norm": 10.811299324035645, "learning_rate": 4.797000000000001e-06, "loss": 7.7418, "mean_token_accuracy": 0.5370370447635651, "num_tokens": 3443610.0, "step": 204 }, { "epoch": 0.041, "grad_norm": 9.57394027709961, "learning_rate": 4.796e-06, "loss": 8.6595, "mean_token_accuracy": 0.45967741310596466, "num_tokens": 3462909.0, "step": 205 }, { "epoch": 0.0412, "grad_norm": 12.8336181640625, "learning_rate": 4.795e-06, "loss": 8.6943, "mean_token_accuracy": 0.45967741310596466, "num_tokens": 3477620.0, "step": 206 }, { "epoch": 0.0414, "grad_norm": 11.37842845916748, "learning_rate": 4.794e-06, "loss": 8.5963, "mean_token_accuracy": 0.4539627134799957, "num_tokens": 3496675.0, "step": 207 }, { "epoch": 0.0416, "grad_norm": 12.427331924438477, "learning_rate": 4.7930000000000005e-06, "loss": 8.67, "mean_token_accuracy": 0.3821548819541931, "num_tokens": 3515975.0, "step": 208 }, { "epoch": 0.0418, "grad_norm": 10.371417045593262, "learning_rate": 4.792000000000001e-06, "loss": 8.4859, "mean_token_accuracy": 0.40784314274787903, "num_tokens": 3535241.0, "step": 209 }, { "epoch": 0.042, "grad_norm": 11.788932800292969, "learning_rate": 4.791e-06, "loss": 8.7143, "mean_token_accuracy": 0.3741379380226135, "num_tokens": 3554502.0, "step": 210 }, { "epoch": 0.0422, "grad_norm": 15.102238655090332, "learning_rate": 4.79e-06, "loss": 8.5368, "mean_token_accuracy": 0.40980392694473267, "num_tokens": 3574437.0, "step": 211 }, { "epoch": 0.0424, "grad_norm": 11.23690128326416, "learning_rate": 4.789e-06, "loss": 8.7326, "mean_token_accuracy": 0.45628078281879425, "num_tokens": 3593696.0, "step": 212 }, { "epoch": 0.0426, "grad_norm": 11.3884859085083, "learning_rate": 4.7880000000000006e-06, "loss": 9.0773, "mean_token_accuracy": 0.4404761791229248, "num_tokens": 3608150.0, "step": 213 }, { "epoch": 0.0428, "grad_norm": 11.106508255004883, "learning_rate": 4.787000000000001e-06, "loss": 9.5678, "mean_token_accuracy": 0.3500000089406967, "num_tokens": 3627724.0, "step": 214 }, { "epoch": 0.043, "grad_norm": 11.691924095153809, "learning_rate": 4.7860000000000004e-06, "loss": 8.2192, "mean_token_accuracy": 0.41582491993904114, "num_tokens": 3642224.0, "step": 215 }, { "epoch": 0.0432, "grad_norm": 13.973259925842285, "learning_rate": 4.785e-06, "loss": 8.9823, "mean_token_accuracy": 0.44195401668548584, "num_tokens": 3660885.0, "step": 216 }, { "epoch": 0.0434, "grad_norm": 11.744901657104492, "learning_rate": 4.784e-06, "loss": 8.698, "mean_token_accuracy": 0.4291125535964966, "num_tokens": 3679634.0, "step": 217 }, { "epoch": 0.0436, "grad_norm": 13.440972328186035, "learning_rate": 4.783000000000001e-06, "loss": 8.3414, "mean_token_accuracy": 0.47999998927116394, "num_tokens": 3689286.0, "step": 218 }, { "epoch": 0.0438, "grad_norm": 9.561469078063965, "learning_rate": 4.782e-06, "loss": 7.6889, "mean_token_accuracy": 0.41898825764656067, "num_tokens": 3710533.0, "step": 219 }, { "epoch": 0.044, "grad_norm": 12.257551193237305, "learning_rate": 4.7810000000000005e-06, "loss": 8.4541, "mean_token_accuracy": 0.5016103088855743, "num_tokens": 3725145.0, "step": 220 }, { "epoch": 0.0442, "grad_norm": 10.785005569458008, "learning_rate": 4.78e-06, "loss": 9.7029, "mean_token_accuracy": 0.4015151560306549, "num_tokens": 3743544.0, "step": 221 }, { "epoch": 0.0444, "grad_norm": 10.52768611907959, "learning_rate": 4.779e-06, "loss": 8.6062, "mean_token_accuracy": 0.40740741789340973, "num_tokens": 3762705.0, "step": 222 }, { "epoch": 0.0446, "grad_norm": 17.872329711914062, "learning_rate": 4.778000000000001e-06, "loss": 7.3673, "mean_token_accuracy": 0.45370370149612427, "num_tokens": 3777166.0, "step": 223 }, { "epoch": 0.0448, "grad_norm": 11.053666114807129, "learning_rate": 4.777e-06, "loss": 9.3428, "mean_token_accuracy": 0.4186507910490036, "num_tokens": 3795909.0, "step": 224 }, { "epoch": 0.045, "grad_norm": 9.966497421264648, "learning_rate": 4.7760000000000005e-06, "loss": 7.9071, "mean_token_accuracy": 0.44636015594005585, "num_tokens": 3815176.0, "step": 225 }, { "epoch": 0.0452, "grad_norm": 12.605799674987793, "learning_rate": 4.775e-06, "loss": 8.3063, "mean_token_accuracy": 0.41692790389060974, "num_tokens": 3834478.0, "step": 226 }, { "epoch": 0.0454, "grad_norm": 9.679677963256836, "learning_rate": 4.774e-06, "loss": 8.7331, "mean_token_accuracy": 0.45628078281879425, "num_tokens": 3853737.0, "step": 227 }, { "epoch": 0.0456, "grad_norm": 27.174549102783203, "learning_rate": 4.773000000000001e-06, "loss": 8.0078, "mean_token_accuracy": 0.4434434473514557, "num_tokens": 3873882.0, "step": 228 }, { "epoch": 0.0458, "grad_norm": 11.591468811035156, "learning_rate": 4.772e-06, "loss": 8.237, "mean_token_accuracy": 0.36707451939582825, "num_tokens": 3888302.0, "step": 229 }, { "epoch": 0.046, "grad_norm": 10.10312271118164, "learning_rate": 4.7710000000000006e-06, "loss": 7.6694, "mean_token_accuracy": 0.5105820149183273, "num_tokens": 3902804.0, "step": 230 }, { "epoch": 0.0462, "grad_norm": 13.620348930358887, "learning_rate": 4.77e-06, "loss": 8.4942, "mean_token_accuracy": 0.421875, "num_tokens": 3922146.0, "step": 231 }, { "epoch": 0.0464, "grad_norm": 11.014819145202637, "learning_rate": 4.769e-06, "loss": 7.637, "mean_token_accuracy": 0.4495798349380493, "num_tokens": 3936372.0, "step": 232 }, { "epoch": 0.0466, "grad_norm": 14.10721492767334, "learning_rate": 4.768000000000001e-06, "loss": 7.4068, "mean_token_accuracy": 0.5078571289777756, "num_tokens": 3946456.0, "step": 233 }, { "epoch": 0.0468, "grad_norm": 13.256854057312012, "learning_rate": 4.767e-06, "loss": 7.9252, "mean_token_accuracy": 0.41277891397476196, "num_tokens": 3965955.0, "step": 234 }, { "epoch": 0.047, "grad_norm": 12.264280319213867, "learning_rate": 4.766000000000001e-06, "loss": 7.6561, "mean_token_accuracy": 0.4913793057203293, "num_tokens": 3980412.0, "step": 235 }, { "epoch": 0.0472, "grad_norm": 11.942499160766602, "learning_rate": 4.765e-06, "loss": 8.0462, "mean_token_accuracy": 0.4900284856557846, "num_tokens": 3999823.0, "step": 236 }, { "epoch": 0.0474, "grad_norm": 10.043482780456543, "learning_rate": 4.7640000000000005e-06, "loss": 7.0965, "mean_token_accuracy": 0.4025973975658417, "num_tokens": 4014414.0, "step": 237 }, { "epoch": 0.0476, "grad_norm": 12.545036315917969, "learning_rate": 4.763000000000001e-06, "loss": 7.667, "mean_token_accuracy": 0.38708220422267914, "num_tokens": 4028922.0, "step": 238 }, { "epoch": 0.0478, "grad_norm": 10.067218780517578, "learning_rate": 4.762e-06, "loss": 7.1117, "mean_token_accuracy": 0.37129031121730804, "num_tokens": 4050175.0, "step": 239 }, { "epoch": 0.048, "grad_norm": 13.74410343170166, "learning_rate": 4.761000000000001e-06, "loss": 7.9069, "mean_token_accuracy": 0.47333332896232605, "num_tokens": 4064811.0, "step": 240 }, { "epoch": 0.0482, "grad_norm": 9.813583374023438, "learning_rate": 4.76e-06, "loss": 6.8128, "mean_token_accuracy": 0.4273170679807663, "num_tokens": 4079317.0, "step": 241 }, { "epoch": 0.0484, "grad_norm": 10.31633472442627, "learning_rate": 4.7590000000000005e-06, "loss": 7.676, "mean_token_accuracy": 0.43584655225276947, "num_tokens": 4094093.0, "step": 242 }, { "epoch": 0.0486, "grad_norm": 13.174894332885742, "learning_rate": 4.758e-06, "loss": 8.5029, "mean_token_accuracy": 0.42592592537403107, "num_tokens": 4113778.0, "step": 243 }, { "epoch": 0.0488, "grad_norm": 11.068340301513672, "learning_rate": 4.757e-06, "loss": 7.6392, "mean_token_accuracy": 0.5028571337461472, "num_tokens": 4125078.0, "step": 244 }, { "epoch": 0.049, "grad_norm": 11.669493675231934, "learning_rate": 4.756000000000001e-06, "loss": 8.0877, "mean_token_accuracy": 0.3697916716337204, "num_tokens": 4144451.0, "step": 245 }, { "epoch": 0.0492, "grad_norm": 12.121454238891602, "learning_rate": 4.755e-06, "loss": 8.3639, "mean_token_accuracy": 0.3320707082748413, "num_tokens": 4164189.0, "step": 246 }, { "epoch": 0.0494, "grad_norm": 11.231935501098633, "learning_rate": 4.7540000000000006e-06, "loss": 8.1039, "mean_token_accuracy": 0.40060852468013763, "num_tokens": 4183560.0, "step": 247 }, { "epoch": 0.0496, "grad_norm": 14.818300247192383, "learning_rate": 4.753e-06, "loss": 8.3393, "mean_token_accuracy": 0.4365079402923584, "num_tokens": 4203160.0, "step": 248 }, { "epoch": 0.0498, "grad_norm": 10.595967292785645, "learning_rate": 4.752e-06, "loss": 8.2095, "mean_token_accuracy": 0.3896551728248596, "num_tokens": 4222502.0, "step": 249 }, { "epoch": 0.05, "grad_norm": 10.283987998962402, "learning_rate": 4.751000000000001e-06, "loss": 7.8432, "mean_token_accuracy": 0.41187499463558197, "num_tokens": 4236999.0, "step": 250 }, { "epoch": 0.0502, "grad_norm": 64.06322479248047, "learning_rate": 4.75e-06, "loss": 8.2479, "mean_token_accuracy": 0.43703703582286835, "num_tokens": 4256296.0, "step": 251 }, { "epoch": 0.0504, "grad_norm": 12.750511169433594, "learning_rate": 4.749000000000001e-06, "loss": 8.0847, "mean_token_accuracy": 0.4107142835855484, "num_tokens": 4270240.0, "step": 252 }, { "epoch": 0.0506, "grad_norm": 19.006969451904297, "learning_rate": 4.748e-06, "loss": 7.6973, "mean_token_accuracy": 0.44636015594005585, "num_tokens": 4284595.0, "step": 253 }, { "epoch": 0.0508, "grad_norm": 11.408003807067871, "learning_rate": 4.7470000000000005e-06, "loss": 7.7927, "mean_token_accuracy": 0.41428571939468384, "num_tokens": 4303855.0, "step": 254 }, { "epoch": 0.051, "grad_norm": 11.841231346130371, "learning_rate": 4.746000000000001e-06, "loss": 7.7046, "mean_token_accuracy": 0.3849431872367859, "num_tokens": 4323122.0, "step": 255 }, { "epoch": 0.0512, "grad_norm": 16.837949752807617, "learning_rate": 4.745e-06, "loss": 7.7556, "mean_token_accuracy": 0.38112305104732513, "num_tokens": 4342022.0, "step": 256 }, { "epoch": 0.0514, "grad_norm": 12.48375415802002, "learning_rate": 4.744000000000001e-06, "loss": 8.2711, "mean_token_accuracy": 0.42356322705745697, "num_tokens": 4361283.0, "step": 257 }, { "epoch": 0.0516, "grad_norm": 12.102031707763672, "learning_rate": 4.743e-06, "loss": 7.6877, "mean_token_accuracy": 0.42748479545116425, "num_tokens": 4375550.0, "step": 258 }, { "epoch": 0.0518, "grad_norm": 10.918121337890625, "learning_rate": 4.7420000000000005e-06, "loss": 7.1491, "mean_token_accuracy": 0.3727777749300003, "num_tokens": 4390013.0, "step": 259 }, { "epoch": 0.052, "grad_norm": 9.632317543029785, "learning_rate": 4.741000000000001e-06, "loss": 6.4275, "mean_token_accuracy": 0.3046218603849411, "num_tokens": 4406480.0, "step": 260 }, { "epoch": 0.0522, "grad_norm": 14.100777626037598, "learning_rate": 4.74e-06, "loss": 7.8382, "mean_token_accuracy": 0.454365074634552, "num_tokens": 4425737.0, "step": 261 }, { "epoch": 0.0524, "grad_norm": 14.973969459533691, "learning_rate": 4.739e-06, "loss": 7.775, "mean_token_accuracy": 0.45428571105003357, "num_tokens": 4440150.0, "step": 262 }, { "epoch": 0.0526, "grad_norm": 10.306495666503906, "learning_rate": 4.738e-06, "loss": 6.7966, "mean_token_accuracy": 0.41637930274009705, "num_tokens": 4454659.0, "step": 263 }, { "epoch": 0.0528, "grad_norm": 11.71308422088623, "learning_rate": 4.7370000000000006e-06, "loss": 7.3197, "mean_token_accuracy": 0.45000000298023224, "num_tokens": 4469178.0, "step": 264 }, { "epoch": 0.053, "grad_norm": 15.400917053222656, "learning_rate": 4.736000000000001e-06, "loss": 7.0285, "mean_token_accuracy": 0.4464285671710968, "num_tokens": 4483833.0, "step": 265 }, { "epoch": 0.0532, "grad_norm": 11.898533821105957, "learning_rate": 4.735e-06, "loss": 7.7648, "mean_token_accuracy": 0.3568965643644333, "num_tokens": 4502729.0, "step": 266 }, { "epoch": 0.0534, "grad_norm": 12.848722457885742, "learning_rate": 4.734e-06, "loss": 7.1883, "mean_token_accuracy": 0.4402298927307129, "num_tokens": 4521713.0, "step": 267 }, { "epoch": 0.0536, "grad_norm": 9.758365631103516, "learning_rate": 4.733e-06, "loss": 6.833, "mean_token_accuracy": 0.39444445073604584, "num_tokens": 4541033.0, "step": 268 }, { "epoch": 0.0538, "grad_norm": 40.770198822021484, "learning_rate": 4.732000000000001e-06, "loss": 7.5144, "mean_token_accuracy": 0.37254129350185394, "num_tokens": 4561917.0, "step": 269 }, { "epoch": 0.054, "grad_norm": 12.640328407287598, "learning_rate": 4.731000000000001e-06, "loss": 7.6694, "mean_token_accuracy": 0.40518517792224884, "num_tokens": 4576371.0, "step": 270 }, { "epoch": 0.0542, "grad_norm": 13.891765594482422, "learning_rate": 4.7300000000000005e-06, "loss": 6.4951, "mean_token_accuracy": 0.48746199905872345, "num_tokens": 4595606.0, "step": 271 }, { "epoch": 0.0544, "grad_norm": 15.260784149169922, "learning_rate": 4.729e-06, "loss": 7.1435, "mean_token_accuracy": 0.4378078728914261, "num_tokens": 4607665.0, "step": 272 }, { "epoch": 0.0546, "grad_norm": 11.873882293701172, "learning_rate": 4.728e-06, "loss": 6.6643, "mean_token_accuracy": 0.4196428507566452, "num_tokens": 4622184.0, "step": 273 }, { "epoch": 0.0548, "grad_norm": 11.417704582214355, "learning_rate": 4.727000000000001e-06, "loss": 7.0497, "mean_token_accuracy": 0.39933258295059204, "num_tokens": 4636646.0, "step": 274 }, { "epoch": 0.055, "grad_norm": 10.876029014587402, "learning_rate": 4.726000000000001e-06, "loss": 6.913, "mean_token_accuracy": 0.41692790389060974, "num_tokens": 4655906.0, "step": 275 }, { "epoch": 0.0552, "grad_norm": 13.294316291809082, "learning_rate": 4.7250000000000005e-06, "loss": 7.6388, "mean_token_accuracy": 0.41338111460208893, "num_tokens": 4675326.0, "step": 276 }, { "epoch": 0.0554, "grad_norm": 13.638504981994629, "learning_rate": 4.724e-06, "loss": 7.3977, "mean_token_accuracy": 0.3982202410697937, "num_tokens": 4694348.0, "step": 277 }, { "epoch": 0.0556, "grad_norm": 15.715910911560059, "learning_rate": 4.723e-06, "loss": 7.6206, "mean_token_accuracy": 0.4444444477558136, "num_tokens": 4713642.0, "step": 278 }, { "epoch": 0.0558, "grad_norm": 15.516464233398438, "learning_rate": 4.722000000000001e-06, "loss": 8.2007, "mean_token_accuracy": 0.424450546503067, "num_tokens": 4732898.0, "step": 279 }, { "epoch": 0.056, "grad_norm": 12.387885093688965, "learning_rate": 4.721e-06, "loss": 7.688, "mean_token_accuracy": 0.3694581240415573, "num_tokens": 4752274.0, "step": 280 }, { "epoch": 0.0562, "grad_norm": 10.614646911621094, "learning_rate": 4.7200000000000005e-06, "loss": 6.3429, "mean_token_accuracy": 0.47413793206214905, "num_tokens": 4761929.0, "step": 281 }, { "epoch": 0.0564, "grad_norm": 21.08198356628418, "learning_rate": 4.719e-06, "loss": 6.1852, "mean_token_accuracy": 0.46727272868156433, "num_tokens": 4771566.0, "step": 282 }, { "epoch": 0.0566, "grad_norm": 13.566423416137695, "learning_rate": 4.718e-06, "loss": 6.3183, "mean_token_accuracy": 0.4772357791662216, "num_tokens": 4792969.0, "step": 283 }, { "epoch": 0.0568, "grad_norm": 11.979252815246582, "learning_rate": 4.717000000000001e-06, "loss": 6.9839, "mean_token_accuracy": 0.4107142835855484, "num_tokens": 4807544.0, "step": 284 }, { "epoch": 0.057, "grad_norm": 10.196455955505371, "learning_rate": 4.716e-06, "loss": 7.2482, "mean_token_accuracy": 0.4211822748184204, "num_tokens": 4826443.0, "step": 285 }, { "epoch": 0.0572, "grad_norm": 10.696710586547852, "learning_rate": 4.715e-06, "loss": 6.7498, "mean_token_accuracy": 0.4461206793785095, "num_tokens": 4840925.0, "step": 286 }, { "epoch": 0.0574, "grad_norm": 12.140236854553223, "learning_rate": 4.714e-06, "loss": 7.5622, "mean_token_accuracy": 0.3907563090324402, "num_tokens": 4860066.0, "step": 287 }, { "epoch": 0.0576, "grad_norm": 9.046162605285645, "learning_rate": 4.7130000000000004e-06, "loss": 6.4519, "mean_token_accuracy": 0.4005681872367859, "num_tokens": 4874533.0, "step": 288 }, { "epoch": 0.0578, "grad_norm": 11.239535331726074, "learning_rate": 4.712000000000001e-06, "loss": 7.0541, "mean_token_accuracy": 0.4107142835855484, "num_tokens": 4888991.0, "step": 289 }, { "epoch": 0.058, "grad_norm": 13.39867115020752, "learning_rate": 4.711e-06, "loss": 7.3852, "mean_token_accuracy": 0.35574713349342346, "num_tokens": 4901961.0, "step": 290 }, { "epoch": 0.0582, "grad_norm": 10.7579984664917, "learning_rate": 4.71e-06, "loss": 7.2866, "mean_token_accuracy": 0.41428571939468384, "num_tokens": 4916815.0, "step": 291 }, { "epoch": 0.0584, "grad_norm": 12.446432113647461, "learning_rate": 4.709e-06, "loss": 7.9278, "mean_token_accuracy": 0.35706017911434174, "num_tokens": 4935892.0, "step": 292 }, { "epoch": 0.0586, "grad_norm": 12.149531364440918, "learning_rate": 4.7080000000000005e-06, "loss": 7.0934, "mean_token_accuracy": 0.4106002599000931, "num_tokens": 4950308.0, "step": 293 }, { "epoch": 0.0588, "grad_norm": 12.248196601867676, "learning_rate": 4.707000000000001e-06, "loss": 7.3677, "mean_token_accuracy": 0.4555555582046509, "num_tokens": 4969605.0, "step": 294 }, { "epoch": 0.059, "grad_norm": 10.024053573608398, "learning_rate": 4.706e-06, "loss": 7.0171, "mean_token_accuracy": 0.4404466450214386, "num_tokens": 4984384.0, "step": 295 }, { "epoch": 0.0592, "grad_norm": 10.064780235290527, "learning_rate": 4.705e-06, "loss": 7.2282, "mean_token_accuracy": 0.3821621537208557, "num_tokens": 4998886.0, "step": 296 }, { "epoch": 0.0594, "grad_norm": 11.68379020690918, "learning_rate": 4.704e-06, "loss": 6.3511, "mean_token_accuracy": 0.45874999463558197, "num_tokens": 5013345.0, "step": 297 }, { "epoch": 0.0596, "grad_norm": 9.920818328857422, "learning_rate": 4.7030000000000005e-06, "loss": 6.4502, "mean_token_accuracy": 0.4145299196243286, "num_tokens": 5026411.0, "step": 298 }, { "epoch": 0.0598, "grad_norm": 9.868770599365234, "learning_rate": 4.702e-06, "loss": 6.6577, "mean_token_accuracy": 0.40407469868659973, "num_tokens": 5046124.0, "step": 299 }, { "epoch": 0.06, "grad_norm": 13.261858940124512, "learning_rate": 4.701e-06, "loss": 6.0468, "mean_token_accuracy": 0.4592592567205429, "num_tokens": 5060802.0, "step": 300 }, { "epoch": 0.0602, "grad_norm": 11.705483436584473, "learning_rate": 4.7e-06, "loss": 6.9336, "mean_token_accuracy": 0.3961039036512375, "num_tokens": 5080148.0, "step": 301 }, { "epoch": 0.0604, "grad_norm": 11.132308006286621, "learning_rate": 4.699e-06, "loss": 6.7934, "mean_token_accuracy": 0.4444444477558136, "num_tokens": 5094604.0, "step": 302 }, { "epoch": 0.0606, "grad_norm": 12.15768051147461, "learning_rate": 4.698000000000001e-06, "loss": 6.1426, "mean_token_accuracy": 0.48148147761821747, "num_tokens": 5109079.0, "step": 303 }, { "epoch": 0.0608, "grad_norm": 13.979254722595215, "learning_rate": 4.697e-06, "loss": 7.0785, "mean_token_accuracy": 0.4423076957464218, "num_tokens": 5121880.0, "step": 304 }, { "epoch": 0.061, "grad_norm": 10.67962646484375, "learning_rate": 4.6960000000000004e-06, "loss": 5.7797, "mean_token_accuracy": 0.43860387802124023, "num_tokens": 5141369.0, "step": 305 }, { "epoch": 0.0612, "grad_norm": 9.39137077331543, "learning_rate": 4.695e-06, "loss": 5.9316, "mean_token_accuracy": 0.4432692378759384, "num_tokens": 5155875.0, "step": 306 }, { "epoch": 0.0614, "grad_norm": 10.971794128417969, "learning_rate": 4.694e-06, "loss": 6.5844, "mean_token_accuracy": 0.4044642895460129, "num_tokens": 5175079.0, "step": 307 }, { "epoch": 0.0616, "grad_norm": 10.385963439941406, "learning_rate": 4.693000000000001e-06, "loss": 6.3867, "mean_token_accuracy": 0.46296297013759613, "num_tokens": 5189415.0, "step": 308 }, { "epoch": 0.0618, "grad_norm": 13.314301490783691, "learning_rate": 4.692e-06, "loss": 6.0294, "mean_token_accuracy": 0.42749999463558197, "num_tokens": 5199112.0, "step": 309 }, { "epoch": 0.062, "grad_norm": 9.319266319274902, "learning_rate": 4.6910000000000005e-06, "loss": 6.3149, "mean_token_accuracy": 0.41097819805145264, "num_tokens": 5231220.0, "step": 310 }, { "epoch": 0.0622, "grad_norm": 9.54952621459961, "learning_rate": 4.69e-06, "loss": 6.8163, "mean_token_accuracy": 0.380952388048172, "num_tokens": 5253011.0, "step": 311 }, { "epoch": 0.0624, "grad_norm": 9.803725242614746, "learning_rate": 4.689e-06, "loss": 6.385, "mean_token_accuracy": 0.4375, "num_tokens": 5272315.0, "step": 312 }, { "epoch": 0.0626, "grad_norm": 9.163276672363281, "learning_rate": 4.688000000000001e-06, "loss": 5.9736, "mean_token_accuracy": 0.3640737682580948, "num_tokens": 5286789.0, "step": 313 }, { "epoch": 0.0628, "grad_norm": 9.91296672821045, "learning_rate": 4.687e-06, "loss": 6.2111, "mean_token_accuracy": 0.3975849747657776, "num_tokens": 5301439.0, "step": 314 }, { "epoch": 0.063, "grad_norm": 11.019015312194824, "learning_rate": 4.6860000000000005e-06, "loss": 6.6715, "mean_token_accuracy": 0.3615301698446274, "num_tokens": 5324389.0, "step": 315 }, { "epoch": 0.0632, "grad_norm": 9.389239311218262, "learning_rate": 4.685000000000001e-06, "loss": 5.7869, "mean_token_accuracy": 0.4253472238779068, "num_tokens": 5334050.0, "step": 316 }, { "epoch": 0.0634, "grad_norm": 9.804286003112793, "learning_rate": 4.684e-06, "loss": 7.3263, "mean_token_accuracy": 0.3896551728248596, "num_tokens": 5353730.0, "step": 317 }, { "epoch": 0.0636, "grad_norm": 18.058841705322266, "learning_rate": 4.683000000000001e-06, "loss": 6.4049, "mean_token_accuracy": 0.43095238506793976, "num_tokens": 5368190.0, "step": 318 }, { "epoch": 0.0638, "grad_norm": 10.300615310668945, "learning_rate": 4.682e-06, "loss": 6.1908, "mean_token_accuracy": 0.44763730466365814, "num_tokens": 5377848.0, "step": 319 }, { "epoch": 0.064, "grad_norm": 8.114563941955566, "learning_rate": 4.681000000000001e-06, "loss": 6.5722, "mean_token_accuracy": 0.39539170265197754, "num_tokens": 5397362.0, "step": 320 }, { "epoch": 0.0642, "grad_norm": 9.576495170593262, "learning_rate": 4.680000000000001e-06, "loss": 5.8224, "mean_token_accuracy": 0.4186507910490036, "num_tokens": 5404613.0, "step": 321 }, { "epoch": 0.0644, "grad_norm": 7.686422348022461, "learning_rate": 4.6790000000000004e-06, "loss": 6.2079, "mean_token_accuracy": 0.3892773985862732, "num_tokens": 5423887.0, "step": 322 }, { "epoch": 0.0646, "grad_norm": 10.824250221252441, "learning_rate": 4.678e-06, "loss": 6.4209, "mean_token_accuracy": 0.380952388048172, "num_tokens": 5438544.0, "step": 323 }, { "epoch": 0.0648, "grad_norm": 22.814104080200195, "learning_rate": 4.677e-06, "loss": 6.3804, "mean_token_accuracy": 0.4807407408952713, "num_tokens": 5453036.0, "step": 324 }, { "epoch": 0.065, "grad_norm": 12.241607666015625, "learning_rate": 4.676000000000001e-06, "loss": 6.9619, "mean_token_accuracy": 0.44413793087005615, "num_tokens": 5472490.0, "step": 325 }, { "epoch": 0.0652, "grad_norm": 9.275449752807617, "learning_rate": 4.675000000000001e-06, "loss": 5.8423, "mean_token_accuracy": 0.44312499463558197, "num_tokens": 5482149.0, "step": 326 }, { "epoch": 0.0654, "grad_norm": 9.22095012664795, "learning_rate": 4.6740000000000005e-06, "loss": 5.6949, "mean_token_accuracy": 0.46666666865348816, "num_tokens": 5491805.0, "step": 327 }, { "epoch": 0.0656, "grad_norm": 9.241523742675781, "learning_rate": 4.673e-06, "loss": 6.0598, "mean_token_accuracy": 0.4404466450214386, "num_tokens": 5505904.0, "step": 328 }, { "epoch": 0.0658, "grad_norm": 17.82420539855957, "learning_rate": 4.672e-06, "loss": 6.6779, "mean_token_accuracy": 0.4031440168619156, "num_tokens": 5525385.0, "step": 329 }, { "epoch": 0.066, "grad_norm": 9.692024230957031, "learning_rate": 4.671000000000001e-06, "loss": 5.7944, "mean_token_accuracy": 0.5252873599529266, "num_tokens": 5540158.0, "step": 330 }, { "epoch": 0.0662, "grad_norm": 7.623928070068359, "learning_rate": 4.670000000000001e-06, "loss": 4.9237, "mean_token_accuracy": 0.44602273404598236, "num_tokens": 5563392.0, "step": 331 }, { "epoch": 0.0664, "grad_norm": 10.33786392211914, "learning_rate": 4.6690000000000005e-06, "loss": 6.147, "mean_token_accuracy": 0.45428571105003357, "num_tokens": 5577885.0, "step": 332 }, { "epoch": 0.0666, "grad_norm": 9.725800514221191, "learning_rate": 4.668e-06, "loss": 6.134, "mean_token_accuracy": 0.49074074625968933, "num_tokens": 5587538.0, "step": 333 }, { "epoch": 0.0668, "grad_norm": 12.834142684936523, "learning_rate": 4.667e-06, "loss": 6.4009, "mean_token_accuracy": 0.448306605219841, "num_tokens": 5606700.0, "step": 334 }, { "epoch": 0.067, "grad_norm": 17.030433654785156, "learning_rate": 4.666000000000001e-06, "loss": 6.5778, "mean_token_accuracy": 0.49074074625968933, "num_tokens": 5627246.0, "step": 335 }, { "epoch": 0.0672, "grad_norm": 7.939487457275391, "learning_rate": 4.665e-06, "loss": 5.705, "mean_token_accuracy": 0.39478765428066254, "num_tokens": 5646622.0, "step": 336 }, { "epoch": 0.0674, "grad_norm": 10.505208015441895, "learning_rate": 4.664000000000001e-06, "loss": 6.2186, "mean_token_accuracy": 0.4741935580968857, "num_tokens": 5665885.0, "step": 337 }, { "epoch": 0.0676, "grad_norm": 9.60438060760498, "learning_rate": 4.663e-06, "loss": 6.4067, "mean_token_accuracy": 0.50063855946064, "num_tokens": 5685143.0, "step": 338 }, { "epoch": 0.0678, "grad_norm": 8.420628547668457, "learning_rate": 4.6620000000000004e-06, "loss": 5.9981, "mean_token_accuracy": 0.48749999701976776, "num_tokens": 5705307.0, "step": 339 }, { "epoch": 0.068, "grad_norm": 9.02039909362793, "learning_rate": 4.661000000000001e-06, "loss": 5.6245, "mean_token_accuracy": 0.44069264829158783, "num_tokens": 5724259.0, "step": 340 }, { "epoch": 0.0682, "grad_norm": 9.671128273010254, "learning_rate": 4.66e-06, "loss": 5.537, "mean_token_accuracy": 0.5000000149011612, "num_tokens": 5738757.0, "step": 341 }, { "epoch": 0.0684, "grad_norm": 9.832850456237793, "learning_rate": 4.659e-06, "loss": 6.0835, "mean_token_accuracy": 0.5, "num_tokens": 5753257.0, "step": 342 }, { "epoch": 0.0686, "grad_norm": 10.410543441772461, "learning_rate": 4.658e-06, "loss": 6.4898, "mean_token_accuracy": 0.517241358757019, "num_tokens": 5769459.0, "step": 343 }, { "epoch": 0.0688, "grad_norm": 11.612724304199219, "learning_rate": 4.6570000000000005e-06, "loss": 6.2025, "mean_token_accuracy": 0.45967741310596466, "num_tokens": 5783916.0, "step": 344 }, { "epoch": 0.069, "grad_norm": 10.020609855651855, "learning_rate": 4.656000000000001e-06, "loss": 6.355, "mean_token_accuracy": 0.4568965584039688, "num_tokens": 5803333.0, "step": 345 }, { "epoch": 0.0692, "grad_norm": 8.622142791748047, "learning_rate": 4.655e-06, "loss": 5.8755, "mean_token_accuracy": 0.49944381415843964, "num_tokens": 5817852.0, "step": 346 }, { "epoch": 0.0694, "grad_norm": 6.586418151855469, "learning_rate": 4.654e-06, "loss": 4.6792, "mean_token_accuracy": 0.4610389471054077, "num_tokens": 5827955.0, "step": 347 }, { "epoch": 0.0696, "grad_norm": 9.011771202087402, "learning_rate": 4.653e-06, "loss": 5.4233, "mean_token_accuracy": 0.525862067937851, "num_tokens": 5842433.0, "step": 348 }, { "epoch": 0.0698, "grad_norm": 11.389575004577637, "learning_rate": 4.6520000000000005e-06, "loss": 6.12, "mean_token_accuracy": 0.5083612203598022, "num_tokens": 5856800.0, "step": 349 }, { "epoch": 0.07, "grad_norm": 7.888221263885498, "learning_rate": 4.651000000000001e-06, "loss": 5.8322, "mean_token_accuracy": 0.504524901509285, "num_tokens": 5870212.0, "step": 350 }, { "epoch": 0.0702, "grad_norm": 9.21318531036377, "learning_rate": 4.65e-06, "loss": 5.8042, "mean_token_accuracy": 0.5434027910232544, "num_tokens": 5889473.0, "step": 351 }, { "epoch": 0.0704, "grad_norm": 19.845186233520508, "learning_rate": 4.649e-06, "loss": 6.2884, "mean_token_accuracy": 0.4252873510122299, "num_tokens": 5900869.0, "step": 352 }, { "epoch": 0.0706, "grad_norm": 10.618582725524902, "learning_rate": 4.648e-06, "loss": 5.9085, "mean_token_accuracy": 0.4852941185235977, "num_tokens": 5919621.0, "step": 353 }, { "epoch": 0.0708, "grad_norm": 6.961612224578857, "learning_rate": 4.6470000000000006e-06, "loss": 5.1376, "mean_token_accuracy": 0.4638047069311142, "num_tokens": 5926888.0, "step": 354 }, { "epoch": 0.071, "grad_norm": 12.083121299743652, "learning_rate": 4.646000000000001e-06, "loss": 6.054, "mean_token_accuracy": 0.5, "num_tokens": 5946144.0, "step": 355 }, { "epoch": 0.0712, "grad_norm": 10.419790267944336, "learning_rate": 4.645e-06, "loss": 6.2987, "mean_token_accuracy": 0.4136960655450821, "num_tokens": 5965843.0, "step": 356 }, { "epoch": 0.0714, "grad_norm": 10.46291732788086, "learning_rate": 4.644e-06, "loss": 5.7, "mean_token_accuracy": 0.5392592549324036, "num_tokens": 5980726.0, "step": 357 }, { "epoch": 0.0716, "grad_norm": 11.251096725463867, "learning_rate": 4.643e-06, "loss": 5.5953, "mean_token_accuracy": 0.5185185372829437, "num_tokens": 5995182.0, "step": 358 }, { "epoch": 0.0718, "grad_norm": 9.273164749145508, "learning_rate": 4.642000000000001e-06, "loss": 5.7778, "mean_token_accuracy": 0.48938991129398346, "num_tokens": 6004839.0, "step": 359 }, { "epoch": 0.072, "grad_norm": 15.386749267578125, "learning_rate": 4.641e-06, "loss": 5.5941, "mean_token_accuracy": 0.49140210449695587, "num_tokens": 6019456.0, "step": 360 }, { "epoch": 0.0722, "grad_norm": 14.047779083251953, "learning_rate": 4.6400000000000005e-06, "loss": 6.0249, "mean_token_accuracy": 0.5166028141975403, "num_tokens": 6038752.0, "step": 361 }, { "epoch": 0.0724, "grad_norm": 7.4056925773620605, "learning_rate": 4.639e-06, "loss": 5.3375, "mean_token_accuracy": 0.5800000131130219, "num_tokens": 6048404.0, "step": 362 }, { "epoch": 0.0726, "grad_norm": 13.957375526428223, "learning_rate": 4.638e-06, "loss": 5.7739, "mean_token_accuracy": 0.561188817024231, "num_tokens": 6067581.0, "step": 363 }, { "epoch": 0.0728, "grad_norm": 10.774636268615723, "learning_rate": 4.637000000000001e-06, "loss": 6.0013, "mean_token_accuracy": 0.4900793582201004, "num_tokens": 6087734.0, "step": 364 }, { "epoch": 0.073, "grad_norm": 8.231881141662598, "learning_rate": 4.636e-06, "loss": 4.9933, "mean_token_accuracy": 0.6004464328289032, "num_tokens": 6102196.0, "step": 365 }, { "epoch": 0.0732, "grad_norm": 8.488526344299316, "learning_rate": 4.6350000000000005e-06, "loss": 5.0155, "mean_token_accuracy": 0.463299423456192, "num_tokens": 6116673.0, "step": 366 }, { "epoch": 0.0734, "grad_norm": 10.228729248046875, "learning_rate": 4.634e-06, "loss": 5.8042, "mean_token_accuracy": 0.5535714328289032, "num_tokens": 6136007.0, "step": 367 }, { "epoch": 0.0736, "grad_norm": 14.825092315673828, "learning_rate": 4.633e-06, "loss": 5.6991, "mean_token_accuracy": 0.5031928569078445, "num_tokens": 6154423.0, "step": 368 }, { "epoch": 0.0738, "grad_norm": 8.982471466064453, "learning_rate": 4.632000000000001e-06, "loss": 5.564, "mean_token_accuracy": 0.5177975445985794, "num_tokens": 6173864.0, "step": 369 }, { "epoch": 0.074, "grad_norm": 20.95796775817871, "learning_rate": 4.631e-06, "loss": 5.8344, "mean_token_accuracy": 0.4821200519800186, "num_tokens": 6193259.0, "step": 370 }, { "epoch": 0.0742, "grad_norm": 35.350887298583984, "learning_rate": 4.6300000000000006e-06, "loss": 5.7328, "mean_token_accuracy": 0.5106837600469589, "num_tokens": 6213732.0, "step": 371 }, { "epoch": 0.0744, "grad_norm": 13.841226577758789, "learning_rate": 4.629e-06, "loss": 5.9662, "mean_token_accuracy": 0.5092592537403107, "num_tokens": 6233169.0, "step": 372 }, { "epoch": 0.0746, "grad_norm": 12.297442436218262, "learning_rate": 4.628e-06, "loss": 5.3639, "mean_token_accuracy": 0.42847076058387756, "num_tokens": 6252724.0, "step": 373 }, { "epoch": 0.0748, "grad_norm": 8.141170501708984, "learning_rate": 4.627000000000001e-06, "loss": 5.1988, "mean_token_accuracy": 0.5410256534814835, "num_tokens": 6267220.0, "step": 374 }, { "epoch": 0.075, "grad_norm": 11.617218017578125, "learning_rate": 4.626e-06, "loss": 6.1014, "mean_token_accuracy": 0.5000000149011612, "num_tokens": 6281873.0, "step": 375 }, { "epoch": 0.0752, "grad_norm": 14.549817085266113, "learning_rate": 4.625000000000001e-06, "loss": 4.9924, "mean_token_accuracy": 0.6129629611968994, "num_tokens": 6296468.0, "step": 376 }, { "epoch": 0.0754, "grad_norm": 8.732035636901855, "learning_rate": 4.624e-06, "loss": 5.3749, "mean_token_accuracy": 0.49193547666072845, "num_tokens": 6316099.0, "step": 377 }, { "epoch": 0.0756, "grad_norm": 11.393025398254395, "learning_rate": 4.6230000000000005e-06, "loss": 5.8518, "mean_token_accuracy": 0.5462962985038757, "num_tokens": 6335358.0, "step": 378 }, { "epoch": 0.0758, "grad_norm": 11.22784423828125, "learning_rate": 4.622e-06, "loss": 5.6366, "mean_token_accuracy": 0.5166666805744171, "num_tokens": 6349895.0, "step": 379 }, { "epoch": 0.076, "grad_norm": 9.452168464660645, "learning_rate": 4.621e-06, "loss": 5.1428, "mean_token_accuracy": 0.48552632331848145, "num_tokens": 6371574.0, "step": 380 }, { "epoch": 0.0762, "grad_norm": 10.894210815429688, "learning_rate": 4.620000000000001e-06, "loss": 5.977, "mean_token_accuracy": 0.484375, "num_tokens": 6389904.0, "step": 381 }, { "epoch": 0.0764, "grad_norm": 13.074360847473145, "learning_rate": 4.619e-06, "loss": 5.3847, "mean_token_accuracy": 0.5892857313156128, "num_tokens": 6409200.0, "step": 382 }, { "epoch": 0.0766, "grad_norm": 11.248861312866211, "learning_rate": 4.6180000000000005e-06, "loss": 5.4468, "mean_token_accuracy": 0.550000011920929, "num_tokens": 6423778.0, "step": 383 }, { "epoch": 0.0768, "grad_norm": 9.336748123168945, "learning_rate": 4.617e-06, "loss": 5.4932, "mean_token_accuracy": 0.5820105969905853, "num_tokens": 6443035.0, "step": 384 }, { "epoch": 0.077, "grad_norm": 9.516325950622559, "learning_rate": 4.616e-06, "loss": 5.4311, "mean_token_accuracy": 0.4709596037864685, "num_tokens": 6459666.0, "step": 385 }, { "epoch": 0.0772, "grad_norm": 7.903703212738037, "learning_rate": 4.615000000000001e-06, "loss": 5.5287, "mean_token_accuracy": 0.541487067937851, "num_tokens": 6478423.0, "step": 386 }, { "epoch": 0.0774, "grad_norm": 13.11430835723877, "learning_rate": 4.614e-06, "loss": 5.3758, "mean_token_accuracy": 0.6235714256763458, "num_tokens": 6492916.0, "step": 387 }, { "epoch": 0.0776, "grad_norm": 14.428086280822754, "learning_rate": 4.6130000000000006e-06, "loss": 5.3445, "mean_token_accuracy": 0.5848214328289032, "num_tokens": 6512237.0, "step": 388 }, { "epoch": 0.0778, "grad_norm": 7.981903553009033, "learning_rate": 4.612e-06, "loss": 5.6573, "mean_token_accuracy": 0.581250011920929, "num_tokens": 6531501.0, "step": 389 }, { "epoch": 0.078, "grad_norm": 11.641196250915527, "learning_rate": 4.611e-06, "loss": 4.5256, "mean_token_accuracy": 0.5635062754154205, "num_tokens": 6545420.0, "step": 390 }, { "epoch": 0.0782, "grad_norm": 7.262872219085693, "learning_rate": 4.610000000000001e-06, "loss": 5.5162, "mean_token_accuracy": 0.5485893487930298, "num_tokens": 6564684.0, "step": 391 }, { "epoch": 0.0784, "grad_norm": 9.247142791748047, "learning_rate": 4.609e-06, "loss": 5.6358, "mean_token_accuracy": 0.5387205481529236, "num_tokens": 6584022.0, "step": 392 }, { "epoch": 0.0786, "grad_norm": 7.343860149383545, "learning_rate": 4.608000000000001e-06, "loss": 5.4758, "mean_token_accuracy": 0.5384615361690521, "num_tokens": 6598518.0, "step": 393 }, { "epoch": 0.0788, "grad_norm": 80.7061767578125, "learning_rate": 4.607e-06, "loss": 5.8124, "mean_token_accuracy": 0.5425287485122681, "num_tokens": 6617265.0, "step": 394 }, { "epoch": 0.079, "grad_norm": 8.777029991149902, "learning_rate": 4.6060000000000005e-06, "loss": 5.63, "mean_token_accuracy": 0.5334528088569641, "num_tokens": 6636418.0, "step": 395 }, { "epoch": 0.0792, "grad_norm": 9.333694458007812, "learning_rate": 4.605000000000001e-06, "loss": 5.5997, "mean_token_accuracy": 0.5315904170274734, "num_tokens": 6656184.0, "step": 396 }, { "epoch": 0.0794, "grad_norm": 8.860589981079102, "learning_rate": 4.604e-06, "loss": 5.1344, "mean_token_accuracy": 0.589756578207016, "num_tokens": 6675525.0, "step": 397 }, { "epoch": 0.0796, "grad_norm": 13.58896541595459, "learning_rate": 4.603000000000001e-06, "loss": 5.2149, "mean_token_accuracy": 0.5039232820272446, "num_tokens": 6693627.0, "step": 398 }, { "epoch": 0.0798, "grad_norm": 9.988759994506836, "learning_rate": 4.602e-06, "loss": 5.6091, "mean_token_accuracy": 0.5166666805744171, "num_tokens": 6708223.0, "step": 399 }, { "epoch": 0.08, "grad_norm": 9.320517539978027, "learning_rate": 4.6010000000000005e-06, "loss": 5.3271, "mean_token_accuracy": 0.5459558963775635, "num_tokens": 6728571.0, "step": 400 }, { "epoch": 0.0802, "grad_norm": 6.697946548461914, "learning_rate": 4.600000000000001e-06, "loss": 4.8994, "mean_token_accuracy": 0.5878571271896362, "num_tokens": 6737950.0, "step": 401 }, { "epoch": 0.0804, "grad_norm": 7.51831579208374, "learning_rate": 4.599e-06, "loss": 5.024, "mean_token_accuracy": 0.5743534564971924, "num_tokens": 6757373.0, "step": 402 }, { "epoch": 0.0806, "grad_norm": 9.05827522277832, "learning_rate": 4.598e-06, "loss": 4.8548, "mean_token_accuracy": 0.5397436022758484, "num_tokens": 6778652.0, "step": 403 }, { "epoch": 0.0808, "grad_norm": 8.396754264831543, "learning_rate": 4.597e-06, "loss": 4.8869, "mean_token_accuracy": 0.5226293057203293, "num_tokens": 6793115.0, "step": 404 }, { "epoch": 0.081, "grad_norm": 7.155496597290039, "learning_rate": 4.5960000000000006e-06, "loss": 5.2171, "mean_token_accuracy": 0.5760368704795837, "num_tokens": 6812376.0, "step": 405 }, { "epoch": 0.0812, "grad_norm": 18.729246139526367, "learning_rate": 4.595000000000001e-06, "loss": 5.3674, "mean_token_accuracy": 0.47413793206214905, "num_tokens": 6826681.0, "step": 406 }, { "epoch": 0.0814, "grad_norm": 5.5656046867370605, "learning_rate": 4.594e-06, "loss": 4.9008, "mean_token_accuracy": 0.5647321343421936, "num_tokens": 6836545.0, "step": 407 }, { "epoch": 0.0816, "grad_norm": 12.937115669250488, "learning_rate": 4.593e-06, "loss": 5.3085, "mean_token_accuracy": 0.5370370447635651, "num_tokens": 6855848.0, "step": 408 }, { "epoch": 0.0818, "grad_norm": 10.064362525939941, "learning_rate": 4.592e-06, "loss": 5.6301, "mean_token_accuracy": 0.5384615361690521, "num_tokens": 6871572.0, "step": 409 }, { "epoch": 0.082, "grad_norm": 11.348885536193848, "learning_rate": 4.591000000000001e-06, "loss": 5.2883, "mean_token_accuracy": 0.5475806295871735, "num_tokens": 6886045.0, "step": 410 }, { "epoch": 0.0822, "grad_norm": 12.09267520904541, "learning_rate": 4.590000000000001e-06, "loss": 5.1195, "mean_token_accuracy": 0.5264367908239365, "num_tokens": 6905344.0, "step": 411 }, { "epoch": 0.0824, "grad_norm": 6.413148403167725, "learning_rate": 4.5890000000000004e-06, "loss": 4.9556, "mean_token_accuracy": 0.613103449344635, "num_tokens": 6919857.0, "step": 412 }, { "epoch": 0.0826, "grad_norm": 14.665050506591797, "learning_rate": 4.588e-06, "loss": 5.375, "mean_token_accuracy": 0.5648148059844971, "num_tokens": 6939271.0, "step": 413 }, { "epoch": 0.0828, "grad_norm": 5.199728012084961, "learning_rate": 4.587e-06, "loss": 4.4413, "mean_token_accuracy": 0.5422222167253494, "num_tokens": 6949145.0, "step": 414 }, { "epoch": 0.083, "grad_norm": 22.955730438232422, "learning_rate": 4.586000000000001e-06, "loss": 4.4207, "mean_token_accuracy": 0.6060605943202972, "num_tokens": 6963199.0, "step": 415 }, { "epoch": 0.0832, "grad_norm": 5.997854709625244, "learning_rate": 4.585e-06, "loss": 4.5337, "mean_token_accuracy": 0.6601382493972778, "num_tokens": 6977679.0, "step": 416 }, { "epoch": 0.0834, "grad_norm": 13.689990997314453, "learning_rate": 4.5840000000000005e-06, "loss": 5.4803, "mean_token_accuracy": 0.6183862388134003, "num_tokens": 6992166.0, "step": 417 }, { "epoch": 0.0836, "grad_norm": 13.06170654296875, "learning_rate": 4.583e-06, "loss": 5.1476, "mean_token_accuracy": 0.6145320236682892, "num_tokens": 7011542.0, "step": 418 }, { "epoch": 0.0838, "grad_norm": 10.056525230407715, "learning_rate": 4.582e-06, "loss": 5.1888, "mean_token_accuracy": 0.5343915373086929, "num_tokens": 7030729.0, "step": 419 }, { "epoch": 0.084, "grad_norm": 6.395605564117432, "learning_rate": 4.581000000000001e-06, "loss": 5.4807, "mean_token_accuracy": 0.5086206793785095, "num_tokens": 7045228.0, "step": 420 }, { "epoch": 0.0842, "grad_norm": 8.128874778747559, "learning_rate": 4.58e-06, "loss": 5.4323, "mean_token_accuracy": 0.5689655244350433, "num_tokens": 7064800.0, "step": 421 }, { "epoch": 0.0844, "grad_norm": 8.527276039123535, "learning_rate": 4.579e-06, "loss": 5.4934, "mean_token_accuracy": 0.545584037899971, "num_tokens": 7083906.0, "step": 422 }, { "epoch": 0.0846, "grad_norm": 5.491787433624268, "learning_rate": 4.578e-06, "loss": 4.8269, "mean_token_accuracy": 0.6545092761516571, "num_tokens": 7094902.0, "step": 423 }, { "epoch": 0.0848, "grad_norm": 6.678435802459717, "learning_rate": 4.577e-06, "loss": 4.7984, "mean_token_accuracy": 0.6296296119689941, "num_tokens": 7104577.0, "step": 424 }, { "epoch": 0.085, "grad_norm": 22.77742576599121, "learning_rate": 4.576000000000001e-06, "loss": 5.7638, "mean_token_accuracy": 0.5648148059844971, "num_tokens": 7124732.0, "step": 425 }, { "epoch": 0.0852, "grad_norm": 9.347881317138672, "learning_rate": 4.575e-06, "loss": 5.2139, "mean_token_accuracy": 0.6139162480831146, "num_tokens": 7143991.0, "step": 426 }, { "epoch": 0.0854, "grad_norm": 5.620838165283203, "learning_rate": 4.574e-06, "loss": 4.6779, "mean_token_accuracy": 0.6503448188304901, "num_tokens": 7158607.0, "step": 427 }, { "epoch": 0.0856, "grad_norm": 6.864266872406006, "learning_rate": 4.573e-06, "loss": 4.8968, "mean_token_accuracy": 0.6183862388134003, "num_tokens": 7173064.0, "step": 428 }, { "epoch": 0.0858, "grad_norm": 22.015869140625, "learning_rate": 4.5720000000000004e-06, "loss": 5.1698, "mean_token_accuracy": 0.5743534564971924, "num_tokens": 7192243.0, "step": 429 }, { "epoch": 0.086, "grad_norm": 8.401832580566406, "learning_rate": 4.571000000000001e-06, "loss": 5.4089, "mean_token_accuracy": 0.5715725719928741, "num_tokens": 7211646.0, "step": 430 }, { "epoch": 0.0862, "grad_norm": 25.050403594970703, "learning_rate": 4.57e-06, "loss": 4.5259, "mean_token_accuracy": 0.6126644611358643, "num_tokens": 7230516.0, "step": 431 }, { "epoch": 0.0864, "grad_norm": 9.29088306427002, "learning_rate": 4.569e-06, "loss": 4.889, "mean_token_accuracy": 0.5278460085391998, "num_tokens": 7249620.0, "step": 432 }, { "epoch": 0.0866, "grad_norm": 15.916458129882812, "learning_rate": 4.568e-06, "loss": 4.9777, "mean_token_accuracy": 0.6257143020629883, "num_tokens": 7264094.0, "step": 433 }, { "epoch": 0.0868, "grad_norm": 6.861588954925537, "learning_rate": 4.5670000000000005e-06, "loss": 5.4239, "mean_token_accuracy": 0.4928571432828903, "num_tokens": 7283359.0, "step": 434 }, { "epoch": 0.087, "grad_norm": 13.225293159484863, "learning_rate": 4.566000000000001e-06, "loss": 4.7175, "mean_token_accuracy": 0.6028921008110046, "num_tokens": 7302888.0, "step": 435 }, { "epoch": 0.0872, "grad_norm": 17.012409210205078, "learning_rate": 4.565e-06, "loss": 5.6497, "mean_token_accuracy": 0.5857143104076385, "num_tokens": 7322148.0, "step": 436 }, { "epoch": 0.0874, "grad_norm": 10.171734809875488, "learning_rate": 4.564e-06, "loss": 4.5458, "mean_token_accuracy": 0.5567567646503448, "num_tokens": 7341333.0, "step": 437 }, { "epoch": 0.0876, "grad_norm": 16.463756561279297, "learning_rate": 4.563e-06, "loss": 4.9291, "mean_token_accuracy": 0.6135416626930237, "num_tokens": 7362605.0, "step": 438 }, { "epoch": 0.0878, "grad_norm": 17.738061904907227, "learning_rate": 4.5620000000000005e-06, "loss": 4.813, "mean_token_accuracy": 0.636904776096344, "num_tokens": 7381865.0, "step": 439 }, { "epoch": 0.088, "grad_norm": 8.94965648651123, "learning_rate": 4.561e-06, "loss": 4.8816, "mean_token_accuracy": 0.5999999940395355, "num_tokens": 7396402.0, "step": 440 }, { "epoch": 0.0882, "grad_norm": 15.025381088256836, "learning_rate": 4.56e-06, "loss": 5.6238, "mean_token_accuracy": 0.5384615361690521, "num_tokens": 7415660.0, "step": 441 }, { "epoch": 0.0884, "grad_norm": 10.704608917236328, "learning_rate": 4.559000000000001e-06, "loss": 5.291, "mean_token_accuracy": 0.5915948152542114, "num_tokens": 7435781.0, "step": 442 }, { "epoch": 0.0886, "grad_norm": 9.916930198669434, "learning_rate": 4.558e-06, "loss": 4.6656, "mean_token_accuracy": 0.6135416626930237, "num_tokens": 7454999.0, "step": 443 }, { "epoch": 0.0888, "grad_norm": 7.206878662109375, "learning_rate": 4.557000000000001e-06, "loss": 4.6639, "mean_token_accuracy": 0.4869281202554703, "num_tokens": 7469396.0, "step": 444 }, { "epoch": 0.089, "grad_norm": 6.467816352844238, "learning_rate": 4.556e-06, "loss": 4.9449, "mean_token_accuracy": 0.5981481373310089, "num_tokens": 7483855.0, "step": 445 }, { "epoch": 0.0892, "grad_norm": 9.08160400390625, "learning_rate": 4.5550000000000004e-06, "loss": 4.6868, "mean_token_accuracy": 0.6126373708248138, "num_tokens": 7498330.0, "step": 446 }, { "epoch": 0.0894, "grad_norm": 6.3002471923828125, "learning_rate": 4.554000000000001e-06, "loss": 5.0805, "mean_token_accuracy": 0.551282063126564, "num_tokens": 7512714.0, "step": 447 }, { "epoch": 0.0896, "grad_norm": 13.898117065429688, "learning_rate": 4.553e-06, "loss": 5.1324, "mean_token_accuracy": 0.5690476298332214, "num_tokens": 7531768.0, "step": 448 }, { "epoch": 0.0898, "grad_norm": 8.279391288757324, "learning_rate": 4.552000000000001e-06, "loss": 4.4569, "mean_token_accuracy": 0.5530530512332916, "num_tokens": 7546253.0, "step": 449 }, { "epoch": 0.09, "grad_norm": 11.584227561950684, "learning_rate": 4.551e-06, "loss": 5.3385, "mean_token_accuracy": 0.581250011920929, "num_tokens": 7565517.0, "step": 450 }, { "epoch": 0.0902, "grad_norm": 8.302013397216797, "learning_rate": 4.5500000000000005e-06, "loss": 4.9839, "mean_token_accuracy": 0.5351213216781616, "num_tokens": 7579402.0, "step": 451 }, { "epoch": 0.0904, "grad_norm": 9.814826965332031, "learning_rate": 4.549000000000001e-06, "loss": 4.6692, "mean_token_accuracy": 0.6079182624816895, "num_tokens": 7598660.0, "step": 452 }, { "epoch": 0.0906, "grad_norm": 6.627237796783447, "learning_rate": 4.548e-06, "loss": 5.4627, "mean_token_accuracy": 0.5092592537403107, "num_tokens": 7613180.0, "step": 453 }, { "epoch": 0.0908, "grad_norm": 11.831220626831055, "learning_rate": 4.547000000000001e-06, "loss": 4.8182, "mean_token_accuracy": 0.5758620798587799, "num_tokens": 7632441.0, "step": 454 }, { "epoch": 0.091, "grad_norm": 7.963436126708984, "learning_rate": 4.546e-06, "loss": 5.1914, "mean_token_accuracy": 0.5578093528747559, "num_tokens": 7651466.0, "step": 455 }, { "epoch": 0.0912, "grad_norm": 23.430757522583008, "learning_rate": 4.5450000000000005e-06, "loss": 5.1092, "mean_token_accuracy": 0.5648148059844971, "num_tokens": 7665848.0, "step": 456 }, { "epoch": 0.0914, "grad_norm": 7.155216693878174, "learning_rate": 4.544000000000001e-06, "loss": 4.5387, "mean_token_accuracy": 0.5334528088569641, "num_tokens": 7680308.0, "step": 457 }, { "epoch": 0.0916, "grad_norm": 9.627572059631348, "learning_rate": 4.543e-06, "loss": 5.2593, "mean_token_accuracy": 0.5478571355342865, "num_tokens": 7694782.0, "step": 458 }, { "epoch": 0.0918, "grad_norm": 7.240375995635986, "learning_rate": 4.542e-06, "loss": 4.444, "mean_token_accuracy": 0.6316666603088379, "num_tokens": 7709405.0, "step": 459 }, { "epoch": 0.092, "grad_norm": 14.054851531982422, "learning_rate": 4.541e-06, "loss": 4.7661, "mean_token_accuracy": 0.6431034505367279, "num_tokens": 7728742.0, "step": 460 }, { "epoch": 0.0922, "grad_norm": 6.709273815155029, "learning_rate": 4.540000000000001e-06, "loss": 5.1866, "mean_token_accuracy": 0.5717073231935501, "num_tokens": 7743526.0, "step": 461 }, { "epoch": 0.0924, "grad_norm": 15.618324279785156, "learning_rate": 4.539000000000001e-06, "loss": 5.0082, "mean_token_accuracy": 0.5, "num_tokens": 7763011.0, "step": 462 }, { "epoch": 0.0926, "grad_norm": 9.383596420288086, "learning_rate": 4.5380000000000004e-06, "loss": 4.7785, "mean_token_accuracy": 0.59375, "num_tokens": 7777515.0, "step": 463 }, { "epoch": 0.0928, "grad_norm": 8.333616256713867, "learning_rate": 4.537e-06, "loss": 4.592, "mean_token_accuracy": 0.522556409239769, "num_tokens": 7792009.0, "step": 464 }, { "epoch": 0.093, "grad_norm": 14.328873634338379, "learning_rate": 4.536e-06, "loss": 4.8537, "mean_token_accuracy": 0.62321937084198, "num_tokens": 7811576.0, "step": 465 }, { "epoch": 0.0932, "grad_norm": 8.614009857177734, "learning_rate": 4.535000000000001e-06, "loss": 5.0954, "mean_token_accuracy": 0.5415282547473907, "num_tokens": 7830643.0, "step": 466 }, { "epoch": 0.0934, "grad_norm": 8.26689338684082, "learning_rate": 4.534000000000001e-06, "loss": 4.742, "mean_token_accuracy": 0.6200265288352966, "num_tokens": 7848085.0, "step": 467 }, { "epoch": 0.0936, "grad_norm": 11.07862377166748, "learning_rate": 4.5330000000000005e-06, "loss": 4.6416, "mean_token_accuracy": 0.5669642984867096, "num_tokens": 7867108.0, "step": 468 }, { "epoch": 0.0938, "grad_norm": 7.7039337158203125, "learning_rate": 4.532e-06, "loss": 4.7395, "mean_token_accuracy": 0.5769230723381042, "num_tokens": 7876728.0, "step": 469 }, { "epoch": 0.094, "grad_norm": 15.2058744430542, "learning_rate": 4.531e-06, "loss": 4.2838, "mean_token_accuracy": 0.5869939625263214, "num_tokens": 7893132.0, "step": 470 }, { "epoch": 0.0942, "grad_norm": 12.777841567993164, "learning_rate": 4.530000000000001e-06, "loss": 4.6823, "mean_token_accuracy": 0.5769230723381042, "num_tokens": 7912476.0, "step": 471 }, { "epoch": 0.0944, "grad_norm": 13.58057975769043, "learning_rate": 4.529000000000001e-06, "loss": 4.4577, "mean_token_accuracy": 0.589756578207016, "num_tokens": 7931920.0, "step": 472 }, { "epoch": 0.0946, "grad_norm": 7.151589393615723, "learning_rate": 4.5280000000000005e-06, "loss": 4.784, "mean_token_accuracy": 0.5804398059844971, "num_tokens": 7946400.0, "step": 473 }, { "epoch": 0.0948, "grad_norm": 7.511891841888428, "learning_rate": 4.527e-06, "loss": 5.1037, "mean_token_accuracy": 0.5648148059844971, "num_tokens": 7960863.0, "step": 474 }, { "epoch": 0.095, "grad_norm": 9.209273338317871, "learning_rate": 4.526e-06, "loss": 4.6556, "mean_token_accuracy": 0.5915948152542114, "num_tokens": 7975547.0, "step": 475 }, { "epoch": 0.0952, "grad_norm": 10.363279342651367, "learning_rate": 4.525000000000001e-06, "loss": 5.0172, "mean_token_accuracy": 0.5742424130439758, "num_tokens": 7991119.0, "step": 476 }, { "epoch": 0.0954, "grad_norm": 7.636615753173828, "learning_rate": 4.524e-06, "loss": 4.952, "mean_token_accuracy": 0.6436781585216522, "num_tokens": 8010377.0, "step": 477 }, { "epoch": 0.0956, "grad_norm": 8.515202522277832, "learning_rate": 4.5230000000000006e-06, "loss": 4.6051, "mean_token_accuracy": 0.6206896305084229, "num_tokens": 8029123.0, "step": 478 }, { "epoch": 0.0958, "grad_norm": 8.392003059387207, "learning_rate": 4.522e-06, "loss": 4.9078, "mean_token_accuracy": 0.6145320236682892, "num_tokens": 8048382.0, "step": 479 }, { "epoch": 0.096, "grad_norm": 4.060763359069824, "learning_rate": 4.521e-06, "loss": 4.5817, "mean_token_accuracy": 0.5833333432674408, "num_tokens": 8058289.0, "step": 480 }, { "epoch": 0.0962, "grad_norm": 10.596019744873047, "learning_rate": 4.520000000000001e-06, "loss": 4.2822, "mean_token_accuracy": 0.5428152531385422, "num_tokens": 8077371.0, "step": 481 }, { "epoch": 0.0964, "grad_norm": 9.166875839233398, "learning_rate": 4.519e-06, "loss": 4.93, "mean_token_accuracy": 0.5862068831920624, "num_tokens": 8096669.0, "step": 482 }, { "epoch": 0.0966, "grad_norm": 9.244916915893555, "learning_rate": 4.518e-06, "loss": 4.3974, "mean_token_accuracy": 0.5298245847225189, "num_tokens": 8111158.0, "step": 483 }, { "epoch": 0.0968, "grad_norm": 5.592848300933838, "learning_rate": 4.517e-06, "loss": 4.9458, "mean_token_accuracy": 0.550000011920929, "num_tokens": 8126046.0, "step": 484 }, { "epoch": 0.097, "grad_norm": 7.911005020141602, "learning_rate": 4.5160000000000005e-06, "loss": 4.9862, "mean_token_accuracy": 0.5861823260784149, "num_tokens": 8156760.0, "step": 485 }, { "epoch": 0.0972, "grad_norm": 22.27390480041504, "learning_rate": 4.515000000000001e-06, "loss": 4.9185, "mean_token_accuracy": 0.5522802919149399, "num_tokens": 8178188.0, "step": 486 }, { "epoch": 0.0974, "grad_norm": 6.784433364868164, "learning_rate": 4.514e-06, "loss": 4.1169, "mean_token_accuracy": 0.6459695100784302, "num_tokens": 8192689.0, "step": 487 }, { "epoch": 0.0976, "grad_norm": 5.436159133911133, "learning_rate": 4.513e-06, "loss": 4.2387, "mean_token_accuracy": 0.6333333253860474, "num_tokens": 8207151.0, "step": 488 }, { "epoch": 0.0978, "grad_norm": 10.865344047546387, "learning_rate": 4.512e-06, "loss": 4.5805, "mean_token_accuracy": 0.6162962913513184, "num_tokens": 8221643.0, "step": 489 }, { "epoch": 0.098, "grad_norm": 18.25914764404297, "learning_rate": 4.5110000000000005e-06, "loss": 4.7056, "mean_token_accuracy": 0.5462962985038757, "num_tokens": 8242886.0, "step": 490 }, { "epoch": 0.0982, "grad_norm": 8.091397285461426, "learning_rate": 4.510000000000001e-06, "loss": 4.7771, "mean_token_accuracy": 0.5669642984867096, "num_tokens": 8262615.0, "step": 491 }, { "epoch": 0.0984, "grad_norm": 5.078823566436768, "learning_rate": 4.509e-06, "loss": 4.7878, "mean_token_accuracy": 0.5173454135656357, "num_tokens": 8275070.0, "step": 492 }, { "epoch": 0.0986, "grad_norm": 10.31528377532959, "learning_rate": 4.508e-06, "loss": 4.5168, "mean_token_accuracy": 0.6481481492519379, "num_tokens": 8289526.0, "step": 493 }, { "epoch": 0.0988, "grad_norm": 8.380623817443848, "learning_rate": 4.507e-06, "loss": 4.5738, "mean_token_accuracy": 0.5497835576534271, "num_tokens": 8306115.0, "step": 494 }, { "epoch": 0.099, "grad_norm": 15.303783416748047, "learning_rate": 4.5060000000000006e-06, "loss": 4.4765, "mean_token_accuracy": 0.5352205336093903, "num_tokens": 8327367.0, "step": 495 }, { "epoch": 0.0992, "grad_norm": 9.432476043701172, "learning_rate": 4.505e-06, "loss": 4.2436, "mean_token_accuracy": 0.5757575631141663, "num_tokens": 8347475.0, "step": 496 }, { "epoch": 0.0994, "grad_norm": 7.746166706085205, "learning_rate": 4.504e-06, "loss": 4.6564, "mean_token_accuracy": 0.5372548997402191, "num_tokens": 8367244.0, "step": 497 }, { "epoch": 0.0996, "grad_norm": 6.070708274841309, "learning_rate": 4.503e-06, "loss": 4.7593, "mean_token_accuracy": 0.5413165390491486, "num_tokens": 8387418.0, "step": 498 }, { "epoch": 0.0998, "grad_norm": 9.557999610900879, "learning_rate": 4.502e-06, "loss": 4.6868, "mean_token_accuracy": 0.6296296119689941, "num_tokens": 8406160.0, "step": 499 }, { "epoch": 0.1, "grad_norm": 18.0493106842041, "learning_rate": 4.501000000000001e-06, "loss": 4.4627, "mean_token_accuracy": 0.5960648059844971, "num_tokens": 8425459.0, "step": 500 } ], "logging_steps": 1, "max_steps": 5000, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 250, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 3.9176829428171366e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }